//+build !noasm //+build !appengine //+build !gccgo // Copyright 2015, Klaus Post, see LICENSE for details. // Based on http://www.snia.org/sites/default/files2/SDC2013/presentations/NewThinking/EthanMiller_Screaming_Fast_Galois_Field%20Arithmetic_SIMD%20Instructions.pdf // and http://jerasure.org/jerasure/gf-complete/tree/master // func galMulSSSE3Xor(low, high, in, out []byte) TEXT ·galMulSSSE3Xor(SB), 7, $0 MOVQ low+0(FP), SI // SI: &low MOVQ high+24(FP), DX // DX: &high MOVOU (SI), X6 // X6 low MOVOU (DX), X7 // X7: high MOVQ $15, BX // BX: low mask MOVQ BX, X8 PXOR X5, X5 MOVQ in+48(FP), SI // R11: &in MOVQ in_len+56(FP), R9 // R9: len(in) MOVQ out+72(FP), DX // DX: &out PSHUFB X5, X8 // X8: lomask (unpacked) SHRQ $4, R9 // len(in) / 16 MOVQ SI, AX MOVQ DX, BX ANDQ $15, AX ANDQ $15, BX CMPQ R9, $0 JEQ done_xor ORQ AX, BX CMPQ BX, $0 JNZ loopback_xor loopback_xor_aligned: MOVOA (SI), X0 // in[x] MOVOA (DX), X4 // out[x] MOVOA X0, X1 // in[x] MOVOA X6, X2 // low copy MOVOA X7, X3 // high copy PSRLQ $4, X1 // X1: high input PAND X8, X0 // X0: low input PAND X8, X1 // X0: high input PSHUFB X0, X2 // X2: mul low part PSHUFB X1, X3 // X3: mul high part PXOR X2, X3 // X3: Result PXOR X4, X3 // X3: Result xor existing out MOVOA X3, (DX) // Store ADDQ $16, SI // in+=16 ADDQ $16, DX // out+=16 SUBQ $1, R9 JNZ loopback_xor_aligned JMP done_xor loopback_xor: MOVOU (SI), X0 // in[x] MOVOU (DX), X4 // out[x] MOVOU X0, X1 // in[x] MOVOU X6, X2 // low copy MOVOU X7, X3 // high copy PSRLQ $4, X1 // X1: high input PAND X8, X0 // X0: low input PAND X8, X1 // X0: high input PSHUFB X0, X2 // X2: mul low part PSHUFB X1, X3 // X3: mul high part PXOR X2, X3 // X3: Result PXOR X4, X3 // X3: Result xor existing out MOVOU X3, (DX) // Store ADDQ $16, SI // in+=16 ADDQ $16, DX // out+=16 SUBQ $1, R9 JNZ loopback_xor done_xor: RET // func galMulSSSE3(low, high, in, out []byte) TEXT ·galMulSSSE3(SB), 7, $0 MOVQ low+0(FP), SI // SI: &low MOVQ high+24(FP), DX // DX: &high MOVOU (SI), X6 // X6 low MOVOU (DX), X7 // X7: high MOVQ $15, BX // BX: low mask MOVQ BX, X8 PXOR X5, X5 MOVQ in+48(FP), SI // R11: &in MOVQ in_len+56(FP), R9 // R9: len(in) MOVQ out+72(FP), DX // DX: &out PSHUFB X5, X8 // X8: lomask (unpacked) MOVQ SI, AX MOVQ DX, BX SHRQ $4, R9 // len(in) / 16 ANDQ $15, AX ANDQ $15, BX CMPQ R9, $0 JEQ done ORQ AX, BX CMPQ BX, $0 JNZ loopback loopback_aligned: MOVOA (SI), X0 // in[x] MOVOA X0, X1 // in[x] MOVOA X6, X2 // low copy MOVOA X7, X3 // high copy PSRLQ $4, X1 // X1: high input PAND X8, X0 // X0: low input PAND X8, X1 // X0: high input PSHUFB X0, X2 // X2: mul low part PSHUFB X1, X3 // X3: mul high part PXOR X2, X3 // X3: Result MOVOA X3, (DX) // Store ADDQ $16, SI // in+=16 ADDQ $16, DX // out+=16 SUBQ $1, R9 JNZ loopback_aligned JMP done loopback: MOVOU (SI), X0 // in[x] MOVOU X0, X1 // in[x] MOVOA X6, X2 // low copy MOVOA X7, X3 // high copy PSRLQ $4, X1 // X1: high input PAND X8, X0 // X0: low input PAND X8, X1 // X0: high input PSHUFB X0, X2 // X2: mul low part PSHUFB X1, X3 // X3: mul high part PXOR X2, X3 // X3: Result MOVOU X3, (DX) // Store ADDQ $16, SI // in+=16 ADDQ $16, DX // out+=16 SUBQ $1, R9 JNZ loopback done: RET // func galMulAVX2Xor(low, high, in, out []byte) TEXT ·galMulAVX2Xor(SB), 7, $0 MOVQ low+0(FP), SI // SI: &low MOVQ high+24(FP), DX // DX: &high MOVQ $15, BX // BX: low mask MOVQ BX, X5 MOVOU (SI), X6 // X6: low MOVOU (DX), X7 // X7: high MOVQ in_len+56(FP), R9 // R9: len(in) VINSERTI128 $1, X6, Y6, Y6 // low VINSERTI128 $1, X7, Y7, Y7 // high VPBROADCASTB X5, Y8 // Y8: lomask (unpacked) SHRQ $5, R9 // len(in) / 32 MOVQ out+72(FP), DX // DX: &out MOVQ in+48(FP), SI // SI: &in TESTQ R9, R9 JZ done_xor_avx2 loopback_xor_avx2: VMOVDQU (SI), Y0 VMOVDQU (DX), Y4 VPSRLQ $4, Y0, Y1 // Y1: high input VPAND Y8, Y0, Y0 // Y0: low input VPAND Y8, Y1, Y1 // Y1: high input VPSHUFB Y0, Y6, Y2 // Y2: mul low part VPSHUFB Y1, Y7, Y3 // Y3: mul high part VPXOR Y3, Y2, Y3 // Y3: Result VPXOR Y4, Y3, Y4 // Y4: Result VMOVDQU Y4, (DX) ADDQ $32, SI // in+=32 ADDQ $32, DX // out+=32 SUBQ $1, R9 JNZ loopback_xor_avx2 done_xor_avx2: VZEROUPPER RET // func galMulAVX2(low, high, in, out []byte) TEXT ·galMulAVX2(SB), 7, $0 MOVQ low+0(FP), SI // SI: &low MOVQ high+24(FP), DX // DX: &high MOVQ $15, BX // BX: low mask MOVQ BX, X5 MOVOU (SI), X6 // X6: low MOVOU (DX), X7 // X7: high MOVQ in_len+56(FP), R9 // R9: len(in) VINSERTI128 $1, X6, Y6, Y6 // low VINSERTI128 $1, X7, Y7, Y7 // high VPBROADCASTB X5, Y8 // Y8: lomask (unpacked) SHRQ $5, R9 // len(in) / 32 MOVQ out+72(FP), DX // DX: &out MOVQ in+48(FP), SI // SI: &in TESTQ R9, R9 JZ done_avx2 loopback_avx2: VMOVDQU (SI), Y0 VPSRLQ $4, Y0, Y1 // Y1: high input VPAND Y8, Y0, Y0 // Y0: low input VPAND Y8, Y1, Y1 // Y1: high input VPSHUFB Y0, Y6, Y2 // Y2: mul low part VPSHUFB Y1, Y7, Y3 // Y3: mul high part VPXOR Y3, Y2, Y4 // Y4: Result VMOVDQU Y4, (DX) ADDQ $32, SI // in+=32 ADDQ $32, DX // out+=32 SUBQ $1, R9 JNZ loopback_avx2 done_avx2: VZEROUPPER RET // func sSE2XorSlice(in, out []byte) TEXT ·sSE2XorSlice(SB), 7, $0 MOVQ in+0(FP), SI // SI: &in MOVQ in_len+8(FP), R9 // R9: len(in) MOVQ out+24(FP), DX // DX: &out SHRQ $4, R9 // len(in) / 16 CMPQ R9, $0 JEQ done_xor_sse2 loopback_xor_sse2: MOVOU (SI), X0 // in[x] MOVOU (DX), X1 // out[x] PXOR X0, X1 MOVOU X1, (DX) ADDQ $16, SI // in+=16 ADDQ $16, DX // out+=16 SUBQ $1, R9 JNZ loopback_xor_sse2 done_xor_sse2: RET // func galMulAVX2Xor_64(low, high, in, out []byte) TEXT ·galMulAVX2Xor_64(SB), 7, $0 MOVQ low+0(FP), SI // SI: &low MOVQ high+24(FP), DX // DX: &high MOVQ $15, BX // BX: low mask MOVQ BX, X5 MOVOU (SI), X6 // X6: low MOVOU (DX), X7 // X7: high MOVQ in_len+56(FP), R9 // R9: len(in) VINSERTI128 $1, X6, Y6, Y6 // low VINSERTI128 $1, X7, Y7, Y7 // high VPBROADCASTB X5, Y8 // Y8: lomask (unpacked) SHRQ $6, R9 // len(in) / 64 MOVQ out+72(FP), DX // DX: &out MOVQ in+48(FP), SI // SI: &in TESTQ R9, R9 JZ done_xor_avx2_64 loopback_xor_avx2_64: VMOVDQU (SI), Y0 VMOVDQU 32(SI), Y10 VMOVDQU (DX), Y4 VMOVDQU 32(DX), Y14 VPSRLQ $4, Y0, Y1 // Y1: high input VPSRLQ $4, Y10, Y11 // Y11: high input 2 VPAND Y8, Y0, Y0 // Y0: low input VPAND Y8, Y10, Y10 // Y10: low input 2 VPAND Y8, Y1, Y1 // Y11: high input VPAND Y8, Y11, Y11 // Y11: high input 2 VPSHUFB Y0, Y6, Y2 // Y2: mul low part VPSHUFB Y10, Y6, Y12 // Y12: mul low part 2 VPSHUFB Y1, Y7, Y3 // Y3: mul high part VPSHUFB Y11, Y7, Y13 // Y13: mul high part 2 VPXOR Y3, Y2, Y3 // Y3: Result VPXOR Y13, Y12, Y13 // Y13: Result 2 VPXOR Y4, Y3, Y4 // Y4: Result VPXOR Y14, Y13, Y14 // Y4: Result 2 VMOVDQU Y4, (DX) VMOVDQU Y14, 32(DX) ADDQ $64, SI // in+=64 ADDQ $64, DX // out+=64 SUBQ $1, R9 JNZ loopback_xor_avx2_64 done_xor_avx2_64: VZEROUPPER RET // func galMulAVX2_64(low, high, in, out []byte) TEXT ·galMulAVX2_64(SB), 7, $0 MOVQ low+0(FP), SI // SI: &low MOVQ high+24(FP), DX // DX: &high MOVQ $15, BX // BX: low mask MOVQ BX, X5 MOVOU (SI), X6 // X6: low MOVOU (DX), X7 // X7: high MOVQ in_len+56(FP), R9 // R9: len(in) VINSERTI128 $1, X6, Y6, Y6 // low VINSERTI128 $1, X7, Y7, Y7 // high VPBROADCASTB X5, Y8 // Y8: lomask (unpacked) SHRQ $6, R9 // len(in) / 64 MOVQ out+72(FP), DX // DX: &out MOVQ in+48(FP), SI // SI: &in TESTQ R9, R9 JZ done_avx2_64 loopback_avx2_64: VMOVDQU (SI), Y0 VMOVDQU 32(SI), Y10 VPSRLQ $4, Y0, Y1 // Y1: high input VPSRLQ $4, Y10, Y11 // Y11: high input 2 VPAND Y8, Y0, Y0 // Y0: low input VPAND Y8, Y10, Y10 // Y10: low input VPAND Y8, Y1, Y1 // Y1: high input VPAND Y8, Y11, Y11 // Y11: high input 2 VPSHUFB Y0, Y6, Y2 // Y2: mul low part VPSHUFB Y10, Y6, Y12 // Y12: mul low part 2 VPSHUFB Y1, Y7, Y3 // Y3: mul high part VPSHUFB Y11, Y7, Y13 // Y13: mul high part 2 VPXOR Y3, Y2, Y4 // Y4: Result VPXOR Y13, Y12, Y14 // Y14: Result 2 VMOVDQU Y4, (DX) VMOVDQU Y14, 32(DX) ADDQ $64, SI // in+=64 ADDQ $64, DX // out+=64 SUBQ $1, R9 JNZ loopback_avx2_64 done_avx2_64: VZEROUPPER RET // func sSE2XorSlice_64(in, out []byte) TEXT ·sSE2XorSlice_64(SB), 7, $0 MOVQ in+0(FP), SI // SI: &in MOVQ in_len+8(FP), R9 // R9: len(in) MOVQ out+24(FP), DX // DX: &out SHRQ $6, R9 // len(in) / 64 CMPQ R9, $0 JEQ done_xor_sse2_64 loopback_xor_sse2_64: MOVOU (SI), X0 // in[x] MOVOU 16(SI), X2 // in[x] MOVOU 32(SI), X4 // in[x] MOVOU 48(SI), X6 // in[x] MOVOU (DX), X1 // out[x] MOVOU 16(DX), X3 // out[x] MOVOU 32(DX), X5 // out[x] MOVOU 48(DX), X7 // out[x] PXOR X0, X1 PXOR X2, X3 PXOR X4, X5 PXOR X6, X7 MOVOU X1, (DX) MOVOU X3, 16(DX) MOVOU X5, 32(DX) MOVOU X7, 48(DX) ADDQ $64, SI // in+=64 ADDQ $64, DX // out+=64 SUBQ $1, R9 JNZ loopback_xor_sse2_64 done_xor_sse2_64: RET