aboutsummaryrefslogtreecommitdiff
path: root/src/simd/internal/simd_test/binary_test.go
diff options
context:
space:
mode:
Diffstat (limited to 'src/simd/internal/simd_test/binary_test.go')
-rw-r--r--src/simd/internal/simd_test/binary_test.go361
1 files changed, 361 insertions, 0 deletions
diff --git a/src/simd/internal/simd_test/binary_test.go b/src/simd/internal/simd_test/binary_test.go
new file mode 100644
index 0000000000..04dca3e2e2
--- /dev/null
+++ b/src/simd/internal/simd_test/binary_test.go
@@ -0,0 +1,361 @@
+// Copyright 2025 The Go Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style
+// license that can be found in the LICENSE file.
+
+//go:build goexperiment.simd && amd64
+
+package simd_test
+
+import (
+ "simd"
+ "testing"
+)
+
+func TestAdd(t *testing.T) {
+ testFloat32x4Binary(t, simd.Float32x4.Add, addSlice[float32])
+ testFloat32x8Binary(t, simd.Float32x8.Add, addSlice[float32])
+ testFloat64x2Binary(t, simd.Float64x2.Add, addSlice[float64])
+ testFloat64x4Binary(t, simd.Float64x4.Add, addSlice[float64])
+
+ testInt16x16Binary(t, simd.Int16x16.Add, addSlice[int16])
+ testInt16x8Binary(t, simd.Int16x8.Add, addSlice[int16])
+ testInt32x4Binary(t, simd.Int32x4.Add, addSlice[int32])
+ testInt32x8Binary(t, simd.Int32x8.Add, addSlice[int32])
+ testInt64x2Binary(t, simd.Int64x2.Add, addSlice[int64])
+ testInt64x4Binary(t, simd.Int64x4.Add, addSlice[int64])
+ testInt8x16Binary(t, simd.Int8x16.Add, addSlice[int8])
+ testInt8x32Binary(t, simd.Int8x32.Add, addSlice[int8])
+
+ testUint16x16Binary(t, simd.Uint16x16.Add, addSlice[uint16])
+ testUint16x8Binary(t, simd.Uint16x8.Add, addSlice[uint16])
+ testUint32x4Binary(t, simd.Uint32x4.Add, addSlice[uint32])
+ testUint32x8Binary(t, simd.Uint32x8.Add, addSlice[uint32])
+ testUint64x2Binary(t, simd.Uint64x2.Add, addSlice[uint64])
+ testUint64x4Binary(t, simd.Uint64x4.Add, addSlice[uint64])
+ testUint8x16Binary(t, simd.Uint8x16.Add, addSlice[uint8])
+ testUint8x32Binary(t, simd.Uint8x32.Add, addSlice[uint8])
+
+ if simd.X86.AVX512() {
+ testFloat32x16Binary(t, simd.Float32x16.Add, addSlice[float32])
+ testFloat64x8Binary(t, simd.Float64x8.Add, addSlice[float64])
+ testInt8x64Binary(t, simd.Int8x64.Add, addSlice[int8])
+ testInt16x32Binary(t, simd.Int16x32.Add, addSlice[int16])
+ testInt32x16Binary(t, simd.Int32x16.Add, addSlice[int32])
+ testInt64x8Binary(t, simd.Int64x8.Add, addSlice[int64])
+ testUint8x64Binary(t, simd.Uint8x64.Add, addSlice[uint8])
+ testUint16x32Binary(t, simd.Uint16x32.Add, addSlice[uint16])
+ testUint32x16Binary(t, simd.Uint32x16.Add, addSlice[uint32])
+ testUint64x8Binary(t, simd.Uint64x8.Add, addSlice[uint64])
+ }
+}
+
+func TestSub(t *testing.T) {
+ testFloat32x4Binary(t, simd.Float32x4.Sub, subSlice[float32])
+ testFloat32x8Binary(t, simd.Float32x8.Sub, subSlice[float32])
+ testFloat64x2Binary(t, simd.Float64x2.Sub, subSlice[float64])
+ testFloat64x4Binary(t, simd.Float64x4.Sub, subSlice[float64])
+
+ testInt16x16Binary(t, simd.Int16x16.Sub, subSlice[int16])
+ testInt16x8Binary(t, simd.Int16x8.Sub, subSlice[int16])
+ testInt32x4Binary(t, simd.Int32x4.Sub, subSlice[int32])
+ testInt32x8Binary(t, simd.Int32x8.Sub, subSlice[int32])
+ testInt64x2Binary(t, simd.Int64x2.Sub, subSlice[int64])
+ testInt64x4Binary(t, simd.Int64x4.Sub, subSlice[int64])
+ testInt8x16Binary(t, simd.Int8x16.Sub, subSlice[int8])
+ testInt8x32Binary(t, simd.Int8x32.Sub, subSlice[int8])
+
+ testUint16x16Binary(t, simd.Uint16x16.Sub, subSlice[uint16])
+ testUint16x8Binary(t, simd.Uint16x8.Sub, subSlice[uint16])
+ testUint32x4Binary(t, simd.Uint32x4.Sub, subSlice[uint32])
+ testUint32x8Binary(t, simd.Uint32x8.Sub, subSlice[uint32])
+ testUint64x2Binary(t, simd.Uint64x2.Sub, subSlice[uint64])
+ testUint64x4Binary(t, simd.Uint64x4.Sub, subSlice[uint64])
+ testUint8x16Binary(t, simd.Uint8x16.Sub, subSlice[uint8])
+ testUint8x32Binary(t, simd.Uint8x32.Sub, subSlice[uint8])
+
+ if simd.X86.AVX512() {
+ testFloat32x16Binary(t, simd.Float32x16.Sub, subSlice[float32])
+ testFloat64x8Binary(t, simd.Float64x8.Sub, subSlice[float64])
+ testInt8x64Binary(t, simd.Int8x64.Sub, subSlice[int8])
+ testInt16x32Binary(t, simd.Int16x32.Sub, subSlice[int16])
+ testInt32x16Binary(t, simd.Int32x16.Sub, subSlice[int32])
+ testInt64x8Binary(t, simd.Int64x8.Sub, subSlice[int64])
+ testUint8x64Binary(t, simd.Uint8x64.Sub, subSlice[uint8])
+ testUint16x32Binary(t, simd.Uint16x32.Sub, subSlice[uint16])
+ testUint32x16Binary(t, simd.Uint32x16.Sub, subSlice[uint32])
+ testUint64x8Binary(t, simd.Uint64x8.Sub, subSlice[uint64])
+ }
+}
+
+func TestMax(t *testing.T) {
+ // testFloat32x4Binary(t, simd.Float32x4.Max, maxSlice[float32]) // nan is wrong
+ // testFloat32x8Binary(t, simd.Float32x8.Max, maxSlice[float32]) // nan is wrong
+ // testFloat64x2Binary(t, simd.Float64x2.Max, maxSlice[float64]) // nan is wrong
+ // testFloat64x4Binary(t, simd.Float64x4.Max, maxSlice[float64]) // nan is wrong
+
+ testInt16x16Binary(t, simd.Int16x16.Max, maxSlice[int16])
+ testInt16x8Binary(t, simd.Int16x8.Max, maxSlice[int16])
+ testInt32x4Binary(t, simd.Int32x4.Max, maxSlice[int32])
+ testInt32x8Binary(t, simd.Int32x8.Max, maxSlice[int32])
+
+ if simd.X86.AVX512() {
+ testInt64x2Binary(t, simd.Int64x2.Max, maxSlice[int64])
+ testInt64x4Binary(t, simd.Int64x4.Max, maxSlice[int64])
+ }
+
+ testInt8x16Binary(t, simd.Int8x16.Max, maxSlice[int8])
+ testInt8x32Binary(t, simd.Int8x32.Max, maxSlice[int8])
+
+ testUint16x16Binary(t, simd.Uint16x16.Max, maxSlice[uint16])
+ testUint16x8Binary(t, simd.Uint16x8.Max, maxSlice[uint16])
+ testUint32x4Binary(t, simd.Uint32x4.Max, maxSlice[uint32])
+ testUint32x8Binary(t, simd.Uint32x8.Max, maxSlice[uint32])
+
+ if simd.X86.AVX512() {
+ testUint64x2Binary(t, simd.Uint64x2.Max, maxSlice[uint64])
+ testUint64x4Binary(t, simd.Uint64x4.Max, maxSlice[uint64])
+ }
+
+ testUint8x16Binary(t, simd.Uint8x16.Max, maxSlice[uint8])
+ testUint8x32Binary(t, simd.Uint8x32.Max, maxSlice[uint8])
+
+ if simd.X86.AVX512() {
+ // testFloat32x16Binary(t, simd.Float32x16.Max, maxSlice[float32]) // nan is wrong
+ // testFloat64x8Binary(t, simd.Float64x8.Max, maxSlice[float64]) // nan is wrong
+ testInt8x64Binary(t, simd.Int8x64.Max, maxSlice[int8])
+ testInt16x32Binary(t, simd.Int16x32.Max, maxSlice[int16])
+ testInt32x16Binary(t, simd.Int32x16.Max, maxSlice[int32])
+ testInt64x8Binary(t, simd.Int64x8.Max, maxSlice[int64])
+ testUint8x64Binary(t, simd.Uint8x64.Max, maxSlice[uint8])
+ testUint16x32Binary(t, simd.Uint16x32.Max, maxSlice[uint16])
+ testUint32x16Binary(t, simd.Uint32x16.Max, maxSlice[uint32])
+ testUint64x8Binary(t, simd.Uint64x8.Max, maxSlice[uint64])
+ }
+}
+
+func TestMin(t *testing.T) {
+ // testFloat32x4Binary(t, simd.Float32x4.Min, minSlice[float32]) // nan is wrong
+ // testFloat32x8Binary(t, simd.Float32x8.Min, minSlice[float32]) // nan is wrong
+ // testFloat64x2Binary(t, simd.Float64x2.Min, minSlice[float64]) // nan is wrong
+ // testFloat64x4Binary(t, simd.Float64x4.Min, minSlice[float64]) // nan is wrong
+
+ testInt16x16Binary(t, simd.Int16x16.Min, minSlice[int16])
+ testInt16x8Binary(t, simd.Int16x8.Min, minSlice[int16])
+ testInt32x4Binary(t, simd.Int32x4.Min, minSlice[int32])
+ testInt32x8Binary(t, simd.Int32x8.Min, minSlice[int32])
+
+ if simd.X86.AVX512() {
+ testInt64x2Binary(t, simd.Int64x2.Min, minSlice[int64])
+ testInt64x4Binary(t, simd.Int64x4.Min, minSlice[int64])
+ }
+
+ testInt8x16Binary(t, simd.Int8x16.Min, minSlice[int8])
+ testInt8x32Binary(t, simd.Int8x32.Min, minSlice[int8])
+
+ testUint16x16Binary(t, simd.Uint16x16.Min, minSlice[uint16])
+ testUint16x8Binary(t, simd.Uint16x8.Min, minSlice[uint16])
+ testUint32x4Binary(t, simd.Uint32x4.Min, minSlice[uint32])
+ testUint32x8Binary(t, simd.Uint32x8.Min, minSlice[uint32])
+
+ if simd.X86.AVX512() {
+ testUint64x2Binary(t, simd.Uint64x2.Min, minSlice[uint64])
+ testUint64x4Binary(t, simd.Uint64x4.Min, minSlice[uint64])
+ }
+
+ testUint8x16Binary(t, simd.Uint8x16.Min, minSlice[uint8])
+ testUint8x32Binary(t, simd.Uint8x32.Min, minSlice[uint8])
+
+ if simd.X86.AVX512() {
+ // testFloat32x16Binary(t, simd.Float32x16.Min, minSlice[float32]) // nan is wrong
+ // testFloat64x8Binary(t, simd.Float64x8.Min, minSlice[float64]) // nan is wrong
+ testInt8x64Binary(t, simd.Int8x64.Min, minSlice[int8])
+ testInt16x32Binary(t, simd.Int16x32.Min, minSlice[int16])
+ testInt32x16Binary(t, simd.Int32x16.Min, minSlice[int32])
+ testInt64x8Binary(t, simd.Int64x8.Min, minSlice[int64])
+ testUint8x64Binary(t, simd.Uint8x64.Min, minSlice[uint8])
+ testUint16x32Binary(t, simd.Uint16x32.Min, minSlice[uint16])
+ testUint32x16Binary(t, simd.Uint32x16.Min, minSlice[uint32])
+ testUint64x8Binary(t, simd.Uint64x8.Min, minSlice[uint64])
+ }
+}
+
+func TestAnd(t *testing.T) {
+ testInt16x16Binary(t, simd.Int16x16.And, andSlice[int16])
+ testInt16x8Binary(t, simd.Int16x8.And, andSlice[int16])
+ testInt32x4Binary(t, simd.Int32x4.And, andSlice[int32])
+ testInt32x8Binary(t, simd.Int32x8.And, andSlice[int32])
+ testInt64x2Binary(t, simd.Int64x2.And, andSlice[int64])
+ testInt64x4Binary(t, simd.Int64x4.And, andSlice[int64])
+ testInt8x16Binary(t, simd.Int8x16.And, andSlice[int8])
+ testInt8x32Binary(t, simd.Int8x32.And, andSlice[int8])
+
+ testUint16x16Binary(t, simd.Uint16x16.And, andSlice[uint16])
+ testUint16x8Binary(t, simd.Uint16x8.And, andSlice[uint16])
+ testUint32x4Binary(t, simd.Uint32x4.And, andSlice[uint32])
+ testUint32x8Binary(t, simd.Uint32x8.And, andSlice[uint32])
+ testUint64x2Binary(t, simd.Uint64x2.And, andSlice[uint64])
+ testUint64x4Binary(t, simd.Uint64x4.And, andSlice[uint64])
+ testUint8x16Binary(t, simd.Uint8x16.And, andSlice[uint8])
+ testUint8x32Binary(t, simd.Uint8x32.And, andSlice[uint8])
+
+ if simd.X86.AVX512() {
+ // testInt8x64Binary(t, simd.Int8x64.And, andISlice[int8]) // missing
+ // testInt16x32Binary(t, simd.Int16x32.And, andISlice[int16]) // missing
+ testInt32x16Binary(t, simd.Int32x16.And, andSlice[int32])
+ testInt64x8Binary(t, simd.Int64x8.And, andSlice[int64])
+ // testUint8x64Binary(t, simd.Uint8x64.And, andISlice[uint8]) // missing
+ // testUint16x32Binary(t, simd.Uint16x32.And, andISlice[uint16]) // missing
+ testUint32x16Binary(t, simd.Uint32x16.And, andSlice[uint32])
+ testUint64x8Binary(t, simd.Uint64x8.And, andSlice[uint64])
+ }
+}
+
+func TestAndNot(t *testing.T) {
+ testInt16x16Binary(t, simd.Int16x16.AndNot, andNotSlice[int16])
+ testInt16x8Binary(t, simd.Int16x8.AndNot, andNotSlice[int16])
+ testInt32x4Binary(t, simd.Int32x4.AndNot, andNotSlice[int32])
+ testInt32x8Binary(t, simd.Int32x8.AndNot, andNotSlice[int32])
+ testInt64x2Binary(t, simd.Int64x2.AndNot, andNotSlice[int64])
+ testInt64x4Binary(t, simd.Int64x4.AndNot, andNotSlice[int64])
+ testInt8x16Binary(t, simd.Int8x16.AndNot, andNotSlice[int8])
+ testInt8x32Binary(t, simd.Int8x32.AndNot, andNotSlice[int8])
+
+ testUint16x16Binary(t, simd.Uint16x16.AndNot, andNotSlice[uint16])
+ testUint16x8Binary(t, simd.Uint16x8.AndNot, andNotSlice[uint16])
+ testUint32x4Binary(t, simd.Uint32x4.AndNot, andNotSlice[uint32])
+ testUint32x8Binary(t, simd.Uint32x8.AndNot, andNotSlice[uint32])
+ testUint64x2Binary(t, simd.Uint64x2.AndNot, andNotSlice[uint64])
+ testUint64x4Binary(t, simd.Uint64x4.AndNot, andNotSlice[uint64])
+ testUint8x16Binary(t, simd.Uint8x16.AndNot, andNotSlice[uint8])
+ testUint8x32Binary(t, simd.Uint8x32.AndNot, andNotSlice[uint8])
+
+ if simd.X86.AVX512() {
+ testInt8x64Binary(t, simd.Int8x64.AndNot, andNotSlice[int8])
+ testInt16x32Binary(t, simd.Int16x32.AndNot, andNotSlice[int16])
+ testInt32x16Binary(t, simd.Int32x16.AndNot, andNotSlice[int32])
+ testInt64x8Binary(t, simd.Int64x8.AndNot, andNotSlice[int64])
+ testUint8x64Binary(t, simd.Uint8x64.AndNot, andNotSlice[uint8])
+ testUint16x32Binary(t, simd.Uint16x32.AndNot, andNotSlice[uint16])
+ testUint32x16Binary(t, simd.Uint32x16.AndNot, andNotSlice[uint32])
+ testUint64x8Binary(t, simd.Uint64x8.AndNot, andNotSlice[uint64])
+ }
+}
+
+func TestXor(t *testing.T) {
+ testInt16x16Binary(t, simd.Int16x16.Xor, xorSlice[int16])
+ testInt16x8Binary(t, simd.Int16x8.Xor, xorSlice[int16])
+ testInt32x4Binary(t, simd.Int32x4.Xor, xorSlice[int32])
+ testInt32x8Binary(t, simd.Int32x8.Xor, xorSlice[int32])
+ testInt64x2Binary(t, simd.Int64x2.Xor, xorSlice[int64])
+ testInt64x4Binary(t, simd.Int64x4.Xor, xorSlice[int64])
+ testInt8x16Binary(t, simd.Int8x16.Xor, xorSlice[int8])
+ testInt8x32Binary(t, simd.Int8x32.Xor, xorSlice[int8])
+
+ testUint16x16Binary(t, simd.Uint16x16.Xor, xorSlice[uint16])
+ testUint16x8Binary(t, simd.Uint16x8.Xor, xorSlice[uint16])
+ testUint32x4Binary(t, simd.Uint32x4.Xor, xorSlice[uint32])
+ testUint32x8Binary(t, simd.Uint32x8.Xor, xorSlice[uint32])
+ testUint64x2Binary(t, simd.Uint64x2.Xor, xorSlice[uint64])
+ testUint64x4Binary(t, simd.Uint64x4.Xor, xorSlice[uint64])
+ testUint8x16Binary(t, simd.Uint8x16.Xor, xorSlice[uint8])
+ testUint8x32Binary(t, simd.Uint8x32.Xor, xorSlice[uint8])
+
+ if simd.X86.AVX512() {
+ // testInt8x64Binary(t, simd.Int8x64.Xor, andISlice[int8]) // missing
+ // testInt16x32Binary(t, simd.Int16x32.Xor, andISlice[int16]) // missing
+ testInt32x16Binary(t, simd.Int32x16.Xor, xorSlice[int32])
+ testInt64x8Binary(t, simd.Int64x8.Xor, xorSlice[int64])
+ // testUint8x64Binary(t, simd.Uint8x64.Xor, andISlice[uint8]) // missing
+ // testUint16x32Binary(t, simd.Uint16x32.Xor, andISlice[uint16]) // missing
+ testUint32x16Binary(t, simd.Uint32x16.Xor, xorSlice[uint32])
+ testUint64x8Binary(t, simd.Uint64x8.Xor, xorSlice[uint64])
+ }
+}
+
+func TestOr(t *testing.T) {
+ testInt16x16Binary(t, simd.Int16x16.Or, orSlice[int16])
+ testInt16x8Binary(t, simd.Int16x8.Or, orSlice[int16])
+ testInt32x4Binary(t, simd.Int32x4.Or, orSlice[int32])
+ testInt32x8Binary(t, simd.Int32x8.Or, orSlice[int32])
+ testInt64x2Binary(t, simd.Int64x2.Or, orSlice[int64])
+ testInt64x4Binary(t, simd.Int64x4.Or, orSlice[int64])
+ testInt8x16Binary(t, simd.Int8x16.Or, orSlice[int8])
+ testInt8x32Binary(t, simd.Int8x32.Or, orSlice[int8])
+
+ testUint16x16Binary(t, simd.Uint16x16.Or, orSlice[uint16])
+ testUint16x8Binary(t, simd.Uint16x8.Or, orSlice[uint16])
+ testUint32x4Binary(t, simd.Uint32x4.Or, orSlice[uint32])
+ testUint32x8Binary(t, simd.Uint32x8.Or, orSlice[uint32])
+ testUint64x2Binary(t, simd.Uint64x2.Or, orSlice[uint64])
+ testUint64x4Binary(t, simd.Uint64x4.Or, orSlice[uint64])
+ testUint8x16Binary(t, simd.Uint8x16.Or, orSlice[uint8])
+ testUint8x32Binary(t, simd.Uint8x32.Or, orSlice[uint8])
+
+ if simd.X86.AVX512() {
+ // testInt8x64Binary(t, simd.Int8x64.Or, andISlice[int8]) // missing
+ // testInt16x32Binary(t, simd.Int16x32.Or, andISlice[int16]) // missing
+ testInt32x16Binary(t, simd.Int32x16.Or, orSlice[int32])
+ testInt64x8Binary(t, simd.Int64x8.Or, orSlice[int64])
+ // testUint8x64Binary(t, simd.Uint8x64.Or, andISlice[uint8]) // missing
+ // testUint16x32Binary(t, simd.Uint16x32.Or, andISlice[uint16]) // missing
+ testUint32x16Binary(t, simd.Uint32x16.Or, orSlice[uint32])
+ testUint64x8Binary(t, simd.Uint64x8.Or, orSlice[uint64])
+ }
+}
+
+func TestMul(t *testing.T) {
+ testFloat32x4Binary(t, simd.Float32x4.Mul, mulSlice[float32])
+ testFloat32x8Binary(t, simd.Float32x8.Mul, mulSlice[float32])
+ testFloat64x2Binary(t, simd.Float64x2.Mul, mulSlice[float64])
+ testFloat64x4Binary(t, simd.Float64x4.Mul, mulSlice[float64])
+
+ testInt16x16Binary(t, simd.Int16x16.Mul, mulSlice[int16])
+ testInt16x8Binary(t, simd.Int16x8.Mul, mulSlice[int16])
+ testInt32x4Binary(t, simd.Int32x4.Mul, mulSlice[int32])
+ testInt32x8Binary(t, simd.Int32x8.Mul, mulSlice[int32])
+
+ // testInt8x16Binary(t, simd.Int8x16.Mul, mulSlice[int8]) // nope
+ // testInt8x32Binary(t, simd.Int8x32.Mul, mulSlice[int8])
+
+ // TODO we should be able to do these, there's no difference between signed/unsigned Mul
+ // testUint16x16Binary(t, simd.Uint16x16.Mul, mulSlice[uint16])
+ // testUint16x8Binary(t, simd.Uint16x8.Mul, mulSlice[uint16])
+ // testUint32x4Binary(t, simd.Uint32x4.Mul, mulSlice[uint32])
+ // testUint32x8Binary(t, simd.Uint32x8.Mul, mulSlice[uint32])
+ // testUint64x2Binary(t, simd.Uint64x2.Mul, mulSlice[uint64])
+ // testUint64x4Binary(t, simd.Uint64x4.Mul, mulSlice[uint64])
+
+ // testUint8x16Binary(t, simd.Uint8x16.Mul, mulSlice[uint8]) // nope
+ // testUint8x32Binary(t, simd.Uint8x32.Mul, mulSlice[uint8])
+
+ if simd.X86.AVX512() {
+ testInt64x2Binary(t, simd.Int64x2.Mul, mulSlice[int64]) // avx512 only
+ testInt64x4Binary(t, simd.Int64x4.Mul, mulSlice[int64])
+
+ testFloat32x16Binary(t, simd.Float32x16.Mul, mulSlice[float32])
+ testFloat64x8Binary(t, simd.Float64x8.Mul, mulSlice[float64])
+
+ // testInt8x64Binary(t, simd.Int8x64.Mul, mulSlice[int8]) // nope
+ testInt16x32Binary(t, simd.Int16x32.Mul, mulSlice[int16])
+ testInt32x16Binary(t, simd.Int32x16.Mul, mulSlice[int32])
+ testInt64x8Binary(t, simd.Int64x8.Mul, mulSlice[int64])
+ // testUint8x64Binary(t, simd.Uint8x64.Mul, mulSlice[uint8]) // nope
+
+ // TODO signed should do the job
+ // testUint16x32Binary(t, simd.Uint16x32.Mul, mulSlice[uint16])
+ // testUint32x16Binary(t, simd.Uint32x16.Mul, mulSlice[uint32])
+ // testUint64x8Binary(t, simd.Uint64x8.Mul, mulSlice[uint64])
+ }
+}
+
+func TestDiv(t *testing.T) {
+ testFloat32x4Binary(t, simd.Float32x4.Div, divSlice[float32])
+ testFloat32x8Binary(t, simd.Float32x8.Div, divSlice[float32])
+ testFloat64x2Binary(t, simd.Float64x2.Div, divSlice[float64])
+ testFloat64x4Binary(t, simd.Float64x4.Div, divSlice[float64])
+
+ if simd.X86.AVX512() {
+ testFloat32x16Binary(t, simd.Float32x16.Div, divSlice[float32])
+ testFloat64x8Binary(t, simd.Float64x8.Div, divSlice[float64])
+ }
+}