|
|
|
|
|
|
|
|
|
|
|
#include <algorithm> |
|
#include <cmath> |
|
#include <cstddef> |
|
#include <cstdint> |
|
#include <cstdlib> |
|
#include <vector> |
|
|
|
#include <gtest/gtest.h> |
|
|
|
#include <xnnpack/aligned-allocator.h> |
|
#include <xnnpack/common.h> |
|
#include <xnnpack/math-stubs.h> |
|
|
|
|
|
constexpr int kBlockSize = 1024; |
|
|
|
|
|
namespace { |
|
|
|
uint64_t Sqrt(uint64_t n) { |
|
if (n == 0) { |
|
return n; |
|
} |
|
|
|
uint64_t x0 = n >> 1; |
|
uint64_t x1 = (x0 + n / x0) >> 1; |
|
do { |
|
x0 = x1; |
|
x1 = (x0 + n / x0) >> 1; |
|
} while (x1 < x0); |
|
|
|
|
|
if (int64_t(x0 * x0 + x0 - n) < 0) { |
|
x0 += 1; |
|
} |
|
return x0; |
|
} |
|
|
|
} |
|
|
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_CVTSATU32F64, min_mantissa_exact_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x0010000000000000) << s; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_cvtsatu32f64(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_CVTSATU32F64, min_mantissa_min_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x001FFFFFFFFFFFFF) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_cvtsatu32f64(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_CVTSATU32F64, min_mantissa_max_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x0020000000000001) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_cvtsatu32f64(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_CVTSATU32F64, max_mantissa_exact_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x001FFFFFFFFFFFFF) << s; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_cvtsatu32f64(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_CVTSATU32F64, max_mantissa_min_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x003FFFFFFFFFFFFD) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_cvtsatu32f64(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_CVTSATU32F64, max_mantissa_max_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x003FFFFFFFFFFFFF) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_cvtsatu32f64(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_CVTSATU32F64, largest_inputs) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint64_t i = 0; i < kBlockSize; i++) { |
|
inputs[i] = -i; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_cvtsatu32f64(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_CVTSATU32F64, double_rounding) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint64_t n = UINT64_C(33554432); n <= UINT64_C(4294967295); n += kBlockSize) { |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t t = std::min(n + uint64_t(i), UINT64_C(4294967295)); |
|
inputs[i] = t * t + t; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_cvtsatu32f64(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
} |
|
|
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_LLRINT, min_mantissa_exact_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x0010000000000000) << s; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_LLRINT, min_mantissa_min_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x001FFFFFFFFFFFFF) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_LLRINT, min_mantissa_max_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x0020000000000001) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_LLRINT, max_mantissa_exact_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x001FFFFFFFFFFFFF) << s; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_LLRINT, max_mantissa_min_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x003FFFFFFFFFFFFD) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_LLRINT, max_mantissa_max_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x003FFFFFFFFFFFFF) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_LLRINT, largest_inputs) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint64_t i = 0; i < kBlockSize; i++) { |
|
inputs[i] = -i; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU32_SQRT_LLRINT, double_rounding) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint64_t n = UINT64_C(33554432); n <= UINT64_C(4294967295); n += kBlockSize) { |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t t = std::min(n + uint64_t(i), UINT64_C(4294967295)); |
|
inputs[i] = t * t + t; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu32_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
} |
|
|
|
|
|
TEST(SQRT__SCALAR_CVTU64_SQRT_LLRINT, min_mantissa_exact_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x0010000000000000) << s; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu64_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU64_SQRT_LLRINT, min_mantissa_min_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x001FFFFFFFFFFFFF) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu64_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU64_SQRT_LLRINT, min_mantissa_max_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x0020000000000001) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu64_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU64_SQRT_LLRINT, max_mantissa_exact_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x001FFFFFFFFFFFFF) << s; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu64_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU64_SQRT_LLRINT, max_mantissa_min_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x003FFFFFFFFFFFFD) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu64_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU64_SQRT_LLRINT, max_mantissa_max_input) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint32_t s = std::min<uint32_t>(i + 1, 11); |
|
inputs[i] = UINT64_C(0x003FFFFFFFFFFFFF) << (s - 1); |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu64_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU64_SQRT_LLRINT, largest_inputs) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint64_t i = 0; i < kBlockSize; i++) { |
|
inputs[i] = -i; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu64_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
|
|
TEST(SQRT__SCALAR_CVTU64_SQRT_LLRINT, double_rounding) { |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> inputs(kBlockSize); |
|
std::vector<uint64_t, AlignedAllocator<uint64_t, 64>> outputs(kBlockSize); |
|
for (uint64_t n = UINT64_C(33554432); n <= UINT64_C(4294967295); n += kBlockSize) { |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t t = std::min(n + uint64_t(i), UINT64_C(4294967295)); |
|
inputs[i] = t * t + t; |
|
} |
|
xnn_math_u64_sqrt__scalar_cvtu64_sqrt_llrint(kBlockSize * sizeof(uint64_t), inputs.data(), outputs.data()); |
|
for (uint32_t i = 0; i < kBlockSize; i++) { |
|
const uint64_t input = inputs[i]; |
|
const uint64_t output = outputs[i]; |
|
const uint64_t reference_output = Sqrt(input); |
|
ASSERT_EQ(output, reference_output) << "input: " << input; |
|
} |
|
} |
|
} |
|
|