Lines Matching full:outputs

34     std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize);  in TEST()  local
36 …xnn_math_f32_expminus__neonfma_rr2_lut64_p2(kBlockSize * sizeof(float), inputs.data(), outputs.dat… in TEST()
38 ASSERT_EQ(reference_output, outputs[0]) in TEST()
41 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
48 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
50 …xnn_math_f32_expminus__neonfma_rr2_lut64_p2(kBlockSize * sizeof(float), inputs.data(), outputs.dat… in TEST()
52 ASSERT_EQ(reference_output, outputs[0]) in TEST()
55 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
62 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
67 …math_f32_expminus__neonfma_rr2_lut64_p2(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
70 ASSERT_EQ(reference_output, float_as_uint32(outputs[i])) in TEST()
73 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
82 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
87 …math_f32_expminus__neonfma_rr2_lut64_p2(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
89 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
91 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
100 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
105 …math_f32_expminus__neonfma_rr2_lut64_p2(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
107 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
109 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
121 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
123 …th_f32_expminus__neonfma_rr2_lut2048_p1(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
125 ASSERT_EQ(reference_output, outputs[0]) in TEST()
128 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
135 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
137 …th_f32_expminus__neonfma_rr2_lut2048_p1(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
139 ASSERT_EQ(reference_output, outputs[0]) in TEST()
142 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
149 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
154 …th_f32_expminus__neonfma_rr2_lut2048_p1(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
157 ASSERT_EQ(reference_output, float_as_uint32(outputs[i])) in TEST()
160 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
169 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
174 …th_f32_expminus__neonfma_rr2_lut2048_p1(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
176 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
178 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
187 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
192 …th_f32_expminus__neonfma_rr2_lut2048_p1(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
194 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
196 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
208 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
210 … xnn_math_f32_expminus__neonfma_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
212 ASSERT_EQ(reference_output, outputs[0]) in TEST()
215 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
222 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
224 … xnn_math_f32_expminus__neonfma_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
226 ASSERT_EQ(reference_output, outputs[0]) in TEST()
229 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
236 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
241 … xnn_math_f32_expminus__neonfma_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
244 ASSERT_EQ(reference_output, float_as_uint32(outputs[i])) in TEST()
247 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
256 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
261 … xnn_math_f32_expminus__neonfma_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
263 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
265 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
274 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
279 … xnn_math_f32_expminus__neonfma_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
281 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
283 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
295 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
297 xnn_math_f32_expminus__avx2_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
299 ASSERT_EQ(reference_output, outputs[0]) in TEST()
302 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
309 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
311 xnn_math_f32_expminus__avx2_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
313 ASSERT_EQ(reference_output, outputs[0]) in TEST()
316 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
323 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
328 xnn_math_f32_expminus__avx2_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
331 ASSERT_EQ(reference_output, float_as_uint32(outputs[i])) in TEST()
334 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
343 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
348 xnn_math_f32_expminus__avx2_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
350 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
352 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
361 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
366 xnn_math_f32_expminus__avx2_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
368 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
370 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
380 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
382 xnn_math_f32_expminus__sse2_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
384 ASSERT_EQ(reference_output, outputs[0]) in TEST()
387 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
392 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
394 xnn_math_f32_expminus__sse2_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
396 ASSERT_EQ(reference_output, outputs[0]) in TEST()
399 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
404 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
409 xnn_math_f32_expminus__sse2_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
412 ASSERT_EQ(reference_output, float_as_uint32(outputs[i])) in TEST()
415 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
422 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
427 xnn_math_f32_expminus__sse2_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
429 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
431 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
438 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
443 xnn_math_f32_expminus__sse2_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
445 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
447 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
456 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
458 …xnn_math_f32_expminus__scalar_rr2_lut64_p2(kBlockSize * sizeof(float), inputs.data(), outputs.data… in TEST()
460 ASSERT_EQ(reference_output, outputs[0]) in TEST()
463 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
468 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
470 …xnn_math_f32_expminus__scalar_rr2_lut64_p2(kBlockSize * sizeof(float), inputs.data(), outputs.data… in TEST()
472 ASSERT_EQ(reference_output, outputs[0]) in TEST()
475 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
480 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
485 …xnn_math_f32_expminus__scalar_rr2_lut64_p2(kBlockSize * sizeof(float), inputs.data(), outputs.data… in TEST()
488 ASSERT_EQ(reference_output, float_as_uint32(outputs[i])) in TEST()
491 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
498 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
503 …xnn_math_f32_expminus__scalar_rr2_lut64_p2(kBlockSize * sizeof(float), inputs.data(), outputs.data… in TEST()
505 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
507 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
514 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
519 …xnn_math_f32_expminus__scalar_rr2_lut64_p2(kBlockSize * sizeof(float), inputs.data(), outputs.data… in TEST()
521 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
523 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
531 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
533 …xnn_math_f32_expminus__scalar_rr2_lut2048_p1(kBlockSize * sizeof(float), inputs.data(), outputs.da… in TEST()
535 ASSERT_EQ(reference_output, outputs[0]) in TEST()
538 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
543 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
545 …xnn_math_f32_expminus__scalar_rr2_lut2048_p1(kBlockSize * sizeof(float), inputs.data(), outputs.da… in TEST()
547 ASSERT_EQ(reference_output, outputs[0]) in TEST()
550 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
555 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
560 …xnn_math_f32_expminus__scalar_rr2_lut2048_p1(kBlockSize * sizeof(float), inputs.data(), outputs.da… in TEST()
563 ASSERT_EQ(reference_output, float_as_uint32(outputs[i])) in TEST()
566 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
573 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
578 …xnn_math_f32_expminus__scalar_rr2_lut2048_p1(kBlockSize * sizeof(float), inputs.data(), outputs.da… in TEST()
580 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
582 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
589 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
594 …xnn_math_f32_expminus__scalar_rr2_lut2048_p1(kBlockSize * sizeof(float), inputs.data(), outputs.da… in TEST()
596 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
598 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
606 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
608 xnn_math_f32_expminus__scalar_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
610 ASSERT_EQ(reference_output, outputs[0]) in TEST()
613 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
618 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
620 xnn_math_f32_expminus__scalar_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
622 ASSERT_EQ(reference_output, outputs[0]) in TEST()
625 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[0]); in TEST()
630 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
635 xnn_math_f32_expminus__scalar_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
638 ASSERT_EQ(reference_output, float_as_uint32(outputs[i])) in TEST()
641 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
648 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
653 xnn_math_f32_expminus__scalar_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
655 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
657 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()
664 std::vector<float, AlignedAllocator<float, 64>> outputs(kBlockSize); in TEST() local
669 xnn_math_f32_expminus__scalar_rr2_p5(kBlockSize * sizeof(float), inputs.data(), outputs.data()); in TEST()
671 ASSERT_TRUE(std::isnan(outputs[i])) in TEST()
673 …", optimized = 0x" << std::hex << std::setw(8) << std::setfill('0') << float_as_uint32(outputs[i]); in TEST()