/aosp_15_r20/external/XNNPACK/src/f32-raddexpminusmax/gen/ |
H A D | avx2-p5-x72.c | 82 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72() local
|
H A D | avx2-p5-x72-acc3.c | 84 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72_acc3() local
|
H A D | avx2-p5-x80-acc5.c | 89 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5() local
|
H A D | avx2-p5-x80.c | 85 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80() local
|
/aosp_15_r20/external/XNNPACK/src/f16-raddstoreexpminusmax/gen/ |
H A D | neonfp16arith-rr2-p2-x64.c | 70 const float16x8_t vs2 = vreinterpretq_f16_s16(vshlq_n_s16(vreinterpretq_s16_f16(vn2), 10)); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x64() local
|
H A D | avx2-rr1-p2-x72-acc3.c | 75 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f16_raddstoreexpminusmax_ukernel__avx2_rr1_p2_x72_acc3() local
|
H A D | avx2-rr1-p2-x72.c | 73 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f16_raddstoreexpminusmax_ukernel__avx2_rr1_p2_x72() local
|
/aosp_15_r20/external/XNNPACK/src/f16-vsigmoid/gen/ |
H A D | vsigmoid-neonfp16arith-rr2-p2-nr1fma-x40.c | 58 const float16x8_t vs2 = vreinterpretq_f16_s16(vshlq_n_s16(vreinterpretq_s16_f16(vn2), 10)); in xnn_f16_vsigmoid_ukernel__neonfp16arith_rr2_p2_nr1fma_x40() local
|
H A D | vsigmoid-avx2-rr1-p2-rcp-x48.c | 63 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f16_vsigmoid_ukernel__avx2_rr1_p2_rcp_x48() local
|
H A D | vsigmoid-neonfp16arith-rr2-p2-nr1recps-x40.c | 58 const float16x8_t vs2 = vreinterpretq_f16_s16(vshlq_n_s16(vreinterpretq_s16_f16(vn2), 10)); in xnn_f16_vsigmoid_ukernel__neonfp16arith_rr2_p2_nr1recps_x40() local
|
H A D | vsigmoid-avx2-rr1-p2-div-x64.c | 69 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f16_vsigmoid_ukernel__avx2_rr1_p2_div_x64() local
|
/aosp_15_r20/external/XNNPACK/src/f32-vsigmoid/gen/ |
H A D | vsigmoid-avx2-rr1-p5-nr2fma-x48.c | 63 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x48() local
|
H A D | vsigmoid-avx2-rr1-p5-nr2fma-x56.c | 66 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x56() local
|
H A D | vsigmoid-avx2-rr1-p5-nr1fma-x56.c | 66 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x56() local
|
H A D | vsigmoid-avx-rr2-p5-nr2-x40.c | 68 const __m256 vs2 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs_lo2), vs_hi2, 1); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_nr2_x40() local
|
/aosp_15_r20/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
H A D | avx2-rr1-p5-x72.c | 73 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72() local
|
H A D | avx2-rr1-p5-x72-acc3.c | 75 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72_acc3() local
|
H A D | avx2-rr1-p5-x80.c | 76 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80() local
|
/aosp_15_r20/external/deqp/external/openglcts/modules/common/ |
H A D | glcUniformBlockTests.cpp | 601 std::string vs2("layout (std140) uniform Data { highp int x; mediump int y; } myData;\n" in iterate() local 712 std::string vs2("precision highp float;\n" in iterate() local
|
/aosp_15_r20/external/XNNPACK/src/f32-raddextexp/gen/ |
H A D | avx2-p5-x80.c | 197 …const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(vdelta_… in xnn_f32_raddextexp_ukernel__avx2_p5_x80() local
|
/aosp_15_r20/external/XNNPACK/src/f32-velu/gen/ |
H A D | velu-avx2-rr1-lut4-p4-perm-x56.c | 90 __m256 vs2 = _mm256_castsi256_ps(_mm256_add_epi32(vl2, ven2)); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x56() local
|
H A D | velu-avx2-rr1-lut16-p3-gather-x48.c | 92 __m256 vs2 = _mm256_castsi256_ps(_mm256_add_epi32(vl2, ven2)); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x48() local
|
H A D | velu-avx2-rr1-p6-x72.c | 75 __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72() local
|
/aosp_15_r20/external/XNNPACK/src/f32-vscaleexpminusmax/gen/ |
H A D | avx2-p5-x80.c | 87 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80() local
|
H A D | avx2-p5-x88.c | 90 const __m256 vs2 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn2), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88() local
|