/aosp_15_r20/external/XNNPACK/src/f32-raddextexp/gen/ |
H A D | avx2-p5-x72.c | 192 …const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(vdelta_… in xnn_f32_raddextexp_ukernel__avx2_p5_x72() local
|
H A D | avx2-p5-x80.c | 203 …const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(vdelta_… in xnn_f32_raddextexp_ukernel__avx2_p5_x80() local
|
/aosp_15_r20/external/XNNPACK/src/f32-vscaleextexp/gen/ |
H A D | avx2-p5-x72.c | 196 …const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve8, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x72() local
|
H A D | avx2-p5-x80.c | 208 …const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve8, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x80() local
|
/aosp_15_r20/external/XNNPACK/src/f32-vscaleexpminusmax/gen/ |
H A D | avx2-p5-x72.c | 90 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x72() local
|
H A D | avx2-p5-x80.c | 93 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80() local
|
H A D | avx2-p5-x88.c | 96 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88() local
|
/aosp_15_r20/external/XNNPACK/src/f32-raddexpminusmax/gen/ |
H A D | avx2-p5-x72.c | 88 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72() local
|
H A D | avx2-p5-x72-acc3.c | 90 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72_acc3() local
|
H A D | avx2-p5-x80-acc5.c | 95 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5() local
|
H A D | avx2-p5-x80.c | 91 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80() local
|
H A D | avx2-p5-x80-acc2.c | 92 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc2() local
|
/aosp_15_r20/external/XNNPACK/src/f16-raddstoreexpminusmax/gen/ |
H A D | avx2-rr1-p2-x72-acc3.c | 81 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f16_raddstoreexpminusmax_ukernel__avx2_rr1_p2_x72_acc3() local
|
H A D | avx2-rr1-p2-x72.c | 79 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f16_raddstoreexpminusmax_ukernel__avx2_rr1_p2_x72() local
|
H A D | avx2-rr1-p2-x80-acc2.c | 83 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f16_raddstoreexpminusmax_ukernel__avx2_rr1_p2_x80_acc2() local
|
H A D | avx2-rr1-p2-x80-acc5.c | 86 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f16_raddstoreexpminusmax_ukernel__avx2_rr1_p2_x80_acc5() local
|
H A D | neonfp16arith-rr2-p2-x72.c | 79 const float16x8_t vs8 = vreinterpretq_f16_s16(vshlq_n_s16(vreinterpretq_s16_f16(vn8), 10)); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x72() local
|
H A D | neonfp16arith-rr2-p2-x72-acc3.c | 81 const float16x8_t vs8 = vreinterpretq_f16_s16(vshlq_n_s16(vreinterpretq_s16_f16(vn8), 10)); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x72_acc3() local
|
H A D | avx2-rr1-p2-x80.c | 82 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f16_raddstoreexpminusmax_ukernel__avx2_rr1_p2_x80() local
|
/aosp_15_r20/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
H A D | avx2-rr1-p5-x72.c | 79 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72() local
|
H A D | avx2-rr1-p5-x72-acc3.c | 81 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72_acc3() local
|
H A D | avx2-rr1-p5-x80.c | 82 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80() local
|
H A D | avx2-rr1-p5-x80-acc2.c | 83 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc2() local
|
H A D | avx2-rr1-p5-x80-acc5.c | 86 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc5() local
|
/aosp_15_r20/external/XNNPACK/src/f32-velu/gen/ |
H A D | velu-avx2-rr1-p6-x72.c | 87 __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72() local
|