/aosp_15_r20/external/XNNPACK/src/qs8-igemm/gen/ |
H A D | 4x16c8-minmax-rndnu-neon-mull.c | 333 const int32x4_t vsum3x89 = vpaddq_s32(vacc3x8, vacc3x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mull() local 457 const int32x2_t vsum3x89 = vpadd_s32(vpsum3x8, vpsum3x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mull() local
|
H A D | 4x16c16-minmax-rndnu-neon-mlal.c | 397 const int32x4_t vsum3x89 = vpaddq_s32(vacc3x8, vacc3x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c16__neon_mlal() local 520 const int32x2_t vsum3x89 = vpadd_s32(vpsum3x8, vpsum3x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c16__neon_mlal() local
|
H A D | 4x16c8-minmax-rndnu-neon-mlal.c | 572 const int32x4_t vsum3x89 = vpaddq_s32(vacc3x8, vacc3x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() local 696 const int32x2_t vsum3x89 = vpadd_s32(vpsum3x8, vpsum3x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() local
|
H A D | 4x16c4s2-minmax-rndnu-neon-mull.c | 335 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4s2__neon_mull() local
|
H A D | 4x16c4-minmax-rndnu-neon-mull-ld1r.c | 431 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld1r() local
|
H A D | 4x16c4-minmax-rndnu-neon-mull-dup.c | 427 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_dup() local
|
H A D | 4x16c4-minmax-rndnu-neon-mull-ld2r.c | 427 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld2r() local
|
H A D | 4x16c4s2-minmax-rndnu-neon-mlal.c | 580 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4s2__neon_mlal() local
|
H A D | 4x16c4-minmax-rndnu-neon-mlal-dup.c | 683 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_dup() local
|
H A D | 4x16c4-minmax-rndnu-neon-mlal-ld1r.c | 695 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r() local
|
H A D | 4x16c4-minmax-rndnu-neon-mlal-ld2r.c | 683 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld2r() local
|
/aosp_15_r20/external/XNNPACK/src/qs8-gemm/gen/ |
H A D | 4x16c8-minmax-rndnu-neon-mull.c | 313 const int32x4_t vsum3x89 = vpaddq_s32(vacc3x8, vacc3x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mull() local 437 const int32x2_t vsum3x89 = vpadd_s32(vpsum3x8, vpsum3x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mull() local
|
H A D | 4x16c16-minmax-rndnu-neon-mlal.c | 377 const int32x4_t vsum3x89 = vpaddq_s32(vacc3x8, vacc3x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c16__neon_mlal() local 500 const int32x2_t vsum3x89 = vpadd_s32(vpsum3x8, vpsum3x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c16__neon_mlal() local
|
H A D | 4x16c8-minmax-rndnu-neon-mlal.c | 552 const int32x4_t vsum3x89 = vpaddq_s32(vacc3x8, vacc3x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() local 676 const int32x2_t vsum3x89 = vpadd_s32(vpsum3x8, vpsum3x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() local
|
H A D | 4x16c4s2-minmax-rndnu-neon-mull.c | 315 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4s2__neon_mull() local
|
H A D | 4x16c4-minmax-rndnu-neon-mull-dup.c | 408 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mull_dup() local
|
H A D | 4x16c4-minmax-rndnu-neon-mull-ld1r.c | 412 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld1r() local
|
H A D | 4x16c4-minmax-rndnu-neon-mull-ld2r.c | 408 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld2r() local
|
H A D | 4x16c4s2-minmax-rndnu-neon-mlal.c | 560 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4s2__neon_mlal() local
|
H A D | 4x16c4-minmax-rndnu-neon-mlal-ld1r.c | 676 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r() local
|
H A D | 4x16c4-minmax-rndnu-neon-mlal-dup.c | 664 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_dup() local
|
H A D | 4x16c4-minmax-rndnu-neon-mlal-ld2r.c | 664 const int32x2_t vsum3x89 = vpadd_s32(vget_low_s32(vacc3x89), vget_high_s32(vacc3x89)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld2r() local
|