/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/src/q8avgpool/ |
H A D | mp8x9p8q-neon.c | 103 const size_t address_increment = k - 8; in pytorch_q8avgpool_ukernel_mp8x9p8q__neon() local 207 const size_t address_increment = k - 8; in pytorch_q8avgpool_ukernel_mp8x9p8q__neon() local 356 const size_t address_increment = k - 8; in pytorch_q8avgpool_ukernel_mp8x9p8q__neon() local
|
H A D | up8x9-neon.c | 151 const size_t address_increment = k - 8; in pytorch_q8avgpool_ukernel_up8x9__neon() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/src/u8maxpool/ |
H A D | 16x9p8q-sse2.c | 108 const size_t address_increment = k - 16; in pytorch_u8maxpool_ukernel_16x9p8q__sse2() local 215 const size_t address_increment = k - 16; in pytorch_u8maxpool_ukernel_16x9p8q__sse2() local
|
H A D | 16x9p8q-neon.c | 105 const size_t address_increment = k - 16; in pytorch_u8maxpool_ukernel_16x9p8q__neon() local 212 const size_t address_increment = k - 16; in pytorch_u8maxpool_ukernel_16x9p8q__neon() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/src/x8zip/ |
H A D | x2-neon.c | 31 const size_t address_increment = n - 8; in pytorch_qnnp_x8zip_x2__neon() local
|
H A D | x3-neon.c | 33 const size_t address_increment = n - 8; in pytorch_qnnp_x8zip_x3__neon() local
|
H A D | x4-neon.c | 36 const size_t address_increment = n - 8; in pytorch_qnnp_x8zip_x4__neon() local
|
H A D | x2-sse2.c | 32 const size_t address_increment = n - 16; in pytorch_qnnp_x8zip_x2__sse2() local
|
H A D | x4-sse2.c | 46 const size_t address_increment = n - 16; in pytorch_qnnp_x8zip_x4__sse2() local
|
H A D | xm-neon.c | 102 const size_t address_increment = k - 8; in pytorch_qnnp_x8zip_xm__neon() local
|
H A D | x3-sse2.c | 109 const size_t address_increment = n - 16; in pytorch_qnnp_x8zip_x3__sse2() local
|
/aosp_15_r20/external/XNNPACK/src/x8-zip/ |
H A D | x2-neon.c | 32 const size_t address_increment = n - 8; in xnn_x8_zip_x2_ukernel__neon() local
|
H A D | x3-neon.c | 34 const size_t address_increment = n - 8; in xnn_x8_zip_x3_ukernel__neon() local
|
H A D | x4-neon.c | 36 const size_t address_increment = n - 8; in xnn_x8_zip_x4_ukernel__neon() local
|
H A D | x2-sse2.c | 37 const size_t address_increment = n - 16; in xnn_x8_zip_x2_ukernel__sse2() local
|
H A D | x4-sse2.c | 51 const size_t address_increment = n - 16; in xnn_x8_zip_x4_ukernel__sse2() local
|
H A D | xm-neon.c | 74 const size_t address_increment = k - 8; in xnn_x8_zip_xm_ukernel__neon() local
|
H A D | x3-sse2.c | 79 const size_t address_increment = n - 16; in xnn_x8_zip_x3_ukernel__sse2() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/src/q8gavgpool/ |
H A D | up8x7-neon.c | 130 const size_t address_increment = n - 8; in pytorch_q8gavgpool_ukernel_up8x7__neon() local
|
H A D | mp8x7p7q-neon.c | 228 const size_t address_increment = n - 8; in pytorch_q8gavgpool_ukernel_mp8x7p7q__neon() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/ |
H A D | Normalization.cuh | 975 int address_increment = inner_loop_stride * stride; in batch_norm_collect_statistics_channels_last_kernel() local 1118 int address_increment = inner_loop_stride * stride; in batch_norm_transform_input_channels_last_kernel() local 1205 int address_increment = inner_loop_stride * stride; in batch_norm_backward_reduce_channels_last_kernel() local 1356 int address_increment = inner_loop_stride * stride; in batch_norm_backward_elemt_channels_last_kernel_impl() local
|
/aosp_15_r20/external/XNNPACK/src/amalgam/ |
H A D | sse2.c | 13960 const size_t address_increment = n - 16; in xnn_x8_zip_x2_ukernel__sse2() local 14045 const size_t address_increment = n - 16; in xnn_x8_zip_x3_ukernel__sse2() local 14142 const size_t address_increment = n - 16; in xnn_x8_zip_x4_ukernel__sse2() local
|