/aosp_15_r20/frameworks/av/media/module/codecs/m4v_h263/enc/src/ |
H A D | fastquant_inline.h | 41 __inline int32 coeff_quant(int32 coeff, int32 q_scale, int32 shift) in coeff_quant() 187 __inline int32 coeff_quant(int32 coeff, int32 q_scale, int32 shift) in coeff_quant() 282 __inline int32 coeff_quant(int32 coeff, int32 q_scale, int32 shift) in coeff_quant() 449 __inline int32 coeff_quant(int32 coeff, int32 q_scale, int32 shift) in coeff_quant()
|
/aosp_15_r20/external/libaom/aom_dsp/x86/ |
H A D | adaptive_quantize_avx2.c | 24 __m256i *shift) { in load_b_values_avx2() 75 const __m256i *shift) { in calculate_qcoeff_avx2() 108 __m256i zbin, round, quant, dequant, shift; in aom_quantize_b_adaptive_avx2() local
|
H A D | intrapred_avx2.c | 1086 __m256i b, res, shift; in highbd_dr_prediction_z1_4xN_internal_avx2() local 1155 __m256i b, res, shift; in highbd_dr_prediction_32bit_z1_4xN_internal_avx2() local 1242 __m256i b, res, res1, shift; in highbd_dr_prediction_32bit_z1_8xN_internal_avx2() local 1327 __m256i b, res, res1, shift; in highbd_dr_prediction_z1_8xN_internal_avx2() local 1435 __m256i shift = _mm256_srli_epi32( in highbd_dr_prediction_32bit_z1_16xN_internal_avx2() local 1515 __m256i shift = in highbd_dr_prediction_z1_16xN_internal_avx2() local 1591 __m256i shift = in highbd_dr_prediction_32bit_z1_32xN_internal_avx2() local 1688 __m256i shift = in highbd_dr_prediction_z1_32xN_internal_avx2() local 1784 __m256i shift = _mm256_srli_epi32( in highbd_dr_prediction_32bit_z1_64xN_avx2() local 1883 __m256i shift = in highbd_dr_prediction_z1_64xN_avx2() local [all …]
|
H A D | quantize_x86.h | 20 const int16_t *shift_ptr, __m128i *shift) { in load_b_values() 41 const __m128i quant, const __m128i shift) { in calculate_qcoeff() 52 const __m128i *shift, in calculate_qcoeff_log_scale()
|
/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/ |
H A D | roll_op.cc | 42 const Tensor& shift = context->input(1); in Compute() local 136 const int shift = dim_size[i] - threshold[i]; in DoRoll() local 223 const int shift = dim_size[i] - threshold[i]; in DoRollWithMemcpy() local
|
/aosp_15_r20/external/guava/android/guava/src/com/google/common/math/ |
H A D | BigIntegerMath.java | 284 int shift = (log2 - DoubleUtils.SIGNIFICAND_BITS) & ~1; // even! in sqrtFloor() local 402 int shift = Long.numberOfTrailingZeros(product); in factorial() local 403 product >>= shift; in factorial() local
|
/aosp_15_r20/external/guava/guava/src/com/google/common/math/ |
H A D | BigIntegerMath.java | 284 int shift = (log2 - DoubleUtils.SIGNIFICAND_BITS) & ~1; // even! in sqrtFloor() local 402 int shift = Long.numberOfTrailingZeros(product); in factorial() local 403 product >>= shift; in factorial() local
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/internal/optimized/integer_ops/ |
H A D | mean.h | 30 int32 multiplier, int32 shift, int32 bias, in MeanImpl() 172 int32 shift; member 214 int32 multiplier, shift; in Mean() local
|
/aosp_15_r20/external/libaom/av1/common/arm/ |
H A D | highbd_inv_txfm_neon.c | 166 static inline void round_shift_4x4(int32x4_t *in, int shift) { in round_shift_4x4() 176 static void round_shift_8x8(int32x4_t *in, int shift) { in round_shift_8x8() 746 int fliplr, int flipud, int shift, int bd) { in write_buffer_4x4() 832 const int8_t *shift = av1_inv_txfm_shift_ls[TX_4X4]; in av1_inv_txfm2d_add_4x4_neon() local 1380 int fliplr, int flipud, int shift, int bd) { in write_buffer_8x8() 1427 const int8_t *shift = av1_inv_txfm_shift_ls[TX_8X8]; in av1_inv_txfm2d_add_8x8_neon() local 4768 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in av1_inv_txfm2d_add_4x8_neon() local 4820 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in av1_inv_txfm2d_add_8x4_neon() local 4867 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in av1_inv_txfm2d_add_4x16_neon() local 4921 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in av1_inv_txfm2d_add_16x4_neon() local [all …]
|
H A D | av1_inv_txfm_neon.c | 3582 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in lowbd_inv_txfm2d_add_idtx_neon() local 3634 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in lowbd_inv_txfm2d_add_v_identity_neon() local 3701 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in lowbd_inv_txfm2d_add_h_identity_neon() local 3760 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in lowbd_inv_txfm2d_add_4x4_neon() local 3824 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in lowbd_inv_txfm2d_add_4x8_neon() local 3890 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in lowbd_inv_txfm2d_add_8x4_neon() local 3956 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in lowbd_inv_txfm2d_add_4x16_neon() local 4021 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in lowbd_inv_txfm2d_add_16x4_neon() local 4085 const int8_t *shift = av1_inv_txfm_shift_ls[tx_size]; in lowbd_inv_txfm2d_add_no_identity_neon() local
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
H A D | neon_tensor_utils.h | 107 const int8_t* input_to_gate_weights, int32_t multiplier, int32_t shift, in MatrixBatchVectorMultiplyAccumulate() 117 const int8_t* input_to_gate_weights, int32_t multiplier, int32_t shift, in MatrixBatchVectorMultiplyAccumulate() 197 int n_input, int shift, int16_t* output) { in CwiseMul() 202 int32_t multiplier, int shift, int n_batch, int n_input, in CwiseMul() 236 int shift, int16_t* result) { in VectorBatchVectorCwiseProductAccumulate()
|
H A D | sse_tensor_utils.h | 121 const int8_t* input_to_gate_weights, int32_t multiplier, int32_t shift, in MatrixBatchVectorMultiplyAccumulate() 131 const int8_t* input_to_gate_weights, int32_t multiplier, int32_t shift, in MatrixBatchVectorMultiplyAccumulate() 210 int n_input, int shift, int16_t* output) { in CwiseMul() 215 int32_t multiplier, int32_t shift, int32_t n_batch, in CwiseMul() 251 int shift, int16_t* result) { in VectorBatchVectorCwiseProductAccumulate()
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
H A D | portable_tensor_utils.h | 140 const int8_t* input_to_gate_weights, int32_t multiplier, int32_t shift, in MatrixBatchVectorMultiplyAccumulate() 150 const int8_t* input_to_gate_weights, int32_t multiplier, int32_t shift, in MatrixBatchVectorMultiplyAccumulate() 229 int n_input, int shift, int16_t* output) { in CwiseMul() 234 int32_t multiplier, int32_t shift, int32_t n_batch, in CwiseMul() 263 int shift, int16_t* result) { in VectorBatchVectorCwiseProductAccumulate()
|
/aosp_15_r20/external/compiler-rt/lib/builtins/ |
H A D | floatunsisf.c | 37 const int shift = significandBits - exponent; in ARM_EABI_FNALIAS() local 40 const int shift = exponent - significandBits; in ARM_EABI_FNALIAS() local
|
H A D | floatsisf.c | 45 const int shift = significandBits - exponent; in ARM_EABI_FNALIAS() local 48 const int shift = exponent - significandBits; in ARM_EABI_FNALIAS() local
|
/aosp_15_r20/external/ComputeLibrary/src/core/NEON/kernels/detail/ |
H A D | NEColorConvertHelper.inl | 420 constexpr auto shift = yuyv ? 0 : 1; local 466 constexpr auto shift = uv ? 0 : 1; local 612 constexpr auto shift = yuyv ? 0 : 1; local 715 constexpr auto shift = uv ? 0 : 1; local 764 constexpr auto shift = yuyv ? 0 : 1; local 824 constexpr auto shift = uv ? 0 : 1; local
|
/aosp_15_r20/external/libopus/silk/mips/ |
H A D | sigproc_fix_mipsr1.h | 42 static inline int silk_LSHIFT_SAT32(int a, int shift) in silk_LSHIFT_SAT32() 52 static inline int silk_RSHIFT_ROUND(int a, int shift) in silk_RSHIFT_ROUND()
|
/aosp_15_r20/external/XNNPACK/test/ |
H A D | u64-u32-vsqrtshift.cc | 34 TEST(U64_U32_VSQRTSHIFT__SCALAR_CVTU32_SQRT_CVTU32F64_X1, shift) { in TEST() argument 35 for (uint32_t shift = 0; shift < 32; shift++) { in TEST() local
|
/aosp_15_r20/external/libaom/av1/encoder/ |
H A D | txb_rdopt_utils.h | 49 int shift, const qm_val_t *qmatrix, in get_coeff_dist() 197 int shift, tran_low_t *qc_low, in get_qc_dqc_low() 207 static inline void update_coeff_eob_fast(int *eob, int shift, in update_coeff_eob_fast()
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | algebraic_simplifier.cc | 1744 if (std::unique_ptr<HloInstruction> shift = in HandleDivide() local 1750 if (std::unique_ptr<HloInstruction> shift = in HandleDivide() local 1756 if (std::unique_ptr<HloInstruction> shift = in HandleDivide() local 1762 if (std::unique_ptr<HloInstruction> shift = in HandleDivide() local 1768 if (std::unique_ptr<HloInstruction> shift = in HandleDivide() local 1774 if (std::unique_ptr<HloInstruction> shift = in HandleDivide() local 1780 if (std::unique_ptr<HloInstruction> shift = in HandleDivide() local 1786 if (std::unique_ptr<HloInstruction> shift = in HandleDivide() local 4135 if (std::unique_ptr<HloInstruction> shift = in HandleRemainder() local 4141 if (std::unique_ptr<HloInstruction> shift = TryRemainderToAnd<int16_t>( in HandleRemainder() local [all …]
|
/aosp_15_r20/external/vixl/src/aarch64/ |
H A D | logic-aarch64.cc | 1489 int shift) { in shl() 1500 int shift) { in sshll() 1512 int shift) { in sshll2() 1524 int shift = LaneSizeInBitsFromFormat(vform) / 2; in shll() local 1532 int shift = LaneSizeInBitsFromFormat(vform) / 2; in shll2() local 1540 int shift) { in ushll() 1552 int shift) { in ushll2() 1653 int shift) { in sli() 1670 int shift) { in sqshl() 1681 int shift) { in uqshl() [all …]
|
/aosp_15_r20/external/coreboot/src/device/ |
H A D | i2c.c | 7 uint8_t mask, uint8_t shift) in i2c_read_field() 21 uint8_t mask, uint8_t shift) in i2c_write_field()
|
/aosp_15_r20/bionic/linker/ |
H A D | linker_sleb128.h | 49 size_t shift = 0; in pop_front() local 80 size_t shift = 0; in pop_front() local
|
/aosp_15_r20/prebuilts/rust/linux-x86/1.81.0.u1/lib/rustlib/src/rust/library/core/src/num/dec2flt/ |
H A D | slow.rs | 51 let shift = get_shift(n); in parse_long_mantissa() localVariable 60 let shift = if d.decimal_point == 0 { in parse_long_mantissa() localVariable
|
/aosp_15_r20/prebuilts/rust/linux-x86/1.81.0/lib/rustlib/src/rust/library/core/src/num/dec2flt/ |
H A D | slow.rs | 51 let shift = get_shift(n); in parse_long_mantissa() localVariable 60 let shift = if d.decimal_point == 0 { in parse_long_mantissa() localVariable
|