/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/ |
H A D | transpose_conv.cc | 353 TfLiteTensor* scratch_buffer; in Prepare() local 356 &scratch_buffer)); in Prepare() 358 scratch_buffer->type = kTfLiteInt64; in Prepare() 360 scratch_buffer->type = kTfLiteInt32; in Prepare() 363 scratch_buffer->allocation_type = kTfLiteDynamic; in Prepare() 365 SetTensorToDynamic(scratch_buffer); in Prepare() 368 ResizeTensor(context, output_shape, scratch_buffer)); in Prepare() 440 TfLiteTensor* output, TfLiteTensor* scratch_buffer) { in EvalQuantized() argument 469 GetTensorData<int32_t>(scratch_buffer)); in EvalQuantized() 479 GetTensorData<int32>(col2im), GetTensorData<int32>(scratch_buffer), in EvalQuantized() [all …]
|
H A D | batch_matmul.cc | 157 TfLiteTensor* scratch_buffer; in InitializeTemporaries() local 159 context, GetTemporarySafe(context, node, /*index=*/0, &scratch_buffer)); in InitializeTemporaries() 168 scratch_buffer->type = op_context->lhs->type; in InitializeTemporaries() 169 scratch_buffer->allocation_type = kTfLiteArenaRw; in InitializeTemporaries() 170 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, scratch_buffer, in InitializeTemporaries() 180 TfLiteTensor* scratch_buffer; in InitializeTemporaries() local 182 context, GetTemporarySafe(context, node, /*index=*/1, &scratch_buffer)); in InitializeTemporaries() 183 scratch_buffer->name = "BatchMatMul_scratch_buffer"; in InitializeTemporaries() 195 scratch_buffer->allocation_type = kTfLiteArenaRwPersistent; in InitializeTemporaries() 197 scratch_buffer->allocation_type = kTfLiteArenaRw; in InitializeTemporaries() [all …]
|
/aosp_15_r20/external/libgav1/src/tile/ |
H A D | tile.cc | 601 TileScratchBuffer* const scratch_buffer) { in ProcessSuperBlockRow() argument 603 assert(scratch_buffer != nullptr); in ProcessSuperBlockRow() 607 if (!ProcessSuperBlock(row4x4, column4x4, scratch_buffer, in ProcessSuperBlockRow() 627 int row4x4, TileScratchBuffer* scratch_buffer); 630 int row4x4, TileScratchBuffer* scratch_buffer); 645 std::unique_ptr<TileScratchBuffer> scratch_buffer = in ParseAndDecode() local 647 if (scratch_buffer == nullptr) { in ParseAndDecode() 656 row4x4, scratch_buffer.get())) { in ParseAndDecode() 661 tile_scratch_buffer_pool_->Release(std::move(scratch_buffer)); in ParseAndDecode() 668 std::unique_ptr<TileScratchBuffer> scratch_buffer = in Parse() local [all …]
|
H A D | prediction.cc | 537 if (!block.scratch_buffer->cfl_luma_buffer_valid) { in ChromaFromLumaPrediction() 541 block.scratch_buffer->cfl_luma_buffer, in ChromaFromLumaPrediction() 546 block.scratch_buffer->cfl_luma_buffer_valid = true; in ChromaFromLumaPrediction() 553 buffer_[plane].columns(), block.scratch_buffer->cfl_luma_buffer, in ChromaFromLumaPrediction() 618 prediction[0] = block.scratch_buffer->prediction_buffer[0]; in CompoundInterPrediction() 619 prediction[1] = block.scratch_buffer->prediction_buffer[1]; in CompoundInterPrediction() 622 prediction[0] = block.scratch_buffer->compound_prediction_buffer_8bpp[0]; in CompoundInterPrediction() 623 prediction[1] = block.scratch_buffer->compound_prediction_buffer_8bpp[1]; in CompoundInterPrediction() 735 candidate_column, block.scratch_buffer->prediction_buffer[index], in InterPrediction() 772 block.scratch_buffer->prediction_buffer[0], in InterPrediction() [all …]
|
/aosp_15_r20/external/sdv/vsomeip/third_party/boost/numeric/ublas/include/boost/numeric/ublas/opencl/ |
D | misc.hpp | 23 compute::vector<T> scratch_buffer(v.size(), queue.get_context()); in a_sum() local 33 scratch_buffer.begin().get_buffer().get(), in a_sum() 46 scratch_buffer.begin().get_buffer().get(), in a_sum() 59 scratch_buffer.begin().get_buffer().get(), in a_sum() 72 scratch_buffer.begin().get_buffer().get(), in a_sum() 113 compute::vector<T> scratch_buffer(2*v.size(), queue.get_context()); in norm_2() local 123 scratch_buffer.begin().get_buffer().get(), in norm_2() 136 scratch_buffer.begin().get_buffer().get(), in norm_2() 149 scratch_buffer.begin().get_buffer().get(), in norm_2() 162 scratch_buffer.begin().get_buffer().get(), in norm_2()
|
/aosp_15_r20/external/libgav1/src/ |
H A D | tile_scratch_buffer.h | 147 std::unique_ptr<TileScratchBuffer> scratch_buffer(new (std::nothrow) in Get() 149 if (scratch_buffer == nullptr || !scratch_buffer->Init(bitdepth_)) { in Get() 152 return scratch_buffer; in Get() 157 void Release(std::unique_ptr<TileScratchBuffer> scratch_buffer) { in Release() argument 159 buffers_.Push(std::move(scratch_buffer)); in Release()
|
H A D | tile.h | 118 bool ProcessSuperBlockRow(int row4x4, TileScratchBuffer* scratch_buffer); 296 int row4x4_start, int column4x4_start, TileScratchBuffer* scratch_buffer, 299 TileScratchBuffer* scratch_buffer, 306 TileScratchBuffer* scratch_buffer); 310 TileScratchBuffer* scratch_buffer, ResidualPtr* residual); 312 void ClearBlockDecoded(TileScratchBuffer* scratch_buffer, int row4x4, 315 TileScratchBuffer* scratch_buffer, 788 TileScratchBuffer* const scratch_buffer, ResidualPtr* residual) in Block() 797 scratch_buffer(scratch_buffer), in Block() 938 TileScratchBuffer* const scratch_buffer; member [all …]
|
H A D | frame_scratch_buffer.h | 113 std::unique_ptr<FrameScratchBuffer> scratch_buffer(new (std::nothrow) in Get() 115 return scratch_buffer; in Get() 118 void Release(std::unique_ptr<FrameScratchBuffer> scratch_buffer) { in Release() argument 120 buffers_.Push(std::move(scratch_buffer)); in Release()
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/internal/reference/integer_ops/ |
H A D | transpose_conv.h | 33 int32_t* scratch_buffer) { in TransposeConv() argument 65 memset(scratch_buffer, 0, num_elements * sizeof(int32_t)); in TransposeConv() 90 scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv() 106 int32_t acc = scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv() 133 Scalar* scratch_buffer) { in TransposeConv() argument 163 memset(scratch_buffer, 0, num_elements * sizeof(Scalar)); in TransposeConv() 188 scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv() 204 Scalar acc = scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv()
|
/aosp_15_r20/external/mesa3d/src/gallium/drivers/radeonsi/ |
H A D | si_cp_dma.c | 207 if (!sctx->scratch_buffer || sctx->scratch_buffer->b.b.width0 < scratch_size) { in si_cp_dma_realign_engine() 208 si_resource_reference(&sctx->scratch_buffer, NULL); in si_cp_dma_realign_engine() 209 sctx->scratch_buffer = si_aligned_buffer_create(&sctx->screen->b, in si_cp_dma_realign_engine() 213 if (!sctx->scratch_buffer) in si_cp_dma_realign_engine() 219 si_cp_dma_prepare(sctx, &sctx->scratch_buffer->b.b, &sctx->scratch_buffer->b.b, size, size, in si_cp_dma_realign_engine() 222 va = sctx->scratch_buffer->gpu_address; in si_cp_dma_realign_engine()
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/V1_2/ |
D | layer_norm_lstm.mod.py | 86 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", variable 104 [scratch_buffer, output_state_out, cell_state_out, output]) 183 scratch_buffer: [0] * (n_batch * n_cell * 4), 258 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", variable 276 [scratch_buffer, output_state_out, cell_state_out, output]) 352 scratch_buffer: [0] * (n_batch * n_cell * 3),
|
D | lstm2_state_float16.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT16", "{%d, %d}" % (n_batch, n_cell * … variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 132 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 3) ],
|
D | lstm3_state2_float16.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT16", "{%d, %d}" % (n_batch, (n_cell *… variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 643 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|
D | lstm2_float16.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT16", "{%d, %d}" % (n_batch, n_cell * … variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 132 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 3) ],
|
/aosp_15_r20/external/ComputeLibrary/src/runtime/CL/functions/ |
H A D | CLLSTMLayer.cpp | 65 …ICLTensor *scratch_buffer, ICLTensor *output_state_out, ICLTensor *cell_state_out, ICLTensor *outp… in configure() argument 69 …e_bias, cell_bias, output_gate_bias, output_state_in, cell_state_in, scratch_buffer, output_state_… in configure() 78 …ICLTensor *scratch_buffer, ICLTensor *output_state_out, ICLTensor *cell_state_out, ICLTensor *outp… in configure() argument 86 scratch_buffer, output_state_out, cell_state_out, output); in configure() 89 …e_bias, cell_bias, output_gate_bias, output_state_in, cell_state_in, scratch_buffer, output_state_… in configure() 104 … scratch_buffer->info(), output_state_out->info(), cell_state_out->info(), output->info(), in configure() 393 _concat_scratch_buffer.configure(compile_context, scratch_inputs, scratch_buffer, Window::DimX); in configure() 405 …const ITensorInfo *scratch_buffer, const ITensorInfo *output_state_out, const ITensorInfo *cell_st… in validate() argument 413 scratch_buffer, output_state_out, cell_state_out, output); in validate() 422 … scratch_buffer, output_state_out, cell_state_out, output); in validate() [all …]
|
/aosp_15_r20/external/ComputeLibrary/src/runtime/NEON/functions/ |
H A D | NELSTMLayer.cpp | 62 … ITensor *scratch_buffer, ITensor *output_state_out, ITensor *cell_state_out, ITensor *output, in configure() argument 70 scratch_buffer, output_state_out, cell_state_out, output); in configure() 76 scratch_buffer, output_state_out, cell_state_out, output, in configure() 91 … scratch_buffer->info(), output_state_out->info(), cell_state_out->info(), output->info(), in configure() 368 _concat_scratch_buffer.configure(scratch_inputs, scratch_buffer, Window::DimX); in configure() 380 …const ITensorInfo *scratch_buffer, const ITensorInfo *output_state_out, const ITensorInfo *cell_st… in validate() argument 388 scratch_buffer, output_state_out, cell_state_out, output); in validate() 397 … scratch_buffer, output_state_out, cell_state_out, output); in validate() 412 ARM_COMPUTE_RETURN_ERROR_ON(scratch_buffer->num_dimensions() > 2); in validate() 416 ARM_COMPUTE_RETURN_ERROR_ON(cell_bias->dimension(0) * 4 != scratch_buffer->dimension(0) in validate() [all …]
|
/aosp_15_r20/hardware/google/pixel/thermal/virtualtemp_estimator/ |
D | virtualtemp_estimator_data.h | 80 scratch_buffer = nullptr; in VtEstimatorTFLiteData() 105 float *scratch_buffer; member 128 if (scratch_buffer) { in ~VtEstimatorTFLiteData() 129 delete scratch_buffer; in ~VtEstimatorTFLiteData() local
|
/aosp_15_r20/external/libxaac/decoder/drc_src/ |
H A D | impd_drc_main_td_process.c | 177 FLOAT32 *scratch_buffer = (FLOAT32 *)p_obj_drc->pp_mem[1]; in impd_process_time_domain() local 181 scratch_buffer + i * (p_obj_drc->str_config.frame_size + 32); in impd_process_time_domain() 190 FLOAT32 *scratch_buffer = (FLOAT32 *)p_obj_drc->pp_mem[1]; in impd_process_time_domain() local 194 scratch_buffer + i * (p_obj_drc->str_config.frame_size + 32); in impd_process_time_domain() 210 FLOAT32 *scratch_buffer = (FLOAT32 *)p_obj_drc->pp_mem[1]; in impd_process_time_domain() local 214 scratch_buffer + i * (p_obj_drc->str_config.frame_size + 32); in impd_process_time_domain()
|
/aosp_15_r20/external/mesa3d/src/imagination/vulkan/ |
H A D | pvr_spm.c | 127 struct pvr_spm_scratch_buffer *scratch_buffer; in pvr_spm_scratch_buffer_alloc() local 142 scratch_buffer = vk_alloc(&device->vk.alloc, in pvr_spm_scratch_buffer_alloc() 143 sizeof(*scratch_buffer), in pvr_spm_scratch_buffer_alloc() 146 if (!scratch_buffer) { in pvr_spm_scratch_buffer_alloc() 152 *scratch_buffer = (struct pvr_spm_scratch_buffer){ in pvr_spm_scratch_buffer_alloc() 157 *buffer_out = scratch_buffer; in pvr_spm_scratch_buffer_alloc() 634 framebuffer->scratch_buffer->bo->vma->dev_addr; in pvr_spm_init_eot_state() 918 framebuffer->scratch_buffer->bo->vma->dev_addr; in pvr_spm_init_bgobj_state()
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/V1_0/ |
D | lstm2_state.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, n_cell * … variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 132 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 3) ],
|
D | lstm_state.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell *… variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 140 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|
D | lstm3_state.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell *… variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 643 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
H A D | transpose_conv.h | 122 uint8_t* im2col_data, int32_t* scratch_buffer) { in TransposeConv() argument 157 memset(scratch_buffer, 0, num_elements * sizeof(int32_t)); in TransposeConv() 182 scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv() 198 int32_t acc = scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv()
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/V1_1/ |
D | lstm2_state2_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, n_cell * … variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 133 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 3) ],
|
D | lstm_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell *… variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 141 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|