/aosp_15_r20/external/ComputeLibrary/src/runtime/CL/functions/ |
H A D | CLLSTMLayerQuantized.cpp | 61 …const ICLTensor *input_to_input_weights, const ICLTensor *input_to_forget_weights, const ICLTensor… in configure() argument 67 …configure(CLKernelLibrary::get().get_compile_context(), input, input_to_input_weights, input_to_fo… in configure() 73 …const ICLTensor *input_to_input_weights, const ICLTensor *input_to_forget_weights, const ICLTensor… in configure() argument 79 …ARM_COMPUTE_ERROR_ON_NULLPTR(input, input_to_input_weights, input_to_forget_weights, input_to_cell… in configure() 83 …ARM_COMPUTE_LOG_PARAMS(input, input_to_input_weights, input_to_forget_weights, input_to_cell_weigh… in configure() 87 …ARM_COMPUTE_ERROR_THROW_ON(CLLSTMLayerQuantized::validate(input->info(), input_to_input_weights->i… in configure() 94 const int output_size = input_to_input_weights->info()->dimension(1); in configure() 96 …const QuantizationInfo qweights = input_to_input_weights->info()->quantization_info(); // Weights … in configure() 101 _input_to_input_weights = input_to_input_weights; in configure() 116 inputs_weights_vector.emplace_back(input_to_input_weights); in configure() [all …]
|
/aosp_15_r20/external/ComputeLibrary/src/runtime/NEON/functions/ |
H A D | NELSTMLayerQuantized.cpp | 61 …const ITensor *input_to_input_weights, const ITensor *input_to_forget_weights, const ITensor *inpu… in configure() argument 67 …ARM_COMPUTE_ERROR_ON_NULLPTR(input, input_to_input_weights, input_to_forget_weights, input_to_cell… in configure() 71 …ARM_COMPUTE_ERROR_THROW_ON(NELSTMLayerQuantized::validate(input->info(), input_to_input_weights->i… in configure() 76 …ARM_COMPUTE_LOG_PARAMS(input, input_to_input_weights, input_to_forget_weights, input_to_cell_weigh… in configure() 82 const int output_size = input_to_input_weights->info()->dimension(1); in configure() 84 …const QuantizationInfo qweights = input_to_input_weights->info()->quantization_info(); // Weights … in configure() 89 _input_to_input_weights = input_to_input_weights; in configure() 103 …std::vector<const ITensor *> inputs_weights_vector{ input_to_input_weights, input_to_forget_weight… in configure() 249 …const ITensorInfo *input_to_input_weights, const ITensorInfo *input_to_forget_weights, const ITens… in validate() argument 255 …ARM_COMPUTE_RETURN_ERROR_ON_NULLPTR(input, input_to_input_weights, input_to_forget_weights, input_… in validate() [all …]
|
/aosp_15_r20/external/ComputeLibrary/tests/validation/CL/ |
H A D | LSTMLayerQuantized.cpp | 93 …auto input_to_input_weights = create_tensor<CLTensor>(input_weights_shape, DataType::QASYMM8,… in TEST_SUITE() local 117 …lstmq.configure(&input, &input_to_input_weights, &input_to_forget_weights, &input_to_cell_weights,… in TEST_SUITE() 122 input_to_input_weights.allocator()->allocate(); in TEST_SUITE() 138 fill_tensor(input_to_input_weights, std::vector<uint8_t>{ 47, 168, in TEST_SUITE() 232 …auto input_to_input_weights = create_tensor<CLTensor>(input_weights_shape, DataType::QASYMM8,… in TEST_CASE() local 256 …lstmq.configure(&input, &input_to_input_weights, &input_to_forget_weights, &input_to_cell_weights,… in TEST_CASE() 261 input_to_input_weights.allocator()->allocate(); in TEST_CASE() 277 … fill_tensor(input_to_input_weights, std::vector<uint8_t>{ 141, 89, 200, 180, 46, 50, 87, 128, in TEST_CASE() 472 …auto input_to_input_weights = create_tensor<CLTensor>(input_weights_shape, DataType::QASYMM8,… in TEST_SUITE() local 496 …lstmq.configure(&input, &input_to_input_weights, &input_to_forget_weights, &input_to_cell_weights,… in TEST_SUITE() [all …]
|
/aosp_15_r20/external/ComputeLibrary/tests/validation/NEON/ |
H A D | LSTMLayerQuantized.cpp | 100 …auto input_to_input_weights = create_tensor<Tensor>(input_weights_shape, DataType::QASYMM8, 1… in TEST_SUITE() local 124 …lstmq.configure(&input, &input_to_input_weights, &input_to_forget_weights, &input_to_cell_weights,… in TEST_SUITE() 129 input_to_input_weights.allocator()->allocate(); in TEST_SUITE() 145 fill_tensor(input_to_input_weights, std::vector<uint8_t>{ 47, 168, in TEST_SUITE() 239 …auto input_to_input_weights = create_tensor<Tensor>(input_weights_shape, DataType::QASYMM8, 1… in TEST_CASE() local 263 …lstmq.configure(&input, &input_to_input_weights, &input_to_forget_weights, &input_to_cell_weights,… in TEST_CASE() 268 input_to_input_weights.allocator()->allocate(); in TEST_CASE() 284 … fill_tensor(input_to_input_weights, std::vector<uint8_t>{ 141, 89, 200, 180, 46, 50, 87, 128, in TEST_CASE() 479 …auto input_to_input_weights = create_tensor<Tensor>(input_weights_shape, DataType::QASYMM8, 1… in TEST_SUITE() local 503 …lstmq.configure(&input, &input_to_input_weights, &input_to_forget_weights, &input_to_cell_weights,… in TEST_SUITE() [all …]
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/V1_2/ |
D | layer_norm_lstm.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", variable 95 "LSTM", input, input_to_input_weights, input_to_forget_weights, 108 input_to_input_weights: [ 201 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", variable 267 "LSTM", input, input_to_input_weights, input_to_forget_weights, 280 input_to_input_weights: [],
|
D | quantized_lstm.mod.py | 33 input_to_input_weights = Input("inputToInputWeights", InputWeightsType) variable 62 input_to_input_weights, 80 input_to_input_weights: [146, 250, 235, 171, 10, 218, 171, 108], 120 input_to_input_weights = Parameter("inputToInputWeights", InputWeightsType, variable 158 model = model.Operation("QUANTIZED_16BIT_LSTM", input_, input_to_input_weights,
|
D | lstm2_state_float16.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT16", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 96 input0 = {input_to_input_weights:[],
|
D | lstm3_state2_float16.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT16", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 96 input0 = {input_to_input_weights: [
|
D | unidirectional_sequence_lstm_layer_norm_cifg_peephole.mod.py | 32 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", variable 93 "UNIDIRECTIONAL_SEQUENCE_LSTM", input, input_to_input_weights, input_to_forget_weights, 105 input_to_input_weights: [],
|
D | lstm2_float16.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT16", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 96 input0 = {input_to_input_weights:[],
|
D | lstm_state_float16.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT16", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 97 input0 = {input_to_input_weights: [-0.45018822, -0.02338299, -0.0870589, -0.34550029, 0.04266912, …
|
D | lstm_state2_float16.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT16", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 97 input0 = {input_to_input_weights: [-0.45018822, -0.02338299, -0.0870589, -0.34550029, 0.04266912, …
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/ |
H A D | lstm.cc | 182 const TfLiteTensor* input_to_input_weights = in PopulateQuantizedLstmParams8x8_16() local 237 const bool use_cifg = (input_to_input_weights == nullptr); in PopulateQuantizedLstmParams8x8_16() 307 input_to_input_weight_scale = input_to_input_weights->params.scale; in PopulateQuantizedLstmParams8x8_16() 470 const TfLiteTensor* input_to_input_weights = in PopulateQuantizedLstmParams8x8_8() local 541 const bool use_cifg = (input_to_input_weights == nullptr); in PopulateQuantizedLstmParams8x8_8() 612 input_to_input_weight_ptr = input_to_input_weights->data.int8; in PopulateQuantizedLstmParams8x8_8() 615 input_to_input_weight_scale = input_to_input_weights->params.scale; in PopulateQuantizedLstmParams8x8_8() 876 const TfLiteTensor* input_to_input_weights = in CheckInputTensorDimensions() local 878 const bool use_cifg = (input_to_input_weights == nullptr); in CheckInputTensorDimensions() 880 TF_LITE_ENSURE_EQ(context, input_to_input_weights->dims->size, 2); in CheckInputTensorDimensions() [all …]
|
H A D | unidirectional_sequence_lstm.cc | 96 const TfLiteTensor* input_to_input_weights = GetOptionalInputTensor( in PopulateQuantizedLstmParams8x8_16() local 156 const bool use_cifg = (input_to_input_weights == nullptr); in PopulateQuantizedLstmParams8x8_16() 229 input_to_input_weight_scale = input_to_input_weights->params.scale; in PopulateQuantizedLstmParams8x8_16() 429 const TfLiteTensor* input_to_input_weights = GetOptionalInputTensor( in CheckInputTensorDimensions() local 431 if (input_to_input_weights != nullptr) { in CheckInputTensorDimensions() 432 TF_LITE_ENSURE_EQ(context, input_to_input_weights->dims->size, 2); in CheckInputTensorDimensions() 433 TF_LITE_ENSURE_EQ(context, input_to_input_weights->dims->data[0], n_cell); in CheckInputTensorDimensions() 434 TF_LITE_ENSURE_EQ(context, input_to_input_weights->dims->data[1], n_input); in CheckInputTensorDimensions() 488 ((input_to_input_weights != nullptr) && in CheckInputTensorDimensions() 490 ((input_to_input_weights == nullptr) && in CheckInputTensorDimensions() [all …]
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/V1_3/ |
D | qlstm_projection.mod.py | 31 input_to_input_weights = Input("input_to_input_weights", InputWeightsType) variable 84 "QUANTIZED_LSTM", input, input_to_input_weights, input_to_forget_weights, 98 input_to_input_weights: [ 159 input_to_input_weights: [],
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/V1_0/ |
D | lstm2_state.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 96 input0 = {input_to_input_weights:[],
|
D | lstm_state.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 97 input0 = {input_to_input_weights: [-0.45018822, -0.02338299, -0.0870589, -0.34550029, 0.04266912, …
|
D | lstm3_state.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 96 input0 = {input_to_input_weights: [
|
D | lstm2.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 96 input0 = {input_to_input_weights:[],
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/V1_1/ |
D | lstm2_state2_relaxed.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 97 input0 = {input_to_input_weights:[],
|
D | lstm_relaxed.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 98 input0 = {input_to_input_weights: [-0.45018822, -0.02338299, -0.0870589, -0.34550029, 0.04266912, …
|
D | lstm2_state_relaxed.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 97 input0 = {input_to_input_weights:[],
|
D | lstm_state_relaxed.mod.py | 29 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", "{%d, %d}" % (n_cell, n_… variable 66 input_to_input_weights, 98 input0 = {input_to_input_weights: [-0.45018822, -0.02338299, -0.0870589, -0.34550029, 0.04266912, …
|
/aosp_15_r20/packages/modules/NeuralNetworks/tools/test_generator/tests/P_backward_compatibility_float/ |
D | lstm_float.mod.py | 25 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", "{%d, %d}" % (n_cell, n_… variable 62 input_to_input_weights, 94 input0 = {input_to_input_weights: [-0.45018822, -0.02338299, -0.0870589, -0.34550029, 0.04266912, …
|
/aosp_15_r20/packages/modules/NeuralNetworks/tools/test_generator/tests/P_vts_backward_compatibility_float/ |
D | lstm_float.mod.py | 25 input_to_input_weights = Input("input_to_input_weights", "TENSOR_FLOAT32", "{%d, %d}" % (n_cell, n_… variable 62 input_to_input_weights, 94 input0 = {input_to_input_weights: [-0.45018822, -0.02338299, -0.0870589, -0.34550029, 0.04266912, …
|