/aosp_15_r20/external/executorch/runtime/executor/ |
H A D | method_meta.cpp | 135 auto tensor_value = s_plan_->values()->Get(input_index)->val_as_Tensor(); in input_tensor_meta() local 138 tensor_value->sizes()->data(), tensor_value->sizes()->size()), in input_tensor_meta() 140 tensor_value->dim_order()->data(), tensor_value->dim_order()->size()), in input_tensor_meta() 141 static_cast<exec_aten::ScalarType>(tensor_value->scalar_type()), in input_tensor_meta() 142 tensor_value->allocation_info() != nullptr || in input_tensor_meta() 143 tensor_value->data_buffer_idx() != in input_tensor_meta() 176 auto tensor_value = s_plan_->values()->Get(output_index)->val_as_Tensor(); in output_tensor_meta() local 180 tensor_value->sizes()->data(), tensor_value->sizes()->size()), in output_tensor_meta() 182 tensor_value->dim_order()->data(), tensor_value->dim_order()->size()), in output_tensor_meta() 183 static_cast<exec_aten::ScalarType>(tensor_value->scalar_type()), in output_tensor_meta() [all …]
|
/aosp_15_r20/external/executorch/backends/xnnpack/runtime/ |
H A D | XNNCompiler.cpp | 163 const fb_xnnpack::XNNTensorValue* tensor_value, in getConstantDataPtr() argument 166 auto buffer_idx = tensor_value->constant_buffer_idx(); in getConstantDataPtr() 198 const fb_xnnpack::XNNTensorValue* tensor_value = nullptr; in defineTensor() local 203 tensor_value = value->xvalue_union_as_XNNTensorValue(); in defineTensor() 208 tensor_value = qtensor_value->tensor_value(); in defineTensor() 221 tensor_value != nullptr, in defineTensor() 227 std::vector<size_t> dims_data = flatbufferDimsToVector(tensor_value->dims()); in defineTensor() 235 getConstantDataPtr(tensor_value, flatbuffer_graph, constant_data_ptr); in defineTensor() 239 auto dq_datatype = getDataType(tensor_value->dq_datatype()); in defineTensor() 250 (tensor_value->flags() & XNN_VALUE_FLAG_EXTERNAL_INPUT), in defineTensor() [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/functorch/ |
H A D | BatchRulesRandomness.cpp | 131 auto [tensor_value, tensor_bdim] = unwrapTensorAtLevel(tensor, cur_level); in unary_pointwise_random_batch_rule() 132 tensor_value = moveBatchDimToFront(tensor_value, tensor_bdim); in unary_pointwise_random_batch_rule() 136 auto shape = tensor_value.sizes(); in unary_pointwise_random_batch_rule() 142 tensor_value = tensor_value.expand_symint(shapeVec); in unary_pointwise_random_batch_rule() 144 auto out = Func(tensor_value, std::forward<ExtraArgs>(extra_args)...); in unary_pointwise_random_batch_rule() 159 auto [tensor_value, tensor_bdim] = unwrapTensorAtLevel(self, cur_level); in tensor_like_random_batch_rule() 160 tensor_value = moveBatchDimToFront(tensor_value, tensor_bdim); in tensor_like_random_batch_rule() 163 tensor_value = tensor_value[0]; in tensor_like_random_batch_rule() 165 auto shape = tensor_value.sizes(); in tensor_like_random_batch_rule() 169 tensor_value = tensor_value.expand_symint(shapeVec); in tensor_like_random_batch_rule() [all …]
|
H A D | BatchRulesBinaryOps.cpp | 63 auto [tensor_value, tensor_bdim] = unwrapTensorAtLevel(tensor, cur_level); in apply() 69 auto shape = tensor_value.sizes(); in apply() 75 tensor_value = tensor_value.expand_symint(shapeVec); in apply() 80 return Func(tensor_value, other_value, std::forward<T>(extra_args)...); in apply() 83 tensor_value, tensor_bdim, other_value, other_bdim, in apply()
|
H A D | BatchRulesHelper.h | 147 auto [tensor_value, tensor_bdim] = unwrapTensorAtLevel(ivalue.toTensor(), cur_level); 148 tensor_inputs.emplace_back(tensor_value, tensor_bdim); 223 const auto& tensor_value = std::get<0>(unpacked); 226 auto candidate_batch_size = tensor_value.size(*tensor_bdim);
|
/aosp_15_r20/external/tensorflow/tensorflow/python/debug/lib/ |
H A D | debug_events_monitors.py | 149 tensor_value, 170 size = np.size(tensor_value) 171 if not size or not np.issubdtype(tensor_value.dtype, np.floating): 173 is_inf = np.isinf(tensor_value) 175 np.logical_and(is_inf, np.less(tensor_value, 0.0))) 177 np.logical_and(is_inf, np.greater(tensor_value, 0.0))) 178 num_nan = np.count_nonzero(np.isnan(tensor_value)) 267 for output_slot, tensor_value in enumerate(tensor_values): 269 tensor_value, execution.wall_time, execution.op_type, output_slot, 290 tensor_value = ( [all …]
|
H A D | debug_v2_ops_test.py | 118 tensor_value = tensor_util.MakeNdarray(trace.tensor_proto) 119 self.assertAllClose(tensor_value, [9.0, 16.0]) 129 tensor_value = tensor_util.MakeNdarray(trace.tensor_proto) 130 self.assertAllClose(tensor_value, [np.sqrt(3.0), 2.0]) 220 tensor_value = tensor_util.MakeNdarray(trace.tensor_proto) 221 self.assertAllClose(tensor_value, [9.0, 16.0])
|
H A D | debug_events_monitors_test.py | 347 self, tensor_value, dtype, expected_size, expected_num_neg_inf, argument 352 np.array(tensor_value, dtype=dtype) 440 self, tensor_value, dtype, expected_size, expected_num_neg_inf, argument 444 tensor_value, dtype=dtype)
|
H A D | debug_data.py | 125 tensor_value = InconvertibleTensorProto(tensor_proto) 128 tensor_value = tensor_util.MakeNdarray(tensor_proto) 130 tensor_value = InconvertibleTensorProto(tensor_proto) 133 tensor_value = InconvertibleTensorProto(tensor_proto, False) 135 return tensor_value
|
/aosp_15_r20/external/tensorflow/tensorflow/core/transforms/shape_inference/ |
H A D | pass.cc | 112 DenseElementsAttr tensor_value; in TryToCacheResultsTensorValue() local 114 tensor_value = DenseElementsAttr::get( in TryToCacheResultsTensorValue() 117 tensor_value = DenseElementsAttr::get( in TryToCacheResultsTensorValue() 120 cached_tensor_values_[op->getResult(0)] = tensor_value; in TryToCacheResultsTensorValue() 125 DenseElementsAttr tensor_value; in TryToCacheResultsTensorValue() local 127 tensor_value = in TryToCacheResultsTensorValue() 131 tensor_value = DenseElementsAttr::get( in TryToCacheResultsTensorValue() 135 cached_tensor_values_[op->getResult(0)] = tensor_value; in TryToCacheResultsTensorValue() 146 DenseElementsAttr tensor_value; in TryToCacheResultsTensorValue() local 148 tensor_value = DenseElementsAttr::get<int>( in TryToCacheResultsTensorValue() [all …]
|
/aosp_15_r20/external/pytorch/torch/csrc/jit/backends/xnnpack/compiler/ |
H A D | xnn_compiler.cpp | 46 auto tensor_value = value->xvalue_as_XNNTensorValue(); in compileModel() local 49 for (auto dim : *tensor_value->dims()) { in compileModel() 55 auto buffer_idx = tensor_value->constant_buffer_idx(); in compileModel() 62 /*num_dims=*/tensor_value->num_dims(), in compileModel() 65 /*external_id=*/tensor_value->external_id(), in compileModel() 66 /*flags=*/tensor_value->flags(), in compileModel() 72 remapped_ids.emplace(std::make_pair(tensor_value->id_out(), id)); in compileModel()
|
/aosp_15_r20/external/pytorch/torch/_dynamo/variables/ |
H A D | optimizer.py | 289 def wrap_tensor(self, tx: "InstructionTranslator", tensor_value): argument 299 if tensor_value in self.tensor_to_source: 301 mark_static_address(tensor_value) 302 builder = VariableBuilder(tx, self.tensor_to_source[tensor_value]) 304 elif tensor_value in self.grad_to_source: 305 builder = VariableBuilder(tx, self.grad_to_source[tensor_value]) 308 mark_static_address(tensor_value) 310 global_name = tx.store_global_weakref_by_id(GLOBAL_KEY_PREFIX, tensor_value) 314 result = builder(tensor_value)
|
/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/ |
H A D | tensor_array_ops.cc | 422 const Tensor* tensor_value; in Compute() local 424 OP_REQUIRES_OK(ctx, ctx->input("value", &tensor_value)); in Compute() 436 ctx, tensor_value->dtype() == tensor_array->ElemType(), in Compute() 440 DataTypeString(tensor_value->dtype()), ".")); in Compute() 442 tensor_array->WriteOrAggregate<Device, T>(ctx, index, tensor_value); in Compute() 1031 const Tensor* tensor_value; in Compute() local 1032 OP_REQUIRES_OK(ctx, ctx->input("value", &tensor_value)); in Compute() 1033 TensorShape element_shape(tensor_value->shape()); in Compute() 1041 ctx, tensor_value->dtype() == tensor_array->ElemType(), in Compute() 1045 DataTypeString(tensor_value->dtype()), ".")); in Compute() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/ops/ |
H A D | collective_ops_gpu_test.py | 144 tensor_value = [0.1, 1.1, 2.1, 3.1, 4.1, 5.1, 6.1, 7.1] 152 t = constant_op.constant(tensor_value) 156 t = constant_op.constant(tensor_value) 162 self.assertAllClose(result, tensor_value, rtol=1e-5, atol=1e-5) 167 tensor_value = [0.1, 1.1, 2.1, 3.1, 4.1, 5.1, 6.1, 7.1] 175 t = constant_op.constant(tensor_value) 185 tensor_value = [0.1, 1.1, 2.1, 3.1, 4.1, 5.1, 6.1, 7.1] 193 t = constant_op.constant(tensor_value)
|
/aosp_15_r20/external/federated-compute/fcp/tensorflow/ |
H A D | append_slices_test.py | 50 for (tensor_name, tensor_value) in zip(tensor_names, tensor_values): 54 data=[tensor_value], 147 for tensor_value in tensor_values: 151 data=[tensor_value], 166 for (tensor_name, tensor_value) in zip(tensor_names, tensor_values): 170 data=[tensor_value],
|
/aosp_15_r20/external/ComputeLibrary/tests/validation/CL/ |
H A D | FillBorder.cpp | 67 const uint8_t tensor_value = distribution_u8(generator); variable 81 std::uniform_int_distribution<uint8_t> distribution{ tensor_value, tensor_value }; 96 validate(CLAccessor(src), &tensor_value);
|
/aosp_15_r20/external/ComputeLibrary/tests/validation/NEON/ |
H A D | FillBorder.cpp | 66 const uint8_t tensor_value = distribution_u8(generator); variable 80 std::uniform_int_distribution<uint32_t> distribution{ tensor_value, tensor_value }; 95 validate(Accessor(src), &tensor_value);
|
/aosp_15_r20/external/tensorflow/tensorflow/core/profiler/internal/ |
H A D | tfprof_tensor_test.cc | 64 EXPECT_GT(root.children(0).tensor_value().value_double_size(), 10); in TEST_F() 66 EXPECT_GT(root.children(1).tensor_value().value_double_size(), 10); in TEST_F() 68 EXPECT_EQ(root.children(2).tensor_value().value_double_size(), 1); in TEST_F()
|
/aosp_15_r20/external/tensorflow/tensorflow/python/framework/ |
H A D | constant_op.py | 282 tensor_value = attr_value_pb2.AttrValue() 283 tensor_value.tensor.CopyFrom( 287 dtype_value = attr_value_pb2.AttrValue(type=tensor_value.tensor.dtype) 288 attrs = {"value": tensor_value, "dtype": dtype_value}
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/tools/benchmark/ |
H A D | benchmark_test.cc | 118 StringRef tensor_value = GetString(input_tensor, tensor_dim_index); in CheckInputTensorValue() local 119 EXPECT_TRUE(absl::equal(tensor_value.str, tensor_value.str + tensor_value.len, in CheckInputTensorValue()
|
/aosp_15_r20/external/eigen/unsupported/test/ |
H A D | cxx11_tensor_block_io.cpp | 257 float tensor_value = tensor_tensor(d0, d1, d2); in test_block_io_copy_using_reordered_dimensions_do_not_squeeze() local 258 VERIFY_IS_EQUAL(block_value, tensor_value); in test_block_io_copy_using_reordered_dimensions_do_not_squeeze() 306 float tensor_value = tensor_tensor(d0, d1, d2, d3); in test_block_io_copy_using_reordered_dimensions_squeeze() local 307 VERIFY_IS_EQUAL(block_value, tensor_value); in test_block_io_copy_using_reordered_dimensions_squeeze()
|
/aosp_15_r20/external/pytorch/torch/csrc/profiler/standalone/ |
H A D | execution_trace_observer.cpp | 360 std::string tensor_shape, tensor_stride, tensor_type, tensor_value; in convertIValue() local 393 tensor_value = fmt::format( in convertIValue() 402 tensor_shape, tensor_stride, tensor_type, tensor_value); in convertIValue() 454 std::string tensor_value = getScalarValue(val); in convertIValue() local 457 tensor_shape, tensor_stride, tensor_type, tensor_value); in convertIValue()
|
/aosp_15_r20/external/tensorflow/tensorflow/core/tfrt/fallback/ |
H A D | op_kernel_runner.h | 113 for (const auto& tensor_value : tensor_values) { in OpKernelRunState() local 114 input_tf_tensors.push_back(*tensor_value.tensor); in OpKernelRunState()
|
/aosp_15_r20/external/tensorflow/tensorflow/python/checkpoint/ |
H A D | functional_saver.py | 66 tensor_value = tensor.tensor 70 if tensor_value is not None: 72 tensors.append(tensor_value)
|
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/layers/legacy_rnn/ |
H A D | rnn_cell_wrapper_impl.py | 115 tensor_value = ops.convert_to_tensor_v2_with_dispatch(v) 116 const_value = tensor_util.constant_value(tensor_value) 117 return (tensor_value, const_value)
|