/aosp_15_r20/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
H A D | data_format_ops.cc | 108 const TensorShape input_tensor_shape = ctx->InputShape(0); in Compile() local 109 int input_rank = input_tensor_shape.dims(); in Compile() 113 input_tensor_shape.DebugString())); in Compile() 114 const int dim0 = input_tensor_shape.dim_size(0); in Compile() 121 input_tensor_shape.num_elements() == spatial_dim_count || in Compile() 122 input_tensor_shape.num_elements() == full_dim_count, in Compile() 126 input_tensor_shape.DebugString())); in Compile() 129 input_tensor_shape.dim_size(0) == spatial_dim_count || in Compile() 130 input_tensor_shape.dim_size(0) == full_dim_count, in Compile() 135 input_tensor_shape.DebugString())); in Compile() [all …]
|
H A D | conv_ops.cc | 106 TensorShape input_tensor_shape; in Compile() local 108 ctx, ctx->ConstantInputAsShape(0, &input_tensor_shape, in Compile() 111 TensorShapeToXLAShape(ctx->input_xla_type(1), input_tensor_shape); in Compile()
|
H A D | batchtospace_op.cc | 25 DataType input_dtype, const TensorShape& input_tensor_shape, in BatchToSpace() argument 28 const int input_rank = input_tensor_shape.dims(); in BatchToSpace() 30 input_tensor_shape.dim_sizes(); in BatchToSpace()
|
H A D | spacetobatch_op.cc | 26 DataType input_dtype, const TensorShape& input_tensor_shape, in SpaceToBatch() argument 29 const int input_rank = input_tensor_shape.dims(); in SpaceToBatch() 31 input_tensor_shape.dim_sizes(); in SpaceToBatch()
|
H A D | conv_op_helpers.cc | 142 TensorShape input_tensor_shape, filter_tensor_shape, in ConvBackpropComputeDimensionsV2XlaShapes() local 144 TF_RETURN_IF_ERROR(XLAShapeToTensorShape(input_shape, &input_tensor_shape)); in ConvBackpropComputeDimensionsV2XlaShapes() 149 label, num_spatial_dims, input_tensor_shape, filter_tensor_shape, in ConvBackpropComputeDimensionsV2XlaShapes() 407 TensorShape input_tensor_shape, filter_tensor_shape, output_tensor_shape; in MakeXlaBackpropFilterConvOp() local 409 TF_RETURN_IF_ERROR(XLAShapeToTensorShape(input_shape, &input_tensor_shape)); in MakeXlaBackpropFilterConvOp()
|
/aosp_15_r20/external/tensorflow/tensorflow/dtensor/mlir/ |
H A D | collectives.cc | 455 llvm::ArrayRef<int64_t> input_tensor_shape = input_tensor_type.getShape(); in EmitHaloExchange() local 460 if (input_tensor_shape[split_dim_index] < halo_size) in EmitHaloExchange() 472 for (const auto& size_and_index : llvm::enumerate(input_tensor_shape)) { in EmitHaloExchange() 517 begin_left[split_dim_index] = input_tensor_shape[split_dim_index] - halo_size; in EmitHaloExchange() 521 llvm::SmallVector<int64_t, 4> size(input_tensor_shape.begin(), in EmitHaloExchange() 522 input_tensor_shape.end()); in EmitHaloExchange() 576 llvm::SmallVector<int64_t, 4> final_shape(input_tensor_shape.begin(), in EmitHaloExchange() 577 input_tensor_shape.end()); in EmitHaloExchange()
|
/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/mkl/ |
H A D | mkl_pooling_ops_common.h | 516 const TensorShape& input_tensor_shape) { in InitMklPoolParameters() argument 519 this->data_format_tf_, input_tensor_shape); in InitMklPoolParameters() 628 TensorShape input_tensor_shape = input_tensor.shape(); in ConfigureInput() local 635 ? TFShapeToMklDnnDimsInNCHW(input_tensor_shape, in ConfigureInput() 637 : TFShapeToMklDnnDimsInNCDHW(input_tensor_shape, in ConfigureInput() 654 input_tensor_shape); in ConfigureInput()
|
H A D | mkl_maxpooling_op.cc | 67 TensorShape input_tensor_shape = input_tensor.shape(); in Compute() local 69 input_tensor_shape); in Compute() 106 input_tensor_shape, this->data_format_tf_), in Compute() 110 input_tensor_shape, this->data_format_tf_), in Compute()
|
H A D | mkl_avgpooling_op.cc | 65 TensorShape input_tensor_shape = input_tensor.shape(); in Compute() local 67 input_tensor_shape); in Compute()
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/mlir/tensorflow/transforms/ |
H A D | shape_inference.cc | 1619 auto input_tensor_shape = in PrecheckForXlaConvV2Op() local 1624 if (input_tensor_shape.size() <= 2) { in PrecheckForXlaConvV2Op() 1626 << "input tensor argument is " << input_tensor_shape.size() in PrecheckForXlaConvV2Op() 1638 if (input_tensor_shape.size() != kernel_tensor_shape.size()) { in PrecheckForXlaConvV2Op() 1664 auto input_features = input_tensor_shape[dnums.input_feature_dimension()]; in PrecheckForXlaConvV2Op() 1665 auto input_batch = input_tensor_shape[dnums.input_batch_dimension()]; in PrecheckForXlaConvV2Op() 1749 auto input_tensor_shape = input_tensor.getType().cast<RankedTensorType>(); in InferShapeForXlaConvV2Op() local 1750 for (auto i = 0; i < input_tensor_shape.getShape().size(); ++i) { in InferShapeForXlaConvV2Op() 1752 << input_tensor_shape.getShape()[i]); in InferShapeForXlaConvV2Op() 1753 input_tensor_dims_vec.push_back(input_tensor_shape.getShape()[i]); in InferShapeForXlaConvV2Op() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/core/ir/importexport/ |
H A D | convert_tensor.cc | 207 TensorShape input_tensor_shape(input_tensor.tensor_shape()); in ConvertTensorProto() local 211 input_tensor_shape.num_elements() > 1) { in ConvertTensorProto() 224 for (auto dim : input_tensor_shape) original_dimensions.push_back(dim.size); in ConvertTensorProto()
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/mlir/tensorflow/utils/ |
H A D | convert_tensor.cc | 209 TensorShape input_tensor_shape(input_tensor.tensor_shape()); in ConvertTensorProto() local 211 input_tensor_shape.num_elements() > 1) { in ConvertTensorProto() 224 for (auto dim : input_tensor_shape) original_dimensions.push_back(dim.size); in ConvertTensorProto()
|
/aosp_15_r20/external/tensorflow/tensorflow/dtensor/mlir/expansions/ |
H A D | conv_spmd_expander.cc | 116 auto input_tensor_shape = input_tensor_type.getShape(); in PadInputOnUnshardedDim() local 124 llvm::SmallVector<int64_t, 4> paddings_shape(input_tensor_shape.begin(), in PadInputOnUnshardedDim() 125 input_tensor_shape.end()); in PadInputOnUnshardedDim()
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | hlo_instruction_test.cc | 1583 Shape input_tensor_shape = ShapeUtil::MakeShape(F32, {50, 49, 48, 47, 46}); in TEST_F() local 1591 HloInstruction::CreateParameter(0, input_tensor_shape, "input_tensor")); in TEST_F() 1619 Shape input_tensor_shape = ShapeUtil::MakeShape(F32, {50, 49, 48, 47, 46}); in TEST_F() local 1627 HloInstruction::CreateParameter(0, input_tensor_shape, "input_tensor")); in TEST_F() 1655 Shape input_tensor_shape = ShapeUtil::MakeShape(F32, {50, 49, 48, 47, 46}); in TEST_F() local 1663 HloInstruction::CreateParameter(0, input_tensor_shape, "input_tensor")); in TEST_F() 1683 input_tensor_shape, input, scatter_indices, scatter_updates, in TEST_F()
|
/aosp_15_r20/external/tensorflow/tensorflow/core/grappler/optimizers/ |
H A D | constant_folding.cc | 2795 TensorShapeProto* input_tensor_shape, TensorShapeProto* output_tensor_shape, in IsReductionCandidateForSimplification() argument 2809 *input_tensor_shape = input_props.shape(); in IsReductionCandidateForSimplification() 2811 for (int i = 0; i < input_tensor_shape->dim_size(); ++i) { in IsReductionCandidateForSimplification() 2812 if (input_tensor_shape->dim(i).size() < 0) { in IsReductionCandidateForSimplification() 2822 TensorShape(*input_tensor_shape).num_elements(); in IsReductionCandidateForSimplification() 2890 TensorShapeProto input_tensor_shape, output_tensor_shape; in SimplifyReduction() local 2892 *node, properties, &input_tensor_shape, &output_tensor_shape, in SimplifyReduction() 2918 *node, input_tensor_shape, keep_dims, reduction_indices_vector); in SimplifyReduction()
|
H A D | constant_folding.h | 229 TensorShapeProto* input_tensor_shape,
|
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/layers/ |
H A D | core.py | 888 input_tensor_shape = tf_utils.convert_shapes(input_shape, to_tuples=False) 889 batch_size = nest.flatten(input_tensor_shape)[0][0] if input_shape else None
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/ |
H A D | xla_data.proto | 530 // for (k : [0, input_tensor_shape.rank))
|