/aosp_15_r20/external/executorch/backends/vulkan/runtime/utils/ |
H A D | StorageUtils.h | 77 enum class GPUMemoryLayout : uint8_t { enum 92 static constexpr GPUMemoryLayout kWidthPacked = 93 GPUMemoryLayout::TENSOR_WIDTH_PACKED; 95 static constexpr GPUMemoryLayout kHeightPacked = 96 GPUMemoryLayout::TENSOR_HEIGHT_PACKED; 98 static constexpr GPUMemoryLayout kChannelsPacked = 99 GPUMemoryLayout::TENSOR_CHANNELS_PACKED; 102 T to_packed_dim(const GPUMemoryLayout layout) { in to_packed_dim() 117 const GPUMemoryLayout layout) {
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/vulkan/api/ |
H A D | Tensor.cpp | 61 const api::GPUMemoryLayout memory_layout, in calc_strides() 66 case api::GPUMemoryLayout::TENSOR_WIDTH_PACKED: in calc_strides() 69 case api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED: in calc_strides() 94 const api::GPUMemoryLayout memory_layout, in calc_gpu_sizes() 124 case api::GPUMemoryLayout::TENSOR_WIDTH_PACKED: in calc_gpu_sizes() 131 case api::GPUMemoryLayout::TENSOR_HEIGHT_PACKED: in calc_gpu_sizes() 138 case api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED: in calc_gpu_sizes() 156 const api::GPUMemoryLayout memory_layout) { in create_image_extents() 174 case api::GPUMemoryLayout::TENSOR_WIDTH_PACKED: in create_image_extents() 178 case api::GPUMemoryLayout::TENSOR_HEIGHT_PACKED: in create_image_extents() [all …]
|
H A D | Tensor.h | 36 const api::GPUMemoryLayout gpu_memory_layout, 88 const api::GPUMemoryLayout gpu_memory_layout, 103 const api::GPUMemoryLayout memory_layout = 104 api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED, 115 const api::GPUMemoryLayout memory_layout = 116 api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED); 141 api::GPUMemoryLayout memory_layout_; 256 inline api::GPUMemoryLayout gpu_memory_layout() const { in gpu_memory_layout()
|
H A D | Types.h | 182 enum class GPUMemoryLayout : uint32_t { enum
|
/aosp_15_r20/external/executorch/backends/vulkan/runtime/graph/ |
H A D | ComputeGraph.h | 104 const utils::GPUMemoryLayout memory_layout); 116 const utils::GPUMemoryLayout memory_layout); 316 inline utils::GPUMemoryLayout estimate_memory_layout_of( 420 utils::GPUMemoryLayout suggested_memory_layout( 440 const utils::GPUMemoryLayout memory_layout, 460 const utils::GPUMemoryLayout memory_layout, 484 const utils::GPUMemoryLayout memory_layout); 492 const utils::GPUMemoryLayout memory_layout); 594 const utils::GPUMemoryLayout memory_layout, 623 const utils::GPUMemoryLayout memory_layout,
|
H A D | ComputeGraph.cpp | 59 const utils::GPUMemoryLayout memory_layout) in VALUE_PTR_CLASS_IMPL() 82 const utils::GPUMemoryLayout memory_layout) in TmpTensor() 166 utils::GPUMemoryLayout ComputeGraph::suggested_memory_layout( in suggested_memory_layout() 241 const utils::GPUMemoryLayout memory_layout, in add_tensor() 272 const utils::GPUMemoryLayout memory_layout, in add_tensor() 281 const utils::GPUMemoryLayout memory_layout) { in add_tensor_like() 287 const utils::GPUMemoryLayout memory_layout) { in add_tensor_like()
|
H A D | GraphConfig.h | 29 utils::GPUMemoryLayout memory_layout_override; 40 void set_memory_layout_override(utils::GPUMemoryLayout memory_layout);
|
H A D | GraphConfig.cpp | 74 utils::GPUMemoryLayout memory_layout) { in set_memory_layout_override()
|
/aosp_15_r20/external/executorch/backends/vulkan/test/ |
H A D | vulkan_compute_api_test.cpp | 106 const utils::GPUMemoryLayout layout, in get_reference_strides() 1908 utils::GPUMemoryLayout src_layout, in test_clone() 1910 utils::GPUMemoryLayout dst_layout) { in test_clone() 1942 std::vector<std::pair<utils::GPUMemoryLayout, utils::GPUMemoryLayout>> cases{ in TEST() 1968 utils::GPUMemoryLayout memory_layout = in TEST() 1969 utils::GPUMemoryLayout::TENSOR_CHANNELS_PACKED; in TEST() 2050 utils::GPUMemoryLayout memory_layout = in TEST() 2051 utils::GPUMemoryLayout::TENSOR_CHANNELS_PACKED; in TEST() 2113 utils::GPUMemoryLayout memory_layout = in TEST() 2114 utils::GPUMemoryLayout::TENSOR_CHANNELS_PACKED; in TEST() [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/vulkan/ops/ |
H A D | Convert.h | 19 inline api::GPUMemoryLayout get_gpu_memory_layout( in get_gpu_memory_layout() 25 return api::GPUMemoryLayout::TENSOR_WIDTH_PACKED; in get_gpu_memory_layout() 27 return api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED; in get_gpu_memory_layout() 42 return api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED; in get_gpu_memory_layout()
|
H A D | Mm.cpp | 39 api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED) { in pack_inputs_using_width_packing() 44 v_input.gpu_memory_layout() == api::GPUMemoryLayout::TENSOR_WIDTH_PACKED, in pack_inputs_using_width_packing() 71 api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED) { in pack_weights_using_height_packing() 78 api::GPUMemoryLayout::TENSOR_HEIGHT_PACKED, in pack_weights_using_height_packing() 684 v_input.gpu_memory_layout() == api::GPUMemoryLayout::TENSOR_WIDTH_PACKED, in run_addmm_context() 689 api::GPUMemoryLayout::TENSOR_HEIGHT_PACKED, in run_addmm_context() 795 api::GPUMemoryLayout::TENSOR_WIDTH_PACKED, in run_baddbmm_context() 800 api::GPUMemoryLayout::TENSOR_HEIGHT_PACKED, in run_baddbmm_context()
|
H A D | Factory.cpp | 27 : api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED, in _empty_affine_quantized() 45 : api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED, in empty_memory_format()
|
H A D | Copy.cpp | 300 case api::GPUMemoryLayout::TENSOR_WIDTH_PACKED: in from_vulkan() 303 case api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED: in from_vulkan()
|
H A D | Convolution.cpp | 805 api::GPUMemoryLayout::TENSOR_CHANNELS_PACKED) { in pack_weights_using_width_packing() 810 v_weight.gpu_memory_layout() == api::GPUMemoryLayout::TENSOR_WIDTH_PACKED, in pack_weights_using_width_packing()
|
/aosp_15_r20/external/executorch/backends/vulkan/runtime/api/containers/ |
H A D | Tensor.h | 185 const utils::GPUMemoryLayout memory_layout = utils::kChannelsPacked, 193 const utils::GPUMemoryLayout memory_layout = utils::kChannelsPacked); 398 utils::GPUMemoryLayout estimate_memory_layout() const;
|
H A D | Tensor.cpp | 16 const utils::GPUMemoryLayout memory_layout) { in calculate_sizes() 440 const utils::GPUMemoryLayout memory_layout, in vTensor() 487 const utils::GPUMemoryLayout memory_layout) in vTensor() 601 utils::GPUMemoryLayout vTensor::estimate_memory_layout() const { in estimate_memory_layout()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/vulkan/impl/ |
H A D | Packing.cpp | 301 api::GPUMemoryLayout target_layout, in channel_image_repacking() 353 api::GPUMemoryLayout::TENSOR_HEIGHT_PACKED, in convert_image_channels_packed_to_height_packed() 360 api::GPUMemoryLayout::TENSOR_WIDTH_PACKED, in convert_image_channels_packed_to_width_packed()
|
/aosp_15_r20/external/executorch/backends/vulkan/runtime/ |
H A D | VulkanBackend.cpp | 106 utils::GPUMemoryLayout get_memory_layout( in get_memory_layout() 138 utils::GPUMemoryLayout memory_layout = in get_graph_config() 139 static_cast<utils::GPUMemoryLayout>(value_as_int); in get_graph_config() 195 utils::GPUMemoryLayout memory_layout = in add_tensor_to_graph()
|
/aosp_15_r20/external/executorch/backends/vulkan/serialization/ |
H A D | schema.fbs | 33 // Describes how memory should be laid out in GPU memory. See the GPUMemoryLayout 35 // entry must match the corresponding entry in utils::GPUMemoryLayout.
|
/aosp_15_r20/external/executorch/backends/vulkan/test/utils/ |
H A D | test_utils.h | 25 utils::GPUMemoryLayout::TENSOR_CHANNELS_PACKED, \ 34 utils::GPUMemoryLayout::TENSOR_WIDTH_PACKED, \
|
/aosp_15_r20/external/executorch/backends/vulkan/runtime/graph/ops/impl/ |
H A D | Staging.h | 50 const utils::GPUMemoryLayout layout,
|
H A D | Staging.cpp | 148 const utils::GPUMemoryLayout layout, in prepack_standard()
|
H A D | MatMul.cpp | 177 const utils::GPUMemoryLayout mat2_layout = in add_matmul_optimized_node()
|
H A D | Linear.cpp | 162 const utils::GPUMemoryLayout mat2_layout = in add_addmm_optimized_node()
|
H A D | Convolution.cpp | 92 const utils::GPUMemoryLayout memory_layout) { in prepack_biases()
|