/aosp_15_r20/external/pytorch/torch/csrc/jit/passes/ |
H A D | metal_rewrite.cpp | 55 %packed_weight_bias = metal_prepack::conv2d_prepack( in insertPrePackedConv2dOp() 93 %packed_weight_bias = metal_prepack::conv2d_prepack( in fuseReluWithPackedOps() 105 …%packed_weight_bias: __torch__.torch.classes.metal.Conv2dOpContext = metal_prepack::conv2d_prepack( in fuseReluWithPackedOps() 125 %packed_weight_bias = metal_prepack::conv2d_prepack( in fuseReluWithPackedOps() 163 …%packed_weight_bias: __torch__.torch.classes.metal.Conv2dOpContext = metal_prepack::conv2d_prepack( in fuseHardtanhWithPackedOps() 172 %packed_weight_bias = metal_prepack::conv2d_prepack( in fuseHardtanhWithPackedOps() 185 %packed_weight_bias = metal_prepack::conv2d_prepack( in fuseHardtanhWithPackedOps() 230 Symbol::fromQualString("metal_prepack::conv2d_prepack")) || in metalFoldPrePackingOps()
|
H A D | mkldnn_rewrite.cpp | 51 Symbol::fromQualString("mkldnn_prepacked::conv2d_prepack"), 1); in insertPrePackedConvOpForNode() 118 %packed_weight_bias = mkldnn_prepacked::conv2d_prepack( in FuseReluWithPackedOps() 129 …cked_weight_bias : __torch__.torch.classes.mkldnn.ConvOpContext = mkldnn_prepacked::conv2d_prepack( in FuseReluWithPackedOps() 160 Symbol::fromQualString("mkldnn_prepacked::conv2d_prepack")); in PrePackingOpsFolder()
|
/aosp_15_r20/external/pytorch/test/ |
H A D | test_metal.py | 81 "metal_prepack::conv2d_prepack": 1, 103 "metal_prepack::conv2d_prepack": 1, 109 pattern_count_map["metal_prepack::conv2d_prepack"] = -1 142 "metal_prepack::conv2d_prepack": 1, 146 pattern_count_map["metal_prepack::conv2d_prepack"] = -1
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/metal/ |
H A D | MetalPrepackOpRegister.cpp | 80 TORCH_SELECTIVE_SCHEMA("metal_prepack::conv2d_prepack(Tensor W, Tensor? B, int[2] stride, " in TORCH_LIBRARY() 95 static c10::intrusive_ptr<Conv2dOpContext> conv2d_prepack( in conv2d_prepack() function 126 m.impl(TORCH_SELECTIVE_NAME("metal_prepack::conv2d_prepack"), TORCH_FN(conv2d_prepack)); in TORCH_LIBRARY_IMPL()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/ |
H A D | qconv_prepack.cpp | 252 "quantized::conv2d_prepack (qnnpack): expected bias to be 1-dimensional " in prepack() 705 "Didn't find engine for operation quantized::conv2d_prepack ", in _run() 830 // conv_prepack is deprecated, please use conv2d_prepack for 2D conv. in TORCH_LIBRARY_IMPL() 833 …m.impl(TORCH_SELECTIVE_NAME("quantized::conv2d_prepack"), TORCH_FN(QConvPackWeightInt8<2>::run_con… in TORCH_LIBRARY_IMPL() 843 …m.impl(TORCH_SELECTIVE_NAME("_quantized::conv2d_prepack"), TORCH_FN(QConvPackWeightInt8<2>::run_co… in TORCH_LIBRARY_IMPL()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/mkldnn/ |
H A D | RegisterMkldnnOpContextClass.cpp | 84 …"mkldnn_prepacked::conv2d_prepack(Tensor W, Tensor? B, int[2] stride, int[2] padding, int[2] dilat… in TORCH_LIBRARY() 92 TORCH_SELECTIVE_NAME("mkldnn_prepacked::conv2d_prepack"), in TORCH_LIBRARY_IMPL()
|
/aosp_15_r20/external/pytorch/torch/csrc/jit/passes/quantization/ |
H A D | finalize.cpp | 161 n->kind() == Symbol::fromQualString("quantized::conv2d_prepack") || in FoldQuantizedPrepackingOps() 178 n->kind() == Symbol::fromQualString("quantized::conv2d_prepack") || in RegisterPrePackingParams()
|
H A D | register_packed_params.cpp | 18 n->kind() == Symbol::fromQualString("quantized::conv2d_prepack") || in isPrepackNode()
|
H A D | quantization_patterns.h | 1205 …_torch__.torch.classes.quantized.Conv2dPackedParamsBase = quantized::conv2d_prepack(%w_quant, %b, … in conv_prepack_unpack_patterns()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/ |
H A D | library.cpp | 77 // conv_prepack is deprecated, please use conv2d_prepack for 2D conv. in TORCH_LIBRARY() 80 …m.def(TORCH_SELECTIVE_SCHEMA("quantized::conv2d_prepack(Tensor weight, Tensor? bias, int[] stride,… in TORCH_LIBRARY() 236 …m.def(TORCH_SELECTIVE_SCHEMA("_quantized::conv2d_prepack(Tensor weight, Tensor? bias, int[] stride… in TORCH_LIBRARY()
|
/aosp_15_r20/external/pytorch/test/mobile/model_test/ |
H A D | coverage.yaml | 673 - quantized::conv2d_prepack 1032 quantized::conv2d_prepack: 14
|
H A D | model_ops.yaml | 424 quantized::conv2d_prepack: 14
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cudnn/ |
H A D | ConvPrepack.cpp | 210 …m.impl(TORCH_SELECTIVE_NAME("quantized::conv2d_prepack"), TORCH_FN(QConvPackWeightInt8Cudnn<2>::ru… in TORCH_LIBRARY_IMPL()
|
/aosp_15_r20/external/pytorch/torch/ao/nn/quantized/modules/ |
H A D | conv.py | 568 self._packed_params = torch.ops.quantized.conv2d_prepack( 572 self._packed_params = torch.ops.quantized.conv2d_prepack(
|
/aosp_15_r20/external/pytorch/test/quantization/core/ |
H A D | test_quantized_op.py | 2845 w_packed = torch.ops.quantized.conv2d_prepack(qw, bias_float, strides, pads, dilations, 1) 5176 qconv_prepack = torch.ops.quantized.conv2d_prepack 5257 qconv_prepack = torch.ops.quantized.conv2d_prepack 5320 qconv_prepack = torch.ops.quantized.conv2d_prepack 5377 qconv_prepack = torch.ops.quantized.conv2d_prepack 5474 qconv, torch.ops.quantized.conv2d_prepack, conv_op, batch_size, 5557 qconv, torch.ops.quantized.conv2d_prepack, conv_op, batch_size, 5634 …weight_prepacked = torch.ops.quantized.conv2d_prepack(weight_int8, None, stride, padding, dilation… 6069 qconv_prepack = torch.ops.quantized.conv2d_prepack 6597 w_packed = torch.ops.quantized.conv2d_prepack(
|
/aosp_15_r20/external/pytorch/torch/ao/nn/quantized/ |
H A D | functional.py | 311 packed_params = torch.ops.quantized.conv2d_prepack(
|
/aosp_15_r20/external/pytorch/torch/ao/quantization/fx/ |
H A D | utils.py | 159 torch.nn.functional.conv2d: torch.ops.quantized.conv2d_prepack,
|
H A D | _lower_to_native_backend.py | 350 torch._ops.ops.quantized.conv2d_prepack,
|
/aosp_15_r20/external/executorch/backends/cadence/aot/ |
H A D | fuse_ops.py | 413 else exir_ops.edge.quantized.conv2d_prepack(*args)
|
/aosp_15_r20/external/pytorch/torch/csrc/jit/tensorexpr/ |
H A D | lowerings.cpp | 1977 …{"quantized::conv2d_prepack(Tensor weight, Tensor? bias, int[] stride, int[] padding, int[] dilati… in nnc_lowerings_lazy_registration()
|
/aosp_15_r20/external/pytorch/test/quantization/jit/ |
H A D | test_quantize_jit.py | 1699 # and conv2d_prepack is folded
|
/aosp_15_r20/external/pytorch/test/quantization/fx/ |
H A D | test_quantize_fx.py | 1099 ns.call_function(torch.ops.quantized.conv2d_prepack),
|
/aosp_15_r20/external/pytorch/aten/src/ATen/test/ |
H A D | vulkan_api_test.cpp | 1667 TEST_F(VulkanAPITest, conv2d_prepack) { in TEST_F() argument
|