Home
last modified time | relevance | path

Searched full:conv2d_prepack (Results 1 – 23 of 23) sorted by relevance

/aosp_15_r20/external/pytorch/torch/csrc/jit/passes/
H A Dmetal_rewrite.cpp55 %packed_weight_bias = metal_prepack::conv2d_prepack( in insertPrePackedConv2dOp()
93 %packed_weight_bias = metal_prepack::conv2d_prepack( in fuseReluWithPackedOps()
105 …%packed_weight_bias: __torch__.torch.classes.metal.Conv2dOpContext = metal_prepack::conv2d_prepack( in fuseReluWithPackedOps()
125 %packed_weight_bias = metal_prepack::conv2d_prepack( in fuseReluWithPackedOps()
163 …%packed_weight_bias: __torch__.torch.classes.metal.Conv2dOpContext = metal_prepack::conv2d_prepack( in fuseHardtanhWithPackedOps()
172 %packed_weight_bias = metal_prepack::conv2d_prepack( in fuseHardtanhWithPackedOps()
185 %packed_weight_bias = metal_prepack::conv2d_prepack( in fuseHardtanhWithPackedOps()
230 Symbol::fromQualString("metal_prepack::conv2d_prepack")) || in metalFoldPrePackingOps()
H A Dmkldnn_rewrite.cpp51 Symbol::fromQualString("mkldnn_prepacked::conv2d_prepack"), 1); in insertPrePackedConvOpForNode()
118 %packed_weight_bias = mkldnn_prepacked::conv2d_prepack( in FuseReluWithPackedOps()
129 …cked_weight_bias : __torch__.torch.classes.mkldnn.ConvOpContext = mkldnn_prepacked::conv2d_prepack( in FuseReluWithPackedOps()
160 Symbol::fromQualString("mkldnn_prepacked::conv2d_prepack")); in PrePackingOpsFolder()
/aosp_15_r20/external/pytorch/test/
H A Dtest_metal.py81 "metal_prepack::conv2d_prepack": 1,
103 "metal_prepack::conv2d_prepack": 1,
109 pattern_count_map["metal_prepack::conv2d_prepack"] = -1
142 "metal_prepack::conv2d_prepack": 1,
146 pattern_count_map["metal_prepack::conv2d_prepack"] = -1
/aosp_15_r20/external/pytorch/aten/src/ATen/native/metal/
H A DMetalPrepackOpRegister.cpp80 TORCH_SELECTIVE_SCHEMA("metal_prepack::conv2d_prepack(Tensor W, Tensor? B, int[2] stride, " in TORCH_LIBRARY()
95 static c10::intrusive_ptr<Conv2dOpContext> conv2d_prepack( in conv2d_prepack() function
126 m.impl(TORCH_SELECTIVE_NAME("metal_prepack::conv2d_prepack"), TORCH_FN(conv2d_prepack)); in TORCH_LIBRARY_IMPL()
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/
H A Dqconv_prepack.cpp252 "quantized::conv2d_prepack (qnnpack): expected bias to be 1-dimensional " in prepack()
705 "Didn't find engine for operation quantized::conv2d_prepack ", in _run()
830 // conv_prepack is deprecated, please use conv2d_prepack for 2D conv. in TORCH_LIBRARY_IMPL()
833 …m.impl(TORCH_SELECTIVE_NAME("quantized::conv2d_prepack"), TORCH_FN(QConvPackWeightInt8<2>::run_con… in TORCH_LIBRARY_IMPL()
843 …m.impl(TORCH_SELECTIVE_NAME("_quantized::conv2d_prepack"), TORCH_FN(QConvPackWeightInt8<2>::run_co… in TORCH_LIBRARY_IMPL()
/aosp_15_r20/external/pytorch/aten/src/ATen/native/mkldnn/
H A DRegisterMkldnnOpContextClass.cpp84 …"mkldnn_prepacked::conv2d_prepack(Tensor W, Tensor? B, int[2] stride, int[2] padding, int[2] dilat… in TORCH_LIBRARY()
92 TORCH_SELECTIVE_NAME("mkldnn_prepacked::conv2d_prepack"), in TORCH_LIBRARY_IMPL()
/aosp_15_r20/external/pytorch/torch/csrc/jit/passes/quantization/
H A Dfinalize.cpp161 n->kind() == Symbol::fromQualString("quantized::conv2d_prepack") || in FoldQuantizedPrepackingOps()
178 n->kind() == Symbol::fromQualString("quantized::conv2d_prepack") || in RegisterPrePackingParams()
H A Dregister_packed_params.cpp18 n->kind() == Symbol::fromQualString("quantized::conv2d_prepack") || in isPrepackNode()
H A Dquantization_patterns.h1205 …_torch__.torch.classes.quantized.Conv2dPackedParamsBase = quantized::conv2d_prepack(%w_quant, %b, … in conv_prepack_unpack_patterns()
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/
H A Dlibrary.cpp77 // conv_prepack is deprecated, please use conv2d_prepack for 2D conv. in TORCH_LIBRARY()
80 …m.def(TORCH_SELECTIVE_SCHEMA("quantized::conv2d_prepack(Tensor weight, Tensor? bias, int[] stride,… in TORCH_LIBRARY()
236 …m.def(TORCH_SELECTIVE_SCHEMA("_quantized::conv2d_prepack(Tensor weight, Tensor? bias, int[] stride… in TORCH_LIBRARY()
/aosp_15_r20/external/pytorch/test/mobile/model_test/
H A Dcoverage.yaml673 - quantized::conv2d_prepack
1032 quantized::conv2d_prepack: 14
H A Dmodel_ops.yaml424 quantized::conv2d_prepack: 14
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cudnn/
H A DConvPrepack.cpp210 …m.impl(TORCH_SELECTIVE_NAME("quantized::conv2d_prepack"), TORCH_FN(QConvPackWeightInt8Cudnn<2>::ru… in TORCH_LIBRARY_IMPL()
/aosp_15_r20/external/pytorch/torch/ao/nn/quantized/modules/
H A Dconv.py568 self._packed_params = torch.ops.quantized.conv2d_prepack(
572 self._packed_params = torch.ops.quantized.conv2d_prepack(
/aosp_15_r20/external/pytorch/test/quantization/core/
H A Dtest_quantized_op.py2845 w_packed = torch.ops.quantized.conv2d_prepack(qw, bias_float, strides, pads, dilations, 1)
5176 qconv_prepack = torch.ops.quantized.conv2d_prepack
5257 qconv_prepack = torch.ops.quantized.conv2d_prepack
5320 qconv_prepack = torch.ops.quantized.conv2d_prepack
5377 qconv_prepack = torch.ops.quantized.conv2d_prepack
5474 qconv, torch.ops.quantized.conv2d_prepack, conv_op, batch_size,
5557 qconv, torch.ops.quantized.conv2d_prepack, conv_op, batch_size,
5634 …weight_prepacked = torch.ops.quantized.conv2d_prepack(weight_int8, None, stride, padding, dilation…
6069 qconv_prepack = torch.ops.quantized.conv2d_prepack
6597 w_packed = torch.ops.quantized.conv2d_prepack(
/aosp_15_r20/external/pytorch/torch/ao/nn/quantized/
H A Dfunctional.py311 packed_params = torch.ops.quantized.conv2d_prepack(
/aosp_15_r20/external/pytorch/torch/ao/quantization/fx/
H A Dutils.py159 torch.nn.functional.conv2d: torch.ops.quantized.conv2d_prepack,
H A D_lower_to_native_backend.py350 torch._ops.ops.quantized.conv2d_prepack,
/aosp_15_r20/external/executorch/backends/cadence/aot/
H A Dfuse_ops.py413 else exir_ops.edge.quantized.conv2d_prepack(*args)
/aosp_15_r20/external/pytorch/torch/csrc/jit/tensorexpr/
H A Dlowerings.cpp1977 …{"quantized::conv2d_prepack(Tensor weight, Tensor? bias, int[] stride, int[] padding, int[] dilati… in nnc_lowerings_lazy_registration()
/aosp_15_r20/external/pytorch/test/quantization/jit/
H A Dtest_quantize_jit.py1699 # and conv2d_prepack is folded
/aosp_15_r20/external/pytorch/test/quantization/fx/
H A Dtest_quantize_fx.py1099 ns.call_function(torch.ops.quantized.conv2d_prepack),
/aosp_15_r20/external/pytorch/aten/src/ATen/test/
H A Dvulkan_api_test.cpp1667 TEST_F(VulkanAPITest, conv2d_prepack) { in TEST_F() argument