Searched refs:use_cpp_packed_gemm_template (Results 1 – 4 of 4) sorted by relevance
/aosp_15_r20/external/pytorch/torch/_inductor/ |
H A D | quantized_lowerings.py | 16 from .utils import use_aten_gemm_kernels, use_cpp_packed_gemm_template 71 if use_cpp_packed_gemm_template(aten_layout, mat1, mat2, mat2_transposed=True):
|
H A D | mkldnn_lowerings.py | 28 from .utils import use_aten_gemm_kernels, use_cpp_packed_gemm_template, use_max_autotune 185 if use_cpp_packed_gemm_template(layout, x, transposed_w): 250 if use_cpp_packed_gemm_template(layout, x, transposed_w): 543 and use_cpp_packed_gemm_template(layout, x, packed_weight) 822 and use_cpp_packed_gemm_template(layout, x, packed_weight) 1050 if use_cpp_packed_gemm_template(layout, x, transposed_w):
|
H A D | utils.py | 1215 def use_cpp_packed_gemm_template(layout, mat1, mat2, mat2_transposed=False): function
|
/aosp_15_r20/external/pytorch/torch/_inductor/kernel/ |
H A D | mm.py | 36 use_cpp_packed_gemm_template, 188 if use_cpp_packed_gemm_template(layout, mat1, mat2): 413 if use_cpp_packed_gemm_template(layout, mat1, mat2):
|