/aosp_15_r20/external/pytorch/benchmarks/functional_autograd_benchmark/ |
H A D | vision_models.py | 74 num_encoder_layers = 6 81 num_encoder_layers=num_encoder_layers,
|
H A D | torchvision_models.py | 502 num_encoder_layers=6, argument 516 hidden_dim, nheads, num_encoder_layers, num_decoder_layers
|
/aosp_15_r20/external/pytorch/torch/csrc/api/include/torch/nn/options/ |
H A D | transformer.h | 30 int64_t num_encoder_layers, 41 TORCH_ARG(int64_t, num_encoder_layers) = 6;
|
/aosp_15_r20/external/pytorch/torch/csrc/api/src/nn/options/ |
H A D | transformer.cpp | 44 int64_t num_encoder_layers, in TransformerOptions() argument 48 num_encoder_layers_(num_encoder_layers), in TransformerOptions()
|
/aosp_15_r20/external/executorch/exir/tests/ |
H A D | transformer.py | 87 def __init__(self, inp_vocab_size=10, model_dim=32, num_encoder_layers=2): argument 94 EncoderLayer(embed_dim=self.model_dim) for _ in range(num_encoder_layers)
|
/aosp_15_r20/external/pytorch/torch/nn/modules/ |
H A D | transformer.py | 103 num_encoder_layers: int = 6, 140 encoder_layer, num_encoder_layers, encoder_norm
|
/aosp_15_r20/external/pytorch/test/mobile/model_test/ |
H A D | quantization_ops.py | 76 d_model=2, nhead=2, num_encoder_layers=1, num_decoder_layers=1
|
H A D | nn_ops.py | 253 d_model=2, nhead=2, num_encoder_layers=1, num_decoder_layers=1
|
/aosp_15_r20/external/pytorch/test/cpp/api/ |
H A D | transformer.cpp | 1376 .num_encoder_layers(2) in transformer_test_helper() 1505 .num_encoder_layers(2) in TEST_F()
|
/aosp_15_r20/external/pytorch/torch/csrc/api/src/nn/modules/ |
H A D | transformer.cpp | 381 options.num_encoder_layers()) in reset()
|
/aosp_15_r20/external/pytorch/test/distributed/fsdp/ |
H A D | test_fsdp_flatten_params.py | 58 num_encoder_layers=2,
|
/aosp_15_r20/external/pytorch/torch/testing/_internal/ |
H A D | common_fsdp.py | 255 num_encoder_layers=2,
|
H A D | common_modules.py | 2579 num_encoder_layers=1, num_decoder_layers=1, 2601 num_encoder_layers=1, num_decoder_layers=1,
|
/aosp_15_r20/external/pytorch/test/nn/ |
H A D | test_module_hooks.py | 872 d_model=4, nhead=2, num_encoder_layers=2, num_decoder_layers=2
|
/aosp_15_r20/external/pytorch/test/ |
H A D | test_nn.py | 2821 num_encoder_layers = 4 2833 transformer = nn.Transformer(d_model, nhead, num_encoder_layers, num_decoder_layers, 3525 num_encoder_layers = 2 3548 model = getattr(nn, model_name)(d_model, nhead, num_encoder_layers, 3626 model = getattr(nn, model_name)(d_model, wrong_nhead, num_encoder_layers, 3671 model = getattr(nn, model_name)(d_model, nhead, num_encoder_layers, num_decoder_layers, 3675 model = getattr(nn, model_name)(d_model, nhead, num_encoder_layers, num_decoder_layers, 8755 … transformer_model = nn.Transformer(nhead=16, num_encoder_layers=12, dtype=torch.double).to(device)
|
H A D | test_jit.py | 15041 num_encoder_layers = 2 15051 transformer = nn.Transformer(d_model, nhead, num_encoder_layers,
|