Skip to content

Commit c253bb4

Browse files
huvunvidiaartbataev
authored andcommitted
Update megatron_t5_pretraining.py (NVIDIA-NeMo#10952)
Signed-off-by: Huy Vu <86480512+huvunvidia@users.noreply.github.com>
1 parent 714c826 commit c253bb4

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

tests/collections/llm/megatron_t5_pretraining.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -59,16 +59,16 @@ def get_args():
5959
paths=args.data_path,
6060
seq_length=512,
6161
seq_length_dec=128,
62-
micro_batch_size=args.devices,
63-
global_batch_size=2 * args.devices,
62+
micro_batch_size=64,
63+
global_batch_size=512,
6464
seed=1234,
6565
tokenizer=tokenizer,
6666
split="99982,9,9",
6767
index_mapping_dir=args.index_mapping_dir,
6868
)
6969
t5_config = llm.t5.model.t5.T5Config(
70-
num_layers=args.devices,
71-
encoder_num_layers=args.devices,
70+
num_layers=12,
71+
encoder_num_layers=12,
7272
hidden_size=768,
7373
ffn_hidden_size=3072,
7474
num_attention_heads=12,

0 commit comments

Comments
 (0)