mtasic85 commited on
Commit
d26da25
·
1 Parent(s): e15928d

pretrain core dataset

Browse files
Files changed (1) hide show
  1. scripts/pretrain-core-model.yaml +1 -1
scripts/pretrain-core-model.yaml CHANGED
@@ -61,7 +61,7 @@ train:
61
  global_batch_size: 512
62
 
63
  # Number of samples per data-parallel rank (type: int, default: 4)
64
- micro_batch_size: 4
65
 
66
  # Number of iterations with learning rate warmup active (type: int, default: 2000)
67
  lr_warmup_steps: 500
 
61
  global_batch_size: 512
62
 
63
  # Number of samples per data-parallel rank (type: int, default: 4)
64
+ micro_batch_size: 2
65
 
66
  # Number of iterations with learning rate warmup active (type: int, default: 2000)
67
  lr_warmup_steps: 500