mtasic85 commited on
Commit
49d664b
·
1 Parent(s): e9901c9

pretrain core 4

Browse files
Files changed (1) hide show
  1. scripts/pretrain_core_model_4.yaml +2 -2
scripts/pretrain_core_model_4.yaml CHANGED
@@ -60,10 +60,10 @@ train:
60
  # Number of samples between optimizer steps across data-parallel ranks (type: int, default: 512)
61
  # global_batch_size: 512
62
  # global_batch_size: 256
63
- global_batch_size: 32
64
 
65
  # Number of samples per data-parallel rank (type: int, default: 4)
66
- micro_batch_size: 4
67
 
68
  # Number of iterations with learning rate warmup active (type: int, default: 2000)
69
  lr_warmup_steps: 0
 
60
  # Number of samples between optimizer steps across data-parallel ranks (type: int, default: 512)
61
  # global_batch_size: 512
62
  # global_batch_size: 256
63
+ global_batch_size: 16
64
 
65
  # Number of samples per data-parallel rank (type: int, default: 4)
66
+ micro_batch_size: 1
67
 
68
  # Number of iterations with learning rate warmup active (type: int, default: 2000)
69
  lr_warmup_steps: 0