diff --git a/config/llama_7b_with_olmo_config.yaml b/config/llama_7b_with_olmo_config.yaml index 0b5bc4067..9864a000f 100644 --- a/config/llama_7b_with_olmo_config.yaml +++ b/config/llama_7b_with_olmo_config.yaml @@ -15,10 +15,6 @@ trainer: project: "marin" tags: ["dolma", "olmo", "llama"] - checkpointer: - keep: - - every: 250 - mp: p=f32,c=bfloat16 train_batch_size: 2048 num_train_steps: 750000 # 3,000,000,000,000 / 4,000,000 = 750,000 @@ -31,5 +27,3 @@ optimizer: weight_decay: 0.1 min_lr_ratio: 0.1 warmup: 0.01 - -data_shuffle: true