Skip to content

Commit

Permalink
Merge pull request #196 from OpenAccess-AI-Collective/openllama-ft-co…
Browse files Browse the repository at this point in the history
…nfig

tweak config to work
  • Loading branch information
winglian authored Jun 13, 2023
2 parents 068fc48 + 2ba4ae8 commit 06652c1
Showing 1 changed file with 6 additions and 5 deletions.
11 changes: 6 additions & 5 deletions examples/openllama-3b/config.yml
Original file line number Diff line number Diff line change
Expand Up @@ -26,17 +26,18 @@ wandb_watch:
wandb_run_id:
wandb_log_model:
output_dir: ./openllama-out
batch_size: 16
micro_batch_size: 4
gradient_accumulation_steps: 1
micro_batch_size: 1
num_epochs: 3
optimizer: adamw_bnb_8bit
torchdistx_path:
lr_scheduler: cosine
learning_rate: 0.0002
learning_rate: 0.00001
train_on_inputs: false
group_by_length: false
float16: true
bf16: false
fp16: true
fp16: false
tf32: false
gradient_checkpointing: true
early_stopping_patience:
Expand All @@ -52,7 +53,7 @@ eval_steps: 50
save_steps:
debug:
deepspeed:
weight_decay: 0.0
weight_decay: 0.1
fsdp:
fsdp_config:
special_tokens:
Expand Down

0 comments on commit 06652c1

Please sign in to comment.