Skip to content

Commit

Permalink
Merge Gemma recipe with full finetune (#668)
Browse files Browse the repository at this point in the history
  • Loading branch information
RdoubleA authored Apr 15, 2024
1 parent ff594c2 commit 3f93b25
Show file tree
Hide file tree
Showing 11 changed files with 14 additions and 572 deletions.
8 changes: 1 addition & 7 deletions recipes/configs/gemma/2B_full.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
#
# This config assumes that you've run the following command before launching
# this run:
# tune download google/gemma-2b --hf-token <HF_TOKEN> --output-dir /tmp/gemma2
# tune download google/gemma-2b --hf-token <HF_TOKEN> --output-dir /tmp/gemma --ignore-patterns ""
#
# To launch on 4 devices, run the following command from root:
# tune run --nnodes 1 --nproc_per_node 4 full_finetune_distributed --config gemma/2B_full
Expand Down Expand Up @@ -42,11 +42,6 @@ checkpointer:
recipe_checkpoint: null
output_dir: /tmp/gemma
model_type: GEMMA
share_weights:
share_weights: True
weight_tying_config: {
"output": "tok_embeddings"
}
resume_from_checkpoint: False

# Fine-tuning arguments
Expand All @@ -60,7 +55,6 @@ loss:
max_steps_per_epoch: null
gradient_accumulation_steps: 1


# Training env
device: cuda

Expand Down
Loading

0 comments on commit 3f93b25

Please sign in to comment.