Skip to content

Commit

Permalink
address comments
Browse files Browse the repository at this point in the history
  • Loading branch information
ebsmothers committed Apr 18, 2024
1 parent a93368e commit ff63068
Show file tree
Hide file tree
Showing 6 changed files with 23 additions and 23 deletions.
6 changes: 3 additions & 3 deletions docs/source/tutorials/llama3.rst
Original file line number Diff line number Diff line change
Expand Up @@ -141,7 +141,7 @@ Next, we modify ``custom_eval_config.yaml`` to include the fine-tuned checkpoint
.. code-block:: yaml
checkpointer:
_component_: torchtune.utils.FullModelHFCheckpointer
_component_: torchtune.utils.FullModelMetaCheckpointer
# directory with the checkpoint files
# this should match the output_dir specified during
Expand Down Expand Up @@ -190,7 +190,7 @@ Now we modify ``custom_generation_config.yaml`` to point to our checkpoint and t
.. code-block:: yaml
checkpointer:
_component_: torchtune.utils.FullModelHFCheckpointer
_component_: torchtune.utils.FullModelMetaCheckpointer
# directory with the checkpoint files
# this should match the output_dir specified during
Expand Down Expand Up @@ -244,7 +244,7 @@ And update ``custom_quantization_config.yaml`` with the following:
.. code-block:: yaml
checkpointer:
_component_: torchtune.utils.FullModelHFCheckpointer
_component_: torchtune.utils.FullModelMetaCheckpointer
# directory with the checkpoint files
# this should match the output_dir specified during
Expand Down
8 changes: 4 additions & 4 deletions recipes/configs/llama3/8B_full.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
#
# This config assumes that you've run the following command before launching
# this run:
# tune download meta-llama/Meta-Llama-3-8B --output-dir /tmp/Llama-3-8b-hf --hf-token <HF_TOKEN>
# tune download meta-llama/Meta-Llama-3-8B --output-dir /tmp/Meta-Llama-3-8B --hf-token <HF_TOKEN>
#
# To launch on 4 devices, run the following command from root:
# tune run --nproc_per_node 4 full_finetune_distributed --config llama3/8B_full
Expand All @@ -21,7 +21,7 @@
# Tokenizer
tokenizer:
_component_: torchtune.models.llama3.llama3_tokenizer
path: /tmp/Llama-3-8b-hf/original/tokenizer.model
path: /tmp/Meta-Llama-3-8B/original/tokenizer.model

# Dataset
dataset:
Expand All @@ -36,12 +36,12 @@ model:

checkpointer:
_component_: torchtune.utils.FullModelMetaCheckpointer
checkpoint_dir: /tmp/Llama-3-8b-hf/original/
checkpoint_dir: /tmp/Meta-Llama-3-8B/original/
checkpoint_files: [
consolidated.00.pth
]
recipe_checkpoint: null
output_dir: /tmp/Llama-3-8b-hf/
output_dir: /tmp/Meta-Llama-3-8B/
model_type: LLAMA3
resume_from_checkpoint: False

Expand Down
8 changes: 4 additions & 4 deletions recipes/configs/llama3/8B_full_single_device.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
#
# This config assumes that you've run the following command before launching
# this run:
# tune download meta-llama/Meta-Llama-3-8B --output-dir /tmp/Llama-3-8b-hf --hf-token <HF_TOKEN>
# tune download meta-llama/Meta-Llama-3-8B --output-dir /tmp/Meta-Llama-3-8B --hf-token <HF_TOKEN>
#
# The default config uses an optimizer from bitsandbytes. If you do not have it installed,
# you can install it with
Expand All @@ -23,7 +23,7 @@
# Tokenizer
tokenizer:
_component_: torchtune.models.llama3.llama3_tokenizer
path: /tmp/Llama-3-8b-hf/original/tokenizer.model
path: /tmp/Meta-Llama-3-8B/original/tokenizer.model

# Dataset
dataset:
Expand All @@ -38,12 +38,12 @@ model:

checkpointer:
_component_: torchtune.utils.FullModelMetaCheckpointer
checkpoint_dir: /tmp/Llama-3-8b-hf/original/
checkpoint_dir: /tmp/Meta-Llama-3-8B/original/
checkpoint_files: [
consolidated.00.pth
]
recipe_checkpoint: null
output_dir: /tmp/Llama-3-8b-hf/
output_dir: /tmp/Meta-Llama-3-8B/
model_type: LLAMA3
resume_from_checkpoint: False

Expand Down
8 changes: 4 additions & 4 deletions recipes/configs/llama3/8B_lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
#
# This config assumes that you've run the following command before launching
# this run:
# tune download meta-llama/Meta-Llama-3-8B --output-dir /tmp/Llama-3-8b-hf --hf-token <HF_TOKEN>
# tune download meta-llama/Meta-Llama-3-8B --output-dir /tmp/Meta-Llama-3-8B --hf-token <HF_TOKEN>
#
# To launch on 2 devices, run the following command from root:
# tune run --nproc_per_node 2 lora_finetune_distributed --config llama3/8B_lora
Expand All @@ -20,7 +20,7 @@
# Tokenizer
tokenizer:
_component_: torchtune.models.llama3.llama3_tokenizer
path: /tmp/Llama-3-8b-hf/original/tokenizer.model
path: /tmp/Meta-Llama-3-8B/original/tokenizer.model

# Model Arguments
model:
Expand All @@ -33,12 +33,12 @@ model:

checkpointer:
_component_: torchtune.utils.FullModelMetaCheckpointer
checkpoint_dir: /tmp/Llama-3-8b-hf/original/
checkpoint_dir: /tmp/Meta-Llama-3-8B/original/
checkpoint_files: [
consolidated.00.pth
]
recipe_checkpoint: null
output_dir: /tmp/Llama-3-8b-hf/
output_dir: /tmp/Meta-Llama-3-8B/
model_type: LLAMA3
resume_from_checkpoint: False

Expand Down
8 changes: 4 additions & 4 deletions recipes/configs/llama3/8B_lora_single_device.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
#
# This config assumes that you've run the following command before launching
# this run:
# tune download meta-llama/Meta-Llama-3-8B --output-dir /tmp/Llama-3-8b-hf --hf-token <HF_TOKEN>
# tune download meta-llama/Meta-Llama-3-8B --output-dir /tmp/Meta-Llama-3-8B --hf-token <HF_TOKEN>
#
# To launch on a single device, run the following command from root:
# tune run lora_finetune_single_device --config llama3/8B_lora_single_device
Expand All @@ -28,16 +28,16 @@ model:
# Tokenizer
tokenizer:
_component_: torchtune.models.llama3.llama3_tokenizer
path: /tmp/Llama-3-8b-hf/original/tokenizer.model
path: /tmp/Meta-Llama-3-8B/original/tokenizer.model

checkpointer:
_component_: torchtune.utils.FullModelMetaCheckpointer
checkpoint_dir: /tmp/Llama-3-8b-hf/original/
checkpoint_dir: /tmp/Meta-Llama-3-8B/original/
checkpoint_files: [
consolidated.00.pth
]
recipe_checkpoint: null
output_dir: /tmp/Llama-3-8b-hf/
output_dir: /tmp/Meta-Llama-3-8B/
model_type: LLAMA3
resume_from_checkpoint: False

Expand Down
8 changes: 4 additions & 4 deletions recipes/configs/llama3/8B_qlora_single_device.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
#
# This config assumes that you've run the following command before launching
# this run:
# tune download meta-llama/Meta-Llama-3-8B --output-dir /tmp/Llama-3-8b-hf --hf-token <HF_TOKEN>
# tune download meta-llama/Meta-Llama-3-8B --output-dir /tmp/Meta-Llama-3-8B --hf-token <HF_TOKEN>
#
# To launch on a single device, run the following command from root:
# tune run lora_finetune_single_device --config llama3/8B_qlora_single_device
Expand All @@ -27,16 +27,16 @@ model:
# Tokenizer
tokenizer:
_component_: torchtune.models.llama3.llama3_tokenizer
path: /tmp/Llama-3-8b-hf/original/tokenizer.model
path: /tmp/Meta-Llama-3-8B/original/tokenizer.model

checkpointer:
_component_: torchtune.utils.FullModelMetaCheckpointer
checkpoint_dir: /tmp/Llama-3-8b-hf/original/
checkpoint_dir: /tmp/Meta-Llama-3-8B/original/
checkpoint_files: [
consolidated.00.pth
]
recipe_checkpoint: null
output_dir: /tmp/Llama-3-8b-hf/
output_dir: /tmp/Meta-Llama-3-8B/
model_type: LLAMA3
resume_from_checkpoint: False

Expand Down

0 comments on commit ff63068

Please sign in to comment.