File tree Expand file tree Collapse file tree 14 files changed +69
-49
lines changed Expand file tree Collapse file tree 14 files changed +69
-49
lines changed Original file line number Diff line number Diff line change @@ -72,7 +72,8 @@ torchtune provides the following finetuning recipes for training on one or more
7272| DoRA/QDoRA Finetuning | ✅ | ✅ | ❌ | [ lora_finetune_single_device] ( recipes/lora_finetune_single_device.py ) <br > [ lora_finetune_distributed] ( recipes/lora_finetune_distributed.py ) | [ Llama3 8B QDoRA single-device] ( recipes/configs/llama3/8B_qdora_single_device.yaml ) <br > [ Llama3 8B DoRA distributed] ( recipes/configs/llama3/8B_dora.yaml )
7373| Quantization-Aware Training | ❌ | ✅ | ❌ | [ qat_distributed] ( recipes/qat_distributed.py ) | [ Llama3 8B QAT] ( recipes/configs/llama3/8B_qat_full.yaml )
7474| Quantization-Aware Training and LoRA Finetuning | ❌ | ✅ | ❌ | [ qat_lora_finetune_distributed] ( recipes/qat_lora_finetune_distributed.py ) | [ Llama3 8B QAT] ( recipes/configs/llama3/8B_qat_lora.yaml )
75- | Direct Preference Optimization | ✅ | ✅ | ❌ | [ lora_dpo_single_device] ( recipes/lora_dpo_single_device.py ) <br > [ lora_dpo_distributed] ( recipes/lora_dpo_distributed.py ) | [ Llama2 7B single-device] ( recipes/configs/llama2/7B_lora_dpo_single_device.yaml ) <br > [ Llama2 7B distributed] ( recipes/configs/llama2/7B_lora_dpo.yaml )
75+ | Direct Preference Optimization: Full Finetuning | ❌ | ✅ | ❌ | [ full_dpo_distributed] ( recipes/full_dpo_distributed.py ) | [ Llama3.1 8B DPO] ( recipes/configs/llama3_1/8B_full_dpo.yaml )
76+ | LoRA Direct Preference Optimization | ✅ | ✅ | ❌ | [ lora_dpo_single_device] ( recipes/lora_dpo_single_device.py ) <br > [ lora_dpo_distributed] ( recipes/lora_dpo_distributed.py ) | [ Llama3.1 8B single-device] ( recipes/configs/llama3_1/8B_lora_dpo_single_device.yaml ) <br > [ Llama3.1 8B distributed] ( recipes/configs/llama3_1/8B_lora_dpo.yaml )
7677| Proximal Policy Optimization | ✅ | ❌ | ❌ | [ ppo_full_finetune_single_device] ( recipes/ppo_full_finetune_single_device.py ) | [ Mistral 7B] ( recipes/configs/mistral/7B_full_ppo_low_memory.yaml )
7778| LoRA Knowledge Distillation | ✅ | ✅ | ❌ | [ knowledge_distillation_single_device] ( recipes/knowledge_distillation_single_device.py ) <br > [ knowledge_distillation_distributed] ( recipes/knowledge_distillation_distributed.py ) | [ Qwen2 1.5B -> 0.5B single-device] ( recipes/configs/qwen2/1.5B_to_0.5B_KD_lora_single_device.yaml ) <br > [ Qwen2 1.5B -> 0.5B distributed] ( recipes/configs/qwen2/1.5B_to_0.5B_KD_lora_distributed.yaml )
7879
Original file line number Diff line number Diff line change 33#
44# This config assumes that you've run the following command before launching
55# this run:
6- # tune download Qwen/Qwen2.5-14B-Instruct --output-dir /tmp/Qwen2_5 -14B-Instruct
6+ # tune download Qwen/Qwen2.5-14B-Instruct --output-dir /tmp/Qwen2.5 -14B-Instruct
77#
88# To launch on a single device, run the following command from root:
99# tune run lora_finetune_single_device --config qwen2_5/14B_lora_single_device
@@ -30,13 +30,13 @@ model:
3030
3131tokenizer :
3232 _component_ : torchtune.models.qwen2_5.qwen2_5_tokenizer
33- path : /tmp/Qwen2_5 -14B-Instruct/vocab.json
34- merges_file : /tmp/Qwen2_5 -14B-Instruct/merges.txt
33+ path : /tmp/Qwen2.5 -14B-Instruct/vocab.json
34+ merges_file : /tmp/Qwen2.5 -14B-Instruct/merges.txt
3535 max_seq_len : null
3636
3737checkpointer :
3838 _component_ : torchtune.training.FullModelHFCheckpointer
39- checkpoint_dir : /tmp/Qwen2_5 -14B-Instruct
39+ checkpoint_dir : /tmp/Qwen2.5 -14B-Instruct
4040 checkpoint_files :
4141 filename_format : model-{}-of-{}.safetensors
4242 max_filename : " 00008"
Original file line number Diff line number Diff line change 33#
44# This config assumes that you've run the following command before launching
55# this run:
6- # tune download Qwen/Qwen2.5-32B-Instruct --output-dir /tmp/Qwen2_5 -32B-Instruct
6+ # tune download Qwen/Qwen2.5-32B-Instruct --output-dir /tmp/Qwen2.5 -32B-Instruct
77#
88# To launch on 8 devices, run the following command from root:
99# tune run --nnodes 1 --nproc_per_node 8 lora_finetune_distributed --config qwen2_5/32B_lora
@@ -28,13 +28,13 @@ model:
2828
2929tokenizer :
3030 _component_ : torchtune.models.qwen2_5.qwen2_5_tokenizer
31- path : /tmp/Qwen2_5 -32B-Instruct/vocab.json
32- merges_file : /tmp/Qwen2_5 -32B-Instruct/merges.txt
31+ path : /tmp/Qwen2.5 -32B-Instruct/vocab.json
32+ merges_file : /tmp/Qwen2.5 -32B-Instruct/merges.txt
3333 max_seq_len : null
3434
3535checkpointer :
3636 _component_ : torchtune.training.FullModelHFCheckpointer
37- checkpoint_dir : /tmp/Qwen2_5 -32B-Instruct
37+ checkpoint_dir : /tmp/Qwen2.5 -32B-Instruct
3838 checkpoint_files :
3939 filename_format : model-{}-of-{}.safetensors
4040 max_filename : " 00017"
Original file line number Diff line number Diff line change 33#
44# This config assumes that you've run the following command before launching
55# this run:
6- # tune download Qwen/Qwen2.5-3B-Instruct --output-dir /tmp/Qwen2_5 -3B-Instruct
6+ # tune download Qwen/Qwen2.5-3B-Instruct --output-dir /tmp/Qwen2.5 -3B-Instruct
77#
88# To launch on 2 devices, run the following command from root:
99# tune run --nnodes 1 --nproc_per_node 2 full_finetune_distributed --config qwen2_5/3B_full
@@ -22,8 +22,8 @@ output_dir: /tmp/torchtune/qwen2_5_3B/full # /tmp may be deleted by your system.
2222# Tokenizer
2323tokenizer :
2424 _component_ : torchtune.models.qwen2_5.qwen2_5_tokenizer
25- path : /tmp/Qwen2_5 -3B-Instruct/vocab.json
26- merges_file : /tmp/Qwen2_5 -3B-Instruct/merges.txt
25+ path : /tmp/Qwen2.5 -3B-Instruct/vocab.json
26+ merges_file : /tmp/Qwen2.5 -3B-Instruct/merges.txt
2727 max_seq_len : null
2828
2929# Dataset
3939
4040checkpointer :
4141 _component_ : torchtune.training.FullModelHFCheckpointer
42- checkpoint_dir : /tmp/Qwen2_5 -3B-Instruct
42+ checkpoint_dir : /tmp/Qwen2.5 -3B-Instruct
4343 checkpoint_files : [
4444 model-00001-of-00002.safetensors,
4545 model-00002-of-00002.safetensors,
Original file line number Diff line number Diff line change 33#
44# This config assumes that you've run the following command before launching
55# this run:
6- # tune download Qwen/Qwen2.5-3B-Instruct --output-dir /tmp/Qwen2_5 -3B-Instruct
6+ # tune download Qwen/Qwen2.5-3B-Instruct --output-dir /tmp/Qwen2.5 -3B-Instruct
77#
88# The default config uses an optimizer from bitsandbytes. If you do not have it installed,
99# you can install it with
@@ -24,8 +24,8 @@ output_dir: /tmp/torchtune/qwen2_5_3B/full_single_device # /tmp may be deleted b
2424# Tokenizer
2525tokenizer :
2626 _component_ : torchtune.models.qwen2_5.qwen2_5_tokenizer
27- path : /tmp/Qwen2_5 -3B-Instruct/vocab.json
28- merges_file : /tmp/Qwen2_5 -3B-Instruct/merges.txt
27+ path : /tmp/Qwen2.5 -3B-Instruct/vocab.json
28+ merges_file : /tmp/Qwen2.5 -3B-Instruct/merges.txt
2929 max_seq_len : null
3030
3131# Dataset
4141
4242checkpointer :
4343 _component_ : torchtune.training.FullModelHFCheckpointer
44- checkpoint_dir : /tmp/Qwen2_5 -3B-Instruct
44+ checkpoint_dir : /tmp/Qwen2.5 -3B-Instruct
4545 checkpoint_files : [
4646 model-00001-of-00002.safetensors,
4747 model-00002-of-00002.safetensors,
Original file line number Diff line number Diff line change 33#
44# This config assumes that you've run the following command before launching
55# this run:
6- # tune download Qwen/Qwen2.5-3B-Instruct --output-dir /tmp/Qwen2_5 -3B-Instruct
6+ # tune download Qwen/Qwen2.5-3B-Instruct --output-dir /tmp/Qwen2.5 -3B-Instruct
77#
88# To launch on 2 devices, run the following command from root:
99# tune run --nnodes 1 --nproc_per_node 2 lora_finetune_distributed --config qwen2_5/3B_lora
@@ -30,13 +30,13 @@ model:
3030
3131tokenizer :
3232 _component_ : torchtune.models.qwen2_5.qwen2_5_tokenizer
33- path : /tmp/Qwen2_5 -3B-Instruct/vocab.json
34- merges_file : /tmp/Qwen2_5 -3B-Instruct/merges.txt
33+ path : /tmp/Qwen2.5 -3B-Instruct/vocab.json
34+ merges_file : /tmp/Qwen2.5 -3B-Instruct/merges.txt
3535 max_seq_len : null
3636
3737checkpointer :
3838 _component_ : torchtune.training.FullModelHFCheckpointer
39- checkpoint_dir : /tmp/Qwen2_5 -3B-Instruct
39+ checkpoint_dir : /tmp/Qwen2.5 -3B-Instruct
4040 checkpoint_files : [
4141 model-00001-of-00002.safetensors,
4242 model-00002-of-00002.safetensors,
Original file line number Diff line number Diff line change 33#
44# This config assumes that you've run the following command before launching
55# this run:
6- # tune download Qwen/Qwen2.5-3B-Instruct --output-dir /tmp/Qwen2_5 -3B-Instruct
6+ # tune download Qwen/Qwen2.5-3B-Instruct --output-dir /tmp/Qwen2.5 -3B-Instruct
77#
88# To launch on a single device, run the following command from root:
99# tune run lora_finetune_single_device --config qwen2_5/3B_lora_single_device
@@ -29,13 +29,13 @@ model:
2929
3030tokenizer :
3131 _component_ : torchtune.models.qwen2_5.qwen2_5_tokenizer
32- path : /tmp/Qwen2_5 -3B-Instruct/vocab.json
33- merges_file : /tmp/Qwen2_5 -3B-Instruct/merges.txt
32+ path : /tmp/Qwen2.5 -3B-Instruct/vocab.json
33+ merges_file : /tmp/Qwen2.5 -3B-Instruct/merges.txt
3434 max_seq_len : null
3535
3636checkpointer :
3737 _component_ : torchtune.training.FullModelHFCheckpointer
38- checkpoint_dir : /tmp/Qwen2_5 -3B-Instruct
38+ checkpoint_dir : /tmp/Qwen2.5 -3B-Instruct
3939 checkpoint_files : [
4040 model-00001-of-00002.safetensors,
4141 model-00002-of-00002.safetensors,
Original file line number Diff line number Diff line change 33#
44# This config assumes that you've run the following command before launching
55# this run:
6- # tune download Qwen/Qwen2.5-72B-Instruct --output-dir /tmp/Qwen2_5 -72B-Instruct
6+ # tune download Qwen/Qwen2.5-72B-Instruct --output-dir /tmp/Qwen2.5 -72B-Instruct
77#
88# To launch on 8 devices, run the following command from root:
99# tune run --nnodes 1 --nproc_per_node 8 lora_finetune_distributed --config qwen2_5/72B_lora
@@ -28,13 +28,13 @@ model:
2828
2929tokenizer :
3030 _component_ : torchtune.models.qwen2_5.qwen2_5_tokenizer
31- path : /tmp/Qwen2_5 -72B-Instruct/vocab.json
32- merges_file : /tmp/Qwen2_5 -72B-Instruct/merges.txt
31+ path : /tmp/Qwen2.5 -72B-Instruct/vocab.json
32+ merges_file : /tmp/Qwen2.5 -72B-Instruct/merges.txt
3333 max_seq_len : null
3434
3535checkpointer :
3636 _component_ : torchtune.training.FullModelHFCheckpointer
37- checkpoint_dir : /tmp/Qwen2_5 -72B-Instruct
37+ checkpoint_dir : /tmp/Qwen2.5 -72B-Instruct
3838 checkpoint_files :
3939 filename_format : model-{}-of-{}.safetensors
4040 max_filename : " 00037"
Original file line number Diff line number Diff line change 33#
44# This config assumes that you've run the following command before launching
55# this run:
6- # tune download Qwen/Qwen2.5-7B-Instruct --output-dir /tmp/Qwen2_5 -7B-Instruct
6+ # tune download Qwen/Qwen2.5-7B-Instruct --output-dir /tmp/Qwen2.5 -7B-Instruct
77#
88# To launch on 2 devices, run the following command from root:
99# tune run --nnodes 1 --nproc_per_node 2 full_finetune_distributed --config qwen2_5/7B_full
@@ -22,8 +22,8 @@ output_dir: /tmp/torchtune/qwen2_5_7B/full # /tmp may be deleted by your system.
2222# Tokenizer
2323tokenizer :
2424 _component_ : torchtune.models.qwen2_5.qwen2_5_tokenizer
25- path : /tmp/Qwen2_5 -7B-Instruct/vocab.json
26- merges_file : /tmp/Qwen2_5 -7B-Instruct/merges.txt
25+ path : /tmp/Qwen2.5 -7B-Instruct/vocab.json
26+ merges_file : /tmp/Qwen2.5 -7B-Instruct/merges.txt
2727 max_seq_len : null
2828
2929# Dataset
3939
4040checkpointer :
4141 _component_ : torchtune.training.FullModelHFCheckpointer
42- checkpoint_dir : /tmp/Qwen2_5 -7B-Instruct
42+ checkpoint_dir : /tmp/Qwen2.5 -7B-Instruct
4343 checkpoint_files : [
4444 model-00001-of-00004.safetensors,
4545 model-00002-of-00004.safetensors,
Original file line number Diff line number Diff line change 33#
44# This config assumes that you've run the following command before launching
55# this run:
6- # tune download Qwen/Qwen2.5-7B-Instruct --output-dir /tmp/Qwen2_5 -7B-Instruct
6+ # tune download Qwen/Qwen2.5-7B-Instruct --output-dir /tmp/Qwen2.5 -7B-Instruct
77#
88# The default config uses an optimizer from bitsandbytes. If you do not have it installed,
99# you can install it with
@@ -24,8 +24,8 @@ output_dir: /tmp/torchtune/qwen2_5_7B/full_single_device # /tmp may be deleted b
2424# Tokenizer
2525tokenizer :
2626 _component_ : torchtune.models.qwen2_5.qwen2_5_tokenizer
27- path : /tmp/Qwen2_5 -7B-Instruct/vocab.json
28- merges_file : /tmp/Qwen2_5 -7B-Instruct/merges.txt
27+ path : /tmp/Qwen2.5 -7B-Instruct/vocab.json
28+ merges_file : /tmp/Qwen2.5 -7B-Instruct/merges.txt
2929 max_seq_len : null
3030
3131# Dataset
4141
4242checkpointer :
4343 _component_ : torchtune.training.FullModelHFCheckpointer
44- checkpoint_dir : /tmp/Qwen2_5 -7B-Instruct
44+ checkpoint_dir : /tmp/Qwen2.5 -7B-Instruct
4545 checkpoint_files : [
4646 model-00001-of-00004.safetensors,
4747 model-00002-of-00004.safetensors,
You can’t perform that action at this time.
0 commit comments