You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* refactoring the readme
* continued refining
* more cleanup
* more cleanup
* more cleanup
* more cleanup
* more cleanup
* more refining
* Update README.md
Update README.md
* move the discaimer down
* remove torchtune from main readme
Fix pathing issues for runner commands
* don't use pybindings for et setup
---------
Co-authored-by: Michael Gschwind <[email protected]>
torchchat supports running inference with models fine-tuned using [torchtune](https://github.com/pytorch/torchtune). To do so, we first need to convert the checkpoints into a format supported by torchchat.
4
+
5
+
Below is a simple workflow to run inference on a fine-tuned Llama3 model. For more details on how to fine-tune Llama3, see the instructions [here](https://github.com/pytorch/torchtune?tab=readme-ov-file#llama3)
6
+
7
+
```bash
8
+
# install torchtune
9
+
pip install torchtune
10
+
11
+
# download the llama3 model
12
+
tune download meta-llama/Meta-Llama-3-8B \
13
+
--output-dir ./Meta-Llama-3-8B \
14
+
--hf-token <ACCESS TOKEN>
15
+
16
+
# Run LoRA fine-tuning on a single device. This assumes the config points to <checkpoint_dir> above
17
+
tune run lora_finetune_single_device --config llama3/8B_lora_single_device
18
+
19
+
# convert the fine-tuned checkpoint to a format compatible with torchchat
0 commit comments