Skip to content

Commit 0e37b9a

Browse files
authored
Fix dangling open skip in README.md (#1299)
1 - Fix an extraneous skip end that is out of order with a skip begin. 2 - fix some typos PS: This might cause some README tests to fail, as they have not been run in a long time.
1 parent ecc628d commit 0e37b9a

File tree

1 file changed

+7
-6
lines changed

1 file changed

+7
-6
lines changed

README.md

+7-6
Original file line numberDiff line numberDiff line change
@@ -171,7 +171,7 @@ python3 torchchat.py download llama3.1
171171
<summary>Additional Model Inventory Management Commands</summary>
172172

173173
### Where
174-
This subcommand shows location of a particular model.
174+
This subcommand shows the location of a particular model.
175175
```bash
176176
python3 torchchat.py where llama3.1
177177
```
@@ -216,7 +216,6 @@ This mode generates text based on an input prompt.
216216
python3 torchchat.py generate llama3.1 --prompt "write me a story about a boy and his bear"
217217
```
218218

219-
[skip default]: end
220219

221220
### Server
222221
This mode exposes a REST API for interacting with a model.
@@ -286,14 +285,16 @@ First, follow the steps in the Server section above to start a local server. The
286285
streamlit run torchchat/usages/browser.py
287286
```
288287

288+
[skip default]: end
289+
289290
Use the "Max Response Tokens" slider to limit the maximum number of tokens generated by the model for each response. Click the "Reset Chat" button to remove the message history and start a fresh chat.
290291

291292

292293
## Desktop/Server Execution
293294

294295
### AOTI (AOT Inductor)
295296
[AOTI](https://pytorch.org/blog/pytorch2-2/) compiles models before execution for faster inference. The process creates a [DSO](https://en.wikipedia.org/wiki/Shared_library) model (represented by a file with extension `.so`)
296-
that is then loaded for inference. This can be done with both Python and C++ enviroments.
297+
that is then loaded for inference. This can be done with both Python and C++ environments.
297298

298299
The following example exports and executes the Llama3.1 8B Instruct
299300
model. The first command compiles and performs the actual export.
@@ -308,9 +309,9 @@ python3 torchchat.py export llama3.1 --output-dso-path exportedModels/llama3.1.s
308309
For more details on quantization and what settings to use for your use
309310
case visit our [customization guide](docs/model_customization.md).
310311

311-
### Run in a Python Enviroment
312+
### Run in a Python Environment
312313

313-
To run in a python enviroment, use the generate subcommand like before, but include the dso file.
314+
To run in a python environment, use the generate subcommand like before, but include the dso file.
314315

315316
```
316317
python3 torchchat.py generate llama3.1 --dso-path exportedModels/llama3.1.so --prompt "Hello my name is"
@@ -377,7 +378,7 @@ While ExecuTorch does not focus on desktop inference, it is capable
377378
of doing so. This is handy for testing out PTE
378379
models without sending them to a physical device.
379380

380-
Specifically there are 2 ways of doing so: Pure Python and via a Runner
381+
Specifically, there are 2 ways of doing so: Pure Python and via a Runner
381382

382383
<details>
383384
<summary>Deploying via Python</summary>

0 commit comments

Comments
 (0)