You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* transfer files in DockerGuides from rst to md
* add some dividing lines
* adjust the title hierarchy in docker_cpp_xpu_quickstart.md
* restore
* switch to the correct branch
* small change
---------
Co-authored-by: ATMxsp01 <[email protected]>
Copy file name to clipboardExpand all lines: docs/mddocs/DockerGuides/docker_cpp_xpu_quickstart.md
+29-26
Original file line number
Diff line number
Diff line change
@@ -1,16 +1,16 @@
1
-
##Run llama.cpp/Ollama/Open-WebUI on an Intel GPU via Docker
1
+
# Run llama.cpp/Ollama/Open-WebUI on an Intel GPU via Docker
2
2
3
3
## Quick Start
4
4
5
5
### Install Docker
6
6
7
7
1. Linux Installation
8
8
9
-
Follow the instructions in this [guide](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/DockerGuides/docker_windows_gpu.html#linux) to install Docker on Linux.
9
+
Follow the instructions in this [guide](./docker_windows_gpu.md#linux) to install Docker on Linux.
10
10
11
11
2. Windows Installation
12
12
13
-
For Windows installation, refer to this [guide](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/DockerGuides/docker_windows_gpu.html#install-docker-desktop-for-windows).
13
+
For Windows installation, refer to this [guide](./docker_windows_gpu.md#install-docker-desktop-for-windows).
Choose one of the following methods to start the container:
30
28
31
-
To map the `xpu` into the container, you need to specify `--device=/dev/dri` when booting the container. Select the device you are running(device type:(Max, Flex, Arc, iGPU)). And change the `/path/to/models` to mount the models. `bench_model` is used to benchmark quickly. If want to benchmark, make sure it on the `/path/to/models`
29
+
<details>
30
+
<Summary>For <strong>Linux</strong>:</summary>
32
31
33
-
.. code-block:: bash
32
+
To map the `xpu` into the container, you need to specify `--device=/dev/dri` when booting the container. Select the device you are running(device type:(Max, Flex, Arc, iGPU)). And change the `/path/to/models` to mount the models. `bench_model` is used to benchmark quickly. If want to benchmark, make sure it on the `/path/to/models`
To map the `xpu` into the container, you need to specify `--device=/dev/dri` when booting the container. And change the `/path/to/models` to mount the models. Then add `--privileged` and map the `/usr/lib/wsl` to the docker.
52
+
<details>
53
+
<summary>For <strong>Windows</strong>:</summary>
53
54
54
-
.. code-block:: bash
55
+
To map the `xpu` into the container, you need to specify `--device=/dev/dri` when booting the container. And change the `/path/to/models` to mount the models. Then add `--privileged` and map the `/usr/lib/wsl` to the docker.
After the container is booted, you could get into the container through `docker exec`.
77
80
@@ -126,7 +129,7 @@ llama_print_timings: eval time = xxx ms / 31 runs ( xxx ms per
126
129
llama_print_timings: total time = xxx ms / xxx tokens
127
130
```
128
131
129
-
Please refer to this [documentation](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/llama_cpp_quickstart.html) for more details.
132
+
Please refer to this [documentation](../Quickstart/llama_cpp_quickstart.md) for more details.
130
133
131
134
132
135
### Running Ollama serving with IPEX-LLM on Intel GPU
@@ -194,13 +197,13 @@ Sample output:
194
197
```
195
198
196
199
197
-
Please refer to this [documentation](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/ollama_quickstart.html#pull-model) for more details.
200
+
Please refer to this [documentation](../Quickstart/ollama_quickstart.md#4-pull-model) for more details.
198
201
199
202
200
203
### Running Open WebUI with Intel GPU
201
204
202
205
Start the ollama and load the model first, then use the open-webui to chat.
203
-
If you have difficulty accessing the huggingface repositories, you may use a mirror, e.g. add export HF_ENDPOINT=https://hf-mirror.combefore running bash start.sh.
206
+
If you have difficulty accessing the huggingface repositories, you may use a mirror, e.g. add `export HF_ENDPOINT=https://hf-mirror.com`before running bash start.sh.
204
207
```bash
205
208
cd /llm/scripts/
206
209
bash start-open-webui.sh
@@ -218,4 +221,4 @@ INFO: Uvicorn running on http://0.0.0.0:8080 (Press CTRL+C to quit)
For how to log-in or other guide, Please refer to this [documentation](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/open_webui_with_ollama_quickstart.html) for more details.
224
+
For how to log-in or other guide, Please refer to this [documentation](../Quickstart/open_webui_with_ollama_quickstart.md) for more details.
Copy file name to clipboardExpand all lines: docs/mddocs/DockerGuides/docker_pytorch_inference_gpu.md
+31-38
Original file line number
Diff line number
Diff line change
@@ -2,16 +2,12 @@
2
2
3
3
We can run PyTorch Inference Benchmark, Chat Service and PyTorch Examples on Intel GPUs within Docker (on Linux or WSL).
4
4
5
-
```eval_rst
6
-
.. note::
7
-
8
-
The current Windows + WSL + Docker solution only supports Arc series dGPU. For Windows users with MTL iGPU, it is recommended to install directly via pip install in Miniforge Prompt. Refer to `this guide <https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html>`_.
9
-
10
-
```
5
+
> [!NOTE]
6
+
> The current Windows + WSL + Docker solution only supports Arc series dGPU. For Windows users with MTL iGPU, it is recommended to install directly via pip install in Miniforge Prompt. Refer to [this guide](../Quickstart/install_windows_gpu.md).
11
7
12
8
## Install Docker
13
9
14
-
Follow the [Docker installation Guide](./docker_windows_gpu.html#install-docker) to install docker on either Linux or Windows.
10
+
Follow the [Docker installation Guide](./docker_windows_gpu.md#install-docker) to install docker on either Linux or Windows.
0 commit comments