forked from intel-analytics/ipex-llm
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
update the README and reorganize the docker guides structure. (intel-…
…analytics#11016) * update the README and reorganize the docker guides structure. * modified docker install guide into overview
- Loading branch information
1 parent
8931974
commit 586a151
Showing
8 changed files
with
897 additions
and
828 deletions.
There are no files selected for viewing
Large diffs are not rendered by default.
Oops, something went wrong.
Large diffs are not rendered by default.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
115 changes: 115 additions & 0 deletions
115
docs/readthedocs/source/doc/LLM/Docker/docker_windows_gpu.md
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,115 @@ | ||
# Overview of IPEX-LLM Containers for Intel GPU | ||
|
||
|
||
An IPEX-LLM container is a pre-configured environment that includes all necessary dependencies for running LLMs on Intel GPUs. | ||
|
||
This guide provides general instructions for setting up the IPEX-LLM Docker containers with Intel GPU. It begins with instructions and tips for Docker installation, and then introduce the available IPEX-LLM containers and their uses. | ||
|
||
## Install Docker | ||
|
||
### Linux | ||
|
||
Follow the instructions in the [Offcial Docker Guide](https://www.docker.com/get-started/) to install Docker on Linux. | ||
|
||
|
||
### Windows | ||
|
||
```eval_rst | ||
.. tip:: | ||
The installation requires at least 35GB of free disk space on C drive. | ||
``` | ||
```eval_rst | ||
.. note:: | ||
Detailed installation instructions for Windows, including steps for enabling WSL2, can be found on the [Docker Desktop for Windows installation page](https://docs.docker.com/desktop/install/windows-install/). | ||
``` | ||
|
||
#### Install Docker Desktop for Windows | ||
Follow the instructions in [this guide](https://docs.docker.com/desktop/install/windows-install/) to install **Docker Desktop for Windows**. Restart you machine after the installation is complete. | ||
|
||
#### Install WSL2 | ||
|
||
Follow the instructions in [this guide](https://docs.microsoft.com/en-us/windows/wsl/install) to install **Windows Subsystem for Linux 2 (WSL2)**. | ||
|
||
```eval_rst | ||
.. tip:: | ||
You may verify WSL2 installation by running the command `wsl --list` in PowerShell or Command Prompt. If WSL2 is installed, you will see a list of installed Linux distributions. | ||
``` | ||
|
||
#### Enable Docker integration with WSL2 | ||
|
||
Open **Docker desktop**, and select `Settings`->`Resources`->`WSL integration`->turn on `Ubuntu` button->`Apply & restart`. | ||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/docker_desktop_new.png"> | ||
<img src="https://llm-assets.readthedocs.io/en/latest/_images/docker_desktop_new.png" width=100%; /> | ||
</a> | ||
|
||
```eval_rst | ||
.. tip:: | ||
If you encounter **Docker Engine stopped** when opening Docker Desktop, you can reopen it in administrator mode. | ||
``` | ||
|
||
#### Verify Docker is enabled in WSL2 | ||
|
||
Execute the following commands in PowerShell or Command Prompt to verify that Docker is enabled in WSL2: | ||
```bash | ||
wsl -d Ubuntu # Run Ubuntu WSL distribution | ||
docker version # Check if Docker is enabled in WSL | ||
``` | ||
|
||
You can see the output similar to the following: | ||
|
||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/docker_wsl.png"> | ||
<img src="https://llm-assets.readthedocs.io/en/latest/_images/docker_wsl.png" width=100%; /> | ||
</a> | ||
|
||
```eval_rst | ||
.. tip:: | ||
During the use of Docker in WSL, Docker Desktop needs to be kept open all the time. | ||
``` | ||
|
||
## IPEX-LLM Docker Containers | ||
|
||
We have several docker images available for running LLMs on Intel GPUs. The following table lists the available images and their uses: | ||
|
||
| Image Name | Description | Use Case | | ||
|------------|-------------|----------| | ||
| intelanalytics/ipex-llm-cpu:2.1.0-SNAPSHOT | CPU Inference |For development and running LLMs using llama.cpp, Ollama and Python| | ||
| intelanalytics/ipex-llm-xpu:2.1.0-SNAPSHOT | GPU Inference |For development and running LLMs using llama.cpp, Ollama and Python| | ||
| intelanalytics/ipex-llm-serving-cpu:2.1.0-SNAPSHOT | CPU Serving|For serving multiple users/requests through REST APIs using vLLM/FastChat| | ||
| intelanalytics/ipex-llm-serving-xpu:2.1.0-SNAPSHOT | GPU Serving|For serving multiple users/requests through REST APIs using vLLM/FastChat| | ||
| intelanalytics/ipex-llm-finetune-qlora-cpu-standalone:2.1.0-SNAPSHOT | CPU Finetuning via Docker|For fine-tuning LLMs using QLora/Lora, etc. | | ||
|intelanalytics/ipex-llm-finetune-qlora-cpu-k8s:2.1.0-SNAPSHOT|CPU Finetuning via Kubernetes|For fine-tuning LLMs using QLora/Lora, etc. | | ||
| intelanalytics/ipex-llm-finetune-qlora-xpu:2.1.0-SNAPSHOT| GPU Finetuning|For fine-tuning LLMs using QLora/Lora, etc.| | ||
|
||
We have also provided several quickstarts for various usage scenarios: | ||
- [Run and develop LLM applications in PyTorch](./docker_pytorch_inference_gpu.html) | ||
|
||
... to be added soon. | ||
|
||
## Troubleshooting | ||
|
||
|
||
If your machine has both an integrated GPU (iGPU) and a dedicated GPU (dGPU) like ARC, you may encounter the following issue: | ||
|
||
```bash | ||
Abort was called at 62 line in file: | ||
./shared/source/os_interface/os_interface.h | ||
LIBXSMM_VERSION: main_stable-1.17-3651 (25693763) | ||
LIBXSMM_TARGET: adl [Intel(R) Core(TM) i7-14700K] | ||
Registry and code: 13 MB | ||
Command: python chat.py --model-path /llm/llm-models/chatglm2-6b/ | ||
Uptime: 29.349235 s | ||
Aborted | ||
``` | ||
To resolve this problem, you can disabling the iGPU in Device Manager on Windows as follows: | ||
|
||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/disable_iGPU.png"> | ||
<img src="https://llm-assets.readthedocs.io/en/latest/_images/disable_iGPU.png" width=100%; /> | ||
</a> |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,8 @@ | ||
IPEX-LLM Docker Container User Guides | ||
===================================== | ||
|
||
In this section, you will find guides related to using IPEX-LLM with Docker, covering how to: | ||
|
||
|
||
* `Overview of IPEX-LLM Containers for Intel GPU <./docker_windows_gpu.html>`_ | ||
* `Run PyTorch Inference on an Intel GPU via Docker <./docker_pytorch_inference_gpu.html>`_ |
Oops, something went wrong.