Skip to content
Merged
Show file tree
Hide file tree
Changes from 3 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions docs/source/deployment/frameworks/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -11,4 +11,5 @@ lws
modal
skypilot
triton
open_webui
:::
27 changes: 27 additions & 0 deletions docs/source/deployment/frameworks/open_webui.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,27 @@
(deployment-open_webui)=

# Open WebUI

1. Install the (Docker)[https://docs.docker.com/engine/install/]

2. Start the vLLM server with supported chat completion, e.g.

```console
vllm serve qwen/Qwen1.5-0.5B-Chat
```

1. Start the (Open WebUI)[https://github.com/open-webui/open-webui] docker container (replace the vllm serve host and vllm serve port):

```console
docker run -d -p 3000:8080 \
--name open-webui \
-v open-webui:/app/backend/data \
-e OPENAI_API_BASE_URL=http://<vllm serve host>:<vllm serve port>/v1 \
--restart always \
ghcr.io/open-webui/open-webui:main
```

1. Open it in the browser: <http://open-webui-host:3000/>

On the top of the web page, you can see the model `qwen/Qwen1.5-0.5B-Chat`.
![Spans details](https://imgur.com/a/pm1VRqG)