Skip to content

Commit ce45c0d

Browse files
committed
Fix docs, update command
Signed-off-by: Rafael Vasquez <[email protected]>
1 parent 6f638e9 commit ce45c0d

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

43 files changed

+346
-332
lines changed

docs/README.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,4 +16,5 @@ make html
1616
```bash
1717
python -m http.server -d build/html/
1818
```
19+
1920
Launch your browser and open localhost:8000.

docs/source/api/multimodal/index.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,6 @@ via the `multi_modal_data` field in {class}`vllm.inputs.PromptType`.
1313

1414
Looking to add your own multi-modal model? Please follow the instructions listed [here](#enabling-multimodal-inputs).
1515

16-
1716
## Module Contents
1817

1918
```{eval-rst}

docs/source/api/params.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,4 +19,3 @@ Optional parameters for vLLM APIs.
1919
.. autoclass:: vllm.PoolingParams
2020
:members:
2121
```
22-

docs/source/community/sponsors.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -6,13 +6,15 @@ vLLM is a community project. Our compute resources for development and testing a
66
<!-- Note: Please keep these consistent with README.md. -->
77

88
Cash Donations:
9+
910
- a16z
1011
- Dropbox
1112
- Sequoia Capital
1213
- Skywork AI
1314
- ZhenFund
1415

1516
Compute Resources:
17+
1618
- AMD
1719
- Anyscale
1820
- AWS

docs/source/contributing/overview.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -37,8 +37,6 @@ pytest tests/
3737
Currently, the repository is not fully checked by `mypy`.
3838
```
3939

40-
# Contribution Guidelines
41-
4240
## Issues
4341

4442
If you encounter a bug or have a feature request, please [search existing issues](https://github.com/vllm-project/vllm/issues?q=is%3Aissue) first to see if it has already been reported. If not, please [file a new issue](https://github.com/vllm-project/vllm/issues/new/choose), providing as much relevant information as possible.

docs/source/deployment/docker.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,8 +28,8 @@ memory to share data between processes under the hood, particularly for tensor p
2828
You can build and run vLLM from source via the provided <gh-file:Dockerfile>. To build vLLM:
2929

3030
```console
31-
$ # optionally specifies: --build-arg max_jobs=8 --build-arg nvcc_threads=2
32-
$ DOCKER_BUILDKIT=1 docker build . --target vllm-openai --tag vllm/vllm-openai
31+
# optionally specifies: --build-arg max_jobs=8 --build-arg nvcc_threads=2
32+
DOCKER_BUILDKIT=1 docker build . --target vllm-openai --tag vllm/vllm-openai
3333
```
3434

3535
```{note}

docs/source/deployment/frameworks/cerebrium.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -13,14 +13,14 @@ vLLM can be run on a cloud based GPU machine with [Cerebrium](https://www.cerebr
1313
To install the Cerebrium client, run:
1414

1515
```console
16-
$ pip install cerebrium
17-
$ cerebrium login
16+
pip install cerebrium
17+
cerebrium login
1818
```
1919

2020
Next, create your Cerebrium project, run:
2121

2222
```console
23-
$ cerebrium init vllm-project
23+
cerebrium init vllm-project
2424
```
2525

2626
Next, to install the required packages, add the following to your cerebrium.toml:
@@ -58,10 +58,10 @@ def run(prompts: list[str], temperature: float = 0.8, top_p: float = 0.95):
5858
Then, run the following code to deploy it to the cloud:
5959

6060
```console
61-
$ cerebrium deploy
61+
cerebrium deploy
6262
```
6363

64-
If successful, you should be returned a CURL command that you can call inference against. Just remember to end the url with the function name you are calling (in our case` /run`)
64+
If successful, you should be returned a CURL command that you can call inference against. Just remember to end the url with the function name you are calling (in our case`/run`)
6565

6666
```python
6767
curl -X POST https://api.cortex.cerebrium.ai/v4/p-xxxxxx/vllm/run \

docs/source/deployment/frameworks/dstack.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -13,16 +13,16 @@ vLLM can be run on a cloud based GPU machine with [dstack](https://dstack.ai/),
1313
To install dstack client, run:
1414

1515
```console
16-
$ pip install "dstack[all]
17-
$ dstack server
16+
pip install "dstack[all]
17+
dstack server
1818
```
1919

2020
Next, to configure your dstack project, run:
2121

2222
```console
23-
$ mkdir -p vllm-dstack
24-
$ cd vllm-dstack
25-
$ dstack init
23+
mkdir -p vllm-dstack
24+
cd vllm-dstack
25+
dstack init
2626
```
2727

2828
Next, to provision a VM instance with LLM of your choice (`NousResearch/Llama-2-7b-chat-hf` for this example), create the following `serve.dstack.yml` file for the dstack `Service`:

docs/source/deployment/frameworks/skypilot.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -338,7 +338,7 @@ run: |
338338
sky launch -c gui ./gui.yaml --env ENDPOINT=$(sky serve status --endpoint vllm)
339339
```
340340

341-
2. Then, we can access the GUI at the returned gradio link:
341+
1. Then, we can access the GUI at the returned gradio link:
342342

343343
```console
344344
| INFO | stdout | Running on public URL: https://6141e84201ce0bb4ed.gradio.live

docs/source/deployment/integrations/llamastack.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ vLLM is also available via [Llama Stack](https://github.com/meta-llama/llama-sta
77
To install Llama Stack, run
88

99
```console
10-
$ pip install llama-stack -q
10+
pip install llama-stack -q
1111
```
1212

1313
## Inference using OpenAI Compatible API

0 commit comments

Comments
 (0)