Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[CI/Build] Add markdown linter #11857

Merged
merged 17 commits into from
Jan 12, 2025
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ on:
- "docs/**"

jobs:
sphinx-lint:
doc-lint:
runs-on: ubuntu-latest
strategy:
matrix:
Expand All @@ -29,4 +29,4 @@ jobs:
python -m pip install --upgrade pip
pip install -r requirements-lint.txt
- name: Linting docs
run: tools/sphinx-lint.sh
run: tools/doc-lint.sh
1 change: 1 addition & 0 deletions docs/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,4 +16,5 @@ make html
```bash
python -m http.server -d build/html/
```

Launch your browser and open localhost:8000.
1 change: 0 additions & 1 deletion docs/source/api/model/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,4 +9,3 @@ interfaces_base
interfaces
adapters
```

2 changes: 2 additions & 0 deletions docs/source/community/sponsors.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,13 +6,15 @@ vLLM is a community project. Our compute resources for development and testing a
<!-- Note: Please keep these consistent with README.md. -->

Cash Donations:

- a16z
- Dropbox
- Sequoia Capital
- Skywork AI
- ZhenFund

Compute Resources:

- AMD
- Anyscale
- AWS
Expand Down
2 changes: 0 additions & 2 deletions docs/source/contributing/overview.md
Original file line number Diff line number Diff line change
Expand Up @@ -37,8 +37,6 @@ pytest tests/
Currently, the repository is not fully checked by `mypy`.
```

# Contribution Guidelines

## Issues

If you encounter a bug or have a feature request, please [search existing issues](https://github.com/vllm-project/vllm/issues?q=is%3Aissue) first to see if it has already been reported. If not, please [file a new issue](https://github.com/vllm-project/vllm/issues/new/choose), providing as much relevant information as possible.
Expand Down
4 changes: 2 additions & 2 deletions docs/source/deployment/docker.md
Original file line number Diff line number Diff line change
Expand Up @@ -28,8 +28,8 @@ memory to share data between processes under the hood, particularly for tensor p
You can build and run vLLM from source via the provided <gh-file:Dockerfile>. To build vLLM:

```console
$ # optionally specifies: --build-arg max_jobs=8 --build-arg nvcc_threads=2
$ DOCKER_BUILDKIT=1 docker build . --target vllm-openai --tag vllm/vllm-openai
# optionally specifies: --build-arg max_jobs=8 --build-arg nvcc_threads=2
DOCKER_BUILDKIT=1 docker build . --target vllm-openai --tag vllm/vllm-openai
DarkLight1337 marked this conversation as resolved.
Show resolved Hide resolved
```

```{note}
Expand Down
10 changes: 5 additions & 5 deletions docs/source/deployment/frameworks/cerebrium.md
Original file line number Diff line number Diff line change
Expand Up @@ -13,14 +13,14 @@ vLLM can be run on a cloud based GPU machine with [Cerebrium](https://www.cerebr
To install the Cerebrium client, run:

```console
$ pip install cerebrium
$ cerebrium login
pip install cerebrium
cerebrium login
```

Next, create your Cerebrium project, run:

```console
$ cerebrium init vllm-project
cerebrium init vllm-project
```

Next, to install the required packages, add the following to your cerebrium.toml:
Expand Down Expand Up @@ -58,10 +58,10 @@ def run(prompts: list[str], temperature: float = 0.8, top_p: float = 0.95):
Then, run the following code to deploy it to the cloud:

```console
$ cerebrium deploy
cerebrium deploy
```

If successful, you should be returned a CURL command that you can call inference against. Just remember to end the url with the function name you are calling (in our case` /run`)
If successful, you should be returned a CURL command that you can call inference against. Just remember to end the url with the function name you are calling (in our case`/run`)

```python
curl -X POST https://api.cortex.cerebrium.ai/v4/p-xxxxxx/vllm/run \
Expand Down
10 changes: 5 additions & 5 deletions docs/source/deployment/frameworks/dstack.md
Original file line number Diff line number Diff line change
Expand Up @@ -13,16 +13,16 @@ vLLM can be run on a cloud based GPU machine with [dstack](https://dstack.ai/),
To install dstack client, run:

```console
$ pip install "dstack[all]
$ dstack server
pip install "dstack[all]
dstack server
```

Next, to configure your dstack project, run:

```console
$ mkdir -p vllm-dstack
$ cd vllm-dstack
$ dstack init
mkdir -p vllm-dstack
cd vllm-dstack
dstack init
```

Next, to provision a VM instance with LLM of your choice (`NousResearch/Llama-2-7b-chat-hf` for this example), create the following `serve.dstack.yml` file for the dstack `Service`:
Expand Down
12 changes: 6 additions & 6 deletions docs/source/deployment/frameworks/skypilot.md
Original file line number Diff line number Diff line change
Expand Up @@ -334,12 +334,12 @@ run: |

1. Start the chat web UI:

```console
sky launch -c gui ./gui.yaml --env ENDPOINT=$(sky serve status --endpoint vllm)
```
```console
sky launch -c gui ./gui.yaml --env ENDPOINT=$(sky serve status --endpoint vllm)
```

2. Then, we can access the GUI at the returned gradio link:

```console
| INFO | stdout | Running on public URL: https://6141e84201ce0bb4ed.gradio.live
```
```console
| INFO | stdout | Running on public URL: https://6141e84201ce0bb4ed.gradio.live
```
2 changes: 1 addition & 1 deletion docs/source/deployment/integrations/llamastack.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ vLLM is also available via [Llama Stack](https://github.com/meta-llama/llama-sta
To install Llama Stack, run

```console
$ pip install llama-stack -q
pip install llama-stack -q
```

## Inference using OpenAI Compatible API
Expand Down
Loading
Loading