Skip to content

Commit 0bdbb81

Browse files
rafvasqAlvant
authored andcommitted
[Doc] Use shell code-blocks and fix section headers (vllm-project#9508)
Signed-off-by: Rafael Vasquez <[email protected]> Signed-off-by: Alvant <[email protected]>
1 parent 43b51fb commit 0bdbb81

File tree

3 files changed

+23
-23
lines changed

3 files changed

+23
-23
lines changed

docs/source/getting_started/debugging.rst

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -107,15 +107,15 @@ If GPU/CPU communication cannot be established, you can use the following Python
107107
108108
If you are testing with a single node, adjust ``--nproc-per-node`` to the number of GPUs you want to use:
109109

110-
.. code-block:: shell
110+
.. code-block:: console
111111
112-
NCCL_DEBUG=TRACE torchrun --nproc-per-node=<number-of-GPUs> test.py
112+
$ NCCL_DEBUG=TRACE torchrun --nproc-per-node=<number-of-GPUs> test.py
113113
114114
If you are testing with multi-nodes, adjust ``--nproc-per-node`` and ``--nnodes`` according to your setup and set ``MASTER_ADDR`` to the correct IP address of the master node, reachable from all nodes. Then, run:
115115

116-
.. code-block:: shell
116+
.. code-block:: console
117117
118-
NCCL_DEBUG=TRACE torchrun --nnodes 2 --nproc-per-node=2 --rdzv_backend=c10d --rdzv_endpoint=$MASTER_ADDR test.py
118+
$ NCCL_DEBUG=TRACE torchrun --nnodes 2 --nproc-per-node=2 --rdzv_backend=c10d --rdzv_endpoint=$MASTER_ADDR test.py
119119
120120
If the script runs successfully, you should see the message ``sanity check is successful!``.
121121

docs/source/getting_started/installation.rst

Lines changed: 17 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -7,14 +7,14 @@ Installation
77
vLLM is a Python library that also contains pre-compiled C++ and CUDA (12.1) binaries.
88

99
Requirements
10-
===========================
10+
============
1111

1212
* OS: Linux
13-
* Python: 3.8 -- 3.12
13+
* Python: 3.8 - 3.12
1414
* GPU: compute capability 7.0 or higher (e.g., V100, T4, RTX20xx, A100, L4, H100, etc.)
1515

1616
Install released versions
17-
===========================
17+
=========================
1818

1919
You can install vLLM using pip:
2020

@@ -51,9 +51,9 @@ You can install vLLM using pip:
5151
.. _install-the-latest-code:
5252

5353
Install the latest code
54-
=========================
54+
=======================
5555

56-
LLM inference is a fast-evolving field, and the latest code may contain bug fixes, performance improvements, and new features that are not released yet. To allow users to try the latest code without waiting for the next release, vLLM provides wheels for Linux running on x86 platform with cuda 12 for every commit since v0.5.3. You can download and install the latest one with the following command:
56+
LLM inference is a fast-evolving field, and the latest code may contain bug fixes, performance improvements, and new features that are not released yet. To allow users to try the latest code without waiting for the next release, vLLM provides wheels for Linux running on a x86 platform with CUDA 12 for every commit since ``v0.5.3``. You can download and install it with the following command:
5757

5858
.. code-block:: console
5959
@@ -66,7 +66,7 @@ If you want to access the wheels for previous commits, you can specify the commi
6666
$ export VLLM_COMMIT=33f460b17a54acb3b6cc0b03f4a17876cff5eafd # use full commit hash from the main branch
6767
$ pip install https://vllm-wheels.s3.us-west-2.amazonaws.com/${VLLM_COMMIT}/vllm-1.0.0.dev-cp38-abi3-manylinux1_x86_64.whl
6868
69-
Note that the wheels are built with Python 3.8 abi (see `PEP 425 <https://peps.python.org/pep-0425/>`_ for more details about abi), so **they are compatible with Python 3.8 and later**. The version string in the wheel file name (``1.0.0.dev``) is just a placeholder to have a unified URL for the wheels. The actual versions of wheels are contained in the wheel metadata.
69+
Note that the wheels are built with Python 3.8 ABI (see `PEP 425 <https://peps.python.org/pep-0425/>`_ for more details about ABI), so **they are compatible with Python 3.8 and later**. The version string in the wheel file name (``1.0.0.dev``) is just a placeholder to have a unified URL for the wheels. The actual versions of wheels are contained in the wheel metadata.
7070

7171
Another way to access the latest code is to use the docker images:
7272

@@ -77,17 +77,17 @@ Another way to access the latest code is to use the docker images:
7777
7878
These docker images are used for CI and testing only, and they are not intended for production use. They will be expired after several days.
7979

80-
Latest code can contain bugs and may not be stable. Please use it with caution.
80+
The latest code can contain bugs and may not be stable. Please use it with caution.
8181

8282
.. _build_from_source:
8383

8484
Build from source
85-
==================
85+
=================
8686

8787
.. _python-only-build:
8888

8989
Python-only build (without compilation)
90-
----------------------------------------
90+
---------------------------------------
9191

9292
If you only need to change Python code, you can simply build vLLM without compilation.
9393

@@ -122,22 +122,22 @@ Once you have finished editing or want to install another vLLM wheel, you should
122122
123123
$ python python_only_dev.py --quit-dev
124124
125-
The script with ``--quit-dev`` flag will:
125+
The ``--quit-dev`` flag will:
126126

127127
* Remove the symbolic link from the current directory to the vLLM package.
128128
* Restore the original vLLM package from the backup.
129129

130-
If you update the vLLM wheel and want to rebuild from the source and make further edits, you will need to start `all above <#python-only-build>`_ over again.
130+
If you update the vLLM wheel and rebuild from the source to make further edits, you will need to repeat the `Python-only build <#python-only-build>`_ steps again.
131131

132132
.. note::
133133

134134
There is a possibility that your source code may have a different commit ID compared to the latest vLLM wheel, which could potentially lead to unknown errors.
135-
It is recommended to use the same commit ID for the source code as the vLLM wheel you have installed. Please refer to `the above section <#install-the-latest-code>`_ for instructions on how to install a specified wheel.
135+
It is recommended to use the same commit ID for the source code as the vLLM wheel you have installed. Please refer to `the section above <#install-the-latest-code>`_ for instructions on how to install a specified wheel.
136136

137137
Full build (with compilation)
138-
---------------------------------
138+
-----------------------------
139139

140-
If you want to modify C++ or CUDA code, you'll need to build vLLM from source. This can take several minutes:
140+
If you want to modify C++ or CUDA code, you'll need to build vLLM from source. This can take several minutes:
141141

142142
.. code-block:: console
143143
@@ -153,7 +153,7 @@ If you want to modify C++ or CUDA code, you'll need to build vLLM from source. T
153153

154154

155155
Use an existing PyTorch installation
156-
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
156+
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
157157
There are scenarios where the PyTorch dependency cannot be easily installed via pip, e.g.:
158158

159159
* Building vLLM with PyTorch nightly or a custom PyTorch build.
@@ -171,7 +171,7 @@ To build vLLM using an existing PyTorch installation:
171171
172172
173173
Troubleshooting
174-
~~~~~~~~~~~~~~~~~
174+
~~~~~~~~~~~~~~~
175175

176176
To avoid your system being overloaded, you can limit the number of compilation jobs
177177
to be run simultaneously, via the environment variable ``MAX_JOBS``. For example:
@@ -207,7 +207,7 @@ Here is a sanity check to verify that the CUDA Toolkit is correctly installed:
207207
208208
209209
Unsupported OS build
210-
----------------------
210+
--------------------
211211

212212
vLLM can fully run only on Linux but for development purposes, you can still build it on other systems (for example, macOS), allowing for imports and a more convenient development environment. The binaries will not be compiled and won't work on non-Linux systems.
213213

docs/source/models/vlm.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -247,9 +247,9 @@ A full code example can be found in `examples/openai_api_client_for_multimodal.p
247247

248248
By default, the timeout for fetching images through http url is ``5`` seconds. You can override this by setting the environment variable:
249249

250-
.. code-block:: shell
250+
.. code-block:: console
251251
252-
export VLLM_IMAGE_FETCH_TIMEOUT=<timeout>
252+
$ export VLLM_IMAGE_FETCH_TIMEOUT=<timeout>
253253
254254
.. note::
255255
There is no need to format the prompt in the API request since it will be handled by the server.

0 commit comments

Comments
 (0)