File tree Expand file tree Collapse file tree 3 files changed +30
-4
lines changed
docs/source/getting_started Expand file tree Collapse file tree 3 files changed +30
-4
lines changed Original file line number Diff line number Diff line change 11# The vLLM Dockerfile is used to construct vLLM image that can be directly used
22# to run the OpenAI compatible server.
33
4- FROM ubuntu:20 .04 AS dev
4+ FROM ubuntu:22 .04 AS dev
55
66RUN apt-get update -y && \
77 apt-get install -y python3-pip git
@@ -18,7 +18,7 @@ COPY setup.py /workspace/vllm/
1818# install build requirements
1919RUN PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu" python3 -m pip install -r /workspace/vllm/requirements-build.txt
2020# build vLLM with OpenVINO backend
21- RUN PIP_PRE=1 PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu https://storage.openvinotoolkit.org/simple/wheels/nightly/ " VLLM_TARGET_DEVICE="openvino" python3 -m pip install /workspace/vllm/
21+ RUN PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu https://storage.openvinotoolkit.org/simple/wheels/pre-release " VLLM_TARGET_DEVICE="openvino" python3 -m pip install /workspace/vllm/
2222
2323COPY examples/ /workspace/vllm/examples
2424COPY benchmarks/ /workspace/vllm/benchmarks
Original file line number Diff line number Diff line change @@ -57,7 +57,7 @@ Install from source
5757
5858 .. code-block :: console
5959
60- $ PIP_PRE=1 PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu https://storage.openvinotoolkit.org/simple/wheels/nightly/ " VLLM_TARGET_DEVICE=openvino python -m pip install -v .
60+ $ PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu https://storage.openvinotoolkit.org/simple/wheels/pre-release " VLLM_TARGET_DEVICE=openvino python -m pip install -v .
6161
6262 .. _openvino_backend_performance_tips :
6363
Original file line number Diff line number Diff line change 11# Common dependencies
2- -r requirements-common.txt
2+ # -r requirements-common.txt
3+ # TODO: remove temporary copy of all common dependencies once Optimum Intel will support Transformers >= 4.43.2
4+ cmake >= 3.21
5+ ninja # For faster builds.
6+ psutil
7+ sentencepiece # Required for LLaMA tokenizer.
8+ numpy < 2.0.0
9+ requests
10+ tqdm
11+ py-cpuinfo
12+ transformers < 4.43
13+ tokenizers >= 0.19.1 # Required for Llama 3.
14+ fastapi
15+ aiohttp
16+ openai
17+ uvicorn[standard]
18+ pydantic >= 2.0 # Required for OpenAI server.
19+ pillow # Required for image processing
20+ prometheus_client >= 0.18.0
21+ prometheus-fastapi-instrumentator >= 7.0.0
22+ tiktoken >= 0.6.0 # Required for DBRX tokenizer
23+ lm-format-enforcer == 0.10.3
24+ outlines >= 0.0.43, < 0.1 # Requires torch >= 2.1.0
25+ typing_extensions
26+ filelock >= 3.10.4 # filelock starts to support `mode` argument from 3.10.4
27+ pyzmq
328
429# OpenVINO dependencies
530torch >= 2.1.2
631openvino ~= 2024.3.0.dev
32+ openvino-tokenizers[transformers] ~= 2024.3.0.0.dev
733optimum-intel[openvino] >= 1.18.1
You can’t perform that action at this time.
0 commit comments