Skip to content

Commit a48b9ad

Browse files
authored
[Doc] Update Readme (vllm-project#11)
<!-- Thanks for sending a pull request! BEFORE SUBMITTING, PLEASE READ https://docs.vllm.ai/en/latest/contributing/overview.html --> ### What this PR does / why we need it? Add feature and model support matrix ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? CI test is enough Signed-off-by: wangxiyuan <[email protected]>
1 parent bfccf73 commit a48b9ad

File tree

2 files changed

+47
-1
lines changed

2 files changed

+47
-1
lines changed

README.md

Lines changed: 46 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ By using vLLM Ascend plugin, popular open-source models, including Transformer-l
3939
| Python | >= 3.9 | [3.10](https://www.python.org/downloads/) | Required for vllm |
4040
| CANN | >= 8.0.RC2 | [8.0.RC3](https://www.hiascend.com/developer/download/community/result?module=cann&cann=8.0.0.beta1) | Required for vllm-ascend and torch-npu |
4141
| torch-npu | >= 2.4.0 | [2.5.1rc1](https://gitee.com/ascend/pytorch/releases/tag/v6.0.0.alpha001-pytorch2.5.1) | Required for vllm-ascend |
42-
| torch | >= 2.4.0 | [2.5.1](https://github.com/pytorch/pytorch/releases/tag/v2.5.1) | Required for torch-npu and vllm required |
42+
| torch | >= 2.4.0 | [2.5.1](https://github.com/pytorch/pytorch/releases/tag/v2.5.1) | Required for torch-npu and vllm |
4343

4444
Find more about how to setup your environment in [here](docs/environment.md).
4545

@@ -92,6 +92,51 @@ docker build -t vllm-ascend-dev-image -f ./Dockerfile .
9292

9393
See [Building and Testing](./CONTRIBUTING.md) for more details, which is a step-by-step guide to help you set up development environment, build and test.
9494

95+
## Feature Support Matrix
96+
| Feature | Supported | Note |
97+
|---------|-----------|------|
98+
| Chunked Prefill || Plan in 2025 Q1 |
99+
| Automatic Prefix Caching || Imporve performance in 2025 Q1 |
100+
| LoRA || Plan in 2025 Q1 |
101+
| Prompt adapter |||
102+
| Speculative decoding || Impore accuracy in 2025 Q1|
103+
| Pooling || Plan in 2025 Q1 |
104+
| Enc-dec || Plan in 2025 Q1 |
105+
| Multi Modality | ✅ (LLaVA/Qwen2-vl/Qwen2-audio/internVL)| Add more model support in 2025 Q1 |
106+
| LogProbs |||
107+
| Prompt logProbs |||
108+
| Async output |||
109+
| Multi step scheduler |||
110+
| Best of |||
111+
| Beam search |||
112+
| Guided Decoding || Plan in 2025 Q1 |
113+
114+
## Model Support Matrix
115+
116+
The list here is a subset of the supported models. See [supported_models](docs/supported_models.md) for more details:
117+
| Model | Supported | Note |
118+
|---------|-----------|------|
119+
| Qwen 2.5 |||
120+
| Mistral | | Need test |
121+
| DeepSeek v2.5 | |Need test |
122+
| LLama3.1/3.2 |||
123+
| Gemma-2 | |Need test|
124+
| baichuan | |Need test|
125+
| minicpm | |Need test|
126+
| internlm |||
127+
| ChatGLM |||
128+
| InternVL 2.5 |||
129+
| Qwen2-VL |||
130+
| GLM-4v | |Need test|
131+
| Molomo |||
132+
| LLaVA 1.5 |||
133+
| Mllama | |Need test|
134+
| LLaVA-Next | |Need test|
135+
| LLaVA-Next-Video | |Need test|
136+
| Phi-3-Vison/Phi-3.5-Vison | |Need test|
137+
| Ultravox | |Need test|
138+
| Qwen2-Audio |||
139+
95140
## Contributing
96141
We welcome and value any contributions and collaborations:
97142
- Please let us know if you encounter a bug by [filing an issue](https://github.com/vllm-project/vllm-ascend/issues).

docs/supported_models.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
TBD

0 commit comments

Comments
 (0)