Skip to content

Commit 5801e49

Browse files
authored
[V0 Deprecation] Remove MQLLMEngine (#25019)
Signed-off-by: Woosuk Kwon <[email protected]> Signed-off-by: Woosuk Kwon <[email protected]>
1 parent 58d4c70 commit 5801e49

File tree

15 files changed

+12
-1969
lines changed

15 files changed

+12
-1969
lines changed

.buildkite/test-pipeline.yaml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,6 @@ steps:
4646
mirror_hardwares: [amdexperimental]
4747
source_file_dependencies:
4848
- vllm/
49-
- tests/mq_llm_engine
5049
- tests/async_engine
5150
- tests/test_inputs.py
5251
- tests/test_outputs.py
@@ -57,7 +56,6 @@ steps:
5756
- tests/transformers_utils
5857
commands:
5958
- python3 standalone_tests/lazy_imports.py
60-
- pytest -v -s mq_llm_engine # MQLLMEngine
6159
- pytest -v -s async_engine # AsyncLLMEngine
6260
- pytest -v -s test_inputs.py
6361
- pytest -v -s test_outputs.py

tests/entrypoints/openai/test_lora_resolvers.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,14 +10,14 @@
1010
import pytest
1111

1212
from vllm.config.multimodal import MultiModalConfig
13-
from vllm.engine.multiprocessing.client import MQLLMEngineClient
1413
from vllm.entrypoints.openai.protocol import CompletionRequest, ErrorResponse
1514
from vllm.entrypoints.openai.serving_completion import OpenAIServingCompletion
1615
from vllm.entrypoints.openai.serving_models import (BaseModelPath,
1716
OpenAIServingModels)
1817
from vllm.lora.request import LoRARequest
1918
from vllm.lora.resolver import LoRAResolver, LoRAResolverRegistry
2019
from vllm.transformers_utils.tokenizer import get_tokenizer
20+
from vllm.v1.engine.async_llm import AsyncLLM
2121

2222
MODEL_NAME = "openai-community/gpt2"
2323
BASE_MODEL_PATHS = [BaseModelPath(name=MODEL_NAME, model_path=MODEL_NAME)]
@@ -82,7 +82,7 @@ def register_mock_resolver():
8282
@pytest.fixture
8383
def mock_serving_setup():
8484
"""Provides a mocked engine and serving completion instance."""
85-
mock_engine = MagicMock(spec=MQLLMEngineClient)
85+
mock_engine = MagicMock(spec=AsyncLLM)
8686
mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME)
8787
mock_engine.errored = False
8888

tests/entrypoints/openai/test_serving_chat.py

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -13,12 +13,12 @@
1313
import pytest_asyncio
1414

1515
from vllm.config.multimodal import MultiModalConfig
16-
from vllm.engine.multiprocessing.client import MQLLMEngineClient
1716
from vllm.entrypoints.openai.protocol import ChatCompletionRequest
1817
from vllm.entrypoints.openai.serving_chat import OpenAIServingChat
1918
from vllm.entrypoints.openai.serving_models import (BaseModelPath,
2019
OpenAIServingModels)
2120
from vllm.transformers_utils.tokenizer import get_tokenizer
21+
from vllm.v1.engine.async_llm import AsyncLLM
2222

2323
from ...utils import RemoteOpenAIServer
2424

@@ -276,7 +276,7 @@ def test_async_serving_chat_init():
276276

277277
@pytest.mark.asyncio
278278
async def test_serving_chat_returns_correct_model_name():
279-
mock_engine = MagicMock(spec=MQLLMEngineClient)
279+
mock_engine = MagicMock(spec=AsyncLLM)
280280
mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME)
281281
mock_engine.errored = False
282282

@@ -312,7 +312,7 @@ async def return_model_name(*args):
312312

313313
@pytest.mark.asyncio
314314
async def test_serving_chat_should_set_correct_max_tokens():
315-
mock_engine = MagicMock(spec=MQLLMEngineClient)
315+
mock_engine = MagicMock(spec=AsyncLLM)
316316
mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME)
317317
mock_engine.errored = False
318318

@@ -355,7 +355,7 @@ async def test_serving_chat_should_set_correct_max_tokens():
355355
}
356356

357357
# Reinitialize the engine with new settings
358-
mock_engine = MagicMock(spec=MQLLMEngineClient)
358+
mock_engine = MagicMock(spec=AsyncLLM)
359359
mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME)
360360
mock_engine.errored = False
361361

@@ -410,7 +410,7 @@ async def test_serving_chat_should_set_correct_max_tokens():
410410
}
411411

412412
# Reinitialize the engine with new settings
413-
mock_engine = MagicMock(spec=MQLLMEngineClient)
413+
mock_engine = MagicMock(spec=AsyncLLM)
414414
mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME)
415415
mock_engine.errored = False
416416

@@ -467,7 +467,7 @@ async def test_serving_chat_could_load_correct_generation_config():
467467
"repetition_penalty": 1.05
468468
}
469469

470-
mock_engine = MagicMock(spec=MQLLMEngineClient)
470+
mock_engine = MagicMock(spec=AsyncLLM)
471471
mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME)
472472
mock_engine.errored = False
473473

@@ -523,7 +523,7 @@ async def test_serving_chat_did_set_correct_cache_salt(model_type):
523523
mock_model_config = MockModelConfig()
524524
mock_model_config.hf_config.model_type = model_type
525525

526-
mock_engine = MagicMock(spec=MQLLMEngineClient)
526+
mock_engine = MagicMock(spec=AsyncLLM)
527527
mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME)
528528
mock_engine.errored = False
529529

tests/mq_llm_engine/__init__.py

Whitespace-only changes.

tests/mq_llm_engine/conftest.py

Lines changed: 0 additions & 12 deletions
This file was deleted.

tests/mq_llm_engine/test_abort.py

Lines changed: 0 additions & 69 deletions
This file was deleted.

0 commit comments

Comments
 (0)