File tree Expand file tree Collapse file tree 2 files changed +20
-0
lines changed
tests/model_executor/model_loader/runai_model_streamer Expand file tree Collapse file tree 2 files changed +20
-0
lines changed Original file line number Diff line number Diff line change @@ -541,6 +541,18 @@ steps:
541541 commands :
542542 - pytest -v -s kernels/mamba
543543
544+ - label : RunAI Model Streamer Test # 77s
545+ # Cross continent streaming could take much take much longer.
546+ timeout_in_minutes : 10
547+ torch_nightly : true
548+ mirror_hardwares : [amdexperimental]
549+ source_file_dependencies :
550+ - vllm/engine
551+ - vllm/model_executor/model_loader
552+ - tests/model_executor/model_loader/runai_model_streamer
553+ commands :
554+ - pytest -v -s model_executor/model_loader/runai_model_streamer
555+
544556- label : Model Executor Test # 23min
545557 timeout_in_minutes : 35
546558 mirror_hardwares : [amdexperimental]
Original file line number Diff line number Diff line change 77
88load_format = "runai_streamer"
99test_model = "openai-community/gpt2"
10+ # TODO(amacaskill): Replace with a GKE owned GCS bucket.
11+ test_gcs_model = "gs://vertex-model-garden-public-us/codegemma/codegemma-2b/"
1012
1113prompts = [
1214 "Hello, my name is" ,
@@ -32,3 +34,9 @@ def test_runai_model_loader_download_files(vllm_runner):
3234 with vllm_runner (test_model , load_format = load_format ) as llm :
3335 deserialized_outputs = llm .generate (prompts , sampling_params )
3436 assert deserialized_outputs
37+
38+
39+ def test_runai_model_loader_download_files_gcs (vllm_runner ):
40+ with vllm_runner (test_gcs_model , load_format = load_format ) as llm :
41+ deserialized_outputs = llm .generate (prompts , sampling_params )
42+ assert deserialized_outputs
You can’t perform that action at this time.
0 commit comments