|
9 | 9 |
|
10 | 10 | from vllm.lora.request import LoRARequest |
11 | 11 |
|
12 | | -from ..models.utils import check_outputs_equal |
13 | | - |
14 | 12 | ULTRAVOX_MODEL_NAME = "fixie-ai/ultravox-v0_3" |
15 | 13 | LLMA_MODEL_NAME = "meta-llama/Llama-3.1-8B-Instruct" |
16 | 14 |
|
@@ -61,6 +59,9 @@ def _get_prompt(audio_count, question, placeholder, model_name) -> str: |
61 | 59 |
|
62 | 60 |
|
63 | 61 | def test_ultravox_lora(vllm_runner): |
| 62 | + """ |
| 63 | + TODO: Train an Ultravox LoRA instead of using a Llama LoRA. |
| 64 | + """ |
64 | 65 | llama3_1_8b_chess_lora = llama3_1_8b_chess_lora_path() |
65 | 66 | with TemporaryDirectory() as temp_ultravox_lora_dir: |
66 | 67 | llama3_1_8b_ultravox_chess_lora = mk_llama3_1_8b_ultravox_chess_lora( |
@@ -97,34 +98,11 @@ def test_ultravox_lora(vllm_runner): |
97 | 98 | dtype="bfloat16", |
98 | 99 | max_model_len=4096, |
99 | 100 | ) as vllm_model: |
100 | | - llama_outputs_no_lora: List[Tuple[List[int], |
101 | | - str]] = vllm_model.generate_greedy( |
102 | | - [ |
103 | | - _get_prompt( |
104 | | - 0, PROMPT, |
105 | | - VLLM_PLACEHOLDER, |
106 | | - LLMA_MODEL_NAME) |
107 | | - ], |
108 | | - 256, |
109 | | - ) |
110 | | - llama_outputs: List[Tuple[List[int], |
111 | | - str]] = vllm_model.generate_greedy( |
112 | | - [ |
113 | | - _get_prompt(0, PROMPT, |
114 | | - VLLM_PLACEHOLDER, |
115 | | - LLMA_MODEL_NAME) |
116 | | - ], |
117 | | - 256, |
118 | | - lora_request=LoRARequest( |
119 | | - str(1), 1, llama3_1_8b_chess_lora), |
120 | | - ) |
121 | | - |
122 | | - check_outputs_equal( |
123 | | - outputs_0_lst=ultravox_outputs, |
124 | | - outputs_1_lst=llama_outputs, |
125 | | - name_0="ultravox", |
126 | | - name_1="llama", |
127 | | - ) |
| 101 | + llama_outputs_no_lora: List[Tuple[List[int], str]] = ( |
| 102 | + vllm_model.generate_greedy( |
| 103 | + [_get_prompt(0, PROMPT, VLLM_PLACEHOLDER, LLMA_MODEL_NAME)], |
| 104 | + 256, |
| 105 | + )) |
128 | 106 |
|
129 | 107 | _, llama_no_lora_str = llama_outputs_no_lora[0] |
130 | 108 | _, ultravox_str = ultravox_outputs[0] |
|
0 commit comments