File tree Expand file tree Collapse file tree 4 files changed +17
-7
lines changed
tensorrt_llm/_torch/models Expand file tree Collapse file tree 4 files changed +17
-7
lines changed Original file line number Diff line number Diff line change 1010from .modeling_hyperclovax import HCXVisionForCausalLM
1111from .modeling_llama import LlamaForCausalLM
1212from .modeling_llava_next import LlavaNextModel
13- from .modeling_mistral import MistralForCausalLM
13+ from .modeling_mistral import Mistral3VLM , MistralForCausalLM
1414from .modeling_mixtral import MixtralForCausalLM
1515from .modeling_nemotron import NemotronForCausalLM
1616from .modeling_nemotron_h import NemotronHForCausalLM
3939 "HCXVisionForCausalLM" ,
4040 "LlamaForCausalLM" ,
4141 "LlavaNextModel" ,
42+ "Mistral3VLM" ,
4243 "MistralForCausalLM" ,
4344 "MixtralForCausalLM" ,
4445 "NemotronForCausalLM" ,
Original file line number Diff line number Diff line change @@ -296,6 +296,8 @@ def __init__(
296296
297297 llm_model_config = self ._get_sub_model_config (model_config ,
298298 "text_config" )
299+ # This is necessary for the auto weight mapper to figure out what it needs.
300+ llm_model_config .pretrained_config .architectures = config .architectures
299301 self .llm = MistralForCausalLM (llm_model_config )
300302
301303 self ._device = "cuda"
Original file line number Diff line number Diff line change 44"""
55import copy
66import os
7+ import shlex
78import subprocess
89import tempfile
910import textwrap as tw
@@ -116,12 +117,17 @@ def run_cmd(self,
116117 new_env = os .environ
117118
118119 if caller .__name__ == 'check_output' :
119- result = subprocess .run (call_args ,
120- env = new_env ,
121- check = True ,
122- capture_output = True ,
123- ** kwargs )
124- return result .stdout .decode ('utf-8' )
120+ try :
121+ result = subprocess .run (call_args ,
122+ env = new_env ,
123+ check = True ,
124+ capture_output = True ,
125+ ** kwargs )
126+ return result .stdout .decode ('utf-8' )
127+ except subprocess .CalledProcessError as e :
128+ raise RuntimeError (f"Failed to run `{ shlex .join (e .cmd )} `:\n "
129+ f"Stdout: { e .stdout .decode ()} \n "
130+ f"Stderr: { e .stderr .decode ()} \n " )
125131 else :
126132 print (f"Start subprocess with { caller } ({ call_args } , env={ new_env } )" )
127133 return caller (call_args , env = new_env , ** kwargs )
Original file line number Diff line number Diff line change @@ -193,6 +193,7 @@ l0_h100:
193193 - accuracy/test_llm_api_pytorch.py::TestQwen3_30B_A3B::test_fp8_block_scales[latency]
194194 - accuracy/test_llm_api_pytorch.py::TestLlama3_1_8BInstruct::test_guided_decoding[llguidance]
195195 - test_e2e.py::test_ptp_quickstart_multimodal[gemma-3-27b-it-gemma/gemma-3-27b-it-image-True]
196+ - test_e2e.py::test_ptp_quickstart_multimodal[mistral-small-3.1-24b-instruct-Mistral-Small-3.1-24B-Instruct-2503-image-True]
196197- condition :
197198 ranges :
198199 system_gpu_count :
You can’t perform that action at this time.
0 commit comments