Skip to content
This repository was archived by the owner on Sep 4, 2025. It is now read-only.

Commit 61b7763

Browse files
committed
fix VLLM_HPU_LOG_STEP_GRAPH_COMPILATION
1 parent 972acf3 commit 61b7763

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

vllm/executor/habana_executor.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -96,6 +96,7 @@ def execute_model(
9696
log_cpu_fallbacks = os.environ.get('VLLM_HPU_LOG_STEP_CPU_FALLBACKS', '0') != '0' or log_cpu_fallbacks_all
9797
if log_graph_compilation or log_cpu_fallbacks:
9898
from habana_frameworks.torch.hpu.metrics import metric_localcontext
99+
seq_group_metadata_list = execute_model_req.seq_group_metadata_list
99100
is_prompt = any([seq_group_metadata.is_prompt for seq_group_metadata in seq_group_metadata_list])
100101
max_context_len = max([max([len(v.prompt_token_ids) + len(v.output_token_ids) for v in seq_group_metadata.seq_data.values()]) for seq_group_metadata in seq_group_metadata_list]) # whoa, that's some spicy stuff right here
101102
max_num_blocks = ((max_context_len - 1) // self.cache_config.block_size) + 1

0 commit comments

Comments
 (0)