Skip to content

Commit 0b6ea14

Browse files
Superjomndominicshanshan
authored andcommitted
[None][fix] api stability bug in status label (NVIDIA#7861)
Signed-off-by: Yan Chunwei <[email protected]> Signed-off-by: Wangshanshan <[email protected]>
1 parent 95eeaef commit 0b6ea14

File tree

2 files changed

+5
-3
lines changed

2 files changed

+5
-3
lines changed

tensorrt_llm/llmapi/llm_args.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1292,8 +1292,8 @@ class BaseLlmArgs(StrictBaseModel):
12921292
fail_fast_on_attention_window_too_large: bool = Field(
12931293
default=False,
12941294
description=
1295-
"Fail fast when attention window is too large to fit even a single sequence in the KV cache."
1296-
)
1295+
"Fail fast when attention window is too large to fit even a single sequence in the KV cache.",
1296+
status="prototype")
12971297

12981298
# LoRA arguments
12991299
enable_lora: bool = Field(default=False, description="Enable LoRA.")

tests/unittest/api_stability/api_stability_core.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -510,7 +510,9 @@ def check_status(field_name, reference_status, context=""):
510510
# step 1: check the method status
511511
method = getattr(self.TEST_CLASS, method_name)
512512
if method_name in committed_data.get('methods', {}):
513-
continue
513+
if method_name != "__init__":
514+
continue
515+
# Both committed and non-committed methods have __init__ with different parameters
514516
if method_name != "__init__":
515517
method_status = get_api_status(method)
516518
if method_status is None:

0 commit comments

Comments
 (0)