Skip to content

Commit 2ee4528

Browse files
authored
Move verify_marlin_supported to GPTQMarlinLinearMethod (#8165)
1 parent 9da25a8 commit 2ee4528

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

vllm/model_executor/layers/quantization/gptq_marlin.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -51,10 +51,6 @@ def __init__(self, weight_bits: int, group_size: int, desc_act: bool,
5151

5252
self.quant_type = self.TYPE_MAP[(weight_bits, is_sym)]
5353

54-
# Verify supported on platform.
55-
verify_marlin_supported(quant_type=self.quant_type,
56-
group_size=self.group_size)
57-
5854
def __repr__(self) -> str:
5955
return (f"GPTQMarlinConfig(quant_type={self.quant_type}, "
6056
f"group_size={self.group_size}, "
@@ -153,6 +149,10 @@ class GPTQMarlinLinearMethod(LinearMethodBase):
153149
def __init__(self, quant_config: GPTQMarlinConfig) -> None:
154150
self.quant_config = quant_config
155151

152+
# Verify supported on platform.
153+
verify_marlin_supported(quant_type=self.quant_config.quant_type,
154+
group_size=self.quant_config.group_size)
155+
156156
def create_weights(
157157
self,
158158
layer: torch.nn.Module,

0 commit comments

Comments
 (0)