Skip to content

Commit 87e191f

Browse files
committed
fix format error
Signed-off-by: Jinzhen Lin <[email protected]>
1 parent ac8ae24 commit 87e191f

File tree

2 files changed

+4
-6
lines changed

2 files changed

+4
-6
lines changed

vllm/model_executor/layers/fused_moe/fused_moe.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -557,8 +557,6 @@ def invoke_fused_moe_kernel(A: torch.Tensor,
557557
)
558558

559559
else:
560-
ndim = lambda x: 0 if x is None else x.ndim
561-
562560
fused_moe_kernel[grid](
563561
A,
564562
B,

vllm/model_executor/layers/quantization/moe_quant_int.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -104,11 +104,11 @@ def get_quant_method(self, layer: torch.nn.Module,
104104
return UnquantizedLinearMethod()
105105
elif isinstance(layer, LinearBase):
106106
if self.linear_quant_method == "gptq":
107-
config = GPTQMarlinConfig.from_config(self.full_config)
108-
return GPTQMarlinLinearMethod(config)
107+
gptq_config = GPTQMarlinConfig.from_config(self.full_config)
108+
return GPTQMarlinLinearMethod(gptq_config)
109109
elif self.linear_quant_method == "awq":
110-
config = AWQMarlinConfig.from_config(self.full_config)
111-
return AWQMarlinLinearMethod(config)
110+
awq_config = AWQMarlinConfig.from_config(self.full_config)
111+
return AWQMarlinLinearMethod(awq_config)
112112
else:
113113
raise ValueError("moe_quant_int only support gptq and awq.")
114114
elif isinstance(layer, FusedMoE):

0 commit comments

Comments
 (0)