Skip to content

Commit b98cbdb

Browse files
committed
format
Signed-off-by: jiahanc <[email protected]>
1 parent 5398428 commit b98cbdb

File tree

3 files changed

+2
-11
lines changed

3 files changed

+2
-11
lines changed

vllm/model_executor/layers/fused_moe/flashinfer_trtllm_moe.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
# SPDX-License-Identifier: Apache-2.0
22
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
33

4-
54
import torch
65

76
from vllm.model_executor.layers.fused_moe.utils import moe_kernel_quantize_input

vllm/model_executor/layers/quantization/fp8.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1291,6 +1291,7 @@ def apply(
12911291
# can override fused_experts or cutlass but not rocm or marlin.
12921292
#
12931293
topk_weights, topk_ids, zero_expert_result = select_result
1294+
12941295
if self.rocm_aiter_moe_enabled:
12951296
from vllm.model_executor.layers.fused_moe.rocm_aiter_fused_moe import ( # noqa: E501
12961297
rocm_aiter_fused_experts,

vllm/model_executor/models/qwen3_next.py

Lines changed: 1 addition & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -181,10 +181,10 @@ def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
181181
orig_shape = hidden_states.shape
182182
num_tokens, hidden_dim = hidden_states.shape
183183
hidden_states = hidden_states.view(-1, hidden_dim)
184+
184185
if self.is_sequence_parallel:
185186
hidden_states = sequence_parallel_chunk(hidden_states)
186187

187-
<<<<<<< HEAD
188188
if self.experts.is_internal_router:
189189
# In this case, the gate/router runs inside the FusedMoE class
190190
final_hidden_states = self.experts(
@@ -196,14 +196,6 @@ def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
196196
final_hidden_states = self.experts(
197197
hidden_states=hidden_states, router_logits=router_logits
198198
)
199-
=======
200-
# print(self.prefix_print)
201-
# router_logits: (num_tokens, n_experts)
202-
router_logits, _ = self.gate(hidden_states)
203-
final_hidden_states = self.experts(
204-
hidden_states=hidden_states, router_logits=router_logits
205-
)
206-
>>>>>>> 9d88f1762 (update work)
207199

208200
if self.shared_expert is not None:
209201
final_hidden_states = final_hidden_states[0] + final_hidden_states[1]
@@ -1013,7 +1005,6 @@ def forward(
10131005
{"hidden_states": hidden_states, "residual": residual}
10141006
)
10151007
hidden_states, _ = self.norm(hidden_states, residual)
1016-
# print("="*60)
10171008
return hidden_states
10181009

10191010
def get_expert_mapping(self) -> list[tuple[str, str, int, str]]:

0 commit comments

Comments
 (0)