Skip to content
Merged
Show file tree
Hide file tree
Changes from 3 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 4 additions & 1 deletion vllm/model_executor/models/commandr.py
Original file line number Diff line number Diff line change
Expand Up @@ -47,6 +47,8 @@
from vllm.model_executor.utils import set_weight_attrs
from vllm.sequence import IntermediateTensors

from .interfaces import SupportsLoRA


@torch.compile
def layer_norm_func(hidden_states, weight, variance_epsilon):
Expand Down Expand Up @@ -292,7 +294,8 @@ def forward(
return hidden_states


class CohereForCausalLM(nn.Module):
class CohereForCausalLM(nn.Module, SupportsLoRA):
supports_lora = True

packed_modules_mapping = {
"qkv_proj": [
Expand Down
6 changes: 5 additions & 1 deletion vllm/model_executor/models/jamba.py
Original file line number Diff line number Diff line change
Expand Up @@ -38,6 +38,8 @@
from vllm.worker.model_runner import (_BATCH_SIZES_TO_CAPTURE,
_get_graph_batch_size)

from .interfaces import SupportsLoRA

KVCache = Tuple[torch.Tensor, torch.Tensor]


Expand Down Expand Up @@ -539,7 +541,9 @@ def forward(
return hidden_states


class JambaForCausalLM(nn.Module, HasInnerState):
class JambaForCausalLM(nn.Module, HasInnerState, SupportsLoRA):
supports_lora = True

packed_modules_mapping = {
"qkv_proj": [
"q_proj",
Expand Down