Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
9 changes: 1 addition & 8 deletions vllm/config/model.py
Original file line number Diff line number Diff line change
Expand Up @@ -1726,19 +1726,12 @@ def head_dtype(self) -> torch.dtype:
logger.debug_once("head dtype: %s", head_dtype)
return head_dtype

@property
def hidden_size(self):
if hasattr(self.hf_config, "hidden_size"):
return self.hf_config.hidden_size
text_config = self.hf_config.get_text_config()
return text_config.hidden_size

@property
def embedding_size(self):
dense_modules = try_get_dense_modules(self.model, revision=self.revision)
if dense_modules is not None:
return dense_modules[-1]["out_features"]
return self.hidden_size
return self.get_hidden_size()

def get_and_verify_max_len(self, max_model_len: int):
# Consider max_model_len in tokenizer_config only when
Expand Down
2 changes: 1 addition & 1 deletion vllm/model_executor/models/adapters.py
Original file line number Diff line number Diff line change
Expand Up @@ -301,7 +301,7 @@ def _init_pooler(self, vllm_config: "VllmConfig", prefix: str = ""):
quant_config = vllm_config.quant_config

self.score = ReplicatedLinear(
model_config.hidden_size,
model_config.get_hidden_size(),
text_config.num_labels,
bias=False,
params_dtype=vllm_config.model_config.head_dtype,
Expand Down