Skip to content

Conversation

@prashantgupta24
Copy link
Collaborator

@prashantgupta24 prashantgupta24 commented Sep 15, 2025

Description

Bump upper bound of vllm to 0.10.2

Related Issues

@github-actions
Copy link

👋 Hi! Thank you for contributing to vLLM support on Spyre.
Just a reminder: Make sure that your code passes all the linting checks, otherwise your PR won't be able to be merged. To do so, first install the linting requirements, then run format.sh and commit the changes. This can be done with uv directly:

uv sync --frozen --group lint --active --inexact

Or this can be done with pip:

uv pip compile --group lint > requirements-lint.txt
pip install -r requirements-lint.txt
bash format.sh

Now you are good to go 🚀

@prashantgupta24 prashantgupta24 added the ready Runs the full CI test suite. Only add to PRs once ready to merge to limit public GHA usage label Sep 15, 2025
Signed-off-by: Prashant Gupta <[email protected]>
@prashantgupta24 prashantgupta24 changed the title ⬆️ make vllm >=0.10.1.1,<=0.10.2 ⬆️ bump vllm lower bound and support 0.10.2 Sep 15, 2025

with pytest.raises(BadRequestError,
match="This model's maximum context length is"):
with pytest.raises(BadRequestError, match="maximum context length is"):
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is a bug in vllm upstream - opened a PR vllm-project/vllm#24995

Signed-off-by: Prashant Gupta <[email protected]>
prashantgupta24 and others added 6 commits September 17, 2025 11:21
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Co-authored-by: Joe Runde <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
@prashantgupta24 prashantgupta24 changed the title ⬆️ bump vllm lower bound and support 0.10.2 ⬆️ bump vllm upper bound and support 0.10.2 Sep 17, 2025
@prashantgupta24 prashantgupta24 changed the title ⬆️ bump vllm upper bound and support 0.10.2 ⬆️ bump vllm upper bound to support 0.10.2 Sep 17, 2025
Copy link
Collaborator

@maxdebayser maxdebayser left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The pooling correction works. So now we only need backwards compatibility tests for the cursor stuff.

@prashantgupta24
Copy link
Collaborator Author

prashantgupta24 commented Sep 18, 2025

The pooling correction works. So now we only need backwards compatibility tests for the cursor stuff.

Oh awesome! Do we want to review and get #468 in first so that we're not stuck in a state where we have to be backward compatible with 0.10.2 (which is going to be much harder vs 0.10.1.1)

Edit: merged!

Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
@prashantgupta24
Copy link
Collaborator Author

prashantgupta24 commented Sep 22, 2025

Hold on, scoring is still failing on 0.10.2 :(

Nope, all pass! Ready for review!

@prashantgupta24
Copy link
Collaborator Author

bot:test

1 similar comment
@prashantgupta24
Copy link
Collaborator Author

bot:test

@prashantgupta24
Copy link
Collaborator Author

Some CI tests failing, looking...

@prashantgupta24
Copy link
Collaborator Author

bot:test

@prashantgupta24
Copy link
Collaborator Author

CI tests all pass! Time for a review

Copy link
Collaborator

@rafvasq rafvasq left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

just a typo

Signed-off-by: Prashant Gupta <[email protected]>

if task == "embed":
self.pooler = Pooler.for_embed(pooler_config=pooler_config)
with set_current_vllm_config(self.vllm_config):
Copy link
Collaborator Author

@prashantgupta24 prashantgupta24 Sep 24, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

we need to set_current_vllm_config because now Pooler class needs the vllm config for it to read vllm_config.model_config.head_dtype

@maxdebayser
Copy link
Collaborator

bot:test

@maxdebayser
Copy link
Collaborator

Merging as all non-main tests are passing

@maxdebayser maxdebayser merged commit 5aa87d5 into main Sep 24, 2025
19 of 27 checks passed
@maxdebayser maxdebayser deleted the upstream-versions branch September 24, 2025 20:53
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready Runs the full CI test suite. Only add to PRs once ready to merge to limit public GHA usage

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants