-
-
Notifications
You must be signed in to change notification settings - Fork 11.9k
[misc][core] lazy import outlines #7831
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, please make sure to run full CI as it is required to merge (or just use auto-merge). To run full CI, you can do one of these:
🚀 |
|
Nice this is good fix for now! But the users will still have problem with the actually usage :( |
|
yep this is a temp fix, because I heard many people are suffering from this problem but they don't use outlines at all. for people who really want guided decoding and suffer from the problem, they can use |
|
I'm waiting for users' confirmation to see if it works. |
|
Tested on slurm and this pr works well for me |
|
when will this be released? I'm unable to build from source on my HPC server |
|
@alex2awesome you don't need to wait for the release, we have per-commit wheel released. see https://docs.vllm.ai/en/latest/getting_started/installation.html |
|
That's good to know!! unfortunately, after installing the nightly build, I'm still getting this error. Is there a way to delete/refresh the database? |
your installation might be wrong. if you have the latest commit installed, line 8 should not be this one. see
|
|
Ahh thanks @youkaichao — dumb error on my part, I just copy/pasted the instructions in the docs. The right version to use for anyone coming here is: |
|
Hello, guys!! Something that might worth mention is: the problem only occurs if I run the app with my linux user... I log in with another user, everything works fine |
As I understand it the fix is not yet in 0.5.5. You will have to install from source |
|
Hi @youkaichao , has the fix been implemented with the OpenAI-compatible inference engine yet? I've tested and am OK with loading the model in python, as in: https://docs.vllm.ai/en/latest/getting_started/quickstart.html but when I try to launch an inference engine using: I get the same errors: |
|
@alex2awesome it should also work for the api server. your stack trace is incomplete, and it is unclear if you use guided decoding or not. |
Signed-off-by: LeiWang1999 <[email protected]>
temp fix for #4193
for users who don't use guided decoding but use slurm cluster, this makes their lives eaiser.