You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Merge remote-tracking branch 'upstream/main' into sharded-tensorizer
* upstream/main: (126 commits)
[Bugfix][Frontend] Cleanup "fix chat logprobs" (vllm-project#5026)
[Bugfix] OpenAI entrypoint limits logprobs while ignoring server defined --max-logprobs (vllm-project#5312)
[Misc] Various simplifications and typing fixes (vllm-project#5368)
[ci] Fix Buildkite agent path (vllm-project#5392)
[Doc] Add documentation for FP8 W8A8 (vllm-project#5388)
Bump version to v0.5.0 (vllm-project#5384)
[Docs] Alphabetically sort sponsors (vllm-project#5386)
[Docs] Add Docs on Limitations of VLM Support (vllm-project#5383)
[ci] Mount buildkite agent on Docker container to upload benchmark results (vllm-project#5330)
[ci] Use small_cpu_queue for doc build (vllm-project#5331)
[Bugfix] Fix LLaVA-NeXT (vllm-project#5380)
[Feature][Frontend]: Continued `stream_options` implementation also in CompletionRequest (vllm-project#5319)
[Model] Initial support for LLaVA-NeXT (vllm-project#4199)
[Misc] Improve error message when LoRA parsing fails (vllm-project#5194)
[misc][typo] fix typo (vllm-project#5372)
[Frontend][Misc] Enforce Pixel Values as Input Type for VLMs in API Server (vllm-project#5374)
[Misc] Update to comply with the new `compressed-tensors` config (vllm-project#5350)
[Bugfix] Fix KeyError: 1 When Using LoRA adapters (vllm-project#5164)
[Kernel][Misc] Use TORCH_LIBRARY instead of PYBIND11_MODULE for custom ops (vllm-project#5047)
[mis][ci/test] fix flaky test in test_sharded_state_loader.py (vllm-project#5361)
...
0 commit comments