-
-
Notifications
You must be signed in to change notification settings - Fork 11.7k
Closed
Labels
feature requestNew feature or requestNew feature or request
Description
🚀 The feature, motivation and pitch
Recently, we read a paper where the vLLM team proposed a method called SmartSpec.
We believe that the research, which dynamically adjusts the speculation length in a commercialized LLM serving system, is superior in terms of practicality compared to existing dynamic speculative length studies.
This idea could be applied to the current vLLM speculative decoding with Batch Expansion enabled, and it might also be applicable to future versions of vLLM with Batch Expansion disabled.
(I am curious whether the SmartSpec research was conducted on vLLM with Batch Expansion enabled. 🤔)
I wonder if the SmartSpec method will be implemented into the main repository in the near future.
Alternatives
No response
Additional context
No response
Metadata
Metadata
Assignees
Labels
feature requestNew feature or requestNew feature or request