Skip to content

Conversation

@tomeras91
Copy link
Contributor

The current Jamba implementation doesn't allow use of Cuda Graph with batch sizes that weren't captured. When trying to do that, a RuntimeError is raised with incompatible tensor shapes. This happens because the mamba cache wasn't padded to the Cuda Graph batch size.

This PR fixes this issue, and adds a test to assert it.

tomeras91 added 3 commits July 8, 2024 15:50
…ape and not by number of sequences. This is so we pad the mamba cache to the captured CG batch sizes
@simon-mo simon-mo merged commit ddc369f into vllm-project:main Jul 8, 2024
tlrmchlsmth added a commit to neuralmagic/nm-vllm that referenced this pull request Jul 16, 2024
xjpang pushed a commit to xjpang/vllm that referenced this pull request Jul 24, 2024
@tomeras91 tomeras91 deleted the mamba-cg-cache-padding branch August 12, 2024 15:00
Alvant pushed a commit to compressa-ai/vllm that referenced this pull request Oct 26, 2024
LeiWang1999 pushed a commit to LeiWang1999/vllm-bitblas that referenced this pull request Mar 26, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants