Skip to content

llama : fix KV cache quantization for hybrid Mamba/attention models#1548

Closed
jnovy wants to merge 1 commit intoikawrakow:mainfrom
jnovy:fix-hybrid-kv-cache-quantization
Closed

llama : fix KV cache quantization for hybrid Mamba/attention models#1548
jnovy wants to merge 1 commit intoikawrakow:mainfrom
jnovy:fix-hybrid-kv-cache-quantization

Commits

Commits on Mar 29, 2026