Skip to content

Conversation

@danielhanchen
Copy link
Contributor

@danielhanchen danielhanchen commented Jul 17, 2025

  1. Fixes Gemma 3N failing on torch 2.7 with out of bounds errors - [Bug] Gemma 3n inference fails on Windows #2941 Possibly [Bug] Finetune Gemma-3n throws canUse32BitIndexMath error #2940
  2. Fixes GRPO state2 missing / NaN for vllm >= 0.9.1
  3. Fixes mamba style training for Liquid, Falcon H1 etc
  4. Fixes many other small bugs
  5. Fixes In GRPO special tokens are removed if max_prompt_length is set, else not #2980 - special tokens and pad tokens not handled properly in GRPO
  6. Fixes [Bug] Compile cache miss typing.Any #2874
  7. Inference somewhat faster

Update Unsloth via:

pip install --upgrade --force-reinstall --no-cache-dir --no-deps unsloth unsloth_zoo

@danielhanchen danielhanchen merged commit 7ebde8e into main Jul 17, 2025
3 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

In GRPO special tokens are removed if max_prompt_length is set, else not [Bug] Compile cache miss typing.Any

2 participants