Skip to content

[Bug] Issue while saving to gguf #3133

@B-Ismail

Description

@B-Ismail
---------------------------------------------------------------------------
RuntimeError                              Traceback (most recent call last)
Cell In[23], line 1
----> 1 model.save_pretrained_gguf("mistral25", tokenizer, quantization_method = "q4_k_m")

File ~/finetuning/venv/lib/python3.12/site-packages/unsloth/save.py:1865, in unsloth_save_pretrained_gguf(self, save_directory, tokenizer, quantization_method, first_conversion, push_to_hub, token, private, is_main_process, state_dict, save_function, max_shard_size, safe_serialization, variant, save_peft_format, tags, temporary_location, maximum_memory_usage)
   1862 is_sentencepiece_model = check_if_sentencepiece_model(self)
   1864 # Save to GGUF
-> 1865 all_file_locations, want_full_precision = save_to_gguf(
   1866     model_type, model_dtype, is_sentencepiece_model,
   1867     new_save_directory, quantization_method, first_conversion, makefile,
   1868 )
   1870 # Save Ollama modelfile
   1871 modelfile = create_ollama_modelfile(tokenizer, all_file_locations[0])

File ~/finetuning/venv/lib/python3.12/site-packages/unsloth/save.py:1093, in save_to_gguf(model_type, model_dtype, is_sentencepiece, model_directory, quantization_method, first_conversion, _run_installer)
   1091     quantize_location = "llama.cpp/build/bin/quantize"
   1092 else:
-> 1093     raise RuntimeError(
   1094         "Unsloth: The file 'llama.cpp/llama-quantize' or `llama.cpp/quantize` does not exist.\n"\
   1095         "We've also double checked the building directory under 'llama.cpp/build/bin/'.\n"\
   1096         "But we expect this file to exist! Check if the file exists under llama.cpp and investigate the building process of llama.cpp (make/cmake)!"
   1097     )
   1098 pass
   1100 # See https://github.com/unslothai/unsloth/pull/730
   1101 # Filenames changed again!

RuntimeError: Unsloth: The file 'llama.cpp/llama-quantize' or `llama.cpp/quantize` does not exist.
We've also double checked the building directory under 'llama.cpp/build/bin/'.
But we expect this file to exist! Check if the file exists under llama.cpp and investigate the building process of llama.cpp (make/cmake)!

can you help me identify the issue here .

i'm using :

pip show unsloth
Name: unsloth
Version: 2025.8.4
Summary: 2-5X faster LLM finetuning
Home-page:
Author: Unsloth AI team
Author-email: [email protected]
License:
Location: /home/user7/finetuning/venv/lib/python3.12/site-packages
Requires: accelerate, bitsandbytes, datasets, diffusers, hf_transfer, huggingface_hub, numpy, packaging, peft, protobuf, psutil, sentencepiece, torch, torchvision, tqdm, transformers, triton, trl, tyro, unsloth_zoo, wheel, xformers
Required-by:

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions