-
-
Notifications
You must be signed in to change notification settings - Fork 4.1k
Closed
Labels
Description
---------------------------------------------------------------------------
RuntimeError Traceback (most recent call last)
Cell In[23], line 1
----> 1 model.save_pretrained_gguf("mistral25", tokenizer, quantization_method = "q4_k_m")
File ~/finetuning/venv/lib/python3.12/site-packages/unsloth/save.py:1865, in unsloth_save_pretrained_gguf(self, save_directory, tokenizer, quantization_method, first_conversion, push_to_hub, token, private, is_main_process, state_dict, save_function, max_shard_size, safe_serialization, variant, save_peft_format, tags, temporary_location, maximum_memory_usage)
1862 is_sentencepiece_model = check_if_sentencepiece_model(self)
1864 # Save to GGUF
-> 1865 all_file_locations, want_full_precision = save_to_gguf(
1866 model_type, model_dtype, is_sentencepiece_model,
1867 new_save_directory, quantization_method, first_conversion, makefile,
1868 )
1870 # Save Ollama modelfile
1871 modelfile = create_ollama_modelfile(tokenizer, all_file_locations[0])
File ~/finetuning/venv/lib/python3.12/site-packages/unsloth/save.py:1093, in save_to_gguf(model_type, model_dtype, is_sentencepiece, model_directory, quantization_method, first_conversion, _run_installer)
1091 quantize_location = "llama.cpp/build/bin/quantize"
1092 else:
-> 1093 raise RuntimeError(
1094 "Unsloth: The file 'llama.cpp/llama-quantize' or `llama.cpp/quantize` does not exist.\n"\
1095 "We've also double checked the building directory under 'llama.cpp/build/bin/'.\n"\
1096 "But we expect this file to exist! Check if the file exists under llama.cpp and investigate the building process of llama.cpp (make/cmake)!"
1097 )
1098 pass
1100 # See https://github.com/unslothai/unsloth/pull/730
1101 # Filenames changed again!
RuntimeError: Unsloth: The file 'llama.cpp/llama-quantize' or `llama.cpp/quantize` does not exist.
We've also double checked the building directory under 'llama.cpp/build/bin/'.
But we expect this file to exist! Check if the file exists under llama.cpp and investigate the building process of llama.cpp (make/cmake)!
can you help me identify the issue here .
i'm using :
pip show unsloth
Name: unsloth
Version: 2025.8.4
Summary: 2-5X faster LLM finetuning
Home-page:
Author: Unsloth AI team
Author-email: [email protected]
License:
Location: /home/user7/finetuning/venv/lib/python3.12/site-packages
Requires: accelerate, bitsandbytes, datasets, diffusers, hf_transfer, huggingface_hub, numpy, packaging, peft, protobuf, psutil, sentencepiece, torch, torchvision, tqdm, transformers, triton, trl, tyro, unsloth_zoo, wheel, xformers
Required-by:
Isam-tfares