When I convert Llama model with convert-pth-to-ggml.py, quantize to 4bit, and load it with gpt4all, I get this:
llama_model_load: invalid model file 'ggml-model-q4_0.bin' (bad magic)
Could you implement to support ggml format that gpt4all uses?
Thanks!