File tree Expand file tree Collapse file tree
Expand file tree Collapse file tree Original file line number Diff line number Diff line change 11{
22 "##" : " The 64-bit version is just a placeholder to pass CI checks." ,
3- "version" : " b5273 " ,
3+ "version" : " b5306 " ,
44 "description" : " (OpenCL Backend for Qualcomm Adreno GPU) llama.cpp enables LLM inference with minimal setup and state-of-the-art performance on a wide range of hardware." ,
55 "homepage" : " https://github.com/ggml-org/llama.cpp" ,
66 "license" : " MIT" ,
77 "architecture" : {
88 "64bit" : {
9- "url" : " https://github.com/ggml-org/llama.cpp/releases/download/b5273 /llama-b5273 -bin-win-avx2 -x64.zip" ,
10- "hash" : " b4e67dead9d459336fdb6cabce18f729f687c8420b00cc87f7523dfc5f3f07c8 "
9+ "url" : " https://github.com/ggml-org/llama.cpp/releases/download/b5306 /llama-b5306 -bin-win-cpu -x64.zip" ,
10+ "hash" : " 5e375cb19af4bdb843d37271e4f546a298dd1687572873fe21e342a03a39f93b "
1111 },
1212 "arm64" : {
13- "url" : " https://github.com/ggml-org/llama.cpp/releases/download/b5273 /llama-b5273 -bin-win-llvm-arm64-opencl-adreno.zip" ,
14- "hash" : " 7918e56a2fcf7db9021077b231ef233b3a0e5e9d89842f168983316921375d51 "
13+ "url" : " https://github.com/ggml-org/llama.cpp/releases/download/b5306 /llama-b5306 -bin-win-llvm-arm64-opencl-adreno.zip" ,
14+ "hash" : " 1c28e2047080c9ad7a657d1a08ffbed47c21468fab6bfc77dd1f91ba25aaec3f "
1515 }
1616 },
1717 "bin" : [
3030 " llama-quantize.exe" ,
3131 " llama-tokenize.exe" ,
3232 " llama-tts.exe" ,
33- " llama-mtmd-cli.exe"
33+ " llama-mtmd-cli.exe" ,
34+ [
35+ " rpc-server.exe" ,
36+ " llama-rpc-server"
37+ ]
3438 ],
3539 "checkver" : {
3640 "url" : " https://github.com/ggml-org/llama.cpp/releases.atom" ,
3943 "autoupdate" : {
4044 "architecture" : {
4145 "64bit" : {
42- "url" : " https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-avx2 -x64.zip"
46+ "url" : " https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-cpu -x64.zip"
4347 },
4448 "arm64" : {
4549 "url" : " https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-llvm-arm64-opencl-adreno.zip"
Load Diff This file was deleted.
Original file line number Diff line number Diff line change 3131 " llama-quantize.exe" ,
3232 " llama-tokenize.exe" ,
3333 " llama-tts.exe" ,
34- " llama-mtmd-cli.exe"
34+ " llama-mtmd-cli.exe" ,
35+ [
36+ " rpc-server.exe" ,
37+ " llama-rpc-server"
38+ ]
3539 ],
3640 "checkver" : {
3741 "url" : " https://github.com/ggml-org/llama.cpp/releases.atom" ,
Original file line number Diff line number Diff line change 2525 " llama-quantize.exe" ,
2626 " llama-tokenize.exe" ,
2727 " llama-tts.exe" ,
28- " llama-mtmd-cli.exe"
28+ " llama-mtmd-cli.exe" ,
29+ [
30+ " rpc-server.exe" ,
31+ " llama-rpc-server"
32+ ]
2933 ],
3034 "checkver" : {
3135 "url" : " https://github.com/ggml-org/llama.cpp/releases.atom" ,
Load Diff This file was deleted.
Original file line number Diff line number Diff line change 2525 " llama-quantize.exe" ,
2626 " llama-tokenize.exe" ,
2727 " llama-tts.exe" ,
28- " llama-mtmd-cli.exe"
28+ " llama-mtmd-cli.exe" ,
29+ [
30+ " rpc-server.exe" ,
31+ " llama-rpc-server"
32+ ]
2933 ],
3034 "checkver" : {
3135 "url" : " https://github.com/ggml-org/llama.cpp/releases.atom" ,
Original file line number Diff line number Diff line change 2525 " llama-quantize.exe" ,
2626 " llama-tokenize.exe" ,
2727 " llama-tts.exe" ,
28- " llama-mtmd-cli.exe"
28+ " llama-mtmd-cli.exe" ,
29+ [
30+ " rpc-server.exe" ,
31+ " llama-rpc-server"
32+ ]
2933 ],
3034 "checkver" : {
3135 "url" : " https://github.com/ggml-org/llama.cpp/releases.atom" ,
Original file line number Diff line number Diff line change 2525 " llama-quantize.exe" ,
2626 " llama-tokenize.exe" ,
2727 " llama-tts.exe" ,
28- " llama-mtmd-cli.exe"
28+ " llama-mtmd-cli.exe" ,
29+ [
30+ " rpc-server.exe" ,
31+ " llama-rpc-server"
32+ ]
2933 ],
3034 "checkver" : {
3135 "url" : " https://github.com/ggml-org/llama.cpp/releases.atom" ,
Original file line number Diff line number Diff line change 11{
2- "version" : " b5273 " ,
2+ "version" : " b5306 " ,
33 "description" : " (CPU Backend) llama.cpp enables LLM inference with minimal setup and state-of-the-art performance on a wide range of hardware." ,
44 "homepage" : " https://github.com/ggml-org/llama.cpp" ,
55 "license" : " MIT" ,
66 "architecture" : {
77 "64bit" : {
8- "url" : " https://github.com/ggml-org/llama.cpp/releases/download/b5273 /llama-b5273 -bin-win-avx2 -x64.zip" ,
9- "hash" : " b4e67dead9d459336fdb6cabce18f729f687c8420b00cc87f7523dfc5f3f07c8 "
8+ "url" : " https://github.com/ggml-org/llama.cpp/releases/download/b5306 /llama-b5306 -bin-win-cpu -x64.zip" ,
9+ "hash" : " 5e375cb19af4bdb843d37271e4f546a298dd1687572873fe21e342a03a39f93b "
1010 },
1111 "arm64" : {
12- "url" : " https://github.com/ggml-org/llama.cpp/releases/download/b5273 /llama-b5273 -bin-win-llvm-arm64.zip" ,
13- "hash" : " 07808be5182c99cc6105d2bc646470a479e3befd12ccb811f24b0a9646397d87 "
12+ "url" : " https://github.com/ggml-org/llama.cpp/releases/download/b5306 /llama-b5306 -bin-win-llvm-arm64.zip" ,
13+ "hash" : " d0ab18fca59852b749c50f9014aa2825bce47a6f16361d04a57419e59103699e "
1414 }
1515 },
1616 "bin" : [
2929 " llama-quantize.exe" ,
3030 " llama-tokenize.exe" ,
3131 " llama-tts.exe" ,
32- " llama-mtmd-cli.exe"
32+ " llama-mtmd-cli.exe" ,
33+ [
34+ " rpc-server.exe" ,
35+ " llama-rpc-server"
36+ ]
3337 ],
3438 "checkver" : {
3539 "url" : " https://github.com/ggml-org/llama.cpp/releases.atom" ,
3842 "autoupdate" : {
3943 "architecture" : {
4044 "64bit" : {
41- "url" : " https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-avx2 -x64.zip"
45+ "url" : " https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-cpu -x64.zip"
4246 },
4347 "arm64" : {
4448 "url" : " https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-llvm-arm64.zip"
You can’t perform that action at this time.
0 commit comments