Skip to content

Commit 02fa4a4

Browse files
committed
llama-cpp: update packages
ggml-org/llama.cpp#13220
1 parent 5d4a2e3 commit 02fa4a4

9 files changed

Lines changed: 47 additions & 101 deletions

bucket/llama-cpp-adreno.json

Lines changed: 11 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,17 +1,17 @@
11
{
22
"##": "The 64-bit version is just a placeholder to pass CI checks.",
3-
"version": "b5273",
3+
"version": "b5306",
44
"description": "(OpenCL Backend for Qualcomm Adreno GPU) llama.cpp enables LLM inference with minimal setup and state-of-the-art performance on a wide range of hardware.",
55
"homepage": "https://github.com/ggml-org/llama.cpp",
66
"license": "MIT",
77
"architecture": {
88
"64bit": {
9-
"url": "https://github.com/ggml-org/llama.cpp/releases/download/b5273/llama-b5273-bin-win-avx2-x64.zip",
10-
"hash": "b4e67dead9d459336fdb6cabce18f729f687c8420b00cc87f7523dfc5f3f07c8"
9+
"url": "https://github.com/ggml-org/llama.cpp/releases/download/b5306/llama-b5306-bin-win-cpu-x64.zip",
10+
"hash": "5e375cb19af4bdb843d37271e4f546a298dd1687572873fe21e342a03a39f93b"
1111
},
1212
"arm64": {
13-
"url": "https://github.com/ggml-org/llama.cpp/releases/download/b5273/llama-b5273-bin-win-llvm-arm64-opencl-adreno.zip",
14-
"hash": "7918e56a2fcf7db9021077b231ef233b3a0e5e9d89842f168983316921375d51"
13+
"url": "https://github.com/ggml-org/llama.cpp/releases/download/b5306/llama-b5306-bin-win-llvm-arm64-opencl-adreno.zip",
14+
"hash": "1c28e2047080c9ad7a657d1a08ffbed47c21468fab6bfc77dd1f91ba25aaec3f"
1515
}
1616
},
1717
"bin": [
@@ -30,7 +30,11 @@
3030
"llama-quantize.exe",
3131
"llama-tokenize.exe",
3232
"llama-tts.exe",
33-
"llama-mtmd-cli.exe"
33+
"llama-mtmd-cli.exe",
34+
[
35+
"rpc-server.exe",
36+
"llama-rpc-server"
37+
]
3438
],
3539
"checkver": {
3640
"url": "https://github.com/ggml-org/llama.cpp/releases.atom",
@@ -39,7 +43,7 @@
3943
"autoupdate": {
4044
"architecture": {
4145
"64bit": {
42-
"url": "https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-avx2-x64.zip"
46+
"url": "https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-cpu-x64.zip"
4347
},
4448
"arm64": {
4549
"url": "https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-llvm-arm64-opencl-adreno.zip"

bucket/llama-cpp-avx512.json

Lines changed: 0 additions & 41 deletions
This file was deleted.

bucket/llama-cpp-cuda.json

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,11 @@
3131
"llama-quantize.exe",
3232
"llama-tokenize.exe",
3333
"llama-tts.exe",
34-
"llama-mtmd-cli.exe"
34+
"llama-mtmd-cli.exe",
35+
[
36+
"rpc-server.exe",
37+
"llama-rpc-server"
38+
]
3539
],
3640
"checkver": {
3741
"url": "https://github.com/ggml-org/llama.cpp/releases.atom",

bucket/llama-cpp-hip.json

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,11 @@
2525
"llama-quantize.exe",
2626
"llama-tokenize.exe",
2727
"llama-tts.exe",
28-
"llama-mtmd-cli.exe"
28+
"llama-mtmd-cli.exe",
29+
[
30+
"rpc-server.exe",
31+
"llama-rpc-server"
32+
]
2933
],
3034
"checkver": {
3135
"url": "https://github.com/ggml-org/llama.cpp/releases.atom",

bucket/llama-cpp-kompute.json

Lines changed: 0 additions & 41 deletions
This file was deleted.

bucket/llama-cpp-openblas.json

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,11 @@
2525
"llama-quantize.exe",
2626
"llama-tokenize.exe",
2727
"llama-tts.exe",
28-
"llama-mtmd-cli.exe"
28+
"llama-mtmd-cli.exe",
29+
[
30+
"rpc-server.exe",
31+
"llama-rpc-server"
32+
]
2933
],
3034
"checkver": {
3135
"url": "https://github.com/ggml-org/llama.cpp/releases.atom",

bucket/llama-cpp-sycl.json

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,11 @@
2525
"llama-quantize.exe",
2626
"llama-tokenize.exe",
2727
"llama-tts.exe",
28-
"llama-mtmd-cli.exe"
28+
"llama-mtmd-cli.exe",
29+
[
30+
"rpc-server.exe",
31+
"llama-rpc-server"
32+
]
2933
],
3034
"checkver": {
3135
"url": "https://github.com/ggml-org/llama.cpp/releases.atom",

bucket/llama-cpp-vulkan.json

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,11 @@
2525
"llama-quantize.exe",
2626
"llama-tokenize.exe",
2727
"llama-tts.exe",
28-
"llama-mtmd-cli.exe"
28+
"llama-mtmd-cli.exe",
29+
[
30+
"rpc-server.exe",
31+
"llama-rpc-server"
32+
]
2933
],
3034
"checkver": {
3135
"url": "https://github.com/ggml-org/llama.cpp/releases.atom",

bucket/llama-cpp.json

Lines changed: 11 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,16 +1,16 @@
11
{
2-
"version": "b5273",
2+
"version": "b5306",
33
"description": "(CPU Backend) llama.cpp enables LLM inference with minimal setup and state-of-the-art performance on a wide range of hardware.",
44
"homepage": "https://github.com/ggml-org/llama.cpp",
55
"license": "MIT",
66
"architecture": {
77
"64bit": {
8-
"url": "https://github.com/ggml-org/llama.cpp/releases/download/b5273/llama-b5273-bin-win-avx2-x64.zip",
9-
"hash": "b4e67dead9d459336fdb6cabce18f729f687c8420b00cc87f7523dfc5f3f07c8"
8+
"url": "https://github.com/ggml-org/llama.cpp/releases/download/b5306/llama-b5306-bin-win-cpu-x64.zip",
9+
"hash": "5e375cb19af4bdb843d37271e4f546a298dd1687572873fe21e342a03a39f93b"
1010
},
1111
"arm64": {
12-
"url": "https://github.com/ggml-org/llama.cpp/releases/download/b5273/llama-b5273-bin-win-llvm-arm64.zip",
13-
"hash": "07808be5182c99cc6105d2bc646470a479e3befd12ccb811f24b0a9646397d87"
12+
"url": "https://github.com/ggml-org/llama.cpp/releases/download/b5306/llama-b5306-bin-win-llvm-arm64.zip",
13+
"hash": "d0ab18fca59852b749c50f9014aa2825bce47a6f16361d04a57419e59103699e"
1414
}
1515
},
1616
"bin": [
@@ -29,7 +29,11 @@
2929
"llama-quantize.exe",
3030
"llama-tokenize.exe",
3131
"llama-tts.exe",
32-
"llama-mtmd-cli.exe"
32+
"llama-mtmd-cli.exe",
33+
[
34+
"rpc-server.exe",
35+
"llama-rpc-server"
36+
]
3337
],
3438
"checkver": {
3539
"url": "https://github.com/ggml-org/llama.cpp/releases.atom",
@@ -38,7 +42,7 @@
3842
"autoupdate": {
3943
"architecture": {
4044
"64bit": {
41-
"url": "https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-avx2-x64.zip"
45+
"url": "https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-cpu-x64.zip"
4246
},
4347
"arm64": {
4448
"url": "https://github.com/ggml-org/llama.cpp/releases/download/$version/llama-$version-bin-win-llvm-arm64.zip"

0 commit comments

Comments
 (0)