Skip to content

Commit 897ad17

Browse files
authored
chore(model gallery): add qwen3-coder-30b-a3b-instruct based on model request (#8082)
* chore(model gallery): add qwen3-coder-30b-a3b-instruct based on model request Signed-off-by: rampa3 <[email protected]> * added missing model config import URL Signed-off-by: rampa3 <[email protected]> --------- Signed-off-by: rampa3 <[email protected]>
1 parent 16a18a2 commit 897ad17

File tree

1 file changed

+35
-0
lines changed

1 file changed

+35
-0
lines changed

gallery/index.yaml

Lines changed: 35 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -3822,6 +3822,41 @@
38223822
- filename: boomerang-qwen3-4.9B.Q4_K_M.gguf
38233823
sha256: 11e6c068351d104dee31dd63550e5e2fc9be70467c1cfc07a6f84030cb701537
38243824
uri: huggingface://mradermacher/boomerang-qwen3-4.9B-GGUF/boomerang-qwen3-4.9B.Q4_K_M.gguf
3825+
- !!merge <<: *qwen3
3826+
name: "qwen3-coder-30b-a3b-instruct"
3827+
icon: https://cdn-avatars.huggingface.co/v1/production/uploads/620760a26e3b7210c2ff1943/-s1gyJfvbE1RgO5iBeNOi.png
3828+
url: "github:mudler/LocalAI/gallery/qwen3.yaml@master"
3829+
urls:
3830+
- https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct
3831+
- https://huggingface.co/unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF
3832+
description: |
3833+
Qwen3-Coder is available in multiple sizes. Today, we're excited to introduce Qwen3-Coder-30B-A3B-Instruct. This streamlined model maintains impressive performance and efficiency, featuring the following key enhancements:
3834+
3835+
- Significant Performance among open models on Agentic Coding, Agentic Browser-Use, and other foundational coding tasks.
3836+
- Long-context Capabilities with native support for 256K tokens, extendable up to 1M tokens using Yarn, optimized for repository-scale understanding.
3837+
- Agentic Coding supporting for most platform such as Qwen Code, CLINE, featuring a specially designed function call format.
3838+
3839+
3840+
Model Overview:
3841+
Qwen3-Coder-30B-A3B-Instruct has the following features:
3842+
3843+
- Type: Causal Language Models
3844+
- Training Stage: Pretraining & Post-training
3845+
- Number of Parameters: 30.5B in total and 3.3B activated
3846+
- Number of Layers: 48
3847+
- Number of Attention Heads (GQA): 32 for Q and 4 for KV
3848+
- Number of Experts: 128
3849+
- Number of Activated Experts: 8
3850+
- Context Length: 262,144 natively.
3851+
3852+
NOTE: This model supports only non-thinking mode and does not generate <think></think> blocks in its output. Meanwhile, specifying enable_thinking=False is no longer required.
3853+
overrides:
3854+
parameters:
3855+
model: Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf
3856+
files:
3857+
- filename: Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf
3858+
sha256: fadc3e5f8d42bf7e894a785b05082e47daee4df26680389817e2093056f088ad
3859+
uri: huggingface://unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF/Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf
38253860
- &gemma3
38263861
url: "github:mudler/LocalAI/gallery/gemma.yaml@master"
38273862
name: "gemma-3-27b-it"

0 commit comments

Comments
 (0)