Skip to content

Conversation

@Lpzhan931
Copy link
Contributor

Add a new model openPangu-Embedded-1/7B-V1.1.
Yu can get the the model from model path.

@github-actions github-actions bot added the python python script changes label Nov 2, 2025
@Lpzhan931
Copy link
Contributor Author

This pull request introduces support for the openPangu-Embedded-1/7B-V1.1 models within the llama.cpp.

Usage examples

Convert models to GGUF files:

python convert_hf_to_gguf.py /model/path/openPangu-Embedded-1B-V1.1/
    --outfile pangu_embedded_1B.gguf 
    --outtype f16

python convert_hf_to_gguf.py /model/path/openPangu-Embedded-7B-V1.1/
    --outfile pangu_embedded_7B.gguf 
    --outtype f16

Run the model:

./build/bin/llama-cli -m pangu_embedded_1B.gguf

./build/bin/llama-cli -m pangu_embedded_7B.gguf

References

openPangu-Embedded Models:
Winging/openpangu
openPangu
powered by openPangu(openPangu is a trademark of Huawei Technologies Co., Ltd.)

@Lpzhan931
Copy link
Contributor Author

Hi, @CISC , thanks for reviewing. The model openPangu-Embedded can run correctly and I’ve also uploaded the converted GGUF model to the Hugging Face Hub.

The converted GGUF files can be accessed here: Lpzhan/openPangu-embedded-gguf.

Thank you very much for your time and review!

Below are some screenshots showing the model running successfully:

bench-1 cli-1 cli-2

@Lpzhan931
Copy link
Contributor Author

Hi @CISC,
I’ve addressed your comments and updated the PR.
Please have a look when you have time. Thanks again for your review!

@CISC CISC added the model Model specific label Nov 4, 2025
change the chat-template check condition and some formatting issue

Co-authored-by: Sigbjørn Skjæret <[email protected]>
@CISC CISC merged commit 9f05247 into ggml-org:master Nov 5, 2025
69 of 75 checks passed
gabe-l-hart added a commit to gabe-l-hart/llama.cpp that referenced this pull request Nov 5, 2025
* origin/master: (21 commits)
vulkan: Fix GGML_VULKAN_CHECK_RESULTS to better handle fusion (ggml-org#16919)
examples(gguf): GGUF example outputs (ggml-org#17025)
mtmd: allow QwenVL to process larger image by default (ggml-org#17020)
server : do not default to multiple slots with speculative decoding (ggml-org#17017)
mtmd: improve struct initialization (ggml-org#16981)
docs: Clarify the endpoint that webui uses (ggml-org#17001)
model : add openPangu-Embedded (ggml-org#16941)
ggml webgpu: minor set rows optimization (ggml-org#16810)
sync : ggml
ggml : fix conv2d_dw SVE path (ggml/1380)
CUDA: update ops.md (ggml-org#17005)
opencl: update doc (ggml-org#17011)
refactor: replace sprintf with snprintf for safer string handling in dump functions (ggml-org#16913)
vulkan: remove the need for the dryrun (ggml-org#16826)
server : do context shift only while generating (ggml-org#17000)
readme : update hot topics (ggml-org#17002)
ggml-cpu : bicubic interpolation (ggml-org#16891)
ci : apply model label to models (ggml-org#16994)
chore : fix models indent after refactor (ggml-org#16992)
Fix garbled output with REPACK at high thread counts (ggml-org#16956)
...
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

model Model specific python python script changes

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants