-
Notifications
You must be signed in to change notification settings - Fork 13.5k
Model: add openPangu-Embedded #16941
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
This pull request introduces support for the openPangu-Embedded-1/7B-V1.1 models within the llama.cpp. Usage examplesConvert models to GGUF files: Run the model: ReferencesopenPangu-Embedded Models: |
|
Hi, @CISC , thanks for reviewing. The model openPangu-Embedded can run correctly and I’ve also uploaded the converted GGUF model to the Hugging Face Hub. The converted GGUF files can be accessed here: Lpzhan/openPangu-embedded-gguf. Thank you very much for your time and review! Below are some screenshots showing the model running successfully:
|
58b0d1a to
6b88093
Compare
|
Hi @CISC, |
6b88093 to
5becaad
Compare
change the chat-template check condition and some formatting issue Co-authored-by: Sigbjørn Skjæret <[email protected]>
* origin/master: (21 commits) vulkan: Fix GGML_VULKAN_CHECK_RESULTS to better handle fusion (ggml-org#16919) examples(gguf): GGUF example outputs (ggml-org#17025) mtmd: allow QwenVL to process larger image by default (ggml-org#17020) server : do not default to multiple slots with speculative decoding (ggml-org#17017) mtmd: improve struct initialization (ggml-org#16981) docs: Clarify the endpoint that webui uses (ggml-org#17001) model : add openPangu-Embedded (ggml-org#16941) ggml webgpu: minor set rows optimization (ggml-org#16810) sync : ggml ggml : fix conv2d_dw SVE path (ggml/1380) CUDA: update ops.md (ggml-org#17005) opencl: update doc (ggml-org#17011) refactor: replace sprintf with snprintf for safer string handling in dump functions (ggml-org#16913) vulkan: remove the need for the dryrun (ggml-org#16826) server : do context shift only while generating (ggml-org#17000) readme : update hot topics (ggml-org#17002) ggml-cpu : bicubic interpolation (ggml-org#16891) ci : apply model label to models (ggml-org#16994) chore : fix models indent after refactor (ggml-org#16992) Fix garbled output with REPACK at high thread counts (ggml-org#16956) ...



Add a new model openPangu-Embedded-1/7B-V1.1.
Yu can get the the model from model path.