UPSTREAM PR #17898: model : Qwen3-Next-80B-A3B has 48 layers#506
UPSTREAM PR #17898: model : Qwen3-Next-80B-A3B has 48 layers#506
Conversation
|
Explore the complete analysis inside the Version Insights Pull Request #506 Technical ReviewPR SummaryTitle: UPSTREAM PR #17898: model : Qwen3-Next-80B-A3B has 48 layers Code Changes AnalysisModified File: Change 1 - llm_type_name() function (line 123):
Change 2 - llama_model::load_hparams() function (line 2261):
Performance Impact AssessmentFunction: llm_type_name()
Function: llama_model::load_hparams()
Inference Performance: Power Consumption:
The changes are cosmetic corrections to model metadata with no measurable impact on inference performance or throughput. |
de9b0c0 to
b28744d
Compare
6c677ac to
c39aef9
Compare
Mirrored from ggml-org/llama.cpp#17898
Qwen3-Next-80B-A3B has 48 layers instead of 80, as pointed out by model README and a comment in original PR.
This change should be purely cosmetic, fixes "?B" model names shown by
llama-bench, etc.