UPSTREAM PR #17911: cli: enable jinja by default#515
Conversation
|
Explore the complete analysis inside the Version Insights Pull Request #515 Performance ReviewPR Title: cli: enable jinja by default SummaryThis PR changes the default value of AnalysisCode Changes:
Performance Impact: The observed performance variations occur in STL template instantiations and JSON operations within
Inference Impact: No functions in the core inference pipeline are affected. The following critical functions show zero performance change:
Tokens per second impact: None. The performance variations are isolated to initialization and parameter parsing code paths that execute once at startup, not during token generation. Power Consumption:
The power consumption changes are within measurement noise and reflect the cumulative effect of STL template instantiation differences during parameter initialization, not runtime inference operations. |
c05b224 to
e70bc15
Compare
81e654d to
c785ce2
Compare
Mirrored from ggml-org/llama.cpp#17911
enable jinja by default for: server and CLI
disabled by default for: mtmd-cli and llama-completion