Conversation
Signed-off-by: Wang Yang <yangwang@iscas.ac.cn>
|
Explore the complete analysis inside the Version Insights Performance Analysis Summary: PR #301 - RISC-V CPU Features DetectionAssessmentThis PR introduces RISC-V CPU feature detection infrastructure without modifying core inference logic. Performance analysis shows no measurable impact across all binaries, with power consumption changes below 0.001% (< 1 nJ absolute delta). No performance-critical functions were modified. Change OverviewThe PR adds three components:
These changes are build infrastructure only - no modifications to model loading, tokenization, batch processing, or inference paths. Performance MetricsPower Consumption (Binary-Level):
Function-Level Analysis:
Flame Graph & CFG Analysis:
Code Review FindingsStrengths:
Implementation Quality:
ConclusionThis PR establishes foundation for RISC-V optimization without affecting current performance. The infrastructure enables future performance gains (4-8x on vector operations) when RVV-optimized code paths are utilized on compatible hardware. No action required for existing deployments. |
a89c6ad to
ad5ad9a
Compare
38683c7 to
fa6cdcc
Compare
Mirrored from ggml-org/llama.cpp#17461
This PR introduces the CPU features detection for the RISC-V platform and allows for dynamic backend loading when compiled with -DGGML_BACKEND_DL=ON -DGGML_CPU_ALL_VARIANTS=ON.
1、Build this PR using:
2、Check that there are 2 libggml-cpu*.so files built:
3、Run a test prompt and let me know which library is loaded via:
build/bin/llama-cli -m Qwen3-0.6B-Q4_K_M.gguf -no-cnv --seed 42 -n 50 -p "Write me a dog walking business idea 1. " 2>&1 | lessHelp me paste the first few outputs from the top. It should print something like this at the top and it should run the prompt completely without problems.