Skip to content

UPSTREAM PR #17624: vulkan: set all memory allocations to high priority#373

Open
loci-dev wants to merge 1 commit intomainfrom
upstream-PR17624-branch_jeffbolznv-priority
Open

UPSTREAM PR #17624: vulkan: set all memory allocations to high priority#373
loci-dev wants to merge 1 commit intomainfrom
upstream-PR17624-branch_jeffbolznv-priority

Conversation

@loci-dev
Copy link
Copy Markdown

Mirrored from ggml-org/llama.cpp#17624

For #17605, though I'm not sure whether it'll help.

@loci-review
Copy link
Copy Markdown

loci-review bot commented Nov 30, 2025

Explore the complete analysis inside the Version Insights

Performance Review Summary

PR #373: Vulkan Memory Priority Implementation

This PR adds Vulkan memory priority support by enabling VK_EXT_memory_priority extension and setting all memory allocations to maximum priority (1.0f). The changes modify device initialization and buffer allocation logic in ggml-vulkan.cpp.

Performance Impact: No measurable performance changes detected across all binaries. All metrics show 0.0% change in response time and throughput. The modifications are runtime configuration changes that do not alter computational paths or execution logic.

Power Consumption: No change detected. All binaries maintain baseline power consumption (libggml-cpu.so: 115,347 nJ).

Inference Impact: No impact on tokens per second. Core inference functions (llama_decode, llama_encode, llama_tokenize) are unaffected by these Vulkan backend initialization changes.

@loci-dev loci-dev force-pushed the main branch 27 times, most recently from fa6cdcc to bf57f85 Compare December 3, 2025 10:10
@loci-dev loci-dev force-pushed the main branch 30 times, most recently from 84f6117 to 91eb894 Compare December 7, 2025 22:08
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants