Skip to content

UPSTREAM PR #18079: model-conversion : use CONVERTED_EMBEDDING_MODEL for embedding_verify_logits#588

Open
loci-dev wants to merge 1 commit intomainfrom
upstream-PR18079-branch_danbev-model-conversion-embedding-converted-model-path
Open

UPSTREAM PR #18079: model-conversion : use CONVERTED_EMBEDDING_MODEL for embedding_verify_logits#588
loci-dev wants to merge 1 commit intomainfrom
upstream-PR18079-branch_danbev-model-conversion-embedding-converted-model-path

Conversation

@loci-dev
Copy link
Copy Markdown

Mirrored from ggml-org/llama.cpp#18079

This commit updates the embedding model verification script to use the CONVERTED_EMBEDDING_MODEL environment variable instead of using the EMBEDDING_MODEL_PATH (the original embedding model path) as the basis for the converted model file name.

The motivation for this that currently if the converted embedding model file name differs from the original embedding model directory/name the verification script will look for the wrong .bin files that were generating when running the models.

…_logits

This commit updates the embedding model verification script to use the
CONVERTED_EMBEDDING_MODEL environment variable instead of using the
EMBEDDING_MODEL_PATH (the original embedding model path) as the basis
for the converted model file name.

The motivation for this that currently if the converted embedding model
file name differs from the original embedding model directory/name the
verification script will look for the wrong .bin files that were
generating when running the models.
@loci-review
Copy link
Copy Markdown

loci-review bot commented Dec 16, 2025

Explore the complete analysis inside the Version Insights

Performance Analysis Summary: PR #588

Analysis Scope: Model conversion embedding verification script update
Files Modified: 1 shell script (non-compiled)
Performance Impact: None

This PR modifies a bash script used for embedding verification in the model conversion workflow. The change corrects file path resolution to use the converted model name instead of the original model name. No compiled binaries were modified. All performance metrics (response time, throughput, power consumption) remain unchanged at 0% across all 16 analyzed binaries including libllama.so, libggml-cpu.so, and llama-run. The script modification adds two variable assignments for path resolution with negligible execution overhead (sub-microsecond). No functions in performance-critical areas (llama_decode, llama_encode, ggml_compute_forward, ggml_mul_mat) were affected. Tokens per second for inference workloads remains unaffected.

@loci-dev loci-dev force-pushed the main branch 27 times, most recently from c785ce2 to ab5b02c Compare December 18, 2025 14:10
@loci-dev loci-dev force-pushed the main branch 30 times, most recently from c07a58e to c71ff69 Compare December 24, 2025 12:14
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants