UPSTREAM PR #17425: server : add Anthropic Messages API support#279
UPSTREAM PR #17425: server : add Anthropic Messages API support#279
Conversation
0f2d111 to
8e531a0
Compare
b7c322b to
aa6192d
Compare
|
Explore the complete analysis inside the Version Insights Performance Analysis Summary: PR #279 - Anthropic Messages API SupportAnalysis Scope: Comparison of version a752fcbb (target) vs aab9b31c (baseline) Performance Impact AssessmentThe performance analysis reveals zero measurable impact on core inference functionality. All binaries show 0.0% power consumption change, with no detectable modifications to performance-critical functions. The code changes implement an API adapter layer in the server component without touching inference paths. Power Consumption Analysis: Function-Level Analysis: Tokens Per Second Impact: Code Implementation: Conclusion: |
d516828 to
0a006e7
Compare
Mirrored from ggml-org/llama.cpp#17425
Summary
This PR adds full Anthropic Messages API compatibility to llama-server, enabling drop-in replacement for applications using the Anthropic API. The implementation converts Anthropic's format to OpenAI-compatible internal format, reusing existing inference pipeline without modifying core llama.cpp functionality.
Motivation
Features Implemented
Endpoints:
Functionality:
Architecture:
Testing
Test Coverage: