Skip to content

Commit a1ebaa6

Browse files
committed
typo: add vLLM Prefix Cache & LoRA Adapters links
1 parent f4f803b commit a1ebaa6

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

site-src/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ The following specific terms to this project:
2323
from [Model Serving](https://github.com/kubernetes-sigs/gateway-api-inference-extension/tree/main/docs/proposals/003-model-server-protocol/README.md).
2424
- **Metrics and Capabilities**: Data provided by model serving platforms about
2525
performance, availability and capabilities to optimize routing. Includes
26-
things like [Prefix Cache] status or [LoRA Adapters] availability.
26+
things like [Prefix Cache](https://docs.vllm.ai/en/stable/design/v1/prefix_caching.html) status or [LoRA Adapters](https://docs.vllm.ai/en/stable/features/lora.html) availability.
2727
- **Endpoint Picker(EPP)**: An implementation of an `Inference Scheduler` with additional Routing, Flow, and Request Control layers to allow for sophisticated routing strategies. Additional info on the architecture of the EPP [here](https://github.com/kubernetes-sigs/gateway-api-inference-extension/tree/main/docs/proposals/0683-epp-architecture-proposal).
2828

2929
[Inference Gateway]:#concepts-and-definitions

0 commit comments

Comments
 (0)