We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent a77294e commit 148e877Copy full SHA for 148e877
vllm/attention/ops/triton_decode_attention.py
@@ -425,7 +425,7 @@ def _decode_grouped_att_m_fwd(
425
extra_kargs = {}
426
num_stages = 2
427
if is_hip_:
428
- # https://rocm.docs.amd.com/en/docs-6.2.0/how-to/llm-fine-tuning-optimization/optimizing-triton-kernel.html
+ # https://rocm.docs.amd.com/en/latest/how-to/rocm-for-ai/inference-optimization/workload.html#triton-kernel-performance-optimization
429
# https://github.com/triton-lang/triton/blob/main/third_party/amd/backend/compiler.py
430
extra_kargs = {
431
"waves_per_eu": 1,
0 commit comments