-
-
Notifications
You must be signed in to change notification settings - Fork 11.9k
[Test] Remove old non-varlen FA2 test #28420
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Test] Remove old non-varlen FA2 test #28420
Conversation
Signed-off-by: Matthew Bonanni <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request removes test_flash_attn_with_paged_kv, a test for the obsolete function flash_attn_with_kvcache. The change is a straightforward code cleanup, removing unused test code. The deletion is self-contained and I found no issues with it.
LucasWilkinson
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM; can we remove this on the vllm-flash-attn side too?
@LucasWilkinson are you sure? I realized that this still exists on upstream main, removing it would cause a divergence: https://github.com/Dao-AILab/flash-attention Edit: discussed offline. Created PR: vllm-project/flash-attention#107 |
yewentao256
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, thanks for the work!
Signed-off-by: Matthew Bonanni <[email protected]> Signed-off-by: xuebwang-amd <[email protected]>
Signed-off-by: Matthew Bonanni <[email protected]>
Purpose
This is testing a function that is no longer used anywhere in vLLM, so this PR removes it. This helps eliminate dependence on FA2.
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.