Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[spec decode] [4/N] Move update_flash_attn_metadata to attn backend #7571

Merged
merged 7 commits into from
Aug 16, 2024

Commits on Aug 16, 2024

  1. Configuration menu
    Copy the full SHA
    50b4b5c View commit details
    Browse the repository at this point in the history
  2. format

    SolitaryThinker committed Aug 16, 2024
    Configuration menu
    Copy the full SHA
    1fc74ff View commit details
    Browse the repository at this point in the history
  3. Update vllm/attention/backends/flash_attn.py

    Co-authored-by: Cody Yu <[email protected]>
    SolitaryThinker and comaniac committed Aug 16, 2024
    Configuration menu
    Copy the full SHA
    bdedb2d View commit details
    Browse the repository at this point in the history
  4. comments

    SolitaryThinker committed Aug 16, 2024
    Configuration menu
    Copy the full SHA
    bf9422e View commit details
    Browse the repository at this point in the history
  5. format

    SolitaryThinker committed Aug 16, 2024
    Configuration menu
    Copy the full SHA
    a053513 View commit details
    Browse the repository at this point in the history
  6. ops

    SolitaryThinker committed Aug 16, 2024
    Configuration menu
    Copy the full SHA
    d08be46 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    083eb89 View commit details
    Browse the repository at this point in the history