Skip to content

Restore dual-repo FA3 logic lost during V4 rewrite #7

@2imi9

Description

@2imi9

Problem

Original train.py selected FA3 repo by GPU capability:

  • varunneal/flash-attention-3 for Hopper (SM 9.0)
  • kernels-community/flash-attn3 for others

V4 rewrite hardcoded only varunneal. Fixed in 4f3fbf7.
Note: neither repo supports Blackwell yet — falls through to FlexAttention.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions