Skip to content
This repository was archived by the owner on Nov 17, 2023. It is now read-only.
This repository was archived by the owner on Nov 17, 2023. It is now read-only.

Can We implement Flash attention 2 in MXnet #21222

@rajveer43

Description

@rajveer43

Description

Flash Attention 2 is a library that provides attention operation kernels for faster and more memory efficient inference and training:

References

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions