FeepingCreature 7aceb9f91c
Add --use-flash-attention flag. (#7223)
* Add --use-flash-attention flag.
This is useful on AMD systems, as FA builds are still 10% faster than Pytorch cross-attention.
2025-03-14 03:22:41 -04:00
..
2024-12-18 18:23:28 -05:00
2024-01-14 17:28:31 -05:00
2023-01-16 22:37:14 -05:00