Skip to content

Commit 179851a

Browse files
ayan4m1colinurbs
authored andcommitted
docs: make notice about ampere cards apply to all attn libs, since it does
1 parent 181c62b commit 179851a

File tree

1 file changed

+2
-4
lines changed

1 file changed

+2
-4
lines changed

docs/get_started.md

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -91,14 +91,14 @@ There are a variety of Python libraries which provide an improved "attention" im
9191

9292
This is because SageAttention is faster than flash-attention, which is faster than xformers.
9393

94+
**NOTE**: Do not install an attention library unless you have an RTX 3000-series or greater card, as RTX 1000- and 2000-series are not supported by any of these libraries and will cause generation errors.
95+
9496
### SageAttention
9597

9698
Ensure your virtual environment is active, and follow their [install instructions](https://github.com/thu-ml/SageAttention?tab=readme-ov-file#install-package).
9799

98100
### flash-attention
99101

100-
**NOTE**: Do not use flash-attention unless you have an [Ampere GPU](<https://en.wikipedia.org/wiki/Ampere_(microarchitecture)#Products_using_Ampere>), as it does not support cards released before that architecture.
101-
102102
Ensure your virtual environment is active, and then use the following command to install flash-attention:
103103

104104
```sh
@@ -107,8 +107,6 @@ pip install flash-attn --no-build-isolation
107107

108108
### xformers
109109

110-
**NOTE**: Do not use xformers unless you have an [Ampere GPU](<https://en.wikipedia.org/wiki/Ampere_(microarchitecture)#Products_using_Ampere>), as it does not support cards released before that architecture.
111-
112110
Ensure your virtual environment is active, and then use the following command to install xformers:
113111

114112
```sh

0 commit comments

Comments
 (0)