You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/get_started.md
+2-4Lines changed: 2 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -91,14 +91,14 @@ There are a variety of Python libraries which provide an improved "attention" im
91
91
92
92
This is because SageAttention is faster than flash-attention, which is faster than xformers.
93
93
94
+
**NOTE**: Do not install an attention library unless you have an RTX 3000-series or greater card, as RTX 1000- and 2000-series are not supported by any of these libraries and will cause generation errors.
95
+
94
96
### SageAttention
95
97
96
98
Ensure your virtual environment is active, and follow their [install instructions](https://github.com/thu-ml/SageAttention?tab=readme-ov-file#install-package).
97
99
98
100
### flash-attention
99
101
100
-
**NOTE**: Do not use flash-attention unless you have an [Ampere GPU](<https://en.wikipedia.org/wiki/Ampere_(microarchitecture)#Products_using_Ampere>), as it does not support cards released before that architecture.
101
-
102
102
Ensure your virtual environment is active, and then use the following command to install flash-attention:
**NOTE**: Do not use xformers unless you have an [Ampere GPU](<https://en.wikipedia.org/wiki/Ampere_(microarchitecture)#Products_using_Ampere>), as it does not support cards released before that architecture.
111
-
112
110
Ensure your virtual environment is active, and then use the following command to install xformers:
0 commit comments