We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Windows
CUDA 12.x
3.12
tabbyAPI/backends/exllamav2/utils.py
Line 23 in bd16681
upstream https://github.com/Dao-AILab/flash-attention contains amd support by now
AMD cdna gpus should be considered supported just as well as ampere+
No response
The text was updated successfully, but these errors were encountered:
No branches or pull requests
OS
Windows
GPU Library
CUDA 12.x
Python version
3.12
Describe the bug
tabbyAPI/backends/exllamav2/utils.py
Line 23 in bd16681
Reproduction steps
upstream https://github.com/Dao-AILab/flash-attention contains amd support by now
Expected behavior
AMD cdna gpus should be considered supported just as well as ampere+
Logs
No response
Additional context
No response
Acknowledgements
The text was updated successfully, but these errors were encountered: