Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Improve cc version detection for cublaslt #110

Merged
merged 2 commits into from
Jan 2, 2023

Conversation

BlackHC
Copy link
Contributor

@BlackHC BlackHC commented Dec 29, 2022

The current code fails to detect cublaslt for new cc versions (e.g. 8.9 fails the check).

This fixes issue #97.

@TimDettmers
Copy link
Collaborator

Thank you for this contribution!

@TimDettmers TimDettmers merged commit d9112dc into bitsandbytes-foundation:main Jan 2, 2023
@gururise
Copy link

gururise commented Jan 5, 2023

I don't think this PR quite fixes issue #97, which referenced lack of Int8 Matmul error being thrown on the Nvidia Tesla P40.

Still getting the following error:

=============================================
ERROR: Your GPU does not support Int8 Matmul!
=============================================

python: /home/gene/dockerx/temp/bitsandbytes/csrc/ops.cu:408: int igemmlt(cublasLtHandle_t, int, int, int, const int8_t*, const int8_t*, void*, float*, int, int, int) [with int FORMATB = 3; int DTYPE_OUT = 32; int SCALE_ROWS = 0; cublasLtHandle_t = cublasLtContext*; int8_t = signed char]: Assertion `false' failed.

I created a new issue #118 to make it clear that problem still exists.

@BlackHC
Copy link
Contributor Author

BlackHC commented Jan 10, 2023

@gururise

The following has helped me better understand what was going on, but your error sounds very different to the one I fixed with my PR.

logging.basicConfig(level=logging.DEBUG)
from bitsandbytes.cextension import CUDASetup
CUDASetup.get_instance().print_log_stack()

Best wishes,
Andreas

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants