(my10) C:\Users\TARGET STORE\Desktop\1\flash_attention>python test.py
Traceback (most recent call last):
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\test.py", line 4, in
from flash_mha import MultiheadFlashAttention
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\flash_mha.py", line 10, in
import common
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\common.py", line 4, in
if torch.cuda.is_available():
AttributeError: module 'torch' has no attribute 'cuda'
(my10) C:\Users\TARGET STORE\Desktop\1\flash_attention>python benchmark.py
Traceback (most recent call last):
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\benchmark.py", line 9, in
from flash_mha import MultiheadFlashAttention
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\flash_mha.py", line 10, in
import common
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\common.py", line 4, in
if torch.cuda.is_available():
AttributeError: module 'torch' has no attribute 'cuda'
(my10) C:\Users\TARGET STORE\Desktop\1\flash_attention>
(my10) C:\Users\TARGET STORE\Desktop\1\flash_attention>python test.py
Traceback (most recent call last):
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\test.py", line 4, in
from flash_mha import MultiheadFlashAttention
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\flash_mha.py", line 10, in
import common
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\common.py", line 4, in
if torch.cuda.is_available():
AttributeError: module 'torch' has no attribute 'cuda'
(my10) C:\Users\TARGET STORE\Desktop\1\flash_attention>python benchmark.py
Traceback (most recent call last):
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\benchmark.py", line 9, in
from flash_mha import MultiheadFlashAttention
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\flash_mha.py", line 10, in
import common
File "C:\Users\TARGET STORE\Desktop\1\flash_attention\common.py", line 4, in
if torch.cuda.is_available():
AttributeError: module 'torch' has no attribute 'cuda'
(my10) C:\Users\TARGET STORE\Desktop\1\flash_attention>