-
Notifications
You must be signed in to change notification settings - Fork 70
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issue on run_matmul #135
Comments
Few questions:
I suspect the issue here is in the |
def run_matmul(inC, outC, batch):
Because the shape for torch.matmul() is different with run_matmul(), so I transpose it. |
Is this actual an issue or I misuse it? |
Describe the bug
Currently I'm do some inference on my laptop using this python package. I found an issue described below.
I use the following code to test the result for classical matrix multiple operation:
import torch
from intel_npu_acceleration_library.backend.runtime import run_matmul
a = torch.randn(1000, 1000, dtype=torch.float16)
b = torch.randn(1000, 1000, dtype=torch.float16)
c1 = run_matmul(a, b.transpose(0, 1))
c2 = torch.matmul(a, b)
assert c1.equal(c2)
The difference between c1 and c2 is huge!
I don't know why?
Test environment:
Ubuntu 24.04, Intel Arc Graphics with NPU enabled.
I checkout this commit from github and install it locally: 3e3dee3
Also I have installed required NPU drivers and it can leverage NPU well because the inference is much faster than pure CPU.
The text was updated successfully, but these errors were encountered: