-
Notifications
You must be signed in to change notification settings - Fork 70
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Large batchsize causes Windows Error 0xe06d7363 #136
Comments
I tried to execute the same script to reproduce the output but despite the change in batchsize the output remains the same : "OSError: [WinError -529697949] Windows Error 0xe06d7363", for the similar system setup. |
Can you try the following code?
|
This model is worth compiling and running in graph mode, this is why I suggested to use the |
Yes, this snippet can run successfully and output |
Describe the bug
I tried a simple net on MNIST classification, an error
[WinError -529697949] Windows Error 0xe06d7363
occured when I shifted from xpu to npu, finally found it was caused by a large batchsize which worked normally on xpu but caused error on npu.To Reproduce
Steps to reproduce the behavior:
run the following python snippet with the
batchSize
variable setted to different valuesExpected behavior
bahavior when batchsize = 512
exception
console output
output when batchsize = 416
output when batchsize = 288, 320, 352 or 384
System info:
intel_npu_acceleration_library==1.3.0
,torch==2.5.1
Additional context
console outputs error message when run the snippet when batchSize = 416 for the first time and does not throw exception, but the error message disappear when tried the second time with the same batchSize.
I tried rewrite the network and did not use
nn.Sequance
, the error occured when pass the inputx
to its second convolutional layer after the first pool layer, removing the pool layer makes the exception and error message disappear.The text was updated successfully, but these errors were encountered: