-
Notifications
You must be signed in to change notification settings - Fork 13
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CUDA error: device-side assert triggered #5
Comments
Hello, this is caused by a bug in the code, when it loops through the data, the last batch does not always has the same number as the batch size. Lines 244 to 256 in 5e44ab5
|
Hello @ijorquera, I got the same error. |
Hello, thanks for the project!
I was trying to run the FAIR benchmark using the validation and test set, but in both cases I get the following error:
Traceback (most recent call last): File "/TRUST/test.py", line 136, in <module> trust.test(return_params=True) File "/TRUST/lib/model.py", line 249, in test 'scene': scene_images[visind], RuntimeError: CUDA error: device-side assert triggered
There are also a few lines like this one:
/pytorch/aten/src/ATen/native/cuda/IndexKernel.cu:93: operator(): block: [19,0,0], thread: [93,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed.
I also checked the test_outputs folder and there was 1 folder with albedo texture jpg and a npy file, that for each folder in the fair benchmark val/test. There was also another folder (test_images_vis) with the images of the albedo jpg, the albedo applied in the 3d model, the light shader, the original input and the 3d model with the tex and the light shader.
The output folder makes me think that everything went alright, but I keep getting the error above every time I run the benchmark.
Any help regarding the error will be highly appreciated! Thanks again.
Edit:
The project was installed in a machine with CUDA 11.3 and pytorch 1.10.2 since the rtx 30xx series don't have support for CUDA 10.1
The text was updated successfully, but these errors were encountered: