-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fatal Python error when executing on second cuda device. #103
Comments
Because this was done without import numpy as np
import torch
import cufinufft
data = torch.view_as_complex(
torch.stack((torch.randn(15, 80, 12000), torch.randn(15, 80, 12000)), dim=-1)
)
omega = torch.rand(2, 12000) * 2 * np.pi - np.pi
cufinufft.nufft2d1(
*omega.to("cuda:1"),
data.reshape(-1, 12000).to("cuda:1"),
(320,320),
isign=-1,
) This leads to a slightly different error:
|
I have opened flatironinstitute/finufft#420, where I believe further discussion should take place. Thank you for reporting! |
@gRox167 I believe we have a fix for this in the latest |
I have tried in my code, and it works perfectly fine! Thank you for helping me to publish a issue to the upstream package and provide responsive fix! |
First let me express my gratitude for developers who bring this brilliant package!
The issue that I encountered is really weird. I am working on a server equipped with 2 H100 GPU. pytorch-finufft can execute smoothly on
cuda:0
device and cpu device, but when I usecuda:1
device it came error:The minimal environment is
I am not sure if this is only happen on my server.
I noticed that when I use
cuda:0
, there is only one process created on gpu0, however, if I usecuda:1
, there will be 2 processes created respectively on both of gpu.The text was updated successfully, but these errors were encountered: