You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi! I was trying to run the inference.ipynb notebook, and I got an RuntimeError.
While I ran model = LlamaForCausalLM.from_pretrained(train_config.model_name, device_map="auto", config=config).to(device) , I got RuntimeError: You can't move a model that has some modules offloaded to cpu or disk.
I guess it is related to the device = "cuda" if torch.cuda.is_available() else "cpu" , but my GPU RAM is available and not full.
How should I do to solve it? Thank you.
Best regards,
Maggie
The text was updated successfully, but these errors were encountered:
Hi! I was trying to run the inference.ipynb notebook, and I got an RuntimeError.
While I ran
model = LlamaForCausalLM.from_pretrained(train_config.model_name, device_map="auto", config=config).to(device)
, I got RuntimeError: You can't move a model that has some modules offloaded to cpu or disk.I guess it is related to the
device = "cuda" if torch.cuda.is_available() else "cpu"
, but my GPU RAM is available and not full.How should I do to solve it? Thank you.
Best regards,
Maggie
The text was updated successfully, but these errors were encountered: