We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
When trying to use kernl with default Llama 7B on A100 device, I get this error.
import torch from transformers import LlamaModel, LlamaConfig, LlamaTokenizer, LlamaForCausalLM from kernl.model_optimization import optimize_model config = LlamaConfig() model = LlamaForCausalLM(config).cuda() optimize_model(model) length = 5 input_ids = torch.randint(low=0, high=model.config.vocab_size, size=(1,length)).cuda() with torch.inference_mode(), torch.cuda.amp.autocast(): outputs = model.generate(input_ids=input_ids) print(outputs.shape)
A properly working Llama model.
The following message occurs:
The text was updated successfully, but these errors were encountered:
It is the same for me, please help up to use LLaMA models :)
Sorry, something went wrong.
No branches or pull requests
Description
When trying to use kernl with default Llama 7B on A100 device, I get this error.
Steps to reproduce
Expected Behavior
A properly working Llama model.
Actual Behavior
The following message occurs:
Your environment
Self-service
Code of Conduct
The text was updated successfully, but these errors were encountered: