-
Notifications
You must be signed in to change notification settings - Fork 4.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add inference code #199
base: main
Are you sure you want to change the base?
Add inference code #199
Conversation
Have you meet error like it? @wade3han |
No, I didn't encounter that error. Can you give me more context? |
just use :
|
same problem. |
RuntimeError: CUDA error: CUBLAS_STATUS_INVALID_VALUE when calling same for inferencing both llama-7b-hf and fine-tune model |
Cool! The problem has been fixed. |
Thanks for the code! However, I had some problems when I run the code in my server with three 3090 GPUs with VRAM of 24GB*3. I know |
Tested with own fine-tuned 7B alpaca model
Largely influenced by https://github.com/kriskrisliu/stanford_alpaca/tree/krisliu