Add inference code#199
Conversation
Have you meet error like it? @wade3han |
|
No, I didn't encounter that error. Can you give me more context? |
just use : |
|
same problem. |
|
RuntimeError: CUDA error: CUBLAS_STATUS_INVALID_VALUE when calling same for inferencing both llama-7b-hf and fine-tune model |
|
Cool! The problem has been fixed. |
|
Thanks for the code! However, I had some problems when I run the code in my server with three 3090 GPUs with VRAM of 24GB*3. I know |
Tested with own fine-tuned 7B alpaca model
Largely influenced by https://github.com/kriskrisliu/stanford_alpaca/tree/krisliu