You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have tested it successfully using the INT4 quantized model. The quantized model may have accuracy degradation. But, the output tokens/second is reasonable on ARL-S 285K CPU with 32 GB RAM and 500 GB drive (200 GB for vRAM swap file and 200 GB to store the model). Also, tested successfully on ARL-H 285H iGPU with 64 GB RAM. Thanks.
The text was updated successfully, but these errors were encountered:
DeepSeek R1 Qwen 32B is the one comparable to OpenAI o1 mini. Is it possible to add it in the deepseek-r1 notebook selection?
https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
I have tested it successfully using the INT4 quantized model. The quantized model may have accuracy degradation. But, the output tokens/second is reasonable on ARL-S 285K CPU with 32 GB RAM and 500 GB drive (200 GB for vRAM swap file and 200 GB to store the model). Also, tested successfully on ARL-H 285H iGPU with 64 GB RAM. Thanks.
The text was updated successfully, but these errors were encountered: