-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Quantization of Other Models #25
Comments
Hi, There should not be any problems about quantizing other models. We have LLM and Diffusion model quantizer just for adding specific quantization configurations separately. |
is there any reference where i can convert the DINO model to onnx or other compression format? @synxlin @sxtyzhangzk @sxtyzhangzk it would be very helpful |
hi @Oguzhanercan can you pls help me i need to quantize the DINO model i have tried native approached like using direct torch module to convert to onnx but it failed and later i tried https://github.com/IDEA-Research/detrex this repo to convert to onnx after this my model latency got increased to 2x...may i pls know what approach have you dont to optimize the DINO model...it would be very helpful ....iam stuck in this past 1week.. |
Hi, thanks for your work. I wonder that, why you named this as LLM and Diffusion model quantizer. I cannot see any problem about quantizing other models architectures with this like DINO, ViT etc. for other tasks (object detection, semantic segmentation). What is your insights about it?
The text was updated successfully, but these errors were encountered: