Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] 2080ti can't quant a 12B model #725

Open
3 tasks done
frenzybiscuit opened this issue Jan 28, 2025 · 1 comment
Open
3 tasks done

[BUG] 2080ti can't quant a 12B model #725

frenzybiscuit opened this issue Jan 28, 2025 · 1 comment
Labels
bug Something isn't working

Comments

@frenzybiscuit
Copy link

frenzybiscuit commented Jan 28, 2025

OS

Linux

GPU Library

CUDA 12.x

Python version

3.12

Pytorch version

2.4.0

Model

LatitudeGames/Wayfarer-12B

Describe the bug

When using a 2080ti (11GB VRAM) it gets all the way to the save stage of the measurements and then crashes from running out of memory.

How much VRAM is required to do 12B models?

I'm trying to find a use for my 2080ti, other than sitting around collecting dust.

Image

Reproduction steps

.

Expected behavior

.

Logs

.

Additional context

.

Acknowledgements

  • I have looked for similar issues before submitting this one.
  • I understand that the developers have lives and my issue will be answered when possible.
  • I understand the developers of this program are human, and I will ask my questions politely.
@frenzybiscuit frenzybiscuit added the bug Something isn't working label Jan 28, 2025
@turboderp
Copy link
Member

The head layer is typically the largest layer of the model and it's not uncommon for that to be the point where you run out of VRAM, sadly. During conversion, the quantizer needs to store three copies of the full tensor in FP32 precision, which on its own is a total of 7.5 GB for that model, in addition to everything else the framework is doing. So I'm afraid that the way it's currently implemented, 11 GB just isn't enough VRAM to convert this model.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants