Skip to content

Why not all tensors are quantized in the GGUF format? #2937

Closed Answered by KerfuffleV2
loretoparisi asked this question in Q&A
Discussion options

You must be logged in to vote

One dimensional tensors are tiny so there's no point in quantizing them. Just for example an f32 4096 element 1D tensor is about 16KB while a 4096x4096 Q3_K tensor would be around 5MB if it was exactly 3 bits per element (I think the real number is a bit higher).

Anyway, even in a 40GB 70B model all the f32 tensors probably add up to less than 20MB. It's the multi-dimensional ones that are actually big enough to care about quantizing.

Replies: 3 comments 3 replies

Comment options

You must be logged in to vote
0 replies
Answer selected by loretoparisi
Comment options

You must be logged in to vote
3 replies
@KerfuffleV2
Comment options

@loretoparisi
Comment options

@KerfuffleV2
Comment options

Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants