Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Plans for block-wise FP8 quantization during training? #1411

Open
beccohov opened this issue Jan 15, 2025 · 3 comments
Open

Plans for block-wise FP8 quantization during training? #1411

beccohov opened this issue Jan 15, 2025 · 3 comments

Comments

@beccohov
Copy link

Hi TE team,

I'm interested in whether there are plans to implement block-wise quantization for FP8 training, similar to what's described in papers like "Deepseek V3".

Block quantization could potentially provide better numerical stability and accuracy compared to tensor-wide quantization, especially for outlier values. This could be particularly valuable for large language models where maintaining precision is crucial.

Some specific questions:

  1. Is this feature currently on your roadmap?
  2. If yes, what's the approximate timeline?
  3. If no, are there technical challenges preventing this implementation?

Thank you for your time!

@zigzagcai
Copy link

zigzagcai commented Jan 16, 2025

I have the same interest with block-wise FP8.

@liangzelang
Copy link

ME TOO

@Monekyzoon
Copy link

Monekyzoon commented Jan 17, 2025

In a addition, activation use tile-wise(1 x 128) quantization in DeepSeek-V3.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants