Skip to content

integrate new float8 quantization primitives into AQT #1598

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed

Conversation

danielvegamyhre
Copy link
Contributor

@danielvegamyhre danielvegamyhre commented Jan 22, 2025

Context

Currently, AQT has the method from_hp_to_floatx for float8 quantization, and from_hp_to_fpx for low precision floating point data types like fp6 (technically can support fp1-fp7).

from_hp_to_floatx re-uses from_hp_to_intx, which in turn uses these generic quantization primitives.

Overall, in the current state the float8 path is a bit confusing for developers, due to both the naming ("floatx") and the use of generic functions which include a bunch of params which are unrelated to float8 quantization.

Summary of changes

The goal of this PR stack is to refactor this to have a clean separation of concerns, and simpler internal API surfaces for code using in float8 quantization for inference.

Specifically:

  • Separate quantization primitives for float8
  • Integrate those new quant primitives into AQT <------------------- (this PR)

Note: I will add float8 static quantization in a separate set of PRs.

[ghstack-poisoned]
@danielvegamyhre
Copy link
Contributor Author

danielvegamyhre commented Jan 22, 2025

Copy link

pytorch-bot bot commented Jan 22, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/1598

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 2cac42e with merge base 860da26 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Jan 22, 2025
danielvegamyhre added a commit that referenced this pull request Jan 22, 2025
ghstack-source-id: 9aacd39
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
@danielvegamyhre danielvegamyhre added quantize topic: improvement Use this tag if this PR is an improvement (doesn't fit into any of the other categories) labels Jan 22, 2025
@danielvegamyhre danielvegamyhre requested review from jainapurva and jerryzh168 and removed request for jainapurva January 22, 2025 19:36
target_dtype,
)
fp8_data = _layout.post_process(fp8_data)
tensor_impl_ctr = get_tensor_impl_constructor(type(_layout))
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

does this have multiple options for float8? if not, just call it directly to reduce # of abstractions the code reader needs to know about

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

good point, done!

target_dtype: torch.dtype,
block_size: Tuple[int, ...],
_layout: Layout = PlainLayout(),
):
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

a docblock here should explain the difference between from_hp_to_floatx, from_hp_to_fpx, from_hp_to_float8

danielvegamyhre added a commit that referenced this pull request Jan 23, 2025
ghstack-source-id: 9aacd39
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
[ghstack-poisoned]
danielvegamyhre added a commit that referenced this pull request Jan 23, 2025
ghstack-source-id: c87842f
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
@@ -422,6 +417,39 @@ def from_hp_to_fpx(
tensor_impl = tensor_impl_ctr(floatx_packed, scale, None, _layout)
return cls(tensor_impl, block_size, original_shape, dtype=input_float.dtype)

@classmethod
def from_hp_to_float8(
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Update from_hp_to_floatx with the new float8 logic. For fp1-fp7, we're using from_hp_to_floatx.

[ghstack-poisoned]
danielvegamyhre added a commit that referenced this pull request Jan 23, 2025
ghstack-source-id: c1deeeb
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
[ghstack-poisoned]
danielvegamyhre added a commit that referenced this pull request Jan 23, 2025
ghstack-source-id: c1deeeb
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
danielvegamyhre added a commit that referenced this pull request Jan 23, 2025
ghstack-source-id: c1deeeb
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
[ghstack-poisoned]
danielvegamyhre added a commit that referenced this pull request Jan 24, 2025
ghstack-source-id: 982ea07
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
[ghstack-poisoned]
danielvegamyhre added a commit that referenced this pull request Jan 24, 2025
ghstack-source-id: e0a6b79
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
[ghstack-poisoned]
danielvegamyhre added a commit that referenced this pull request Jan 24, 2025
ghstack-source-id: fccae98
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
[ghstack-poisoned]
danielvegamyhre added a commit that referenced this pull request Jan 24, 2025
ghstack-source-id: 2d1d4f1
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
[ghstack-poisoned]
danielvegamyhre added a commit that referenced this pull request Jan 24, 2025
ghstack-source-id: 0a9fcdb
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
[ghstack-poisoned]
danielvegamyhre added a commit that referenced this pull request Jan 24, 2025
ghstack-source-id: a098c0e
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
[ghstack-poisoned]
danielvegamyhre added a commit that referenced this pull request Jan 24, 2025
ghstack-source-id: b9139dc
ghstack-comment-id: 2608090492
Pull Request resolved: #1598
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. quantize topic: improvement Use this tag if this PR is an improvement (doesn't fit into any of the other categories)
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants