You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Summary:
Adds the workaround from
pytorch/pytorch#141881 to the torchao float8
rowwise recipe, to reduce memory usage when FSDP is on.
Test Plan: tested in torchtitan, LLaMa 3 8B 8H100 training with rowwise
peak memory decreased from 67GiB to 59GiB
Reviewers:
Subscribers:
Tasks:
Tags:
0 commit comments