Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add compiled autograd tutorial #3026

Merged
merged 14 commits into from
Oct 9, 2024
Merged

Conversation

xmfan
Copy link
Member

@xmfan xmfan commented Sep 3, 2024

Fixes #3034

Description

Add tutorial for a PyTorch 2.4 feature

Checklist

  • The issue that is being fixed is referred in the description (see above "Fixes #ISSUE_NUMBER")
  • Only one issue is addressed in this pull request
  • Labels from the issue that this PR is fixing are added to this pull request
  • No unnecessary issues are included into this pull request.

Copy link

pytorch-bot bot commented Sep 3, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/tutorials/3026

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 14c7499 with merge base 19fffda (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@xmfan xmfan changed the title [WIP] Add compiled autograd tutorial Add compiled autograd tutorial Sep 4, 2024
@xmfan xmfan marked this pull request as ready for review September 4, 2024 20:01
@xmfan xmfan requested a review from svekars September 4, 2024 20:02
Copy link
Contributor

@svekars svekars left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you for following the template - I really appreciate it! I added some editorial / formatting suggestions. Can you please take a look and let me know if you have any questions.


######################################################################
# Compiled Autograd addresses certain limitations of AOTAutograd
# ------------
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

should this be a title or just a paragraph?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

paragraph to clarify that it's a new section after "Compiling the forward and backward pass using different flags"

@xmfan xmfan force-pushed the xmfan/ca_tutorial branch from 4b696bb to a9d3bb3 Compare September 6, 2024 18:19
@xmfan xmfan force-pushed the xmfan/ca_tutorial branch from a9d3bb3 to 271b8f2 Compare September 6, 2024 18:20
@xmfan xmfan requested a review from svekars September 6, 2024 18:21
@svekars
Copy link
Contributor

svekars commented Sep 6, 2024

There is still some issue on line 54 that prevents the tutorial from building. Can you check?

@xmfan xmfan mentioned this pull request Sep 7, 2024
4 tasks
@svekars svekars added the 2.5 PR related to version 2.5 label Sep 9, 2024
* Added runtime overhead at the start of the backward for cache lookup
* More prone to recompiles and graph breaks in dynamo due to the larger capture

.. note:: Compiled Autograd is under active development and is not yet compatible with all existing PyTorch features. For the latest status on a particular feature, refer to `Compiled Autograd Landing Page <https://docs.google.com/document/d/11VucFBEewzqgkABIjebZIzMvrXr3BtcY1aGKpX61pJY>`_.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm not sure why it's called Landing Page. Should it be called a Compiled Autograd Roadmap? Also, it feels like it would be better to point to issues on github with a specific label, perhaps, or a Github Project even?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm expecting the content of this doc to change rapidly at the moment, so I'll keep it as a google doc for now. It's also something we did for dynamic shapes and custom operators.

The end goal once the development is more stable, is to move it into the docs

@xmfan xmfan requested a review from svekars September 13, 2024 21:49
@svekars svekars added 2.5 PR related to version 2.5 and removed 2.5 PR related to version 2.5 labels Sep 16, 2024
@svekars svekars added the 2.5 PR related to version 2.5 label Sep 16, 2024
@xmfan
Copy link
Member Author

xmfan commented Sep 23, 2024

Hi @svekars, how can I merge the PR?

@svekars
Copy link
Contributor

svekars commented Sep 23, 2024

@xmfan can you add a customcarditem and an entry in a toctree in the index.rst

@yitingw1
Copy link

Hi, I am concerned about CompiledAutograd's progress. After a series of DDP-related optimization:

  • [DDP] Use compiled_autograd to trace DDP backward allreduce #110662
  • [DDP][PT2D] Allreduce fusion fx pass using concat and all_reduce_coalesced #113209

Does the model have compute-computation overlap when enabled CompiledAutograd? Is there any latest progress related to DDP for CompiledAutograd?
Now can the multi-card model with CompiledAutograd perform better compared with torch.compile using DDPOptimizer? Thanks!

@fegin
Copy link

fegin commented Sep 25, 2024

@yitingw1 When enable CompiledAutograd, we should also enable the new CompiledDDP. Right now it is not automatically enabled. As for the overlapping, the answer is yes if the new CompiledDDP is enabled. The overlapping strategy is the same as the DDPOptimizer and the eager DDP. The difference from the DDPOptimizer is that the new CompiledDDP, with the help from CompiledAutograd, should produce less if not zero graph breaks. However, whether it will perform better or not depending on the models.

def train(model, x):
model = torch.compile(model)
loss = model(x).sum()
with torch._dynamo.compiled_autograd.enable(torch.compile(fullgraph=True)):
Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Do we have to ask user to wrap both the forward and backward inside torch._dynamo.compiled_autograd.enable as both DDP and FSDP require users to do so.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

no, this is the 2nd "more flexible" API

@xmfan
Copy link
Member Author

xmfan commented Sep 25, 2024

@svekars can we do that in a separate PR? we don't even have a compile section right now. and I'd like to add in the torch.compile tutorial as well as the TORCH_LOGS one

@xmfan
Copy link
Member Author

xmfan commented Oct 9, 2024

New description looks good to me!

@svekars svekars merged commit d862a95 into pytorch:main Oct 9, 2024
2 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
2.5 PR related to version 2.5 cla signed new tutorial
Projects
None yet
Development

Successfully merging this pull request may close these issues.

💡 [REQUEST] - Compiled Autograd Tutorial
5 participants