File tree Expand file tree Collapse file tree 2 files changed +26
-0
lines changed Expand file tree Collapse file tree 2 files changed +26
-0
lines changed Original file line number Diff line number Diff line change @@ -16,6 +16,7 @@ PyTorch with each method having their advantages in certain use cases:
1616* `Tensor Parallel (TP) <#learn-tp >`__
1717* `Device Mesh <#device-mesh >`__
1818* `Remote Procedure Call (RPC) distributed training <#learn-rpc >`__
19+ * `Monarch Framework <#learn-monarch >`__
1920* `Custom Extensions <#custom-extensions >`__
2021
2122Read more about these options in `Distributed Overview <https://docs.pytorch.org/tutorials/beginner/dist_overview.html?utm_source=distr_landing >`__.
@@ -159,6 +160,22 @@ Learn RPC
159160 +++
160161 :octicon: `code;1em ` Code
161162
163+ .. _learn-monarch :
164+
165+ Learn Monarch
166+ ----------
167+
168+ .. grid :: 3
169+
170+ .. grid-item-card :: :octicon:`file-code;1em`
171+ Interactive Distributed Applications with Monarch
172+ :link: https://docs.pytorch.org/tutorials/intermediate/monarch_distributed_tutorial.html
173+ :link-type: url
174+
175+ Learn how to use Monarch's actor framework
176+ +++
177+ :octicon: `code;1em ` Code
178+
162179.. _custom-extensions :
163180
164181Custom Extensions
Original file line number Diff line number Diff line change @@ -7,6 +7,7 @@ Welcome to PyTorch Tutorials
77* `Supporting Custom C++ Classes in torch.compile/torch.export <https://docs.pytorch.org/tutorials/advanced/custom_class_pt2.html >`__
88* `Accelerating torch.save and torch.load with GPUDirect Storage <https://docs.pytorch.org/tutorials/unstable/gpu_direct_storage.html >`__
99* `Getting Started with Fully Sharded Data Parallel (FSDP2) <https://docs.pytorch.org/tutorials/intermediate/FSDP_tutorial.html >`__
10+ * `Interactive Distributed Applications with Monarch <https://docs.pytorch.org/tutorials/intermediate/monarch_distributed_tutorial.html >`__
1011
1112.. raw :: html
1213
@@ -688,6 +689,14 @@ Welcome to PyTorch Tutorials
688689 :link: intermediate/monarch_distributed_tutorial.html
689690 :tags: Parallel-and-Distributed-Training
690691
692+
693+ .. customcarditem ::
694+ :header: Interactive Distributed Applications with Monarch
695+ :card_description: Learn how to use Monarch's actor framework with TorchTitan to simplify large-scale distributed training across SLURM clusters.
696+ :image: _static/img/thumbnails/cropped/generic-pytorch-logo.png
697+ :link: intermediate/monarch_distributed_tutorial.html
698+ :tags: Parallel-and-Distributed-Training
699+
691700.. Edge
692701
693702 .. customcarditem ::
You can’t perform that action at this time.
0 commit comments