File tree Expand file tree Collapse file tree 5 files changed +813
-1
lines changed Expand file tree Collapse file tree 5 files changed +813
-1
lines changed Original file line number Diff line number Diff line change 3333 },
3434 "recipes_source/torch_export_aoti_python.py" : {
3535 "needs" : " linux.g5.4xlarge.nvidia.gpu"
36- },
36+ },
3737 "advanced_source/pendulum.py" : {
3838 "needs" : " linux.g5.4xlarge.nvidia.gpu" ,
3939 "_comment" : " need to be here for the compiling_optimizer_lr_scheduler.py to run."
5858 "intermediate_source/scaled_dot_product_attention_tutorial.py" : {
5959 "needs" : " linux.g5.4xlarge.nvidia.gpu"
6060 },
61+ "intermediate_source/transformer_building_blocks.py" : {
62+ "needs" : " linux.g5.4xlarge.nvidia.gpu"
63+ },
6164 "recipes_source/torch_compile_user_defined_triton_kernel_tutorial.py" : {
6265 "needs" : " linux.g5.4xlarge.nvidia.gpu"
6366 },
Original file line number Diff line number Diff line change 2525 "intermediate_source/mnist_train_nas" , # used by ax_multiobjective_nas_tutorial.py
2626 "intermediate_source/fx_conv_bn_fuser" ,
2727 "intermediate_source/_torch_export_nightly_tutorial" , # does not work on release
28+ "intermediate_source/transformer_building_blocks" , # does not work on release
2829 "advanced_source/super_resolution_with_onnxruntime" ,
2930 "advanced_source/usb_semisup_learn" , # fails with CUDA OOM error, should try on a different worker
3031 "prototype_source/fx_graph_mode_ptq_dynamic" ,
Original file line number Diff line number Diff line change 11ACL
22ADI
3+ ALiBi
34AOT
45AOTInductor
56APIs
7980FX's
8081FairSeq
8182Fastpath
83+ FFN
8284FloydHub
8385FloydHub's
8486Frobenius
@@ -127,6 +129,7 @@ Kihyuk
127129Kiuk
128130Kubernetes
129131Kuei
132+ KV
130133LRSchedulers
131134LSTM
132135LSTMs
162165NTK
163166NUMA
164167NaN
168+ NaNs
165169NanoGPT
166170Netron
167171NeurIPS
@@ -231,6 +235,7 @@ Sigmoid
231235SoTA
232236Sohn
233237Spacy
238+ SwiGLU
234239TCP
235240THP
236241TIAToolbox
@@ -276,6 +281,7 @@ Xcode
276281Xeon
277282Yidong
278283YouTube
284+ Zipf
279285accelerometer
280286accuracies
281287activations
305311benchmarked
306312benchmarking
307313bitwise
314+ bool
308315boolean
309316breakpoint
310317broadcasted
333340cuDNN
334341cuda
335342customizable
343+ customizations
336344datafile
337345dataflow
338346dataframe
@@ -377,6 +385,7 @@ fbgemm
377385feedforward
378386finetune
379387finetuning
388+ FlexAttention
380389fp
381390frontend
382391functionalized
431440macos
432441manualSeed
433442matmul
443+ matmuls
434444matplotlib
435445memcpy
436446memset
@@ -446,6 +456,7 @@ modularized
446456mpp
447457mucosa
448458multihead
459+ MultiheadAttention
449460multimodal
450461multimodality
451462multinode
@@ -456,7 +467,11 @@ multithreading
456467namespace
457468natively
458469ndarrays
470+ nheads
459471nightlies
472+ NJT
473+ NJTs
474+ NJT's
460475num
461476numericalize
462477numpy
@@ -532,6 +547,7 @@ runtime
532547runtime
533548runtimes
534549scalable
550+ SDPA
535551sharded
536552softmax
537553sparsified
@@ -591,12 +607,14 @@ tradeoff
591607tradeoffs
592608triton
593609uint
610+ UX
594611umap
595612uncomment
596613uncommented
597614underflowing
598615unfused
599616unimodal
617+ unigram
600618unnormalized
601619unoptimized
602620unparametrized
@@ -618,6 +636,7 @@ warmstarted
618636warmstarting
619637warmup
620638webp
639+ wikitext
621640wsi
622641wsis
623642Meta's
Original file line number Diff line number Diff line change @@ -664,6 +664,14 @@ Welcome to PyTorch Tutorials
664664 :link: beginner/knowledge_distillation_tutorial.html
665665 :tags: Model-Optimization,Image/Video
666666
667+
668+ .. customcarditem ::
669+ :header: Accelerating PyTorch Transformers by replacing nn.Transformer with Nested Tensors and torch.compile()
670+ :card_description: This tutorial goes over recommended best practices for implementing Transformers with native PyTorch.
671+ :image: _static/img/thumbnails/cropped/pytorch-logo.png
672+ :link: intermediate/transformer_building_blocks.html
673+ :tags: Transformer
674+
667675.. Parallel-and-Distributed-Training
668676
669677
You can’t perform that action at this time.
0 commit comments