-
Notifications
You must be signed in to change notification settings - Fork 527
Pull requests: pytorch/torchtitan
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Add einops to requirements.txt
CLA Signed
This label is managed by the Meta Open Source bot.
high priority
#1734
opened Sep 22, 2025 by
ezyang
Loading…
fix: pp grad accumulation is broken
CLA Signed
This label is managed by the Meta Open Source bot.
#1732
opened Sep 22, 2025 by
jdinalt
Loading…
Add support for AC budget API
CLA Signed
This label is managed by the Meta Open Source bot.
#1731
opened Sep 21, 2025 by
tohskai
Loading…
fix: datasets broken import due to HF package and folder name collision
CLA Signed
This label is managed by the Meta Open Source bot.
#1730
opened Sep 21, 2025 by
idoh
Loading…
handle unable to load ft checkpoint
CLA Signed
This label is managed by the Meta Open Source bot.
#1729
opened Sep 19, 2025 by
tushar00jain
Loading…
Fix flux image_generation_test. Change input dim to match encoder
CLA Signed
This label is managed by the Meta Open Source bot.
#1726
opened Sep 19, 2025 by
yosoyjay
Loading…
[DONT REVIEW] debug ac(fsdp) in llama and deepseek
CLA Signed
This label is managed by the Meta Open Source bot.
[torchtitan][replicate] experimenting new replicate integration with torchtitan
CLA Signed
This label is managed by the Meta Open Source bot.
[Do not merge] Reproduce AC(FSDP(moe.experts)) composibility issue
CLA Signed
This label is managed by the Meta Open Source bot.
[CP][RFC] Enable FlexCP for llama3 with parallelize_module
CLA Signed
This label is managed by the Meta Open Source bot.
#1707
opened Sep 12, 2025 by
fegin
Loading…
[WIP] async_tp shape mismatch in rs+mm repro
CLA Signed
This label is managed by the Meta Open Source bot.
#1705
opened Sep 12, 2025 by
IvanKobzarev
Loading…
[mxfp8 moe training] add MX MoE model converter using torchao mxfp8 moe training
CLA Signed
This label is managed by the Meta Open Source bot.
#1701
opened Sep 12, 2025 by
danielvegamyhre
Loading…
grouped expert and shared expert in same graph
CLA Signed
This label is managed by the Meta Open Source bot.
#1693
opened Sep 10, 2025 by
bobrenjc93
•
Draft
[WIP][DSV3] Offload dequantization process to DCP QuantizedHFReader
CLA Signed
This label is managed by the Meta Open Source bot.
[ignore] expert_overlap_compile suggested fixes
CLA Signed
This label is managed by the Meta Open Source bot.
#1687
opened Sep 8, 2025 by
bobrenjc93
Loading…
Separate SAC Wrapping of MoE and Attention Modules to Enable Flex Attention Compilation
CLA Signed
This label is managed by the Meta Open Source bot.
high priority
#1683
opened Sep 5, 2025 by
fegin
Loading…
Fake balanced routing in MoE
CLA Signed
This label is managed by the Meta Open Source bot.
#1670
opened Sep 1, 2025 by
rakkit
Loading…
Use new DeviceMesh unflatten to rewrite parallel_dims
CLA Signed
This label is managed by the Meta Open Source bot.
Support llama3 autoparallel + pipelining
CLA Signed
This label is managed by the Meta Open Source bot.
#1657
opened Aug 28, 2025 by
wconstab
Loading…
code refactor : making key steps modular train_step()
CLA Signed
This label is managed by the Meta Open Source bot.
fb-exported
#1650
opened Aug 28, 2025 by
Shagun-G
Loading…
[WIP] DCP: Dequantization and expert grouping for DSv3
CLA Signed
This label is managed by the Meta Open Source bot.
[DO NOT REVIEW] debug fsdp2 checkpoint for uneven sharding
CLA Signed
This label is managed by the Meta Open Source bot.
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.