-
Notifications
You must be signed in to change notification settings - Fork 4.7k
Pull requests: deepspeedai/DeepSpeed
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
[BUG] Fix: Fix gradient norm calculation and dynamic shape blocking in PP+ZeRO1 collective communication
#7847
opened Feb 12, 2026 by
Thinksky5124
Loading…
Fix ROCm BF16 conversion intrinsics in inference v2 (#7843)
#7846
opened Feb 12, 2026 by
tohtana
Loading…
Fix no-grad grad-fn lookup in ZeRO hook counting on PyTorch 2.3 (#7830)
#7841
opened Feb 10, 2026 by
tohtana
Loading…
Throw error when parameter is modified in GatheredParameters
#7832
opened Feb 5, 2026 by
tohtana
Loading…
fix: Ensure full gradient reduction for Muon with reduce_scatter
#7808
opened Jan 23, 2026 by
nathon-lee
Loading…
Fix bf16 dtype mismatch in ZeRO-3 with zero_quantized_weights
#7792
opened Jan 18, 2026 by
juyterman1000
Loading…
Fix Muon optimizer conflict with gradient clipping in ZeRO 1/2
#7776
opened Jan 12, 2026 by
fy817
Loading…
Fix: ZenFlow Adam integration for updated PyTorch backward flow (#7759)
#7771
opened Jan 11, 2026 by
Antlera
Loading…
Introduce all_reduce_hook to support gradient aggregation across replica groups.
#7764
opened Jan 7, 2026 by
zhengchenyu
Loading…
feat: add parameter-level precision control for BF16 training
#7750
opened Dec 30, 2025 by
nathon-lee
Loading…
Fix Muon optimizer checkpoint resume with bf16 mode
#7748
opened Dec 28, 2025 by
yurekami
Loading…
2 tasks done
Introduce Megatron-style parallel state management
#7726
opened Dec 15, 2025 by
eternalNight
Loading…
4 of 5 tasks
let allgather and alltoall execute in parallel when both attention and MOE used TP
#7723
opened Dec 11, 2025 by
taozhiwei
Loading…
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.