Repository brief

deepspeedai/DeepSpeed

Read the upstream summary on the left, browse the cached forks below it, and load each fork comparison into the right-hand panel.

Cached analysis
cached 2026-03-31T09:41:55.663Z

deepspeedai/DeepSpeed

DeepSpeed is an active Apache-2.0 deep learning optimization library for distributed training and inference. It has a large upstream footprint with 41,948 stars, 4,770 forks, and very recent activity on 2026-03-30, which makes it a high-interest upstream if you care about large-scale model training systems.

GitHub
Stars41,948
Forks4,770
Default branchmaster
Last pushed2026-03-30T20:07:25Z
Best maintainedHabanaAI/DeepSpeed
Closest to upstreamtarxemo/DeepSpeed
Most feature-richB06901052/DeepSpeed
Most opinionatedB06901052/DeepSpeed
Forks

Choose a fork to inspect

10 cached fork briefs
deepspeedai/DeepSpeed · Discofork