GitHub topics: sequence-parallelism
InternLM/InternEvo
InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencies.
Language: Python - Size: 6.78 MB - Last synced at: about 16 hours ago - Pushed at: about 17 hours ago - Stars: 386 - Forks: 64

xrsrke/pipegoose
Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*
Language: Python - Size: 1.26 MB - Last synced at: 12 days ago - Pushed at: over 1 year ago - Stars: 82 - Forks: 18

AlibabaPAI/FlashModels
Fast and easy distributed model training examples.
Language: Python - Size: 42.9 MB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 9 - Forks: 4
