GitHub topics: deepspeed-ulysses
feifeibear/long-context-attention
USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference
Language: Python - Size: 4.61 MB - Last synced at: 2 days ago - Pushed at: about 2 months ago - Stars: 556 - Forks: 64

InternLM/InternEvo
InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencies.
Language: Python - Size: 6.79 MB - Last synced at: 16 days ago - Pushed at: 16 days ago - Stars: 404 - Forks: 70
