An open API service providing repository metadata for many open source software ecosystems.

Topic: "tensor-parallelism"

bigscience-workshop/petals

🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading

Language: Python - Size: 4.06 MB - Last synced at: 10 days ago - Pushed at: 9 months ago - Stars: 9,645 - Forks: 560

InternLM/InternEvo

InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencies.

Language: Python - Size: 6.8 MB - Last synced at: 14 days ago - Pushed at: 14 days ago - Stars: 389 - Forks: 67

kaiyuyue/torchshard

Slicing a PyTorch Tensor Into Parallel Shards

Language: Python - Size: 4.81 MB - Last synced at: 3 days ago - Pushed at: 5 days ago - Stars: 299 - Forks: 15

ai-decentralized/BloomBee

Decentralized LLMs fine-tuning and inference with offloading

Language: Python - Size: 36.6 MB - Last synced at: 11 days ago - Pushed at: 29 days ago - Stars: 91 - Forks: 15

xrsrke/pipegoose

Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*

Language: Python - Size: 1.26 MB - Last synced at: 11 days ago - Pushed at: over 1 year ago - Stars: 82 - Forks: 18

ShinoharaHare/LLM-Training

A distributed training framework for large language models powered by Lightning.

Language: Python - Size: 281 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 19 - Forks: 4

aniquetahir/JORA

JORA: JAX Tensor-Parallel LoRA Library

Language: Python - Size: 6.89 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 18 - Forks: 0

AlibabaPAI/FlashModels

Fast and easy distributed model training examples.

Language: Python - Size: 42.9 MB - Last synced at: 8 months ago - Pushed at: 8 months ago - Stars: 9 - Forks: 4

fattorib/transformer_shmap

Tensor Parallelism with JAX + Shard Map

Language: Python - Size: 85 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0