Topic: "conditional-computation"
lucidrains/st-moe-pytorch
Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch
Language: Python - Size: 178 KB - Last synced at: about 3 hours ago - Pushed at: 11 months ago - Stars: 331 - Forks: 29

thomasverelst/dynconv
Code for Dynamic Convolutions: Exploiting Spatial Sparsity for Faster Inference (CVPR2020)
Language: Cuda - Size: 105 MB - Last synced at: over 1 year ago - Pushed at: over 3 years ago - Stars: 120 - Forks: 14

Adlith/MoE-Jetpack
[NeurIPS 24] MoE Jetpack: From Dense Checkpoints to Adaptive Mixture of Experts for Vision Tasks
Language: Python - Size: 32.3 MB - Last synced at: about 1 month ago - Pushed at: 6 months ago - Stars: 115 - Forks: 1

antonio-f/mixture-of-experts-from-scratch
Mixture of Experts from scratch
Language: Jupyter Notebook - Size: 234 KB - Last synced at: 12 days ago - Pushed at: about 1 year ago - Stars: 6 - Forks: 1

Mixture-AI/Mixture-of-Depths
Google DeepMind: Mixture of Depths Unofficial Implementation.
Language: Python - Size: 15.6 KB - Last synced at: 12 months ago - Pushed at: 12 months ago - Stars: 1 - Forks: 1
