Topic: "adaptive-computation"
koayon/awesome-adaptive-computation
A curated reading list of research in Adaptive Computation, Inference-Time Computation & Mixture of Experts (MoE).
Size: 331 KB - Last synced at: about 22 hours ago - Pushed at: 5 months ago - Stars: 147 - Forks: 9

LINs-lab/DynMoE
[ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models
Language: Python - Size: 57.3 MB - Last synced at: about 1 month ago - Pushed at: 4 months ago - Stars: 89 - Forks: 11

lucidrains/self-reasoning-tokens-pytorch
Exploration into the proposed "Self Reasoning Tokens" by Felipe Bonetto
Language: Python - Size: 81.1 KB - Last synced at: about 1 month ago - Pushed at: about 1 year ago - Stars: 55 - Forks: 4

lucidrains/pause-transformer
Yet another random morning idea to be quickly tried and architecture shared if it works; to allow the transformer to pause for any amount of time on any token
Language: Python - Size: 659 KB - Last synced at: 11 days ago - Pushed at: over 1 year ago - Stars: 53 - Forks: 1

Lee-Gihun/MicroNet_OSI-AI
(NeurIPS-2019 MicroNet Challenge - 3rd Winner) Open source code for "SIPA: A simple framework for efficient networks"
Language: Python - Size: 14.8 MB - Last synced at: about 1 year ago - Pushed at: over 2 years ago - Stars: 18 - Forks: 6

USArmyResearchLab/ARL-Hierarchical-Multiscale-Framework
The ARL Hierarchical MultiScale Framework (ARL-HMS) is a software library for development of multiscale models on heterogeneous high-performance computing systems.
Language: C++ - Size: 1.56 MB - Last synced at: 3 months ago - Pushed at: 11 months ago - Stars: 1 - Forks: 2

CNCLgithub/AdaptiveComputation
Model implementation for "Adaptive computation as a new mechanism of dynamic human attention"
Language: Julia - Size: 1.37 MB - Last synced at: 14 days ago - Pushed at: 14 days ago - Stars: 0 - Forks: 0
