An open API service providing repository metadata for many open source software ecosystems.

GitHub topics: mixture-of-experts

lucidrains/soft-moe-pytorch

Implementation of Soft MoE, proposed by Brain's Vision team, in Pytorch

Language: Python - Size: 1.38 MB - Last synced at: about 19 hours ago - Pushed at: about 1 month ago - Stars: 290 - Forks: 8

learning-at-home/hivemind

Decentralized deep learning in PyTorch. Built to train models on thousands of volunteers across the world.

Language: Python - Size: 12.1 MB - Last synced at: about 16 hours ago - Pushed at: 8 days ago - Stars: 2,176 - Forks: 186

ymcui/Chinese-Mixtral

中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs)

Language: Python - Size: 519 KB - Last synced at: 2 days ago - Pushed at: about 1 year ago - Stars: 603 - Forks: 44

deepspeedai/DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language: Python - Size: 217 MB - Last synced at: 2 days ago - Pushed at: 4 days ago - Stars: 38,300 - Forks: 4,360

eduardzamfir/seemoredetails

[ICML 2024] See More Details: Efficient Image Super-Resolution by Experts Mining

Language: Python - Size: 10.2 MB - Last synced at: 2 days ago - Pushed at: 2 days ago - Stars: 138 - Forks: 2

andersonvc/alpha-pulse

Real-Time SEC 8-K Filing Analyzer for Financial Modeling

Language: Python - Size: 683 KB - Last synced at: 4 days ago - Pushed at: 4 days ago - Stars: 1 - Forks: 0

he-h/ST-MoE-BERT

This repository contains the code for the paper "ST-MoE-BERT: A Spatial-Temporal Mixture-of-Experts Framework for Long-Term Cross-City Mobility Prediction".

Language: Python - Size: 872 KB - Last synced at: 1 day ago - Pushed at: 3 months ago - Stars: 9 - Forks: 3

pjlab-sys4nlp/llama-moe

⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024)

Language: Python - Size: 1.69 MB - Last synced at: 4 days ago - Pushed at: 5 months ago - Stars: 961 - Forks: 56

EfficientMoE/MoE-Infinity

PyTorch library for cost-effective, fast and easy serving of MoE models.

Language: Python - Size: 457 KB - Last synced at: 4 days ago - Pushed at: 4 days ago - Stars: 179 - Forks: 13

LINs-lab/DynMoE

[ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models

Language: Python - Size: 57.3 MB - Last synced at: 3 days ago - Pushed at: 3 months ago - Stars: 89 - Forks: 11

CASE-Lab-UMD/Unified-MoE-Compression

The official implementation of the paper "Towards Efficient Mixture of Experts: A Holistic Study of Compression Techniques (TMLR)".

Language: Python - Size: 47.1 MB - Last synced at: 3 days ago - Pushed at: about 2 months ago - Stars: 67 - Forks: 5

relf/egobox

Efficient global optimization toolbox in Rust: bayesian optimization, mixture of gaussian processes, sampling methods

Language: Rust - Size: 11.9 MB - Last synced at: 7 days ago - Pushed at: 7 days ago - Stars: 105 - Forks: 6

arpita8/Awesome-Mixture-of-Experts-Papers

Survey: A collection of AWESOME papers and resources on the latest research in Mixture of Experts.

Size: 2.21 MB - Last synced at: 5 days ago - Pushed at: 9 months ago - Stars: 115 - Forks: 3

codelion/optillm

Optimizing inference proxy for LLMs

Language: Python - Size: 1.64 MB - Last synced at: 7 days ago - Pushed at: 7 days ago - Stars: 2,209 - Forks: 172

microsoft/Tutel

Tutel MoE: Optimized Mixture-of-Experts Library, Support DeepSeek FP8/FP4

Language: C - Size: 1.18 MB - Last synced at: 6 days ago - Pushed at: 10 days ago - Stars: 818 - Forks: 96

danelpeng/Awesome-Continual-Leaning-with-PTMs

This is a curated list of "Continual Learning with Pretrained Models" research.

Size: 254 KB - Last synced at: 3 days ago - Pushed at: about 2 months ago - Stars: 17 - Forks: 0

AIDC-AI/Parrot

🎉 The code repository for "Parrot: Multilingual Visual Instruction Tuning" in PyTorch.

Language: Python - Size: 25.2 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 36 - Forks: 1

SuperBruceJia/Awesome-Mixture-of-Experts

Awesome Mixture of Experts (MoE): A Curated List of Mixture of Experts (MoE) and Mixture of Multimodal Experts (MoME)

Size: 438 KB - Last synced at: 12 days ago - Pushed at: 4 months ago - Stars: 27 - Forks: 3

alexdrnd/micar-vl-moe

[IJCNN 2025] [Official code] - MicarVLMoE: A modern gated cross-aligned vision-language mixture of experts model for medical image captioning and report generation

Language: Python - Size: 935 KB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 0 - Forks: 0

inferflow/inferflow

Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).

Language: C++ - Size: 1.89 MB - Last synced at: 3 days ago - Pushed at: about 1 year ago - Stars: 243 - Forks: 25

dmis-lab/Monet

[ICLR 2025] Monet: Mixture of Monosemantic Experts for Transformers

Language: Python - Size: 252 KB - Last synced at: 9 days ago - Pushed at: 4 months ago - Stars: 66 - Forks: 3

kyegomez/SwitchTransformers

Implementation of Switch Transformers from the paper: "Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity"

Language: Python - Size: 2.42 MB - Last synced at: 3 days ago - Pushed at: about 1 month ago - Stars: 100 - Forks: 12

clint-kristopher-morris/llm-guided-evolution

LLM Guided Evolution - The Automation of Models Advancing Models

Language: Python - Size: 345 MB - Last synced at: 15 days ago - Pushed at: 15 days ago - Stars: 8 - Forks: 5

SMTorg/smt

Surrogate Modeling Toolbox

Language: Jupyter Notebook - Size: 163 MB - Last synced at: 13 days ago - Pushed at: 16 days ago - Stars: 755 - Forks: 215

reshalfahsi/gpt2moe-instruct

Instruction Fine-tuning of the GPT2MoE Model: GPT-2 with Mixture-of-Experts

Language: Jupyter Notebook - Size: 12 MB - Last synced at: 13 days ago - Pushed at: 16 days ago - Stars: 0 - Forks: 0

AI-14/micar-vl-moe

[IJCNN 2025] [Official code] - MicarVLMoE: A modern gated cross-aligned vision-language mixture of experts model for medical image captioning and report generation

Language: Python - Size: 950 KB - Last synced at: 16 days ago - Pushed at: 16 days ago - Stars: 0 - Forks: 0

ZhenbangDu/DSD

[IEEE TAI] Mixture-of-Experts for Open Set Domain Adaptation: A Dual-Space Detection Approach

Language: Python - Size: 643 KB - Last synced at: 17 days ago - Pushed at: 17 days ago - Stars: 5 - Forks: 0

alexliap/greek_gpt

MoE Decoder Transformer implementation with MLX

Language: Python - Size: 107 KB - Last synced at: 9 days ago - Pushed at: 9 days ago - Stars: 5 - Forks: 1

jaisidhsingh/pytorch-mixtures

One-stop solutions for Mixture of Experts and Mixture of Depth modules in PyTorch.

Language: Python - Size: 366 KB - Last synced at: 18 days ago - Pushed at: 18 days ago - Stars: 22 - Forks: 1

GnanaPrakashSG2004/Concept_Distillation

Framework to distill concepts of a teacher model to a student model. Trying to understand impact on performance and interpretability of the distilled student model obtained from this training paradigm.

Language: Jupyter Notebook - Size: 71.8 MB - Last synced at: 9 days ago - Pushed at: 9 days ago - Stars: 0 - Forks: 0

yanring/Megatron-MoE-ModelZoo

Best practices for testing advanced Mixtral, DeepSeek, and Qwen series MoE models using Megatron Core MoE.

Language: Python - Size: 26.4 KB - Last synced at: 20 days ago - Pushed at: 20 days ago - Stars: 8 - Forks: 1

AmazaspShumik/mtlearn

Multi-Task Learning package built with tensorflow 2 (Multi-Gate Mixture of Experts, Cross-Stitch, Ucertainty Weighting)

Language: Python - Size: 10.1 MB - Last synced at: 10 days ago - Pushed at: over 5 years ago - Stars: 52 - Forks: 6

xmarva/transformer-based-architectures

Breakdown of SoTA transformer-based architectures

Language: Jupyter Notebook - Size: 741 KB - Last synced at: 22 days ago - Pushed at: 22 days ago - Stars: 0 - Forks: 0

nusnlp/moece

The official code of the "Efficient and Interpretable Grammatical Error Correction with Mixture of Experts" paper

Language: Python - Size: 4.83 MB - Last synced at: 23 days ago - Pushed at: 23 days ago - Stars: 5 - Forks: 0

RufelleEmmanuelPactol/Mixture-of-Experts-Transcript-Evaluator

A mixture of experts inspired transcript evaluator using LLM fine-tuning. Contains a routing mechanism to assign specific questions to "experts".

Language: Jupyter Notebook - Size: 1.89 MB - Last synced at: 24 days ago - Pushed at: 24 days ago - Stars: 1 - Forks: 2

xrsrke/pipegoose

Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*

Language: Python - Size: 1.26 MB - Last synced at: 18 days ago - Pushed at: over 1 year ago - Stars: 82 - Forks: 18

koayon/awesome-adaptive-computation

A curated reading list of research in Adaptive Computation, Inference-Time Computation & Mixture of Experts (MoE).

Size: 331 KB - Last synced at: 9 days ago - Pushed at: 4 months ago - Stars: 143 - Forks: 9

korovod/kenotron

Experimental fork of Nanotron, a minimalistic large language model 4D-parallelism training

Language: Python - Size: 12.3 MB - Last synced at: 4 days ago - Pushed at: 29 days ago - Stars: 1 - Forks: 0

cmavro/PackLLM

Pack of LLMs: Model Fusion at Test-Time via Perplexity Optimization

Language: Python - Size: 169 KB - Last synced at: 6 days ago - Pushed at: about 1 year ago - Stars: 10 - Forks: 1

lucidrains/sinkhorn-router-pytorch

Self contained pytorch implementation of a sinkhorn based router, for mixture of experts or otherwise

Language: Python - Size: 27.3 KB - Last synced at: 4 days ago - Pushed at: 9 months ago - Stars: 34 - Forks: 0

dvmazur/mixtral-offloading

Run Mixtral-8x7B models in Colab or consumer desktops

Language: Python - Size: 261 KB - Last synced at: 30 days ago - Pushed at: about 1 year ago - Stars: 2,303 - Forks: 233

PKU-YuanGroup/MoE-LLaVA

Mixture-of-Experts for Large Vision-Language Models

Language: Python - Size: 16.5 MB - Last synced at: about 1 month ago - Pushed at: 5 months ago - Stars: 2,140 - Forks: 134

james-oldfield/muMoE

[NeurIPS'24] Multilinear Mixture of Experts: Scalable Expert Specialization through Factorization

Language: Python - Size: 2.95 MB - Last synced at: 25 days ago - Pushed at: 8 months ago - Stars: 30 - Forks: 1

Adlith/MoE-Jetpack

[NeurIPS 24] MoE Jetpack: From Dense Checkpoints to Adaptive Mixture of Experts for Vision Tasks

Language: Python - Size: 32.3 MB - Last synced at: about 1 month ago - Pushed at: 6 months ago - Stars: 115 - Forks: 1

antonio-f/mixture-of-experts-from-scratch

Mixture of Experts from scratch

Language: Jupyter Notebook - Size: 234 KB - Last synced at: 9 days ago - Pushed at: about 1 year ago - Stars: 6 - Forks: 1

louisbrulenaudet/mergeKit

Tools for merging pretrained Large Language Models and create Mixture of Experts (MoE) from open-source models.

Language: Jupyter Notebook - Size: 13.7 KB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 8 - Forks: 0

kyegomez/LIMoE

Implementation of the "the first large-scale multimodal mixture of experts models." from the paper: "Multimodal Contrastive Learning with LIMoE: the Language-Image Mixture of Experts"

Language: Python - Size: 2.17 MB - Last synced at: 29 days ago - Pushed at: about 1 month ago - Stars: 28 - Forks: 2

OpenSparseLLMs/LLaMA-MoE-v2

🚀 LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training

Language: Python - Size: 2.21 MB - Last synced at: about 1 month ago - Pushed at: 5 months ago - Stars: 78 - Forks: 11

BorealisAI/MMoEEx-MTL

PyTorch Implementation of the Multi-gate Mixture-of-Experts with Exclusivity (MMoEEx)

Language: Python - Size: 31.4 MB - Last synced at: about 1 month ago - Pushed at: almost 4 years ago - Stars: 32 - Forks: 4

lucidrains/st-moe-pytorch

Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch

Language: Python - Size: 178 KB - Last synced at: about 1 month ago - Pushed at: 11 months ago - Stars: 326 - Forks: 28

lucidrains/mixture-of-experts

A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models

Language: Python - Size: 136 KB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 724 - Forks: 55

AviSoori1x/makeMoE

From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)

Language: Jupyter Notebook - Size: 6.96 MB - Last synced at: about 1 month ago - Pushed at: 7 months ago - Stars: 686 - Forks: 73

SkyworkAI/MoH

MoH: Multi-Head Attention as Mixture-of-Head Attention

Language: Python - Size: 5.26 MB - Last synced at: about 1 month ago - Pushed at: 7 months ago - Stars: 233 - Forks: 9

dannyxiaocn/awesome-moe

a repo for moe papers and systems aggregation

Size: 474 KB - Last synced at: 6 days ago - Pushed at: over 3 years ago - Stars: 7 - Forks: 2

Bhazantri/EvoLingua

EvoLingua: A Scalable Mixture-of-Experts Language Model Framework

Language: Python - Size: 36.1 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

efeslab/fiddler

[ICLR'25] Fast Inference of MoE Models with CPU-GPU Orchestration

Language: Python - Size: 1.72 MB - Last synced at: about 1 month ago - Pushed at: 6 months ago - Stars: 203 - Forks: 18

zjukg/MoMoK

[Paper][ICLR 2025] Multiple Heads are Better than One: Mixture of Modality Knowledge Experts for Entity Representation Learning

Language: Python - Size: 6.99 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 24 - Forks: 2

lucidrains/PEER-pytorch

Pytorch implementation of the PEER block from the paper, Mixture of A Million Experts, by Xu Owen He at Deepmind

Language: Python - Size: 271 KB - Last synced at: 4 days ago - Pushed at: 9 months ago - Stars: 123 - Forks: 3

shufangxun/LLaVA-MoD

[ICLR 2025] LLaVA-MoD: Making LLaVA Tiny via MoE-Knowledge Distillation

Language: Python - Size: 3.41 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 120 - Forks: 7

lucidrains/mixture-of-attention

Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts

Language: Python - Size: 34.1 MB - Last synced at: 4 days ago - Pushed at: 7 months ago - Stars: 118 - Forks: 4

YAGI0423/gpt_modules

「Pytorch 기반 GPT 모델 및 모듈 라이브러리」에 대한 내용을 다루고 있습니다.

Language: Python - Size: 343 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

Vignesh010101/Intelligent-Health-LLM-System

An Intelligent Health LLM System for Personalized Medication Guidance and Support.

Language: Jupyter Notebook - Size: 615 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

Nitin-Sagar-B/RaTiO-CoRE

A modular multi-model AI framework demonstrating advanced techniques in semantic knowledge transfer, context management, and collaborative intelligence across diverse language models.

Language: Python - Size: 136 KB - Last synced at: about 1 month ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

YangLing0818/RealCompo

[NeurIPS 2024] RealCompo: Balancing Realism and Compositionality Improves Text-to-Image Diffusion Models

Language: Python - Size: 7.45 MB - Last synced at: about 2 months ago - Pushed at: 6 months ago - Stars: 115 - Forks: 4

alt2177/mllm-public

A framework for merging multiple LMs to improve OOD performance without additional training

Language: Jupyter Notebook - Size: 166 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

ozyurtf/mixture-of-experts

Training two separate expert neural networks and one gater that can switch the expert networks.

Language: Python - Size: 10.1 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 6 - Forks: 0

bwconrad/soft-moe

PyTorch implementation of "From Sparse to Soft Mixtures of Experts"

Language: Python - Size: 344 KB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 53 - Forks: 3

OpenSparseLLMs/CLIP-MoE

CLIP-MoE: Mixture of Experts for CLIP

Language: Python - Size: 2.35 MB - Last synced at: about 1 month ago - Pushed at: 7 months ago - Stars: 29 - Forks: 0

EfficientMoE/MoE-Gen

High-throughput offline inference for MoE models with limited GPUs

Language: Python - Size: 552 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 8 - Forks: 0

AdamG012/moe-paper-models

A sumary of MoE experimental setups across a number of different papers.

Size: 10.7 KB - Last synced at: 8 days ago - Pushed at: about 2 years ago - Stars: 16 - Forks: 1

APWS25/AccelMoE

This repository is for CUDA kernel re-implementation of CPU-based MoE model.

Language: C++ - Size: 687 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 1

eduardzamfir/MoCE-IR

[CVPR 2025] Complexity Experts are Task-Discriminative Learners for Any Image Restoration

Language: Python - Size: 821 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 33 - Forks: 0

vishwapatel14/Skin-Cancer-Classification-Model

Skin Cancer Classification Project for Advanced Machine Learning course in Fall 2024 at The University of Texas at Austin

Language: Jupyter Notebook - Size: 15 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 2

Keefe-Murphy/MEDseq

Mixtures of Exponential-Distance Models for Clustering Longitudinal Life-Course Sequences with Gating Covariates and Sampling Weights

Language: R - Size: 10.1 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 5 - Forks: 0

checkstep/mole-stance

MoLE: Cross-Domain Label-Adaptive Stance Detection

Language: Python - Size: 47.9 KB - Last synced at: about 2 months ago - Pushed at: about 3 years ago - Stars: 17 - Forks: 5

924973292/DeMo

【AAAI2025】DeMo: Decoupled Feature-Based Mixture of Experts for Multi-Modal Object Re-Identification

Language: Python - Size: 17 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 41 - Forks: 2

UNITES-Lab/HEXA-MoE

Official code for the paper "HEXA-MoE: Efficient and Heterogeneous-Aware MoE Acceleration with Zero Computation Redundancy"

Language: Python - Size: 19.2 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 9 - Forks: 1

UNITES-Lab/glider

Official code for the paper "Glider: Global and Local Instruction-Driven Expert Router"

Language: Python - Size: 477 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 7 - Forks: 0

Keefe-Murphy/MoEClust

Gaussian Parsimonious Clustering Models with Gating and Expert Network Covariates

Language: R - Size: 1.67 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 7 - Forks: 0

Luodian/Generalizable-Mixture-of-Experts

GMoE could be the next backbone model for many kinds of generalization task.

Language: Python - Size: 2.04 MB - Last synced at: 6 days ago - Pushed at: about 2 years ago - Stars: 269 - Forks: 35

eduardzamfir/DaAIR

GitHub repository for our project "Efficient Degradation-aware Any Image Restoration"

Size: 15.6 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 30 - Forks: 0

Wuyxin/GraphMETRO

GraphMETRO: Mitigating Complex Graph Distribution Shifts via Mixture of Aligned Experts (NeurIPS 2024)

Language: Python - Size: 36.1 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 21 - Forks: 1

xpmoe/app

Mixture of Experts Framework for Enhanced Explainability of Anxiety States Pre- and Post-Intervention Across Experimental Groups

Size: 9.77 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 3 - Forks: 0

Spico197/MoE-SFT

🍼 Official implementation of Dynamic Data Mixing Maximizes Instruction Tuning for Mixture-of-Experts

Language: Python - Size: 552 KB - Last synced at: about 1 month ago - Pushed at: 8 months ago - Stars: 38 - Forks: 0

discover-Austin/Neural_Network_Mixture_of_Experts

Mixture_of_Experts implemented in an easy to understand and custom Neural Network

Language: Python - Size: 12.7 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

rhymes-ai/Aria

Codebase for Aria - an Open Multimodal Native MoE

Language: Jupyter Notebook - Size: 120 MB - Last synced at: 3 months ago - Pushed at: 4 months ago - Stars: 995 - Forks: 83

mryab/learning-at-home

"Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts" (NeurIPS 2020), original PyTorch implementation

Language: Jupyter Notebook - Size: 272 KB - Last synced at: 14 days ago - Pushed at: over 4 years ago - Stars: 54 - Forks: 1

LoserCheems/WonderfulMatrices

Wonderful Matrices to Build Small Language Models

Language: Python - Size: 8.78 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 43 - Forks: 0

PV-Bhat/RSRC

RSRC Calculator is a practical tool designed to evaluate the efficiency of AI models in the post-scaling era: Recursive Self-Referential Compression (RSRC), this tool computes training efficiency metrics by analyzing factors such as training FLOPs, energy consumption, and model architecture details.

Language: Python - Size: 75.2 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

fkodom/soft-mixture-of-experts

PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)

Language: Python - Size: 152 KB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 71 - Forks: 5

clementetienam/Data_Driven_MPC_Controller_Using_CCR

Language: MATLAB - Size: 11.9 MB - Last synced at: about 1 month ago - Pushed at: over 4 years ago - Stars: 4 - Forks: 2

VITA-Group/Neural-Implicit-Dict

[ICML 2022] "Neural Implicit Dictionary via Mixture-of-Expert Training" by Peihao Wang, Zhiwen Fan, Tianlong Chen, Zhangyang Wang

Language: Python - Size: 958 KB - Last synced at: 25 days ago - Pushed at: over 1 year ago - Stars: 40 - Forks: 1

nitic-nlp-team/webnavix

A continuous generalist web navigation agent that merges individually fine-tuned LLMs as domain experts.

Language: Typst - Size: 56.3 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 2 - Forks: 0

ilyalasy/moe-routing

Analysis of token routing for different implementations of Mixture of Experts

Language: Jupyter Notebook - Size: 882 KB - Last synced at: about 1 month ago - Pushed at: about 1 year ago - Stars: 9 - Forks: 0

dominiquegarmier/grok-pytorch

pytorch implementation of grok

Language: Python - Size: 44.9 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 13 - Forks: 0

dsy109/mixtools

Tools for Analyzing Finite Mixture Models

Language: R - Size: 499 KB - Last synced at: 30 days ago - Pushed at: 11 months ago - Stars: 20 - Forks: 4

AhmedMagdyHendawy/MOORE

Official code of the paper "Multi-Task Reinforcement Learning with Mixture of Orthogonal Experts" at ICLR2024

Language: Python - Size: 813 KB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 7 - Forks: 3

s-chh/PyTorch-Scratch-LLM

Simple and easy to understand PyTorch implementation of Large Language Model (LLM) GPT and LLAMA from scratch with detailed steps. Implemented: Byte-Pair Tokenizer, Rotational Positional Embedding (RoPe), SwishGLU, RMSNorm, Mixture of Experts (MOE). Tested on Taylor Swift song lyrics dataset.

Language: Python - Size: 58.6 KB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 2 - Forks: 0

jyjohnchoi/SMoP

The repository contains the code for our EMNLP 2023 paper "SMoP: Towards Efficient and Effective Prompt Tuning with Sparse Mixture-of-Prompts", written by Joon-Young Choi, Junho Kim, Jun-Hyung Park, Mok-Wing Lam, and SangKeun Lee.

Language: Python - Size: 154 KB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 7 - Forks: 2

ednialzavlare/MixKABRN

This is the repo for the MixKABRN Neural Network (Mixture of Kolmogorov-Arnold Bit Retentive Networks), and an attempt at first adapting it for training on text, and later adjust it for other modalities.

Language: Python - Size: 85.9 KB - Last synced at: 3 months ago - Pushed at: about 1 year ago - Stars: 4 - Forks: 0

Related Keywords
mixture-of-experts 150 deep-learning 30 moe 24 machine-learning 22 llm 22 pytorch 21 large-language-models 20 artificial-intelligence 14 nlp 13 transformer 11 transformers 9 computer-vision 9 ai 9 gaussian-processes 6 language-model 6 neural-networks 6 deep-neural-networks 5 llm-inference 5 llama 5 multi-task-learning 5 efficiency 5 multimodal-large-language-models 4 natural-language-processing 4 vision-transformer 4 transfer-learning 4 generative-ai 4 python 4 ml 4 llms 4 attention 4 ensemble 4 mixtral-8x7b 3 multi-modal 3 huggingface 3 pytorch-implementation 3 gpt 3 instruction-tuning 3 graph-neural-networks 3 unsupervised-learning 3 llms-reasoning 3 keras 3 mixture-models 3 ensemble-learning 3 foundation-models 3 vision-language-models 3 deepseek 3 prompt-tuning 3 inference 3 low-level-vision 3 tensorflow 3 lora 3 conditional-computation 3 neural-network 3 sft 2 feature-pyramid-network 2 ct-scans 2 chest-xrays 2 alignment-strategies 2 model 2 generative-model 2 mixtral 2 deep-reinforcement-learning 2 llms-benchmarking 2 data-parallelism 2 regression-algorithms 2 gpu 2 peft 2 machine-learning-algorithms 2 model-parallelism 2 pipeline-parallelism 2 optimization 2 generalization 2 mixture-of-models 2 mixture-model 2 gpt4 2 distributed-systems 2 training 2 interpretability 2 qwen 2 llama3 2 contrastive-learning 2 fine-tuning 2 mistral 2 clustering 2 anomaly-detection 2 federated-learning 2 mergekit 2 domain-adaptation 2 radiology-report-generation 2 medical-imaging 2 medical-image-captioning 2 agents 2 attention-mechanism 2 peft-fine-tuning-llm 2 multitask-learning 2 agent 2 mixture-of-adapters 2 awesome 2 surrogate-models 2 parameter-efficient-fine-tuning 2