Topic: "small-models"
SqueezeAILab/SqueezeLLM
[ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization
Language: Python - Size: 1.5 MB - Last synced at: 7 days ago - Pushed at: 10 months ago - Stars: 688 - Forks: 45

SqueezeAILab/KVQuant
[NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
Language: Python - Size: 19.8 MB - Last synced at: 12 days ago - Pushed at: 10 months ago - Stars: 350 - Forks: 31

aitomatic/openssa
OpenSSA: Small Specialist Agents based on Domain-Aware Neurosymbolic Agent (DANA) architecture for industrial problem-solving
Language: Python - Size: 44.8 MB - Last synced at: 1 day ago - Pushed at: 7 days ago - Stars: 303 - Forks: 48

logic-OT/Decoder-Only-LLM
This repository features a custom-built decoder-only language model (LLM) with a total of 37 million parameters 🔥. I train the model to be able to ask question from a given context
Language: Jupyter Notebook - Size: 396 KB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 9 - Forks: 3

MCG-NJU/AMD
[CVPR 2024] Asymmetric Masked Distillation for Pre-Training Small Foundation Models
Language: Python - Size: 961 KB - Last synced at: 12 months ago - Pushed at: 12 months ago - Stars: 7 - Forks: 1

WenliangZhoushan/GreaterPrompt
🤩 GreaterPrompt: A Python Toolkit for Prompt Optimization via Small Models
Language: Python - Size: 27.2 MB - Last synced at: 17 days ago - Pushed at: 2 months ago - Stars: 5 - Forks: 0

lolguy-s-AI-stuff/micro-models
AI models that are less than 1000 parameters
Language: C++ - Size: 25.4 KB - Last synced at: over 1 year ago - Pushed at: almost 2 years ago - Stars: 4 - Forks: 0

sfarhat/dapt
Code for "On the Surprising Efficacy of Distillation as an Alternative to Pre-Training Small Models"
Language: Python - Size: 37.1 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 2 - Forks: 0

zhangyifei01/Awesome-Self-supervised-Learning-of-Tiny-Models
Overview of self-supervised learning of tiny models, including distillation-based methods (aks. self-supervised distillation) and non-distillation methods.
Size: 17.6 KB - Last synced at: about 2 years ago - Pushed at: over 2 years ago - Stars: 2 - Forks: 0

ENSTA-U2IS-AI/optuMNIST
Help us define the Pareto front of small models for MNIST classification. Frugal AI.
Language: Python - Size: 7.81 KB - Last synced at: about 1 year ago - Pushed at: almost 2 years ago - Stars: 1 - Forks: 0

dane-meister/TellMeWhy-Context-Injection
Fine-tunes a T5-small model on the TellMeWhy dataset using context injection from a large language model (Gemini) to improve causal reasoning for “why” questions in narratives. Combines efficient training with human and automated evaluations to assess impact.
Language: Jupyter Notebook - Size: 3.54 MB - Last synced at: 4 days ago - Pushed at: 13 days ago - Stars: 0 - Forks: 0

Someboi1681/BobVLM
BobVLM – A 1.5B multimodal model built from scratch and pre-trained on a single P100 GPU capable of image descriptions and moderate question answering. 🤗🎉
Size: 1000 Bytes - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0
