An open API service providing repository metadata for many open source software ecosystems.

Topic: "model-quantization"

Efficient-ML/Awesome-Model-Quantization

A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.

Size: 61.5 MB - Last synced at: 10 days ago - Pushed at: 3 months ago - Stars: 2,084 - Forks: 221

horseee/Awesome-Efficient-LLM

A curated list for Efficient Large Language Models

Language: Python - Size: 62.3 MB - Last synced at: 4 days ago - Pushed at: 25 days ago - Stars: 1,657 - Forks: 134

datawhalechina/awesome-compression

模型压缩的小白入门教程

Size: 302 MB - Last synced at: 11 days ago - Pushed at: 6 months ago - Stars: 274 - Forks: 34

inferflow/inferflow

Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).

Language: C++ - Size: 1.89 MB - Last synced at: 7 days ago - Pushed at: about 1 year ago - Stars: 243 - Forks: 25

Efficient-ML/Awesome-Efficient-AIGC

A list of papers, docs, codes about efficient AIGC. This repo is aimed to provide the info for efficient AIGC research, including language and vision, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.

Size: 63.5 KB - Last synced at: 10 days ago - Pushed at: 3 months ago - Stars: 178 - Forks: 11

sayakpaul/Adventures-in-TensorFlow-Lite

This repository contains notebooks that show the usage of TensorFlow Lite for quantizing deep neural networks.

Language: Jupyter Notebook - Size: 49.1 MB - Last synced at: about 1 month ago - Pushed at: over 2 years ago - Stars: 172 - Forks: 35

RodolfoFerro/psychopathology-fer-assistant

[WINNER! 🏆] Psychopathology FER Assistant. Because mental health matters. My project submission for #TFWorld TF 2.0 Challenge at Devpost.

Language: Jupyter Notebook - Size: 12 MB - Last synced at: over 1 year ago - Pushed at: about 2 years ago - Stars: 67 - Forks: 25

htqin/BiBench

This project is the official implementation of our accepted ICML 2023 paper BiBench: Benchmarking and Analyzing Network Binarization.

Language: Python - Size: 110 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 41 - Forks: 3

htqin/QuantSR

This project is the official implementation of our accepted NeurIPS 2023 (spotlight) paper QuantSR: Accurate Low-bit Quantization for Efficient Image Super-Resolution.

Language: Python - Size: 9.75 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 31 - Forks: 2

seonglae/llama2gptq

Chat to LLaMa 2 that also provides responses with reference documents over vector database. Locally available model using GPTQ 4bit quantization.

Language: Python - Size: 9.48 MB - Last synced at: 7 days ago - Pushed at: over 1 year ago - Stars: 29 - Forks: 0

nbasyl/OFQ

The official implementation of the ICML 2023 paper OFQ-ViT

Language: Python - Size: 640 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 16 - Forks: 0

dcarpintero/ai-engineering

AI Engineering: Annotated NBs to dive into Self-Attention, In-Context Learning, RAG, Knowledge-Graphs, Fine-Tuning, Model Optimization, and many more.

Language: Jupyter Notebook - Size: 11.6 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 6 - Forks: 0

NANEXLABS/Nanex-AI

Enterprise multi-agent framework for secure, borderless data collaboration with zero-trust and federated learning-lightweight edge-ready.

Language: Python - Size: 119 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 4 - Forks: 0

frickyinn/BiDense

PyTorch implementation of "BiDense: Binarization for Dense Prediction," A binary neural network for dense prediction tasks.

Language: Python - Size: 1.21 MB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 3 - Forks: 0

SRDdev/Model-Quantization

Quantization is a technique to reduce the computational and memory costs of running inference by representing the weights and activations with low-precision data types like 8-bit integer (int8) instead of the usual 32-bit floating point (float32).

Language: Jupyter Notebook - Size: 3.16 MB - Last synced at: about 1 year ago - Pushed at: almost 2 years ago - Stars: 3 - Forks: 0

dwain-barnes/LLM-GGUF-Auto-Converter

Automated Jupyter notebook solution for batch converting Large Language Models to GGUF format with multiple quantization options. Built on llama.cpp with HuggingFace integration.

Language: Jupyter Notebook - Size: 13.7 KB - Last synced at: about 2 months ago - Pushed at: 4 months ago - Stars: 1 - Forks: 2

nnilayy/Spresense

Language: C++ - Size: 2.59 MB - Last synced at: almost 2 years ago - Pushed at: almost 2 years ago - Stars: 1 - Forks: 0

Keshavpatel2/local-llm-workbench

🧠 A comprehensive toolkit for benchmarking, optimizing, and deploying local Large Language Models. Includes performance testing tools, optimized configurations for CPU/GPU/hybrid setups, and detailed guides to maximize LLM performance on your hardware.

Language: Shell - Size: 8.79 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

xhay-p/ttPG

Torch and Transformers Playground: Learn and Code Deep Learning using PyTorch and HuggingFace Transformers.

Language: Jupyter Notebook - Size: 154 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

satyampurwar/large-language-models

Unlocking the Power of Generative AI: In-Context Learning, Instruction Fine-Tuning and Reinforcement Learning Fine-Tuning.

Language: Jupyter Notebook - Size: 170 KB - Last synced at: 3 months ago - Pushed at: 7 months ago - Stars: 0 - Forks: 0

Chenguiti6444/Vehicle_Detection_and_Classification_using_Deep_Learning

Fine-tuning Pretrained Deep Learning Models to Classify Low Quality Images of Land Vehicles.

Language: Jupyter Notebook - Size: 1.35 MB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 0 - Forks: 0

dslisleedh/NCNet-flax

Unofficial implementation of NCNet using flax and jax

Language: Python - Size: 131 KB - Last synced at: about 2 years ago - Pushed at: over 2 years ago - Stars: 0 - Forks: 0

Related Topics
model-compression 6 quantization 4 llm 3 cuda 3 large-language-models 3 llama2 2 llama-cpp 2 bert 2 generative-ai 2 knowledge-distillation 2 langchain 2 super-resolution 2 compression 2 tensorflow 2 python 2 pruning 2 awesome 2 binarized-neural-networks 2 transformers 2 deep-learning 2 tensorflow-lite 2 classification 2 model-pruning 2 efficient-deep-learning 2 arduino 1 low-rank-adaptation 1 megacmd 1 memory-management 1 binary-network 1 peft-fine-tuning-llm 1 prompt-engineering 1 proximal-policy-optimization 1 reinforcement-learning-from-ai-feedback 1 reinforcement-learning-from-human-feedback 1 storage-management 1 ai 1 aiagent 1 edge-computing 1 enterprise-ai 1 federated-learning 1 grpc-web 1 iot-security 1 mqtt-protocol 1 multi-agent-framework 1 onnx-runtime 1 zero-trust-security 1 conda-environment 1 encoder-decoder-model 1 encoder-model 1 few-shot-prompting 1 flan-t5 1 instruction-fine-tuning 1 kl-divergence 1 lightweight-neural-network 1 llm-compression 1 pruning-algorithms 1 baichuan2 1 bloom 1 deepseek 1 falcon 1 gemma 1 internlm 1 llamacpp 1 llm-inference 1 m2m100 1 minicpm 1 mistral 1 mixtral 1 mixture-of-experts 1 moe 1 multi-gpu-inference 1 phi-2 1 qwen 1 model-acceleration 1 aigc 1 diffusion-models 1 distillation 1 generative-model 1 kd 1 neural-architecture-search 1 prune 1 tinyml 1 chatai 1 chatbot 1 chatgpt 1 gpt 1 llama-2 1 question-answering 1 rye 1 streamlit-chat 1 efficient-llm 1 language-model 1 gguf 1 huggingface 1 jupyter-notebook 1 quantized-neural-networks 1 assistant-app 1 dash 1 dash-bootstrap-components 1 firebase-realtime-database 1