Ecosyste.ms: Repos

An open API service providing repository metadata for many open source software ecosystems.

GitHub topics: quantization

SqueezeAILab/KVQuant

KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization

Language: Python - Size: 19.7 MB - Last synced: about 1 month ago - Pushed: about 1 month ago - Stars: 187 - Forks: 14

Asad-Ismail/lane_detection

Lane Detection and Classification using Front camera monocular images

Language: Python - Size: 40.9 MB - Last synced: about 1 month ago - Pushed: about 1 year ago - Stars: 3 - Forks: 0

IST-DASLab/marlin

FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.

Language: Python - Size: 758 KB - Last synced: about 1 month ago - Pushed: about 1 month ago - Stars: 314 - Forks: 21

hailo-ai/hailo_model_zoo

The Hailo Model Zoo includes pre-trained models and a full building and evaluation environment

Language: Python - Size: 4.32 MB - Last synced: about 1 month ago - Pushed: about 2 months ago - Stars: 91 - Forks: 28

PaddlePaddle/PaddleSlim

PaddleSlim is an open-source library for deep model compression and architecture search.

Language: Python - Size: 16.3 MB - Last synced: about 1 month ago - Pushed: 2 months ago - Stars: 1,514 - Forks: 347

jy-yuan/KIVI

KIVI: A Tuning-Free Asymmetric 2bit Quantization for KV Cache

Language: Python - Size: 16.8 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 107 - Forks: 5

lmEshoo/quantization

Post-training model quantization using apache's tvm.ai

Language: Jupyter Notebook - Size: 37.9 MB - Last synced: about 2 months ago - Pushed: about 4 years ago - Stars: 1 - Forks: 1

lmEshoo/pruning

model weight pruning

Language: Jupyter Notebook - Size: 72.4 MB - Last synced: about 2 months ago - Pushed: about 4 years ago - Stars: 0 - Forks: 0

intel/auto-round

SOTA Weight-only Quantization Algorithm for LLMs

Language: Python - Size: 8.33 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 48 - Forks: 7

emsquared2/Telecommunications-NTUA

Project assignment for course Introduction to Telecommunications at ECE NTUA

Language: MATLAB - Size: 4.14 MB - Last synced: about 2 months ago - Pushed: about 1 year ago - Stars: 0 - Forks: 0

huawei-noah/Pretrained-Language-Model

Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.

Language: Python - Size: 29 MB - Last synced: about 2 months ago - Pushed: 4 months ago - Stars: 2,953 - Forks: 623

iAmGiG/MadeSmallML

MadeSmallML is an open-source initiative designed to explore model quantization techniques for machine learning models. Our goal is to enable efficient deployment of these models on devices with limited computational resources by reducing model size and computational demands without significantly compromising performance.

Size: 5.86 KB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 0 - Forks: 0

intel/intel-extension-for-pytorch

A Python package for extending the official PyTorch that can easily obtain performance on Intel platform

Language: Python - Size: 92.1 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 1,316 - Forks: 193

neuralmagic/sparsify

ML model optimization product to accelerate inference.

Language: Python - Size: 7.18 MB - Last synced: 14 days ago - Pushed: about 2 months ago - Stars: 315 - Forks: 27

kurianbenoy/Indic-Subtitler

Open source subtitling platform ๐Ÿ’ป for transcribing and translating videos/audios in Indic languages.

Language: Jupyter Notebook - Size: 36.3 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 55 - Forks: 7

OpenGVLab/OmniQuant

[ICLR2024 spotlight] OmniQuant is a simple and powerful quantization technique for LLMs.

Language: Python - Size: 8.17 MB - Last synced: about 2 months ago - Pushed: 2 months ago - Stars: 546 - Forks: 43

huggingface/optimum

๐Ÿš€ Accelerate training and inference of ๐Ÿค— Transformers and ๐Ÿค— Diffusers with easy to use hardware optimization tools

Language: Python - Size: 4.01 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 2,096 - Forks: 356

ModelTC/llmc

llmc is an efficient LLM compression tool with various advanced compression methods, supporting multiple inference backends.

Language: Python - Size: 3.35 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 25 - Forks: 2

archinetai/bitcodes-pytorch

A vector quantization method with binary codes, in PyTorch.

Language: Python - Size: 5.86 KB - Last synced: 9 days ago - Pushed: over 1 year ago - Stars: 6 - Forks: 0

LowinLi/stable-diffusion-streamlit

Quantized stable-diffusion cutting down memory 75%, testing in streamlit, deploying in container

Language: Python - Size: 53.6 MB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 53 - Forks: 7

minseok0809/transformers-compression-practice

Transformers Compression Practice

Language: Jupyter Notebook - Size: 19.9 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 0 - Forks: 0

Spinkoo/Embedded-Yolov7

Compress, Deploy and Inference YOLOv7 on low-cost MCUs

Language: Jupyter Notebook - Size: 53.3 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 2 - Forks: 0

Azazel0203/Sentiment_Analyser

This project employs transformer models for sentiment analysis on primate-themed posts, encompassing data preprocessing, model training, evaluation, and optional quantization. It provides a streamlined workflow for analyzing sentiment in primate-related textual data, offering insights into public perceptions of primates.

Language: Jupyter Notebook - Size: 25.4 KB - Last synced: 23 days ago - Pushed: about 2 months ago - Stars: 2 - Forks: 0

jiho264/2024-RISE-Quantization

Try to get some insights about Quantization

Language: Jupyter Notebook - Size: 6.77 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 0 - Forks: 0

minseok0809/awesome-ai-paper

A curated list of awesome NLP, Computer Vision, Model Compression, XAI, Reinforcement Learning, Security etc Paper

Language: Jupyter Notebook - Size: 2.81 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 0 - Forks: 0

clovaai/frostnet

FrostNet: Towards Quantization-Aware Network Architecture Search

Language: Python - Size: 505 KB - Last synced: 18 days ago - Pushed: 28 days ago - Stars: 106 - Forks: 18

monk1337/NanoPeft

The simplest repository & Neat implementation of different Lora methods for training/fine-tuning Transformer-based models (i.e., BERT, GPTs). [ Research purpose ]

Language: Jupyter Notebook - Size: 68.4 KB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 5 - Forks: 2

fdbtrs/QuantFace

QuantFace: Towards Lightweight Face Recognition by Synthetic Data Low-bit Quantization

Language: Python - Size: 5.47 MB - Last synced: 18 days ago - Pushed: almost 2 years ago - Stars: 30 - Forks: 3

RahulSChand/gpu_poor

Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization

Language: JavaScript - Size: 4.85 MB - Last synced: about 2 months ago - Pushed: 7 months ago - Stars: 595 - Forks: 22

1duo/awesome-ai-infrastructures

Infrastructuresโ„ข for Machine Learning Training/Inference in Production.

Size: 11.8 MB - Last synced: 12 days ago - Pushed: about 5 years ago - Stars: 368 - Forks: 71

elphinkuo/llamaqt.c

Clean C language version of quantizing llama2 model and running quantized llama2 model

Language: C - Size: 455 KB - Last synced: about 2 months ago - Pushed: 9 months ago - Stars: 2 - Forks: 0

aaron-xichen/pytorch-playground

Base pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)

Language: Python - Size: 45.9 KB - Last synced: 2 months ago - Pushed: over 1 year ago - Stars: 2,543 - Forks: 616

mit-han-lab/tinyengine

[NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning; [NeurIPS 2022] MCUNetV3: On-Device Training Under 256KB Memory

Language: C - Size: 235 MB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 717 - Forks: 123

stracini-git/bitwise-weight-training

Training neural network weights in a bitwise fashion: https://arxiv.org/abs/2202.09571

Language: Python - Size: 1.8 MB - Last synced: 2 months ago - Pushed: almost 2 years ago - Stars: 2 - Forks: 0

stracini-git/qnn

Training neural nets with quantized weights on arbitrarily specified bit-depth

Language: Python - Size: 1.2 MB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 1 - Forks: 1

daehwa00/VoiceKey

Published "VoiceKey: Real-time Compression Encoding and Quantization for Voice Authentication Model on Edge Devices" at The 4th Korea Artificial Intelligence Conference.

Language: Python - Size: 2.88 MB - Last synced: 19 days ago - Pushed: 3 months ago - Stars: 1 - Forks: 1

megvii-research/Sparsebit

A model compression and acceleration toolbox based on pytorch.

Language: Python - Size: 7.45 MB - Last synced: about 2 months ago - Pushed: 5 months ago - Stars: 317 - Forks: 39

mit-han-lab/haq

[CVPR 2019, Oral] HAQ: Hardware-Aware Automated Quantization with Mixed Precision

Language: Python - Size: 64.5 KB - Last synced: 2 months ago - Pushed: over 3 years ago - Stars: 350 - Forks: 83

Cydia2018/YOLOv5-Light

provide some new architecture, channel pruning and quantization methods for yolov5

Language: Python - Size: 12.5 MB - Last synced: about 1 month ago - Pushed: 4 months ago - Stars: 29 - Forks: 5

shivpalSW/Optimized-CPU-Implementation-of-Llama2

Optimized CPU Implementation of Llama2-LLM

Language: Python - Size: 703 KB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 0 - Forks: 0

FLHonker/ZAQ-code

CVPR 2021 : Zero-shot Adversarial Quantization (ZAQ)

Language: Python - Size: 188 KB - Last synced: about 2 months ago - Pushed: over 2 years ago - Stars: 64 - Forks: 16

Xiuyu-Li/q-diffusion

[ICCV 2023] Q-Diffusion: Quantizing Diffusion Models.

Language: Python - Size: 5.97 MB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 243 - Forks: 17

briancpark/csc791-025

Computer Science 791-025: Real-Time AI & High-Performance Machine Learning

Language: TeX - Size: 77 MB - Last synced: about 1 month ago - Pushed: 4 months ago - Stars: 1 - Forks: 0

alizameller/Digitial-Signal-Processing

ECE-310 Digital Signal Processing - Cooper Union Fall 2022

Language: MATLAB - Size: 7.81 KB - Last synced: 3 months ago - Pushed: over 1 year ago - Stars: 0 - Forks: 0

GATECH-EIC/Double-Win-Quant

[ICML 2021] "Double-Win Quant: Aggressively Winning Robustness of Quantized DeepNeural Networks via Random Precision Training and Inference" by Yonggan Fu, Qixuan Yu, Meng Li, Vikas Chandra, Yingyan Lin

Language: Python - Size: 947 KB - Last synced: about 1 month ago - Pushed: over 2 years ago - Stars: 12 - Forks: 5

motokimura/pytorch_quantization_fx

An example to quantize MobileNetV2 trained on CIFAR-10 dataset with PyTorch

Language: Python - Size: 86.6 MB - Last synced: 2 months ago - Pushed: over 1 year ago - Stars: 6 - Forks: 3

linkedin/QuantEase

QuantEase, a layer-wise quantization framework, frames the problem as discrete-structured non-convex optimization. Our work leverages Coordinate Descent techniques, offering high-quality solutions without the need for matrix inversion or decomposition.

Language: Python - Size: 209 KB - Last synced: about 2 months ago - Pushed: 3 months ago - Stars: 15 - Forks: 1

laelhalawani/glai

glai - GGUF LLAMA AI - Package for simplified model handling and text generation with Llama models quantized to GGUF format. APIs for downloading and loading models automatically, includes a db with models of various scale and quantizations. With this high level API you need one line to load the model and one to generate text completions.

Language: Python - Size: 208 KB - Last synced: 3 months ago - Pushed: 5 months ago - Stars: 3 - Forks: 0

daemyung/nzip

The deep learning quantization library for PyTorch.

Language: Python - Size: 16.6 KB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 1 - Forks: 0

dsindex/ntagger

reference pytorch code for named entity tagging

Language: Python - Size: 36 MB - Last synced: 2 months ago - Pushed: almost 2 years ago - Stars: 85 - Forks: 13

AnonymousWorld123/Q-Layer

Code for Q-Layer: Quantization Layer for Robust Convolutional Neural Network

Language: Python - Size: 103 KB - Last synced: about 1 month ago - Pushed: over 3 years ago - Stars: 3 - Forks: 0

megvii-research/SSQL-ECCV2022

PyTorch implementation of SSQL (Accepted to ECCV2022 oral presentation)

Language: Python - Size: 211 KB - Last synced: 24 days ago - Pushed: about 1 year ago - Stars: 75 - Forks: 6

megvii-research/FQ-ViT

[IJCAI 2022] FQ-ViT: Post-Training Quantization for Fully Quantized Vision Transformer

Language: Python - Size: 729 KB - Last synced: 3 months ago - Pushed: about 1 year ago - Stars: 259 - Forks: 45

ModelTC/QLLM

[ICLR 2024] This is the official PyTorch implementation of "QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Models"

Language: Python - Size: 1.68 MB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 15 - Forks: 0

jiho264/2024-RISE

Try to gain some insights from Quantization

Language: Jupyter Notebook - Size: 3.69 MB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 0 - Forks: 0

bupt-ai-club/llm-compression-papers

papers of llm compression

Size: 103 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 3 - Forks: 0

Tabrizian/learning-to-quantize

Code for "Adaptive Gradient Quantization for Data-Parallel SGD", published in NeurIPS 2020.

Language: Jupyter Notebook - Size: 12.8 MB - Last synced: about 1 month ago - Pushed: over 3 years ago - Stars: 25 - Forks: 5

smpanaro/apple-silicon-4bit-quant

Supporting code for "LLMs for your iPhone: Whole-Tensor 4 Bit Quantization"

Language: Python - Size: 8.79 KB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 0 - Forks: 0

SRDdev/Model-Quantization

Quantization is a technique to reduce the computational and memory costs of running inference by representing the weights and activations with low-precision data types like 8-bit integer (int8) instead of the usual 32-bit floating point (float32).

Language: Jupyter Notebook - Size: 3.16 MB - Last synced: 24 days ago - Pushed: 10 months ago - Stars: 3 - Forks: 0

microsoft/LQ-Nets

LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks

Language: Python - Size: 28.3 KB - Last synced: about 2 months ago - Pushed: almost 2 years ago - Stars: 235 - Forks: 67

Zhen-Dong/Awesome-Quantization-Papers

List of papers related to neural network quantization in recent AI conferences and journals.

Size: 277 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 245 - Forks: 27

skolai/fewbit

Compression schema for gradients of activations in backward pass

Language: Python - Size: 313 KB - Last synced: 3 months ago - Pushed: 10 months ago - Stars: 36 - Forks: 3

dejwi/image-quantization

Image Color quantization with 3D visuals

Language: Rust - Size: 1.31 MB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 1 - Forks: 0

lucidrains/discrete-key-value-bottleneck-pytorch

Implementation of Discrete Key / Value Bottleneck, in Pytorch

Language: Python - Size: 196 KB - Last synced: 28 days ago - Pushed: 11 months ago - Stars: 87 - Forks: 3

dsindex/iclassifier

reference pytorch code for intent classification

Language: Python - Size: 19 MB - Last synced: 2 months ago - Pushed: about 2 years ago - Stars: 44 - Forks: 15

aredden/torch-bnb-fp4

Faster Pytorch bitsandbytes 4bit fp4 nn.Linear ops

Language: Python - Size: 72.3 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 13 - Forks: 0

wenwei202/terngrad

Ternary Gradients to Reduce Communication in Distributed Deep Learning (TensorFlow)

Language: Python - Size: 5.59 MB - Last synced: about 1 month ago - Pushed: over 5 years ago - Stars: 181 - Forks: 48

xvyaward/owq

Code for the AAAI 2024 Oral paper "OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Models".

Language: Python - Size: 3.03 MB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 30 - Forks: 5

flandrade/communication-system-simulation

Communication system in Matlab โ˜Ž๏ธ

Language: Matlab - Size: 364 KB - Last synced: 2 months ago - Pushed: over 8 years ago - Stars: 13 - Forks: 5

xuanandsix/Tensorrt-int8-quantization-pipline

a simple pipline of int8 quantization based on tensorrt.

Language: Python - Size: 836 KB - Last synced: 3 months ago - Pushed: over 1 year ago - Stars: 27 - Forks: 2

Intelligent-Microsystems-Lab/QuantizedLSTM

Models and training scripts for "LSTMs for Keyword Spotting with ReRAM-based Compute-In-Memory Architectures" (ISCAS 2021).

Language: Shell - Size: 14.8 GB - Last synced: 3 months ago - Pushed: about 3 years ago - Stars: 14 - Forks: 4

Maknee/minigpt4.cpp

Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)

Language: C++ - Size: 2.12 MB - Last synced: 3 months ago - Pushed: 10 months ago - Stars: 532 - Forks: 22

umjammer/vavi-image-sandbox

๐Ÿ–ผ๏ธ Imaging sandbox (HEIF Java ImageIO SPI, filters, swing animation component)

Language: Java - Size: 8.52 MB - Last synced: 30 days ago - Pushed: 3 months ago - Stars: 1 - Forks: 0

0xMartin/BMPEditor

A simple BMP image viewer, converter and editor. App is primarily focused on implementation of own code for working with BMP images

Language: C++ - Size: 12.7 MB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 1 - Forks: 0

laelhalawani/gguf_llama

Wrapper for simplified use of Llama2 GGUF quantized models.

Language: Python - Size: 48.8 KB - Last synced: about 1 month ago - Pushed: 5 months ago - Stars: 2 - Forks: 1

dvmazur/mixtral-offloading

Run Mixtral-8x7B models in Colab or consumer desktops

Language: Python - Size: 261 KB - Last synced: 3 months ago - Pushed: 4 months ago - Stars: 2,093 - Forks: 195

raywan-110/AdaQP

Adaptive Message Quantization and Parallelization for Distributed Full-graph GNN Training

Language: Python - Size: 87.9 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 15 - Forks: 0

sachink1729/Finetuning-Mistral-7B-Chat-Doctor-Huggingface-LoRA-PEFT

Finetuning Mistral-7B into a Medical Chat Doctor using Huggingface ๐Ÿค—+ QLoRA + PEFT.

Language: Jupyter Notebook - Size: 10.2 MB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 2 - Forks: 1

Neural-Dreamers/Forest-Sound-Analysis-on-Edge

A Comparative Analysis of Sound Data Pre-processing and Deep Learning Model Compression Techniques: A Study on Forest Sound Classification

Language: Jupyter Notebook - Size: 49.8 KB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 0 - Forks: 0

JulesBelveze/bert-squeeze

๐Ÿ› ๏ธ Tools for Transformers compression using PyTorch Lightning โšก

Language: Python - Size: 2.41 MB - Last synced: 18 days ago - Pushed: 3 months ago - Stars: 78 - Forks: 10

maherabdelsamie/Active-Time-Theory-Atomic-Clocks

The Secret Inner Workings of Time Exposed by Atomic Clocks

Language: Python - Size: 21.5 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 1 - Forks: 0

inuwamobarak/KOSMOS-2

KOSMOS-2 is designed to handle text and images simultaneously, and redefine the way we perceive and interact with multimodal data, KOSMOS-2 is built on a Transformer-based causal language model architecture, similar to other renowned models like LLaMa-2 and Mistral AI's 7b model.

Language: Jupyter Notebook - Size: 6.97 MB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 1 - Forks: 0

insuofficial/pytorch-quantization

Quantization simulation of neural networks with PyTorch

Language: Python - Size: 33.2 KB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 0 - Forks: 0

maherabdelsamie/Active-Time-Theory-Time-Dilation

Re-interpreting Time Dilation Through the Lens of Active Time Theory

Language: Python - Size: 33.2 KB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 1 - Forks: 0

JosefAlbers/Roy

Roy: A lightweight, model-agnostic framework for crafting advanced multi-agent systems using large language models.

Language: Jupyter Notebook - Size: 321 KB - Last synced: 4 months ago - Pushed: 7 months ago - Stars: 58 - Forks: 2

maherabdelsamie/Active-Time-Theory-Quantum-Tunneling

Simulation of Quantum Tunneling Dynamics Validates Signatures of Active Time

Language: Python - Size: 76.2 KB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 1 - Forks: 0

koulanurag/mmn

Moore Machine Networks (MMN): Learning Finite-State Representations of Recurrent Policy Networks

Language: Python - Size: 115 MB - Last synced: 2 months ago - Pushed: over 1 year ago - Stars: 46 - Forks: 12

microsoft/Lightweight-Low-Resource-NMT

Official code for "Too Brittle To Touch: Comparing the Stability of Quantization and Distillation Towards Developing Lightweight Low-Resource MT Models" to appear in WMT 2022.

Language: Python - Size: 45.9 KB - Last synced: about 2 months ago - Pushed: 8 months ago - Stars: 16 - Forks: 3

ssi-research/eptq

Implementation of EPTQ - an Enhanced Post-Training Quantization algorithm for DNN compression

Language: Python - Size: 99.6 KB - Last synced: 18 days ago - Pushed: 8 months ago - Stars: 4 - Forks: 0

smpanaro/norm-tweaking

Post post-training-quantization (PTQ) method for improving LLMs. Unofficial implementation of https://arxiv.org/abs/2309.02784

Language: Python - Size: 32.2 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 0 - Forks: 0

mrbestnaija/MLOps

My Machine Learning Operations (MLOps) Reference repository

Language: Jupyter Notebook - Size: 236 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 1 - Forks: 0

Kovelja009/mixed_precision_dnn_pytorch

Pytorch implementation of the paper Mixed Precision DNNs: All you need is a good parametrization.

Language: Python - Size: 4.69 MB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 0 - Forks: 0

maherabdelsamie/Active-Time-Theory

The Foundations of Active Time Theory

Language: Python - Size: 125 KB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 1 - Forks: 0

pnucse-capstone/capstone-2023-1-11

์‹ฌ์žฅ์งˆํ™˜ ํ™˜์ž ECG ๋ฐ์ดํ„ฐ ๋ถ„์„์„ ์œ„ํ•œ ๋”ฅ๋Ÿฌ๋‹ ๊ธฐ๋ฒ• ์„ค๊ณ„ ๋ฐ ๊ฒฝ๋Ÿ‰ํ™” ๋ชจ๋ธ ๊ตฌ์ถ•

Language: Jupyter Notebook - Size: 8.22 MB - Last synced: 21 days ago - Pushed: 7 months ago - Stars: 1 - Forks: 0

j-marple-dev/model_compression

PyTorch Model Compression

Language: Python - Size: 31 MB - Last synced: 3 months ago - Pushed: over 1 year ago - Stars: 227 - Forks: 24

stdlib-js/ml-incr-kmeans

Incrementally partition data into `k` clusters.

Language: JavaScript - Size: 4.61 MB - Last synced: 29 days ago - Pushed: about 2 months ago - Stars: 6 - Forks: 0

esimov/colorquant

Go library for color quantization and dithering

Language: Go - Size: 2.26 MB - Last synced: 30 days ago - Pushed: over 3 years ago - Stars: 84 - Forks: 4

tpoisonooo/llama.onnx

LLaMa/RWKV onnx models, quantization and testcase

Language: Python - Size: 1.3 MB - Last synced: 4 months ago - Pushed: 11 months ago - Stars: 303 - Forks: 26

Cheshulko/Quantize-image-rs

Quantize an image using k-means clustering

Language: Rust - Size: 3.18 MB - Last synced: 4 months ago - Pushed: 5 months ago - Stars: 0 - Forks: 0

ZLkanyo009/Yolo-compression-and-deployment-in-FPGA

ๅŸบไบŽFPGA้‡ๅŒ–็š„ไบบ่„ธๅฃ็ฝฉๆฃ€ๆต‹

Language: C - Size: 290 MB - Last synced: 3 months ago - Pushed: almost 3 years ago - Stars: 16 - Forks: 6