Ecosyste.ms: Repos
An open API service providing repository metadata for many open source software ecosystems.
GitHub topics: quantization
SqueezeAILab/KVQuant
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
Language: Python - Size: 19.7 MB - Last synced: about 1 month ago - Pushed: about 1 month ago - Stars: 187 - Forks: 14
Asad-Ismail/lane_detection
Lane Detection and Classification using Front camera monocular images
Language: Python - Size: 40.9 MB - Last synced: about 1 month ago - Pushed: about 1 year ago - Stars: 3 - Forks: 0
IST-DASLab/marlin
FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.
Language: Python - Size: 758 KB - Last synced: about 1 month ago - Pushed: about 1 month ago - Stars: 314 - Forks: 21
hailo-ai/hailo_model_zoo
The Hailo Model Zoo includes pre-trained models and a full building and evaluation environment
Language: Python - Size: 4.32 MB - Last synced: about 1 month ago - Pushed: about 2 months ago - Stars: 91 - Forks: 28
PaddlePaddle/PaddleSlim
PaddleSlim is an open-source library for deep model compression and architecture search.
Language: Python - Size: 16.3 MB - Last synced: about 1 month ago - Pushed: 2 months ago - Stars: 1,514 - Forks: 347
jy-yuan/KIVI
KIVI: A Tuning-Free Asymmetric 2bit Quantization for KV Cache
Language: Python - Size: 16.8 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 107 - Forks: 5
lmEshoo/quantization
Post-training model quantization using apache's tvm.ai
Language: Jupyter Notebook - Size: 37.9 MB - Last synced: about 2 months ago - Pushed: about 4 years ago - Stars: 1 - Forks: 1
lmEshoo/pruning
model weight pruning
Language: Jupyter Notebook - Size: 72.4 MB - Last synced: about 2 months ago - Pushed: about 4 years ago - Stars: 0 - Forks: 0
intel/auto-round
SOTA Weight-only Quantization Algorithm for LLMs
Language: Python - Size: 8.33 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 48 - Forks: 7
emsquared2/Telecommunications-NTUA
Project assignment for course Introduction to Telecommunications at ECE NTUA
Language: MATLAB - Size: 4.14 MB - Last synced: about 2 months ago - Pushed: about 1 year ago - Stars: 0 - Forks: 0
huawei-noah/Pretrained-Language-Model
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
Language: Python - Size: 29 MB - Last synced: about 2 months ago - Pushed: 4 months ago - Stars: 2,953 - Forks: 623
iAmGiG/MadeSmallML
MadeSmallML is an open-source initiative designed to explore model quantization techniques for machine learning models. Our goal is to enable efficient deployment of these models on devices with limited computational resources by reducing model size and computational demands without significantly compromising performance.
Size: 5.86 KB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 0 - Forks: 0
intel/intel-extension-for-pytorch
A Python package for extending the official PyTorch that can easily obtain performance on Intel platform
Language: Python - Size: 92.1 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 1,316 - Forks: 193
neuralmagic/sparsify
ML model optimization product to accelerate inference.
Language: Python - Size: 7.18 MB - Last synced: 14 days ago - Pushed: about 2 months ago - Stars: 315 - Forks: 27
kurianbenoy/Indic-Subtitler
Open source subtitling platform ๐ป for transcribing and translating videos/audios in Indic languages.
Language: Jupyter Notebook - Size: 36.3 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 55 - Forks: 7
OpenGVLab/OmniQuant
[ICLR2024 spotlight] OmniQuant is a simple and powerful quantization technique for LLMs.
Language: Python - Size: 8.17 MB - Last synced: about 2 months ago - Pushed: 2 months ago - Stars: 546 - Forks: 43
huggingface/optimum
๐ Accelerate training and inference of ๐ค Transformers and ๐ค Diffusers with easy to use hardware optimization tools
Language: Python - Size: 4.01 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 2,096 - Forks: 356
ModelTC/llmc
llmc is an efficient LLM compression tool with various advanced compression methods, supporting multiple inference backends.
Language: Python - Size: 3.35 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 25 - Forks: 2
archinetai/bitcodes-pytorch
A vector quantization method with binary codes, in PyTorch.
Language: Python - Size: 5.86 KB - Last synced: 9 days ago - Pushed: over 1 year ago - Stars: 6 - Forks: 0
LowinLi/stable-diffusion-streamlit
Quantized stable-diffusion cutting down memory 75%, testing in streamlit, deploying in container
Language: Python - Size: 53.6 MB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 53 - Forks: 7
minseok0809/transformers-compression-practice
Transformers Compression Practice
Language: Jupyter Notebook - Size: 19.9 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 0 - Forks: 0
Spinkoo/Embedded-Yolov7
Compress, Deploy and Inference YOLOv7 on low-cost MCUs
Language: Jupyter Notebook - Size: 53.3 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 2 - Forks: 0
Azazel0203/Sentiment_Analyser
This project employs transformer models for sentiment analysis on primate-themed posts, encompassing data preprocessing, model training, evaluation, and optional quantization. It provides a streamlined workflow for analyzing sentiment in primate-related textual data, offering insights into public perceptions of primates.
Language: Jupyter Notebook - Size: 25.4 KB - Last synced: 23 days ago - Pushed: about 2 months ago - Stars: 2 - Forks: 0
jiho264/2024-RISE-Quantization
Try to get some insights about Quantization
Language: Jupyter Notebook - Size: 6.77 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 0 - Forks: 0
minseok0809/awesome-ai-paper
A curated list of awesome NLP, Computer Vision, Model Compression, XAI, Reinforcement Learning, Security etc Paper
Language: Jupyter Notebook - Size: 2.81 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 0 - Forks: 0
clovaai/frostnet
FrostNet: Towards Quantization-Aware Network Architecture Search
Language: Python - Size: 505 KB - Last synced: 18 days ago - Pushed: 28 days ago - Stars: 106 - Forks: 18
monk1337/NanoPeft
The simplest repository & Neat implementation of different Lora methods for training/fine-tuning Transformer-based models (i.e., BERT, GPTs). [ Research purpose ]
Language: Jupyter Notebook - Size: 68.4 KB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 5 - Forks: 2
fdbtrs/QuantFace
QuantFace: Towards Lightweight Face Recognition by Synthetic Data Low-bit Quantization
Language: Python - Size: 5.47 MB - Last synced: 18 days ago - Pushed: almost 2 years ago - Stars: 30 - Forks: 3
RahulSChand/gpu_poor
Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization
Language: JavaScript - Size: 4.85 MB - Last synced: about 2 months ago - Pushed: 7 months ago - Stars: 595 - Forks: 22
1duo/awesome-ai-infrastructures
Infrastructuresโข for Machine Learning Training/Inference in Production.
Size: 11.8 MB - Last synced: 12 days ago - Pushed: about 5 years ago - Stars: 368 - Forks: 71
elphinkuo/llamaqt.c
Clean C language version of quantizing llama2 model and running quantized llama2 model
Language: C - Size: 455 KB - Last synced: about 2 months ago - Pushed: 9 months ago - Stars: 2 - Forks: 0
aaron-xichen/pytorch-playground
Base pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)
Language: Python - Size: 45.9 KB - Last synced: 2 months ago - Pushed: over 1 year ago - Stars: 2,543 - Forks: 616
mit-han-lab/tinyengine
[NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning; [NeurIPS 2022] MCUNetV3: On-Device Training Under 256KB Memory
Language: C - Size: 235 MB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 717 - Forks: 123
stracini-git/bitwise-weight-training
Training neural network weights in a bitwise fashion: https://arxiv.org/abs/2202.09571
Language: Python - Size: 1.8 MB - Last synced: 2 months ago - Pushed: almost 2 years ago - Stars: 2 - Forks: 0
stracini-git/qnn
Training neural nets with quantized weights on arbitrarily specified bit-depth
Language: Python - Size: 1.2 MB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 1 - Forks: 1
daehwa00/VoiceKey
Published "VoiceKey: Real-time Compression Encoding and Quantization for Voice Authentication Model on Edge Devices" at The 4th Korea Artificial Intelligence Conference.
Language: Python - Size: 2.88 MB - Last synced: 19 days ago - Pushed: 3 months ago - Stars: 1 - Forks: 1
megvii-research/Sparsebit
A model compression and acceleration toolbox based on pytorch.
Language: Python - Size: 7.45 MB - Last synced: about 2 months ago - Pushed: 5 months ago - Stars: 317 - Forks: 39
mit-han-lab/haq
[CVPR 2019, Oral] HAQ: Hardware-Aware Automated Quantization with Mixed Precision
Language: Python - Size: 64.5 KB - Last synced: 2 months ago - Pushed: over 3 years ago - Stars: 350 - Forks: 83
Cydia2018/YOLOv5-Light
provide some new architecture, channel pruning and quantization methods for yolov5
Language: Python - Size: 12.5 MB - Last synced: about 1 month ago - Pushed: 4 months ago - Stars: 29 - Forks: 5
shivpalSW/Optimized-CPU-Implementation-of-Llama2
Optimized CPU Implementation of Llama2-LLM
Language: Python - Size: 703 KB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 0 - Forks: 0
FLHonker/ZAQ-code
CVPR 2021 : Zero-shot Adversarial Quantization (ZAQ)
Language: Python - Size: 188 KB - Last synced: about 2 months ago - Pushed: over 2 years ago - Stars: 64 - Forks: 16
Xiuyu-Li/q-diffusion
[ICCV 2023] Q-Diffusion: Quantizing Diffusion Models.
Language: Python - Size: 5.97 MB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 243 - Forks: 17
briancpark/csc791-025
Computer Science 791-025: Real-Time AI & High-Performance Machine Learning
Language: TeX - Size: 77 MB - Last synced: about 1 month ago - Pushed: 4 months ago - Stars: 1 - Forks: 0
alizameller/Digitial-Signal-Processing
ECE-310 Digital Signal Processing - Cooper Union Fall 2022
Language: MATLAB - Size: 7.81 KB - Last synced: 3 months ago - Pushed: over 1 year ago - Stars: 0 - Forks: 0
GATECH-EIC/Double-Win-Quant
[ICML 2021] "Double-Win Quant: Aggressively Winning Robustness of Quantized DeepNeural Networks via Random Precision Training and Inference" by Yonggan Fu, Qixuan Yu, Meng Li, Vikas Chandra, Yingyan Lin
Language: Python - Size: 947 KB - Last synced: about 1 month ago - Pushed: over 2 years ago - Stars: 12 - Forks: 5
motokimura/pytorch_quantization_fx
An example to quantize MobileNetV2 trained on CIFAR-10 dataset with PyTorch
Language: Python - Size: 86.6 MB - Last synced: 2 months ago - Pushed: over 1 year ago - Stars: 6 - Forks: 3
linkedin/QuantEase
QuantEase, a layer-wise quantization framework, frames the problem as discrete-structured non-convex optimization. Our work leverages Coordinate Descent techniques, offering high-quality solutions without the need for matrix inversion or decomposition.
Language: Python - Size: 209 KB - Last synced: about 2 months ago - Pushed: 3 months ago - Stars: 15 - Forks: 1
laelhalawani/glai
glai - GGUF LLAMA AI - Package for simplified model handling and text generation with Llama models quantized to GGUF format. APIs for downloading and loading models automatically, includes a db with models of various scale and quantizations. With this high level API you need one line to load the model and one to generate text completions.
Language: Python - Size: 208 KB - Last synced: 3 months ago - Pushed: 5 months ago - Stars: 3 - Forks: 0
daemyung/nzip
The deep learning quantization library for PyTorch.
Language: Python - Size: 16.6 KB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 1 - Forks: 0
dsindex/ntagger
reference pytorch code for named entity tagging
Language: Python - Size: 36 MB - Last synced: 2 months ago - Pushed: almost 2 years ago - Stars: 85 - Forks: 13
AnonymousWorld123/Q-Layer
Code for Q-Layer: Quantization Layer for Robust Convolutional Neural Network
Language: Python - Size: 103 KB - Last synced: about 1 month ago - Pushed: over 3 years ago - Stars: 3 - Forks: 0
megvii-research/SSQL-ECCV2022
PyTorch implementation of SSQL (Accepted to ECCV2022 oral presentation)
Language: Python - Size: 211 KB - Last synced: 24 days ago - Pushed: about 1 year ago - Stars: 75 - Forks: 6
megvii-research/FQ-ViT
[IJCAI 2022] FQ-ViT: Post-Training Quantization for Fully Quantized Vision Transformer
Language: Python - Size: 729 KB - Last synced: 3 months ago - Pushed: about 1 year ago - Stars: 259 - Forks: 45
ModelTC/QLLM
[ICLR 2024] This is the official PyTorch implementation of "QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Models"
Language: Python - Size: 1.68 MB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 15 - Forks: 0
jiho264/2024-RISE
Try to gain some insights from Quantization
Language: Jupyter Notebook - Size: 3.69 MB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 0 - Forks: 0
bupt-ai-club/llm-compression-papers
papers of llm compression
Size: 103 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 3 - Forks: 0
Tabrizian/learning-to-quantize
Code for "Adaptive Gradient Quantization for Data-Parallel SGD", published in NeurIPS 2020.
Language: Jupyter Notebook - Size: 12.8 MB - Last synced: about 1 month ago - Pushed: over 3 years ago - Stars: 25 - Forks: 5
smpanaro/apple-silicon-4bit-quant
Supporting code for "LLMs for your iPhone: Whole-Tensor 4 Bit Quantization"
Language: Python - Size: 8.79 KB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 0 - Forks: 0
SRDdev/Model-Quantization
Quantization is a technique to reduce the computational and memory costs of running inference by representing the weights and activations with low-precision data types like 8-bit integer (int8) instead of the usual 32-bit floating point (float32).
Language: Jupyter Notebook - Size: 3.16 MB - Last synced: 24 days ago - Pushed: 10 months ago - Stars: 3 - Forks: 0
microsoft/LQ-Nets
LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks
Language: Python - Size: 28.3 KB - Last synced: about 2 months ago - Pushed: almost 2 years ago - Stars: 235 - Forks: 67
Zhen-Dong/Awesome-Quantization-Papers
List of papers related to neural network quantization in recent AI conferences and journals.
Size: 277 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 245 - Forks: 27
skolai/fewbit
Compression schema for gradients of activations in backward pass
Language: Python - Size: 313 KB - Last synced: 3 months ago - Pushed: 10 months ago - Stars: 36 - Forks: 3
dejwi/image-quantization
Image Color quantization with 3D visuals
Language: Rust - Size: 1.31 MB - Last synced: 2 months ago - Pushed: 2 months ago - Stars: 1 - Forks: 0
lucidrains/discrete-key-value-bottleneck-pytorch
Implementation of Discrete Key / Value Bottleneck, in Pytorch
Language: Python - Size: 196 KB - Last synced: 28 days ago - Pushed: 11 months ago - Stars: 87 - Forks: 3
dsindex/iclassifier
reference pytorch code for intent classification
Language: Python - Size: 19 MB - Last synced: 2 months ago - Pushed: about 2 years ago - Stars: 44 - Forks: 15
aredden/torch-bnb-fp4
Faster Pytorch bitsandbytes 4bit fp4 nn.Linear ops
Language: Python - Size: 72.3 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 13 - Forks: 0
wenwei202/terngrad
Ternary Gradients to Reduce Communication in Distributed Deep Learning (TensorFlow)
Language: Python - Size: 5.59 MB - Last synced: about 1 month ago - Pushed: over 5 years ago - Stars: 181 - Forks: 48
xvyaward/owq
Code for the AAAI 2024 Oral paper "OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Models".
Language: Python - Size: 3.03 MB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 30 - Forks: 5
flandrade/communication-system-simulation
Communication system in Matlab โ๏ธ
Language: Matlab - Size: 364 KB - Last synced: 2 months ago - Pushed: over 8 years ago - Stars: 13 - Forks: 5
xuanandsix/Tensorrt-int8-quantization-pipline
a simple pipline of int8 quantization based on tensorrt.
Language: Python - Size: 836 KB - Last synced: 3 months ago - Pushed: over 1 year ago - Stars: 27 - Forks: 2
Intelligent-Microsystems-Lab/QuantizedLSTM
Models and training scripts for "LSTMs for Keyword Spotting with ReRAM-based Compute-In-Memory Architectures" (ISCAS 2021).
Language: Shell - Size: 14.8 GB - Last synced: 3 months ago - Pushed: about 3 years ago - Stars: 14 - Forks: 4
Maknee/minigpt4.cpp
Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)
Language: C++ - Size: 2.12 MB - Last synced: 3 months ago - Pushed: 10 months ago - Stars: 532 - Forks: 22
umjammer/vavi-image-sandbox
๐ผ๏ธ Imaging sandbox (HEIF Java ImageIO SPI, filters, swing animation component)
Language: Java - Size: 8.52 MB - Last synced: 30 days ago - Pushed: 3 months ago - Stars: 1 - Forks: 0
0xMartin/BMPEditor
A simple BMP image viewer, converter and editor. App is primarily focused on implementation of own code for working with BMP images
Language: C++ - Size: 12.7 MB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 1 - Forks: 0
laelhalawani/gguf_llama
Wrapper for simplified use of Llama2 GGUF quantized models.
Language: Python - Size: 48.8 KB - Last synced: about 1 month ago - Pushed: 5 months ago - Stars: 2 - Forks: 1
dvmazur/mixtral-offloading
Run Mixtral-8x7B models in Colab or consumer desktops
Language: Python - Size: 261 KB - Last synced: 3 months ago - Pushed: 4 months ago - Stars: 2,093 - Forks: 195
raywan-110/AdaQP
Adaptive Message Quantization and Parallelization for Distributed Full-graph GNN Training
Language: Python - Size: 87.9 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 15 - Forks: 0
sachink1729/Finetuning-Mistral-7B-Chat-Doctor-Huggingface-LoRA-PEFT
Finetuning Mistral-7B into a Medical Chat Doctor using Huggingface ๐ค+ QLoRA + PEFT.
Language: Jupyter Notebook - Size: 10.2 MB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 2 - Forks: 1
Neural-Dreamers/Forest-Sound-Analysis-on-Edge
A Comparative Analysis of Sound Data Pre-processing and Deep Learning Model Compression Techniques: A Study on Forest Sound Classification
Language: Jupyter Notebook - Size: 49.8 KB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 0 - Forks: 0
JulesBelveze/bert-squeeze
๐ ๏ธ Tools for Transformers compression using PyTorch Lightning โก
Language: Python - Size: 2.41 MB - Last synced: 18 days ago - Pushed: 3 months ago - Stars: 78 - Forks: 10
maherabdelsamie/Active-Time-Theory-Atomic-Clocks
The Secret Inner Workings of Time Exposed by Atomic Clocks
Language: Python - Size: 21.5 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 1 - Forks: 0
inuwamobarak/KOSMOS-2
KOSMOS-2 is designed to handle text and images simultaneously, and redefine the way we perceive and interact with multimodal data, KOSMOS-2 is built on a Transformer-based causal language model architecture, similar to other renowned models like LLaMa-2 and Mistral AI's 7b model.
Language: Jupyter Notebook - Size: 6.97 MB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 1 - Forks: 0
insuofficial/pytorch-quantization
Quantization simulation of neural networks with PyTorch
Language: Python - Size: 33.2 KB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 0 - Forks: 0
maherabdelsamie/Active-Time-Theory-Time-Dilation
Re-interpreting Time Dilation Through the Lens of Active Time Theory
Language: Python - Size: 33.2 KB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 1 - Forks: 0
JosefAlbers/Roy
Roy: A lightweight, model-agnostic framework for crafting advanced multi-agent systems using large language models.
Language: Jupyter Notebook - Size: 321 KB - Last synced: 4 months ago - Pushed: 7 months ago - Stars: 58 - Forks: 2
maherabdelsamie/Active-Time-Theory-Quantum-Tunneling
Simulation of Quantum Tunneling Dynamics Validates Signatures of Active Time
Language: Python - Size: 76.2 KB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 1 - Forks: 0
koulanurag/mmn
Moore Machine Networks (MMN): Learning Finite-State Representations of Recurrent Policy Networks
Language: Python - Size: 115 MB - Last synced: 2 months ago - Pushed: over 1 year ago - Stars: 46 - Forks: 12
microsoft/Lightweight-Low-Resource-NMT
Official code for "Too Brittle To Touch: Comparing the Stability of Quantization and Distillation Towards Developing Lightweight Low-Resource MT Models" to appear in WMT 2022.
Language: Python - Size: 45.9 KB - Last synced: about 2 months ago - Pushed: 8 months ago - Stars: 16 - Forks: 3
ssi-research/eptq
Implementation of EPTQ - an Enhanced Post-Training Quantization algorithm for DNN compression
Language: Python - Size: 99.6 KB - Last synced: 18 days ago - Pushed: 8 months ago - Stars: 4 - Forks: 0
smpanaro/norm-tweaking
Post post-training-quantization (PTQ) method for improving LLMs. Unofficial implementation of https://arxiv.org/abs/2309.02784
Language: Python - Size: 32.2 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 0 - Forks: 0
mrbestnaija/MLOps
My Machine Learning Operations (MLOps) Reference repository
Language: Jupyter Notebook - Size: 236 KB - Last synced: 3 months ago - Pushed: 3 months ago - Stars: 1 - Forks: 0
Kovelja009/mixed_precision_dnn_pytorch
Pytorch implementation of the paper Mixed Precision DNNs: All you need is a good parametrization.
Language: Python - Size: 4.69 MB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 0 - Forks: 0
maherabdelsamie/Active-Time-Theory
The Foundations of Active Time Theory
Language: Python - Size: 125 KB - Last synced: 4 months ago - Pushed: 4 months ago - Stars: 1 - Forks: 0
pnucse-capstone/capstone-2023-1-11
์ฌ์ฅ์งํ ํ์ ECG ๋ฐ์ดํฐ ๋ถ์์ ์ํ ๋ฅ๋ฌ๋ ๊ธฐ๋ฒ ์ค๊ณ ๋ฐ ๊ฒฝ๋ํ ๋ชจ๋ธ ๊ตฌ์ถ
Language: Jupyter Notebook - Size: 8.22 MB - Last synced: 21 days ago - Pushed: 7 months ago - Stars: 1 - Forks: 0
j-marple-dev/model_compression
PyTorch Model Compression
Language: Python - Size: 31 MB - Last synced: 3 months ago - Pushed: over 1 year ago - Stars: 227 - Forks: 24
stdlib-js/ml-incr-kmeans
Incrementally partition data into `k` clusters.
Language: JavaScript - Size: 4.61 MB - Last synced: 29 days ago - Pushed: about 2 months ago - Stars: 6 - Forks: 0
esimov/colorquant
Go library for color quantization and dithering
Language: Go - Size: 2.26 MB - Last synced: 30 days ago - Pushed: over 3 years ago - Stars: 84 - Forks: 4
tpoisonooo/llama.onnx
LLaMa/RWKV onnx models, quantization and testcase
Language: Python - Size: 1.3 MB - Last synced: 4 months ago - Pushed: 11 months ago - Stars: 303 - Forks: 26
Cheshulko/Quantize-image-rs
Quantize an image using k-means clustering
Language: Rust - Size: 3.18 MB - Last synced: 4 months ago - Pushed: 5 months ago - Stars: 0 - Forks: 0
ZLkanyo009/Yolo-compression-and-deployment-in-FPGA
ๅบไบFPGA้ๅ็ไบบ่ธๅฃ็ฝฉๆฃๆต
Language: C - Size: 290 MB - Last synced: 3 months ago - Pushed: almost 3 years ago - Stars: 16 - Forks: 6