An open API service providing repository metadata for many open source software ecosystems.

GitHub topics: bitsandbytes

stackav-oss/conch

A "standard library" of Triton kernels.

Language: Python - Size: 438 KB - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 13 - Forks: 1

ParagEkbote/quantized-containerized-models

A project that demonstrates how to deploy AI models within containerized environments using Cog. Ideal for reproducible, scalable and hardware-efficient inference.

Language: Python - Size: 6.47 MB - Last synced at: 8 days ago - Pushed at: 8 days ago - Stars: 0 - Forks: 0

bobazooba/xllm

๐Ÿฆ– Xโ€”LLM: Cutting Edge & Easy LLM Finetuning

Language: Python - Size: 1.81 MB - Last synced at: 3 days ago - Pushed at: over 1 year ago - Stars: 406 - Forks: 21

AtharvaTaras/BankLlama

Code and files used during research for 'Fine Tuned Large Language Model In Banking'

Language: Jupyter Notebook - Size: 8.13 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

gaurisharan/phi2-alpaca-lora-4bit

Fine-tuning Microsoftโ€™s Phi-2 model using QLoRA on the Alpaca-cleaned dataset with 4-bit quantization (bitsandbytes).

Language: Jupyter Notebook - Size: 779 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

edcalderin/huggingface-ragflow

This project implements a classic Retrieval-Augmented Generation (RAG) system using HuggingFace models with quantization techniques. The system processes PDF documents, extracts their content, and enables interactive question-answering through a Streamlit web application.

Language: Python - Size: 120 KB - Last synced at: about 1 month ago - Pushed at: about 2 months ago - Stars: 1 - Forks: 0

NightKing-V/SubtitleLLM_EngtoSin

Experimental Eng->Sin Subtitle Translation Model

Language: Jupyter Notebook - Size: 851 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

lpalbou/model-quantizer

Effortlessly quantize, benchmark, and publish Hugging Face models with cross-platform support for CPU/GPU. Reduce model size by 75% while maintaining performance.

Language: Python - Size: 165 KB - Last synced at: 5 days ago - Pushed at: 6 months ago - Stars: 2 - Forks: 0

Md-Emon-Hasan/Fine-Tuning

End-to-end fine-tuning of Hugging Face models using LoRA, QLoRA, quantization, and PEFT techniques. Optimized for low-memory with efficient model deployment

Language: Jupyter Notebook - Size: 5.53 MB - Last synced at: about 1 month ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

arjunravi26/ai_psychologist

Fine tune llama 2 model with mental health counseling data from huggingface

Language: Jupyter Notebook - Size: 14.5 MB - Last synced at: about 1 month ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

SCCSMARTCODE/Deep-Learning-03-LLM-FineTuning

Scalable and modular framework for fine-tuning large language models (LLMs) using LoRA and QLoRA. Supports 4-bit/8-bit quantization, Hugging Face Transformers, and instruction-tuning workflows across various tasks and datasets. Built for reproducibility, extensibility, and efficient experimentation.

Language: Jupyter Notebook - Size: 66.4 KB - Last synced at: 2 months ago - Pushed at: 4 months ago - Stars: 0 - Forks: 0

AkimfromParis/RAG-Japanese

Open source RAG with Llama Index for Japanese LLM in low resource settting

Language: Jupyter Notebook - Size: 2.54 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 8 - Forks: 3

GodreignElgin/llm-comparision

Jupyter Notebook for LLM compression via quantization (INT8, INT4, FP16) and evaluation using metrics such as ROUGE and BLEU. Facilitates efficient LLM optimization.

Language: Jupyter Notebook - Size: 1.82 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 0 - Forks: 1

ArchitJ6/Llama2-FineTuning

๐Ÿฆ™ Llama2-FineTuning: Fine-tune LLAMA 2 with Custom Datasets Using LoRA and QLoRA Techniques

Language: Jupyter Notebook - Size: 2.83 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 0 - Forks: 0

shaheennabi/Production-Ready-Instruction-Finetuning-of-Meta-Llama-3.2-3B-Instruct-Project

๐ŸŽ‹๐ŸŒฟ๐ŸŒŸ Instruction Fine-Tuning of Meta Llama 3.2-3B Instruct on Kannada Conversations ๐ŸŒŸ๐ŸŒฟ๐ŸŽ‹ Tailoring the model to follow specific instructions in Kannada, enhancing its ability to generate relevant, context-aware responses based on conversational inputs. ๐Ÿš€โœจ Using the Kannada Instruct dataset for fine-tuning! Happy Finetuning๐ŸŽ‡๐ŸŽ‰

Language: Jupyter Notebook - Size: 692 KB - Last synced at: 5 months ago - Pushed at: 7 months ago - Stars: 18 - Forks: 2

dvgodoy/FineTuningLLMs

Official repository of my book "A Hands-On Guide to Fine-Tuning LLMs with PyTorch and Hugging Face"

Language: Jupyter Notebook - Size: 7.48 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 31 - Forks: 7

Bushra-Butt-17/BudgetBuddy-Finance-Chatbot

Budget Buddy is a finance chatbot built using Chainlit and the LLaMA language model. It analyzes PDF documents, such as bank statements and budget reports, to provide personalized financial advice and insights. The chatbot is integrated with Hugging Face for model management, offering an interactive way to manage personal finances.

Language: Python - Size: 19.3 MB - Last synced at: 5 months ago - Pushed at: 9 months ago - Stars: 2 - Forks: 0

403errors/CancerCareAI

An AI-powered system for extracting cancer-related information from patient Electronic Health Record (EHR) notes

Language: Jupyter Notebook - Size: 2.77 MB - Last synced at: 22 days ago - Pushed at: 6 months ago - Stars: 0 - Forks: 0

arham-kk/llama2-qlora-sft

This model is a fine-tuned model based on the "TinyPixel/Llama-2-7B-bf16-sharded" model and "timdettmers/openassistant-guanaco" dataset

Size: 640 KB - Last synced at: 2 days ago - Pushed at: almost 2 years ago - Stars: 3 - Forks: 0

kedir/Specialized-Immigration-Assistant

provide specialized immigration assistance in the field of immigration law using large language model

Language: Jupyter Notebook - Size: 64.5 KB - Last synced at: 2 days ago - Pushed at: almost 2 years ago - Stars: 2 - Forks: 0

Varun0157/quantisation

Experiments in quantisation consisting of quantisation from scratch, bitsandbytes, and llama.cpp. [Assignment 4 of Advanced Natural Language Processing, IIIT-H Monsoon '24]

Language: Python - Size: 395 KB - Last synced at: 25 days ago - Pushed at: 8 months ago - Stars: 0 - Forks: 0

eljandoubi/AI-Photo-Editing-with-Inpainting

A web app that allows you to select a subject and then change its background, OR keep the background and change the subject.

Language: Jupyter Notebook - Size: 10.9 MB - Last synced at: 4 months ago - Pushed at: over 1 year ago - Stars: 8 - Forks: 1

dasdristanta13/LLM-Lora-PEFT_accumulate

LLM-Lora-PEFT_accumulate explores optimizations for Large Language Models (LLMs) using PEFT, LORA, and QLORA. Contribute experiments and implementations to enhance LLM efficiency. Join discussions and push the boundaries of LLM optimization. Let's make LLMs more efficient together!

Language: Jupyter Notebook - Size: 138 KB - Last synced at: 2 days ago - Pushed at: about 2 years ago - Stars: 6 - Forks: 1

antonio-f/Orca2

Orca 2 on Colab

Language: Jupyter Notebook - Size: 137 KB - Last synced at: 5 months ago - Pushed at: over 1 year ago - Stars: 8 - Forks: 1

to-aoki/bitsandbytes

bitsandbytes modify for jetson orin

Language: Python - Size: 1.09 MB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0

bobazooba/shurale

Conversation AI model for open domain dialogs

Language: Python - Size: 64.5 KB - Last synced at: 2 days ago - Pushed at: almost 2 years ago - Stars: 4 - Forks: 1

ryan-air/Alpaca-350M-Fine-Tuned

In this project, I have provided code and a Colaboratory notebook that facilitates the fine-tuning process of an Alpaca 350M parameter model originally developed at Stanford University. The model was adapted using LoRA to run with fewer computational resources and training parameters and used HuggingFace's PEFT library.

Language: Jupyter Notebook - Size: 70.3 KB - Last synced at: about 2 years ago - Pushed at: about 2 years ago - Stars: 2 - Forks: 1

ryan-air/Alpaca-3B-Fine-Tuned

In this project, I have provided code and a Colaboratory notebook that facilitates the fine-tuning process of an Alpaca 3B parameter model originally developed at Stanford University. The model was adapted using LoRA to run with fewer computational resources and training parameters and used HuggingFace's PEFT library.

Language: Jupyter Notebook - Size: 48.8 KB - Last synced at: about 2 years ago - Pushed at: about 2 years ago - Stars: 2 - Forks: 0