GitHub topics: bitsandbytes
stackav-oss/conch
A "standard library" of Triton kernels.
Language: Python - Size: 438 KB - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 13 - Forks: 1

ParagEkbote/quantized-containerized-models
A project that demonstrates how to deploy AI models within containerized environments using Cog. Ideal for reproducible, scalable and hardware-efficient inference.
Language: Python - Size: 6.47 MB - Last synced at: 8 days ago - Pushed at: 8 days ago - Stars: 0 - Forks: 0

bobazooba/xllm
๐ฆ XโLLM: Cutting Edge & Easy LLM Finetuning
Language: Python - Size: 1.81 MB - Last synced at: 3 days ago - Pushed at: over 1 year ago - Stars: 406 - Forks: 21

AtharvaTaras/BankLlama
Code and files used during research for 'Fine Tuned Large Language Model In Banking'
Language: Jupyter Notebook - Size: 8.13 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

gaurisharan/phi2-alpaca-lora-4bit
Fine-tuning Microsoftโs Phi-2 model using QLoRA on the Alpaca-cleaned dataset with 4-bit quantization (bitsandbytes).
Language: Jupyter Notebook - Size: 779 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

edcalderin/huggingface-ragflow
This project implements a classic Retrieval-Augmented Generation (RAG) system using HuggingFace models with quantization techniques. The system processes PDF documents, extracts their content, and enables interactive question-answering through a Streamlit web application.
Language: Python - Size: 120 KB - Last synced at: about 1 month ago - Pushed at: about 2 months ago - Stars: 1 - Forks: 0

NightKing-V/SubtitleLLM_EngtoSin
Experimental Eng->Sin Subtitle Translation Model
Language: Jupyter Notebook - Size: 851 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

lpalbou/model-quantizer
Effortlessly quantize, benchmark, and publish Hugging Face models with cross-platform support for CPU/GPU. Reduce model size by 75% while maintaining performance.
Language: Python - Size: 165 KB - Last synced at: 5 days ago - Pushed at: 6 months ago - Stars: 2 - Forks: 0

Md-Emon-Hasan/Fine-Tuning
End-to-end fine-tuning of Hugging Face models using LoRA, QLoRA, quantization, and PEFT techniques. Optimized for low-memory with efficient model deployment
Language: Jupyter Notebook - Size: 5.53 MB - Last synced at: about 1 month ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

arjunravi26/ai_psychologist
Fine tune llama 2 model with mental health counseling data from huggingface
Language: Jupyter Notebook - Size: 14.5 MB - Last synced at: about 1 month ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

SCCSMARTCODE/Deep-Learning-03-LLM-FineTuning
Scalable and modular framework for fine-tuning large language models (LLMs) using LoRA and QLoRA. Supports 4-bit/8-bit quantization, Hugging Face Transformers, and instruction-tuning workflows across various tasks and datasets. Built for reproducibility, extensibility, and efficient experimentation.
Language: Jupyter Notebook - Size: 66.4 KB - Last synced at: 2 months ago - Pushed at: 4 months ago - Stars: 0 - Forks: 0

AkimfromParis/RAG-Japanese
Open source RAG with Llama Index for Japanese LLM in low resource settting
Language: Jupyter Notebook - Size: 2.54 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 8 - Forks: 3

GodreignElgin/llm-comparision
Jupyter Notebook for LLM compression via quantization (INT8, INT4, FP16) and evaluation using metrics such as ROUGE and BLEU. Facilitates efficient LLM optimization.
Language: Jupyter Notebook - Size: 1.82 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 0 - Forks: 1

ArchitJ6/Llama2-FineTuning
๐ฆ Llama2-FineTuning: Fine-tune LLAMA 2 with Custom Datasets Using LoRA and QLoRA Techniques
Language: Jupyter Notebook - Size: 2.83 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 0 - Forks: 0

shaheennabi/Production-Ready-Instruction-Finetuning-of-Meta-Llama-3.2-3B-Instruct-Project
๐๐ฟ๐ Instruction Fine-Tuning of Meta Llama 3.2-3B Instruct on Kannada Conversations ๐๐ฟ๐ Tailoring the model to follow specific instructions in Kannada, enhancing its ability to generate relevant, context-aware responses based on conversational inputs. ๐โจ Using the Kannada Instruct dataset for fine-tuning! Happy Finetuning๐๐
Language: Jupyter Notebook - Size: 692 KB - Last synced at: 5 months ago - Pushed at: 7 months ago - Stars: 18 - Forks: 2

dvgodoy/FineTuningLLMs
Official repository of my book "A Hands-On Guide to Fine-Tuning LLMs with PyTorch and Hugging Face"
Language: Jupyter Notebook - Size: 7.48 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 31 - Forks: 7

Bushra-Butt-17/BudgetBuddy-Finance-Chatbot
Budget Buddy is a finance chatbot built using Chainlit and the LLaMA language model. It analyzes PDF documents, such as bank statements and budget reports, to provide personalized financial advice and insights. The chatbot is integrated with Hugging Face for model management, offering an interactive way to manage personal finances.
Language: Python - Size: 19.3 MB - Last synced at: 5 months ago - Pushed at: 9 months ago - Stars: 2 - Forks: 0

403errors/CancerCareAI
An AI-powered system for extracting cancer-related information from patient Electronic Health Record (EHR) notes
Language: Jupyter Notebook - Size: 2.77 MB - Last synced at: 22 days ago - Pushed at: 6 months ago - Stars: 0 - Forks: 0

arham-kk/llama2-qlora-sft
This model is a fine-tuned model based on the "TinyPixel/Llama-2-7B-bf16-sharded" model and "timdettmers/openassistant-guanaco" dataset
Size: 640 KB - Last synced at: 2 days ago - Pushed at: almost 2 years ago - Stars: 3 - Forks: 0

kedir/Specialized-Immigration-Assistant
provide specialized immigration assistance in the field of immigration law using large language model
Language: Jupyter Notebook - Size: 64.5 KB - Last synced at: 2 days ago - Pushed at: almost 2 years ago - Stars: 2 - Forks: 0

Varun0157/quantisation
Experiments in quantisation consisting of quantisation from scratch, bitsandbytes, and llama.cpp. [Assignment 4 of Advanced Natural Language Processing, IIIT-H Monsoon '24]
Language: Python - Size: 395 KB - Last synced at: 25 days ago - Pushed at: 8 months ago - Stars: 0 - Forks: 0

eljandoubi/AI-Photo-Editing-with-Inpainting
A web app that allows you to select a subject and then change its background, OR keep the background and change the subject.
Language: Jupyter Notebook - Size: 10.9 MB - Last synced at: 4 months ago - Pushed at: over 1 year ago - Stars: 8 - Forks: 1

dasdristanta13/LLM-Lora-PEFT_accumulate
LLM-Lora-PEFT_accumulate explores optimizations for Large Language Models (LLMs) using PEFT, LORA, and QLORA. Contribute experiments and implementations to enhance LLM efficiency. Join discussions and push the boundaries of LLM optimization. Let's make LLMs more efficient together!
Language: Jupyter Notebook - Size: 138 KB - Last synced at: 2 days ago - Pushed at: about 2 years ago - Stars: 6 - Forks: 1

antonio-f/Orca2
Orca 2 on Colab
Language: Jupyter Notebook - Size: 137 KB - Last synced at: 5 months ago - Pushed at: over 1 year ago - Stars: 8 - Forks: 1

to-aoki/bitsandbytes
bitsandbytes modify for jetson orin
Language: Python - Size: 1.09 MB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0

bobazooba/shurale
Conversation AI model for open domain dialogs
Language: Python - Size: 64.5 KB - Last synced at: 2 days ago - Pushed at: almost 2 years ago - Stars: 4 - Forks: 1

ryan-air/Alpaca-350M-Fine-Tuned
In this project, I have provided code and a Colaboratory notebook that facilitates the fine-tuning process of an Alpaca 350M parameter model originally developed at Stanford University. The model was adapted using LoRA to run with fewer computational resources and training parameters and used HuggingFace's PEFT library.
Language: Jupyter Notebook - Size: 70.3 KB - Last synced at: about 2 years ago - Pushed at: about 2 years ago - Stars: 2 - Forks: 1

ryan-air/Alpaca-3B-Fine-Tuned
In this project, I have provided code and a Colaboratory notebook that facilitates the fine-tuning process of an Alpaca 3B parameter model originally developed at Stanford University. The model was adapted using LoRA to run with fewer computational resources and training parameters and used HuggingFace's PEFT library.
Language: Jupyter Notebook - Size: 48.8 KB - Last synced at: about 2 years ago - Pushed at: about 2 years ago - Stars: 2 - Forks: 0
