GitHub topics: peft
huggingface/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Language: Python - Size: 22.4 MB - Last synced at: about 7 hours ago - Pushed at: about 8 hours ago - Stars: 19,469 - Forks: 2,017

ModelCloud/GPTQModel
LLM model compression/quantization toolkit with hw acceleration support for Nvidia CUDA, AMD ROCm, Intel XPU and Intel/AMD/Apple CPU via HF, vLLM, and SGLang.
Language: Python - Size: 12.3 MB - Last synced at: about 6 hours ago - Pushed at: about 7 hours ago - Stars: 767 - Forks: 111

nabeelshan78/Transformer-Adaptation-Playbook
An empirical study of Transformer adaptation techniques. Pre-training from scratch (MLM), classic fine-tuning, and from-scratch implementations of PEFT methods (LoRA, Adapters). Tuning both encoder (BERT) and decoder (OPT) models.
Language: Jupyter Notebook - Size: 1.45 MB - Last synced at: about 12 hours ago - Pushed at: about 14 hours ago - Stars: 0 - Forks: 0

mxagar/llm_peft_fine_tuning_example
Example project in which a Large Language Model is fine-tuned using PEFT.
Language: Jupyter Notebook - Size: 6.3 MB - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 0 - Forks: 0

stochasticai/xTuring
Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6
Language: Python - Size: 18.4 MB - Last synced at: 2 days ago - Pushed at: 12 months ago - Stars: 2,660 - Forks: 202

InternLM/xtuner
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
Language: Python - Size: 2.14 MB - Last synced at: 2 days ago - Pushed at: 22 days ago - Stars: 4,726 - Forks: 354

TUDB-Labs/mLoRA
An Efficient "Factory" to Build Multiple LoRA Adapters
Language: Python - Size: 11 MB - Last synced at: 2 days ago - Pushed at: 7 months ago - Stars: 337 - Forks: 61

ashishpatel26/LLM-Finetuning
LLM Finetuning with peft
Language: Jupyter Notebook - Size: 3.47 MB - Last synced at: 2 days ago - Pushed at: about 1 month ago - Stars: 2,620 - Forks: 682

Blue-No1/fine-tuning-experiments-v2
LoRA/QLoRA experiments on open-weight LLMs (config, scripts, logs).
Language: Python - Size: 9.77 KB - Last synced at: 3 days ago - Pushed at: 3 days ago - Stars: 0 - Forks: 0

NisaarAgharia/Indian-LawyerGPT
Fine-Tuning Falcon-7B, LLAMA 2 with QLoRA to create an advanced AI model with a profound understanding of the Indian legal context.
Language: Jupyter Notebook - Size: 3.54 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 85 - Forks: 33

hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Language: Python - Size: 53 MB - Last synced at: 3 days ago - Pushed at: 6 days ago - Stars: 56,995 - Forks: 6,986

shreyas27092004/flan-t5-detoxification-ppo
Fine-Tuning FLAN-T5 for Detoxification with PPO and PEFT
Language: Jupyter Notebook - Size: 34.2 KB - Last synced at: 3 days ago - Pushed at: 3 days ago - Stars: 0 - Forks: 0

lxe/simple-llm-finetuner
Simple UI for LLM Model Finetuning
Language: Jupyter Notebook - Size: 1.53 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 2,066 - Forks: 132

yangjianxin1/Firefly
Firefly: 大模型训练工具,支持训练Qwen2.5、Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
Language: Python - Size: 6.24 MB - Last synced at: 3 days ago - Pushed at: 10 months ago - Stars: 6,531 - Forks: 583

X-LANCE/SLAM-LLM
Speech, Language, Audio, Music Processing with Large Language Model
Language: Python - Size: 169 MB - Last synced at: 4 days ago - Pushed at: 28 days ago - Stars: 884 - Forks: 91

jasonvanf/llama-trl
LLaMA-TRL: Fine-tuning LLaMA with PPO and LoRA
Language: Python - Size: 37 MB - Last synced at: 2 days ago - Pushed at: 17 days ago - Stars: 227 - Forks: 23

modelscope/ms-swift
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 500+ LLMs (Qwen3, Qwen3-MoE, Llama4, GLM4.5, InternLM3, DeepSeek-R1, ...) and 200+ MLLMs (Qwen2.5-VL, Qwen2.5-Omni, Qwen2-Audio, InternVL3, Ovis2.5, Llava, GLM4v, Phi4, ...) (AAAI 2025).
Language: Python - Size: 67.9 MB - Last synced at: 5 days ago - Pushed at: 6 days ago - Stars: 9,568 - Forks: 841

cirovitale/text2sql
Fine-tuned LLM system for conversational natural language to SQL translation. Uses LoRA fine-tuning of deepseek-coder-1.3B-instruct on CoSQL dataset with Parameter-Efficient Fine-Tuning (PEFT). Evaluated on Question Match and Interaction Match metrics for multi-turn conversational settings.
Language: Python - Size: 196 MB - Last synced at: 5 days ago - Pushed at: 6 days ago - Stars: 0 - Forks: 0

Joyce94/LLM-RLHF-Tuning
LLM Tuning with PEFT (SFT+RM+PPO+DPO with LoRA)
Language: Python - Size: 22.3 MB - Last synced at: 2 days ago - Pushed at: almost 2 years ago - Stars: 433 - Forks: 20

iamarunbrahma/finetuned-qlora-falcon7b-medical
Finetuning of Falcon-7B LLM using QLoRA on Mental Health Conversational Dataset
Language: Jupyter Notebook - Size: 174 KB - Last synced at: 5 days ago - Pushed at: over 1 year ago - Stars: 261 - Forks: 27

huggingface/lora-fast
Minimal repository to demonstrate fast LoRA inference with Flux family of models.
Language: Python - Size: 26.4 KB - Last synced at: 6 days ago - Pushed at: about 1 month ago - Stars: 22 - Forks: 1

Cre4T3Tiv3/unsloth-llama3-alpaca-lora
Advanced 4-bit QLoRA fine-tuning pipeline for LLaMA 3 8B with production-grade optimization. Memory-efficient training on consumer GPUs for instruction-following specialization. Demonstrates cutting-edge parameter-efficient fine-tuning with Unsloth integration.
Language: Jupyter Notebook - Size: 2.11 MB - Last synced at: 6 days ago - Pushed at: about 2 months ago - Stars: 25 - Forks: 0

WeiminWu2000/GenomeBridge
An Integrated Platform for Tuning and Deploying Genomic Foundation Models
Language: Python - Size: 19.9 MB - Last synced at: 8 days ago - Pushed at: 8 days ago - Stars: 10 - Forks: 1

AmirAAZ818/GPT2-LoRA-QA
Implementation of Low-Rank Adaptation (LoRA) for parameter-efficient fine-tuning of GPT-2 on the SQuAD dataset for question answering, exploring training efficiency, loss masking, and performance metrics like F1 and Exact Match. Final Course project for Deep Learning at University of Kerman, Spring 2025.
Language: Jupyter Notebook - Size: 1.16 MB - Last synced at: 9 days ago - Pushed at: 9 days ago - Stars: 0 - Forks: 0

adithya-s-k/CompanionLLM
CompanionLLM - A framework to finetune LLMs to be your own sentient conversational companion
Language: Jupyter Notebook - Size: 40.1 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 42 - Forks: 5

gyuilLim/GSoC_2025
Implementation and documentation of PEFT methods (LoRA & DoRA) in OpenVINO™ Training Extensions, including experimental results.
Language: Python - Size: 303 KB - Last synced at: 10 days ago - Pushed at: 10 days ago - Stars: 0 - Forks: 0

DONGRYEOLLEE1/PEFT
This repository contains implementations of various Parameter-Efficient Fine-Tuning (PEFT) methods from scratch using PyTorch. The goal is to provide clear, educational implementations that help understand the core concepts and mechanics of different PEFT techniques.
Language: Python - Size: 121 KB - Last synced at: 10 days ago - Pushed at: 10 days ago - Stars: 0 - Forks: 0

PRITHIVSAKTHIUR/Qwen-Image-Diffusion
Generate high-quality images from text prompts using the Qwen diffusion model with an intuitive Gradio web interface.
Language: Python - Size: 52.7 KB - Last synced at: 10 days ago - Pushed at: 10 days ago - Stars: 3 - Forks: 0

zetavg/LLaMA-LoRA-Tuner
UI tool for fine-tuning and testing your own LoRA models base on LLaMA, GPT-J and more. One-click run on Google Colab. + A Gradio ChatGPT-like Chat UI to demonstrate your language models.
Language: Python - Size: 1010 KB - Last synced at: 1 day ago - Pushed at: over 2 years ago - Stars: 475 - Forks: 99

Guitaricet/relora
Official code for ReLoRA from the paper Stack More Layers Differently: High-Rank Training Through Low-Rank Updates
Language: Jupyter Notebook - Size: 1.89 MB - Last synced at: 5 days ago - Pushed at: over 1 year ago - Stars: 462 - Forks: 40

Simplifine-gamedev/Simplifine
🚀 Easy, open-source LLM finetuning with one-line commands, seamless cloud integration, and popular optimization frameworks. ✨
Language: Python - Size: 844 KB - Last synced at: 10 days ago - Pushed at: about 1 year ago - Stars: 93 - Forks: 4

Gawdanzo/Build-a-LLM-model-from-scratch
🚀 Build a complete LLM model from scratch with an easy-to-follow, end-to-end pipeline for data processing, training, and fine-tuning.
Language: Jupyter Notebook - Size: 420 KB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 0 - Forks: 0

shreyas27092004/Generative-AI-Model-Fine-Tuning-Hugging-Face-Transformers-
Fine-tuning a Generative AI model using Hugging Face Transformers. Includes dataset preprocessing, model training, evaluation, and text generation with a custom fine-tuned model.
Language: Jupyter Notebook - Size: 34.2 KB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 0 - Forks: 0

sandyFit/genAI-Nanodegree
A hands-on learning journey through Generative AI. Covers foundational concepts, deep learning, and practical skills in text and image generation using large language models and computer vision. Includes projects like chatbots and AI agents, with a focus on applying models to real-world tasks.
Language: Jupyter Notebook - Size: 8.72 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 1 - Forks: 0

mindspore-courses/step_into_llm
MindSpore online courses: Step into LLM
Language: Jupyter Notebook - Size: 246 MB - Last synced at: 2 days ago - Pushed at: 14 days ago - Stars: 477 - Forks: 123

punpunzaz10/TADFormer
Efficiently implement multi-task learning with TADFormer, a task-adaptive dynamic transformer. Explore the code on GitHub! 🚀🌟
Language: Python - Size: 126 KB - Last synced at: 14 days ago - Pushed at: 14 days ago - Stars: 2 - Forks: 0

Degon3399/XTTS_V2
This repository offers a framework for fine-tuning the XTTS_V2 model, focusing on multilingual text-to-speech applications. It includes tools for both full model fine-tuning and LoRA fine-tuning, along with inference scripts for easy speech synthesis. 🐙🌐
Language: Python - Size: 269 KB - Last synced at: 15 days ago - Pushed at: 15 days ago - Stars: 1 - Forks: 1

beeracs/Llama
Run Llama models in your web browser using JavaScript and WebAssembly. Explore light and dark modes easily. 🌐🐱👤
Language: HTML - Size: 964 KB - Last synced at: 15 days ago - Pushed at: 15 days ago - Stars: 0 - Forks: 0

MRROBOT401/DyT-NoNorm-LLMs-REWILD
Replacing LayerNorm with Dynamic Tanh (DyT) in DistilGPT2 + LoRA, evaluated on RE-WILD, Alpaca, and ShareGPT.
Language: Jupyter Notebook - Size: 21.1 MB - Last synced at: 15 days ago - Pushed at: 15 days ago - Stars: 0 - Forks: 0

sandy1990418/Finetune-Qwen2.5-VL
Fine-tuning Qwen2.5-VL for vision-language tasks | Optimized for Vision understanding | LoRA & PEFT support.
Language: Python - Size: 144 KB - Last synced at: 15 days ago - Pushed at: 7 months ago - Stars: 113 - Forks: 11

ictup/Enhancing-QA-Systems-through-Integrated-Reasoning-over-Knowledge-Bases-and-Large-Language-Models
KG-RAG + ToT + multi-agent LLMs for evidence-grounded QA with Neo4j and fine-tuning; reproducible medical case study & evaluation.
Language: Python - Size: 141 KB - Last synced at: 20 days ago - Pushed at: 20 days ago - Stars: 0 - Forks: 0

ictup/Build-a-LLM-model-from-scratch
LLM pipeline: data→tokenizer→GPT train/eval→instruction FT→sampling. Reproducible, clean configs, RTX-4060 defaults, ready for AMP/LoRA/DDP.
Language: Jupyter Notebook - Size: 511 KB - Last synced at: 20 days ago - Pushed at: 20 days ago - Stars: 1 - Forks: 0

pheonix-19/OpsAI
OpsAI (Operational AI) is an intelligent IT support automation platform that uses AI to automatically categorize tickets, suggest solutions, and route requests to the right teams. Built with advanced NLP and machine learning technologies, it integrates with Jira, Slack, and Freshdesk to streamline operational workflows and improve response times.
Language: Python - Size: 2.43 MB - Last synced at: 21 days ago - Pushed at: 21 days ago - Stars: 0 - Forks: 0

brown-palm/AntGPT
Official code implemtation of paper AntGPT: Can Large Language Models Help Long-term Action Anticipation from Videos?
Language: Python - Size: 28.8 MB - Last synced at: 2 days ago - Pushed at: 12 months ago - Stars: 23 - Forks: 2

jackaduma/Alpaca-LoRA-RLHF-PyTorch
A full pipeline to finetune Alpaca LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the Alpaca architecture. Basically ChatGPT but with Alpaca
Language: Python - Size: 18.7 MB - Last synced at: 2 days ago - Pushed at: over 2 years ago - Stars: 60 - Forks: 6

MAGICS-LAB/GERM
[ICML 2025] Fast and Low-Cost Genomic Foundation Models via Outlier Removal.
Language: Python - Size: 21 MB - Last synced at: 4 days ago - Pushed at: 3 months ago - Stars: 17 - Forks: 2

PRITHIVSAKTHIUR/Sub-Memory-Efficient-Merging-FluxKreaDev
black-forest-labs/FLUX.1-dev and black-forest-labs/FLUX.1-Krea-dev. This merged model integrates the capabilities of the rectified flow transformer FLUX.1-dev, known for competitive prompt following and high-quality outputs, with FLUX.1-Krea-dev, a guidance distilled model emphasizing aesthetics and photorealism.
Language: Python - Size: 30.3 KB - Last synced at: 24 days ago - Pushed at: 25 days ago - Stars: 1 - Forks: 0

jackaduma/ChatGLM-LoRA-RLHF-PyTorch
A full pipeline to finetune ChatGLM LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the ChatGLM architecture. Basically ChatGPT but with ChatGLM
Language: Python - Size: 25.3 MB - Last synced at: 2 days ago - Pushed at: over 2 years ago - Stars: 138 - Forks: 10

elpikaoocle2006/Sub-Memory-Efficient-Merging-FluxKreaDev
Merge advanced image generation models FLUX.1-dev and FLUX.1-Krea-dev for improved text-to-image performance. Explore the repository for seamless integration! 🐙
Language: Python - Size: 17.6 KB - Last synced at: 22 days ago - Pushed at: 23 days ago - Stars: 0 - Forks: 0

Baijiong-Lin/LoRA-Torch
PyTorch Reimplementation of LoRA (featuring with supporting nn.MultiheadAttention in OpenCLIP)
Language: Python - Size: 60.5 KB - Last synced at: 13 days ago - Pushed at: 3 months ago - Stars: 67 - Forks: 7

PRITHIVSAKTHIUR/Sub-Memory-Efficient-Merging-FluxKreaSchnell
A memory-efficient approach to merge FLUX.1-schnell and FLUX.1-Krea-dev transformer models, combining their strengths while maintaining guidance capabilities from the schnell model.
Language: Python - Size: 11.7 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 1 - Forks: 0

arcforge-tune/bmg-lora
LoRA + QLoRA fine‑tuning toolkit optimized for Intel Arc Battlemage GPUs
Language: Python - Size: 131 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

robuno/Title-Generator-with-LLM-QLoRa
Fine-tuning LLMs with LoRA to generate titles from the given abstract, specifically for CS-related papers.
Language: Jupyter Notebook - Size: 254 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 7 - Forks: 0

Aisuko/notebooks
Implementation for the different ML tasks on Kaggle platform with GPUs.
Language: Jupyter Notebook - Size: 159 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 24 - Forks: 4

mominalix/LLM-Finetuning-Pipeline-LoRA-QLoRA
Production-ready pipeline for fine-tuning Large Language Models using LoRA/QLoRA techniques with MLflow experiment tracking,
Language: Python - Size: 70.3 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

PRITHIVSAKTHIUR/Flux-LoRA-DLC
Experience the power of the FLUX.1-dev diffusion model combined with a massive collection of 255+ community-created LoRAs! This Gradio application provides an easy-to-use interface to explore diverse artistic styles directly on top of the FLUX base model.
Language: Python - Size: 1.42 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 12 - Forks: 1

TUDB-Labs/MoE-PEFT
An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT
Language: Python - Size: 7.18 MB - Last synced at: about 1 month ago - Pushed at: 6 months ago - Stars: 107 - Forks: 16

wahabzh/lora-smollm-finetuning
⚙️ LoRA implementation for efficient SmolLM fine-tuning. Achieves comparable performance with only 0.24% trainable parameters.
Language: Jupyter Notebook - Size: 155 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

Ahmadhammam03/Galaxy-Chatbot
🤖 AI-powered chatbot for Galaxy Organisation and Alibaba Cloud Academy
Language: Jupyter Notebook - Size: 8.66 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 1 - Forks: 0

Samarth2001/LLM-Fine-tuning
Parameter-efficient fine-tuning experiments for 7B LLMs on consumer hardware. QLoRA implementations, memory optimization strategies, and reproducible benchmarks for Mistral, Llama-2, and other models on Google Colab T4 GPUs.
Language: Jupyter Notebook - Size: 63.5 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 1

AtharvaTaras/BankLlama
Code and files used during research for 'Fine Tuned Large Language Model In Banking'
Language: Jupyter Notebook - Size: 8.13 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

dev-kanika/llm-chatbot-multimodal-rag-bioasq
Multimodal chatbot project using five LLM-based approaches including RAG, PEFT, and prompt engineering, built on the BioASQ dataset for biomedical QA.
Language: Jupyter Notebook - Size: 544 KB - Last synced at: 24 days ago - Pushed at: about 1 month ago - Stars: 3 - Forks: 1

Showmick119/Fine-Tuning-Open-Source-LLM
Fine-Tuning the open source Code-LLama-7B model for code generation on FastAPI tasks.
Language: Python - Size: 7.78 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 1 - Forks: 0

calpt/awesome-adapter-resources
Collection of Tools and Papers related to Adapters / Parameter-Efficient Transfer Learning/ Fine-Tuning
Language: Python - Size: 213 KB - Last synced at: 6 days ago - Pushed at: over 1 year ago - Stars: 197 - Forks: 11

samay-jain/Fine_tuning_Distilbert_Model_using_LoRA_Low-Rank-Adaptation
Parameter-efficient fine-tuning of DistilBERT using LoRA for sentiment and topic classification, with CLI, API, and interactive chatbot interfaces.
Language: Python - Size: 171 MB - Last synced at: 24 days ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

ROIM1998/APT
[ICML'24 Oral] APT: Adaptive Pruning and Tuning Pretrained Language Models for Efficient Training and Inference
Language: Python - Size: 4.08 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 44 - Forks: 2

CatsMeow492/adaptive-lora-rank-allocation
This repository contains the experimental code for investigating whether jointly optimizing LoRA rank allocation and mixed-precision quantization yields better efficiency-performance trade-offs than existing baselines on laptop-class hardware.
Language: Python - Size: 1.16 MB - Last synced at: 5 days ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

pratzohol/MediTune
Fine-tuning LLMs on MedMCQA medical MCQ dataset
Language: Python - Size: 15.6 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

jackaduma/Vicuna-LoRA-RLHF-PyTorch
A full pipeline to finetune Vicuna LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the Vicuna architecture. Basically ChatGPT but with Vicuna
Language: Python - Size: 18.7 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 219 - Forks: 19

wiseodd/lapeft-bayesopt
Discrete Bayesian optimization with LLMs, PEFT finetuning methods, and the Laplace approximation.
Language: Python - Size: 4.85 MB - Last synced at: 3 days ago - Pushed at: about 1 year ago - Stars: 19 - Forks: 2

louisc-s/QLoRA-Fine-tuning-for-Film-Character-Styled-Responses-from-LLM
Code for fine-tuning Llama2 LLM with custom text dataset to produce film character styled responses
Language: Python - Size: 63.1 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 9 - Forks: 1

Expo314/Gemma-3n-local-training
Lightweight GPU framework for Gemma 3n training and fine-tuning. Ideal for chatbots and Q&A systems. 🚀💻 Optimize performance with LoRA and quantization.
Language: Python - Size: 155 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

llm-db/FineInfer
Deferred Continuous Batching in Resource-Efficient Large Language Model Serving (EuroMLSys 2024)
Language: Python - Size: 53.7 KB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 17 - Forks: 2

km1994/llms_paper
该仓库主要记录 LLMs 算法工程师相关的顶会论文研读笔记(多模态、PEFT、小样本QA问答、RAG、LMMs可解释性、Agents、CoT)
Size: 37.7 MB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 343 - Forks: 14

Rampati/MIT-6.S191-Lab3
In this lab, you will fine-tune a multi-billion parameter language model to generate specific style responses. You'll work with tokenization strategies, prompt templates, and a complete fine-tuning workflow to enhance LLM outputs. 🐙✨
Language: Jupyter Notebook - Size: 43 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 1 - Forks: 0

Aradhye2002/selective-peft-toolkit
Official implementation of the paper "Step-by-Step Unmasking for Parameter-Efficient Fine-tuning of Large Language Models"
Language: Python - Size: 43.9 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 9 - Forks: 0

paulinhok14/fine-tuning-llms
A lightweight toolkit for efficient quantization and LoRA‑based fine‑tuning of LLMs (Large Language Models).
Language: Jupyter Notebook - Size: 795 KB - Last synced at: about 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

Dee66/CodeCraft-AI
Cloud-native, AWS-native platform for Retrieval-Augmented Generation (RAG) and parameter-efficient fine-tuning (PEFT), built with full MLOps and Infrastructure as Code. Features FastAPI, Docker, AWS CDK, ECS Fargate, SageMaker, S3, Secrets Manager, CloudWatch, CI/CD with GitHub Actions, robust security, monitoring, automation for enterprise AI.
Language: Python - Size: 1.45 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

gazelle93/llm-fine-tuning-sft-lora-qlora
Practical examples for fine-tuning large language models (LLMs) with SFT, LoRA, and QLoRA using Hugging Face Transformers and PEFT.
Language: Python - Size: 0 Bytes - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

ankraj1234/MediGuide
Comparing QLoRA, Prompt & Prefix Tuning on Mistral-7B for medical instruction-following
Language: Jupyter Notebook - Size: 7.93 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 1 - Forks: 1

hiyouga/ChatGLM-Efficient-Tuning 📦
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
Language: Python - Size: 194 MB - Last synced at: 2 months ago - Pushed at: almost 2 years ago - Stars: 3,708 - Forks: 478

lijoraju/charitra-retrieval-augmented-qa
Retrieval-Augmented Q&A system for high school Social Studies, fine-tuned using QLoRA and deployed with FastAPI.
Language: Python - Size: 67.4 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

yspkm/lora-ft
This repository contains a collection of scripts I used to experiment with fine-tuning models like Llama and Gemma on various math and commonsense reasoning tasks using LoRA. My main field is AI-based antenna design for 6G, this is my side project for exploring LLM for fun.
Language: Python - Size: 74.2 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

aman-17/MediSOAP
FineTuning LLMs on conversational medical dataset.
Language: Jupyter Notebook - Size: 39.9 MB - Last synced at: 2 days ago - Pushed at: about 1 year ago - Stars: 4 - Forks: 0

Md-Emon-Hasan/Fine-Tuning
End-to-end fine-tuning of Hugging Face models using LoRA, QLoRA, quantization, and PEFT techniques. Optimized for low-memory with efficient model deployment
Language: Jupyter Notebook - Size: 5.53 MB - Last synced at: about 1 month ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

d0tTino/DeepThought-ReThought
A refactored version of the DeepThought Discord bot, focusing on improved architecture, performance, and AI agent capabilities.
Language: Python - Size: 20.3 MB - Last synced at: 7 days ago - Pushed at: 7 days ago - Stars: 1 - Forks: 1

FaNa-AI/MIT-6.S191-Lab3
MIT 6.S191 Lab 3 teaches you how to fine-tune large language models like Gemma 2B, structure prompts, and evaluate outputs using tools like Opik and LFM-40B.
Language: Jupyter Notebook - Size: 43.9 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

computational-cell-analytics/peft-sam
Parameter Efficient Fine-Tuning of Segment Anything Model
Language: Python - Size: 413 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 16 - Forks: 1

BorealisAI/flora-opt
This is the official repository for the paper "Flora: Low-Rank Adapters Are Secretly Gradient Compressors" in ICML 2024.
Language: Python - Size: 672 KB - Last synced at: 1 day ago - Pushed at: about 1 year ago - Stars: 104 - Forks: 5

Jiacheng-Zhu-AIML/AsymmetryLoRA
Preprint: Asymmetry in Low-Rank Adapters of Foundation Models
Language: Python - Size: 51.8 KB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 35 - Forks: 3

MaharshPatelX/HumanTouch
Transform AI-generated text into natural, human-like writing using DoRA fine-tuned Qwen models. Features interactive training modes, Google Colab support, and comprehensive data processing pipeline for 32k context humanization.
Language: Python - Size: 50.8 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

AMfeta99/NLP_LLM
This repository is dedicated to small projects and some theoretical material that I used to get into NLP and LLM in a practical and efficient way.
Language: Jupyter Notebook - Size: 85.2 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

Facico/GOAT-PEFT
[ICML2025] Make LoRA Great Again: Boosting LoRA with Adaptive Singular Values and Mixture-of-Experts Optimization Alignment
Language: Python - Size: 39 MB - Last synced at: 3 months ago - Pushed at: 4 months ago - Stars: 100 - Forks: 11

gokhaneraslan/XTTS_V2
Training XTTS V2 and PEFT LORA Text-to-Speech (TTS)
Language: Python - Size: 275 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

arjunravi26/ai_psychologist
Fine tune llama 2 model with mental health counseling data from huggingface
Language: Jupyter Notebook - Size: 14.5 MB - Last synced at: about 1 month ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

abdussahid26/LLM-Post-training-Techniques
This repository contains implementations of LLM post-training techniques, including SFT, PEFT, RLHF, PPO, DPO, and more.
Language: Jupyter Notebook - Size: 317 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

kriskrisliu/PAT
[AAAI 2025] PAT: Pruning-Aware Tuning for Large Language Models
Language: Python - Size: 30.2 MB - Last synced at: 2 months ago - Pushed at: 7 months ago - Stars: 32 - Forks: 1

PRITHIVSAKTHIUR/GALLO-3XL
High Quality Image Generation Model - Powered with NVIDIA A100
Language: Python - Size: 11.2 MB - Last synced at: 18 days ago - Pushed at: about 1 year ago - Stars: 13 - Forks: 1

zyds/transformers-code
手把手带你实战 Huggingface Transformers 课程视频同步更新在B站与YouTube
Language: Jupyter Notebook - Size: 76.5 MB - Last synced at: 3 months ago - Pushed at: about 1 year ago - Stars: 2,903 - Forks: 398

Ajairajv/Fine-Tuning-a-Generative-AI-Model
Fine-tunes FLAN-T5 for dialogue summarization using full fine-tuning and LoRA-based PEFT, evaluated with ROUGE metrics and human assessment.
Language: Jupyter Notebook - Size: 28.3 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0
