An open API service providing repository metadata for many open source software ecosystems.

GitHub topics: peft

huggingface/peft

🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

Language: Python - Size: 22.4 MB - Last synced at: about 7 hours ago - Pushed at: about 8 hours ago - Stars: 19,469 - Forks: 2,017

ModelCloud/GPTQModel

LLM model compression/quantization toolkit with hw acceleration support for Nvidia CUDA, AMD ROCm, Intel XPU and Intel/AMD/Apple CPU via HF, vLLM, and SGLang.

Language: Python - Size: 12.3 MB - Last synced at: about 6 hours ago - Pushed at: about 7 hours ago - Stars: 767 - Forks: 111

nabeelshan78/Transformer-Adaptation-Playbook

An empirical study of Transformer adaptation techniques. Pre-training from scratch (MLM), classic fine-tuning, and from-scratch implementations of PEFT methods (LoRA, Adapters). Tuning both encoder (BERT) and decoder (OPT) models.

Language: Jupyter Notebook - Size: 1.45 MB - Last synced at: about 12 hours ago - Pushed at: about 14 hours ago - Stars: 0 - Forks: 0

mxagar/llm_peft_fine_tuning_example

Example project in which a Large Language Model is fine-tuned using PEFT.

Language: Jupyter Notebook - Size: 6.3 MB - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 0 - Forks: 0

stochasticai/xTuring

Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6

Language: Python - Size: 18.4 MB - Last synced at: 2 days ago - Pushed at: 12 months ago - Stars: 2,660 - Forks: 202

InternLM/xtuner

An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)

Language: Python - Size: 2.14 MB - Last synced at: 2 days ago - Pushed at: 22 days ago - Stars: 4,726 - Forks: 354

TUDB-Labs/mLoRA

An Efficient "Factory" to Build Multiple LoRA Adapters

Language: Python - Size: 11 MB - Last synced at: 2 days ago - Pushed at: 7 months ago - Stars: 337 - Forks: 61

ashishpatel26/LLM-Finetuning

LLM Finetuning with peft

Language: Jupyter Notebook - Size: 3.47 MB - Last synced at: 2 days ago - Pushed at: about 1 month ago - Stars: 2,620 - Forks: 682

Blue-No1/fine-tuning-experiments-v2

LoRA/QLoRA experiments on open-weight LLMs (config, scripts, logs).

Language: Python - Size: 9.77 KB - Last synced at: 3 days ago - Pushed at: 3 days ago - Stars: 0 - Forks: 0

NisaarAgharia/Indian-LawyerGPT

Fine-Tuning Falcon-7B, LLAMA 2 with QLoRA to create an advanced AI model with a profound understanding of the Indian legal context.

Language: Jupyter Notebook - Size: 3.54 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 85 - Forks: 33

hiyouga/LLaMA-Factory

Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)

Language: Python - Size: 53 MB - Last synced at: 3 days ago - Pushed at: 6 days ago - Stars: 56,995 - Forks: 6,986

shreyas27092004/flan-t5-detoxification-ppo

Fine-Tuning FLAN-T5 for Detoxification with PPO and PEFT

Language: Jupyter Notebook - Size: 34.2 KB - Last synced at: 3 days ago - Pushed at: 3 days ago - Stars: 0 - Forks: 0

lxe/simple-llm-finetuner

Simple UI for LLM Model Finetuning

Language: Jupyter Notebook - Size: 1.53 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 2,066 - Forks: 132

yangjianxin1/Firefly

Firefly: 大模型训练工具,支持训练Qwen2.5、Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型

Language: Python - Size: 6.24 MB - Last synced at: 3 days ago - Pushed at: 10 months ago - Stars: 6,531 - Forks: 583

X-LANCE/SLAM-LLM

Speech, Language, Audio, Music Processing with Large Language Model

Language: Python - Size: 169 MB - Last synced at: 4 days ago - Pushed at: 28 days ago - Stars: 884 - Forks: 91

jasonvanf/llama-trl

LLaMA-TRL: Fine-tuning LLaMA with PPO and LoRA

Language: Python - Size: 37 MB - Last synced at: 2 days ago - Pushed at: 17 days ago - Stars: 227 - Forks: 23

modelscope/ms-swift

Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 500+ LLMs (Qwen3, Qwen3-MoE, Llama4, GLM4.5, InternLM3, DeepSeek-R1, ...) and 200+ MLLMs (Qwen2.5-VL, Qwen2.5-Omni, Qwen2-Audio, InternVL3, Ovis2.5, Llava, GLM4v, Phi4, ...) (AAAI 2025).

Language: Python - Size: 67.9 MB - Last synced at: 5 days ago - Pushed at: 6 days ago - Stars: 9,568 - Forks: 841

cirovitale/text2sql

Fine-tuned LLM system for conversational natural language to SQL translation. Uses LoRA fine-tuning of deepseek-coder-1.3B-instruct on CoSQL dataset with Parameter-Efficient Fine-Tuning (PEFT). Evaluated on Question Match and Interaction Match metrics for multi-turn conversational settings.

Language: Python - Size: 196 MB - Last synced at: 5 days ago - Pushed at: 6 days ago - Stars: 0 - Forks: 0

Joyce94/LLM-RLHF-Tuning

LLM Tuning with PEFT (SFT+RM+PPO+DPO with LoRA)

Language: Python - Size: 22.3 MB - Last synced at: 2 days ago - Pushed at: almost 2 years ago - Stars: 433 - Forks: 20

iamarunbrahma/finetuned-qlora-falcon7b-medical

Finetuning of Falcon-7B LLM using QLoRA on Mental Health Conversational Dataset

Language: Jupyter Notebook - Size: 174 KB - Last synced at: 5 days ago - Pushed at: over 1 year ago - Stars: 261 - Forks: 27

huggingface/lora-fast

Minimal repository to demonstrate fast LoRA inference with Flux family of models.

Language: Python - Size: 26.4 KB - Last synced at: 6 days ago - Pushed at: about 1 month ago - Stars: 22 - Forks: 1

Cre4T3Tiv3/unsloth-llama3-alpaca-lora

Advanced 4-bit QLoRA fine-tuning pipeline for LLaMA 3 8B with production-grade optimization. Memory-efficient training on consumer GPUs for instruction-following specialization. Demonstrates cutting-edge parameter-efficient fine-tuning with Unsloth integration.

Language: Jupyter Notebook - Size: 2.11 MB - Last synced at: 6 days ago - Pushed at: about 2 months ago - Stars: 25 - Forks: 0

WeiminWu2000/GenomeBridge

An Integrated Platform for Tuning and Deploying Genomic Foundation Models

Language: Python - Size: 19.9 MB - Last synced at: 8 days ago - Pushed at: 8 days ago - Stars: 10 - Forks: 1

AmirAAZ818/GPT2-LoRA-QA

Implementation of Low-Rank Adaptation (LoRA) for parameter-efficient fine-tuning of GPT-2 on the SQuAD dataset for question answering, exploring training efficiency, loss masking, and performance metrics like F1 and Exact Match. Final Course project for Deep Learning at University of Kerman, Spring 2025.

Language: Jupyter Notebook - Size: 1.16 MB - Last synced at: 9 days ago - Pushed at: 9 days ago - Stars: 0 - Forks: 0

adithya-s-k/CompanionLLM

CompanionLLM - A framework to finetune LLMs to be your own sentient conversational companion

Language: Jupyter Notebook - Size: 40.1 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 42 - Forks: 5

gyuilLim/GSoC_2025

Implementation and documentation of PEFT methods (LoRA & DoRA) in OpenVINO™ Training Extensions, including experimental results.

Language: Python - Size: 303 KB - Last synced at: 10 days ago - Pushed at: 10 days ago - Stars: 0 - Forks: 0

DONGRYEOLLEE1/PEFT

This repository contains implementations of various Parameter-Efficient Fine-Tuning (PEFT) methods from scratch using PyTorch. The goal is to provide clear, educational implementations that help understand the core concepts and mechanics of different PEFT techniques.

Language: Python - Size: 121 KB - Last synced at: 10 days ago - Pushed at: 10 days ago - Stars: 0 - Forks: 0

PRITHIVSAKTHIUR/Qwen-Image-Diffusion

Generate high-quality images from text prompts using the Qwen diffusion model with an intuitive Gradio web interface.

Language: Python - Size: 52.7 KB - Last synced at: 10 days ago - Pushed at: 10 days ago - Stars: 3 - Forks: 0

zetavg/LLaMA-LoRA-Tuner

UI tool for fine-tuning and testing your own LoRA models base on LLaMA, GPT-J and more. One-click run on Google Colab. + A Gradio ChatGPT-like Chat UI to demonstrate your language models.

Language: Python - Size: 1010 KB - Last synced at: 1 day ago - Pushed at: over 2 years ago - Stars: 475 - Forks: 99

Guitaricet/relora

Official code for ReLoRA from the paper Stack More Layers Differently: High-Rank Training Through Low-Rank Updates

Language: Jupyter Notebook - Size: 1.89 MB - Last synced at: 5 days ago - Pushed at: over 1 year ago - Stars: 462 - Forks: 40

Simplifine-gamedev/Simplifine

🚀 Easy, open-source LLM finetuning with one-line commands, seamless cloud integration, and popular optimization frameworks. ✨

Language: Python - Size: 844 KB - Last synced at: 10 days ago - Pushed at: about 1 year ago - Stars: 93 - Forks: 4

Gawdanzo/Build-a-LLM-model-from-scratch

🚀 Build a complete LLM model from scratch with an easy-to-follow, end-to-end pipeline for data processing, training, and fine-tuning.

Language: Jupyter Notebook - Size: 420 KB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 0 - Forks: 0

shreyas27092004/Generative-AI-Model-Fine-Tuning-Hugging-Face-Transformers-

Fine-tuning a Generative AI model using Hugging Face Transformers. Includes dataset preprocessing, model training, evaluation, and text generation with a custom fine-tuned model.

Language: Jupyter Notebook - Size: 34.2 KB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 0 - Forks: 0

sandyFit/genAI-Nanodegree

A hands-on learning journey through Generative AI. Covers foundational concepts, deep learning, and practical skills in text and image generation using large language models and computer vision. Includes projects like chatbots and AI agents, with a focus on applying models to real-world tasks.

Language: Jupyter Notebook - Size: 8.72 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 1 - Forks: 0

mindspore-courses/step_into_llm

MindSpore online courses: Step into LLM

Language: Jupyter Notebook - Size: 246 MB - Last synced at: 2 days ago - Pushed at: 14 days ago - Stars: 477 - Forks: 123

punpunzaz10/TADFormer

Efficiently implement multi-task learning with TADFormer, a task-adaptive dynamic transformer. Explore the code on GitHub! 🚀🌟

Language: Python - Size: 126 KB - Last synced at: 14 days ago - Pushed at: 14 days ago - Stars: 2 - Forks: 0

Degon3399/XTTS_V2

This repository offers a framework for fine-tuning the XTTS_V2 model, focusing on multilingual text-to-speech applications. It includes tools for both full model fine-tuning and LoRA fine-tuning, along with inference scripts for easy speech synthesis. 🐙🌐

Language: Python - Size: 269 KB - Last synced at: 15 days ago - Pushed at: 15 days ago - Stars: 1 - Forks: 1

beeracs/Llama

Run Llama models in your web browser using JavaScript and WebAssembly. Explore light and dark modes easily. 🌐🐱👤

Language: HTML - Size: 964 KB - Last synced at: 15 days ago - Pushed at: 15 days ago - Stars: 0 - Forks: 0

MRROBOT401/DyT-NoNorm-LLMs-REWILD

Replacing LayerNorm with Dynamic Tanh (DyT) in DistilGPT2 + LoRA, evaluated on RE-WILD, Alpaca, and ShareGPT.

Language: Jupyter Notebook - Size: 21.1 MB - Last synced at: 15 days ago - Pushed at: 15 days ago - Stars: 0 - Forks: 0

sandy1990418/Finetune-Qwen2.5-VL

Fine-tuning Qwen2.5-VL for vision-language tasks | Optimized for Vision understanding | LoRA & PEFT support.

Language: Python - Size: 144 KB - Last synced at: 15 days ago - Pushed at: 7 months ago - Stars: 113 - Forks: 11

ictup/Enhancing-QA-Systems-through-Integrated-Reasoning-over-Knowledge-Bases-and-Large-Language-Models

KG-RAG + ToT + multi-agent LLMs for evidence-grounded QA with Neo4j and fine-tuning; reproducible medical case study & evaluation.

Language: Python - Size: 141 KB - Last synced at: 20 days ago - Pushed at: 20 days ago - Stars: 0 - Forks: 0

ictup/Build-a-LLM-model-from-scratch

LLM pipeline: data→tokenizer→GPT train/eval→instruction FT→sampling. Reproducible, clean configs, RTX-4060 defaults, ready for AMP/LoRA/DDP.

Language: Jupyter Notebook - Size: 511 KB - Last synced at: 20 days ago - Pushed at: 20 days ago - Stars: 1 - Forks: 0

pheonix-19/OpsAI

OpsAI (Operational AI) is an intelligent IT support automation platform that uses AI to automatically categorize tickets, suggest solutions, and route requests to the right teams. Built with advanced NLP and machine learning technologies, it integrates with Jira, Slack, and Freshdesk to streamline operational workflows and improve response times.

Language: Python - Size: 2.43 MB - Last synced at: 21 days ago - Pushed at: 21 days ago - Stars: 0 - Forks: 0

brown-palm/AntGPT

Official code implemtation of paper AntGPT: Can Large Language Models Help Long-term Action Anticipation from Videos?

Language: Python - Size: 28.8 MB - Last synced at: 2 days ago - Pushed at: 12 months ago - Stars: 23 - Forks: 2

jackaduma/Alpaca-LoRA-RLHF-PyTorch

A full pipeline to finetune Alpaca LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the Alpaca architecture. Basically ChatGPT but with Alpaca

Language: Python - Size: 18.7 MB - Last synced at: 2 days ago - Pushed at: over 2 years ago - Stars: 60 - Forks: 6

MAGICS-LAB/GERM

[ICML 2025] Fast and Low-Cost Genomic Foundation Models via Outlier Removal.

Language: Python - Size: 21 MB - Last synced at: 4 days ago - Pushed at: 3 months ago - Stars: 17 - Forks: 2

PRITHIVSAKTHIUR/Sub-Memory-Efficient-Merging-FluxKreaDev

black-forest-labs/FLUX.1-dev and black-forest-labs/FLUX.1-Krea-dev. This merged model integrates the capabilities of the rectified flow transformer FLUX.1-dev, known for competitive prompt following and high-quality outputs, with FLUX.1-Krea-dev, a guidance distilled model emphasizing aesthetics and photorealism.

Language: Python - Size: 30.3 KB - Last synced at: 24 days ago - Pushed at: 25 days ago - Stars: 1 - Forks: 0

jackaduma/ChatGLM-LoRA-RLHF-PyTorch

A full pipeline to finetune ChatGLM LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the ChatGLM architecture. Basically ChatGPT but with ChatGLM

Language: Python - Size: 25.3 MB - Last synced at: 2 days ago - Pushed at: over 2 years ago - Stars: 138 - Forks: 10

elpikaoocle2006/Sub-Memory-Efficient-Merging-FluxKreaDev

Merge advanced image generation models FLUX.1-dev and FLUX.1-Krea-dev for improved text-to-image performance. Explore the repository for seamless integration! 🐙

Language: Python - Size: 17.6 KB - Last synced at: 22 days ago - Pushed at: 23 days ago - Stars: 0 - Forks: 0

Baijiong-Lin/LoRA-Torch

PyTorch Reimplementation of LoRA (featuring with supporting nn.MultiheadAttention in OpenCLIP)

Language: Python - Size: 60.5 KB - Last synced at: 13 days ago - Pushed at: 3 months ago - Stars: 67 - Forks: 7

PRITHIVSAKTHIUR/Sub-Memory-Efficient-Merging-FluxKreaSchnell

A memory-efficient approach to merge FLUX.1-schnell and FLUX.1-Krea-dev transformer models, combining their strengths while maintaining guidance capabilities from the schnell model.

Language: Python - Size: 11.7 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 1 - Forks: 0

arcforge-tune/bmg-lora

LoRA + QLoRA fine‑tuning toolkit optimized for Intel Arc Battlemage GPUs

Language: Python - Size: 131 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

robuno/Title-Generator-with-LLM-QLoRa

Fine-tuning LLMs with LoRA to generate titles from the given abstract, specifically for CS-related papers.

Language: Jupyter Notebook - Size: 254 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 7 - Forks: 0

Aisuko/notebooks

Implementation for the different ML tasks on Kaggle platform with GPUs.

Language: Jupyter Notebook - Size: 159 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 24 - Forks: 4

mominalix/LLM-Finetuning-Pipeline-LoRA-QLoRA

Production-ready pipeline for fine-tuning Large Language Models using LoRA/QLoRA techniques with MLflow experiment tracking,

Language: Python - Size: 70.3 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

PRITHIVSAKTHIUR/Flux-LoRA-DLC

Experience the power of the FLUX.1-dev diffusion model combined with a massive collection of 255+ community-created LoRAs! This Gradio application provides an easy-to-use interface to explore diverse artistic styles directly on top of the FLUX base model.

Language: Python - Size: 1.42 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 12 - Forks: 1

TUDB-Labs/MoE-PEFT

An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT

Language: Python - Size: 7.18 MB - Last synced at: about 1 month ago - Pushed at: 6 months ago - Stars: 107 - Forks: 16

wahabzh/lora-smollm-finetuning

⚙️ LoRA implementation for efficient SmolLM fine-tuning. Achieves comparable performance with only 0.24% trainable parameters.

Language: Jupyter Notebook - Size: 155 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

Ahmadhammam03/Galaxy-Chatbot

🤖 AI-powered chatbot for Galaxy Organisation and Alibaba Cloud Academy

Language: Jupyter Notebook - Size: 8.66 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 1 - Forks: 0

Samarth2001/LLM-Fine-tuning

Parameter-efficient fine-tuning experiments for 7B LLMs on consumer hardware. QLoRA implementations, memory optimization strategies, and reproducible benchmarks for Mistral, Llama-2, and other models on Google Colab T4 GPUs.

Language: Jupyter Notebook - Size: 63.5 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 1

AtharvaTaras/BankLlama

Code and files used during research for 'Fine Tuned Large Language Model In Banking'

Language: Jupyter Notebook - Size: 8.13 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

dev-kanika/llm-chatbot-multimodal-rag-bioasq

Multimodal chatbot project using five LLM-based approaches including RAG, PEFT, and prompt engineering, built on the BioASQ dataset for biomedical QA.

Language: Jupyter Notebook - Size: 544 KB - Last synced at: 24 days ago - Pushed at: about 1 month ago - Stars: 3 - Forks: 1

Showmick119/Fine-Tuning-Open-Source-LLM

Fine-Tuning the open source Code-LLama-7B model for code generation on FastAPI tasks.

Language: Python - Size: 7.78 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 1 - Forks: 0

calpt/awesome-adapter-resources

Collection of Tools and Papers related to Adapters / Parameter-Efficient Transfer Learning/ Fine-Tuning

Language: Python - Size: 213 KB - Last synced at: 6 days ago - Pushed at: over 1 year ago - Stars: 197 - Forks: 11

samay-jain/Fine_tuning_Distilbert_Model_using_LoRA_Low-Rank-Adaptation

Parameter-efficient fine-tuning of DistilBERT using LoRA for sentiment and topic classification, with CLI, API, and interactive chatbot interfaces.

Language: Python - Size: 171 MB - Last synced at: 24 days ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

ROIM1998/APT

[ICML'24 Oral] APT: Adaptive Pruning and Tuning Pretrained Language Models for Efficient Training and Inference

Language: Python - Size: 4.08 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 44 - Forks: 2

CatsMeow492/adaptive-lora-rank-allocation

This repository contains the experimental code for investigating whether jointly optimizing LoRA rank allocation and mixed-precision quantization yields better efficiency-performance trade-offs than existing baselines on laptop-class hardware.

Language: Python - Size: 1.16 MB - Last synced at: 5 days ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

pratzohol/MediTune

Fine-tuning LLMs on MedMCQA medical MCQ dataset

Language: Python - Size: 15.6 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

jackaduma/Vicuna-LoRA-RLHF-PyTorch

A full pipeline to finetune Vicuna LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the Vicuna architecture. Basically ChatGPT but with Vicuna

Language: Python - Size: 18.7 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 219 - Forks: 19

wiseodd/lapeft-bayesopt

Discrete Bayesian optimization with LLMs, PEFT finetuning methods, and the Laplace approximation.

Language: Python - Size: 4.85 MB - Last synced at: 3 days ago - Pushed at: about 1 year ago - Stars: 19 - Forks: 2

louisc-s/QLoRA-Fine-tuning-for-Film-Character-Styled-Responses-from-LLM

Code for fine-tuning Llama2 LLM with custom text dataset to produce film character styled responses

Language: Python - Size: 63.1 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 9 - Forks: 1

Expo314/Gemma-3n-local-training

Lightweight GPU framework for Gemma 3n training and fine-tuning. Ideal for chatbots and Q&A systems. 🚀💻 Optimize performance with LoRA and quantization.

Language: Python - Size: 155 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

llm-db/FineInfer

Deferred Continuous Batching in Resource-Efficient Large Language Model Serving (EuroMLSys 2024)

Language: Python - Size: 53.7 KB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 17 - Forks: 2

km1994/llms_paper

该仓库主要记录 LLMs 算法工程师相关的顶会论文研读笔记(多模态、PEFT、小样本QA问答、RAG、LMMs可解释性、Agents、CoT)

Size: 37.7 MB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 343 - Forks: 14

Rampati/MIT-6.S191-Lab3

In this lab, you will fine-tune a multi-billion parameter language model to generate specific style responses. You'll work with tokenization strategies, prompt templates, and a complete fine-tuning workflow to enhance LLM outputs. 🐙✨

Language: Jupyter Notebook - Size: 43 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 1 - Forks: 0

Aradhye2002/selective-peft-toolkit

Official implementation of the paper "Step-by-Step Unmasking for Parameter-Efficient Fine-tuning of Large Language Models"

Language: Python - Size: 43.9 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 9 - Forks: 0

paulinhok14/fine-tuning-llms

A lightweight toolkit for efficient quantization and LoRA‑based fine‑tuning of LLMs (Large Language Models).

Language: Jupyter Notebook - Size: 795 KB - Last synced at: about 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

Dee66/CodeCraft-AI

Cloud-native, AWS-native platform for Retrieval-Augmented Generation (RAG) and parameter-efficient fine-tuning (PEFT), built with full MLOps and Infrastructure as Code. Features FastAPI, Docker, AWS CDK, ECS Fargate, SageMaker, S3, Secrets Manager, CloudWatch, CI/CD with GitHub Actions, robust security, monitoring, automation for enterprise AI.

Language: Python - Size: 1.45 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

gazelle93/llm-fine-tuning-sft-lora-qlora

Practical examples for fine-tuning large language models (LLMs) with SFT, LoRA, and QLoRA using Hugging Face Transformers and PEFT.

Language: Python - Size: 0 Bytes - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

ankraj1234/MediGuide

Comparing QLoRA, Prompt & Prefix Tuning on Mistral-7B for medical instruction-following

Language: Jupyter Notebook - Size: 7.93 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 1 - Forks: 1

hiyouga/ChatGLM-Efficient-Tuning 📦

Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调

Language: Python - Size: 194 MB - Last synced at: 2 months ago - Pushed at: almost 2 years ago - Stars: 3,708 - Forks: 478

lijoraju/charitra-retrieval-augmented-qa

Retrieval-Augmented Q&A system for high school Social Studies, fine-tuned using QLoRA and deployed with FastAPI.

Language: Python - Size: 67.4 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

yspkm/lora-ft

This repository contains a collection of scripts I used to experiment with fine-tuning models like Llama and Gemma on various math and commonsense reasoning tasks using LoRA. My main field is AI-based antenna design for 6G, this is my side project for exploring LLM for fun.

Language: Python - Size: 74.2 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

aman-17/MediSOAP

FineTuning LLMs on conversational medical dataset.

Language: Jupyter Notebook - Size: 39.9 MB - Last synced at: 2 days ago - Pushed at: about 1 year ago - Stars: 4 - Forks: 0

Md-Emon-Hasan/Fine-Tuning

End-to-end fine-tuning of Hugging Face models using LoRA, QLoRA, quantization, and PEFT techniques. Optimized for low-memory with efficient model deployment

Language: Jupyter Notebook - Size: 5.53 MB - Last synced at: about 1 month ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

d0tTino/DeepThought-ReThought

A refactored version of the DeepThought Discord bot, focusing on improved architecture, performance, and AI agent capabilities.

Language: Python - Size: 20.3 MB - Last synced at: 7 days ago - Pushed at: 7 days ago - Stars: 1 - Forks: 1

FaNa-AI/MIT-6.S191-Lab3

MIT 6.S191 Lab 3 teaches you how to fine-tune large language models like Gemma 2B, structure prompts, and evaluate outputs using tools like Opik and LFM-40B.

Language: Jupyter Notebook - Size: 43.9 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

computational-cell-analytics/peft-sam

Parameter Efficient Fine-Tuning of Segment Anything Model

Language: Python - Size: 413 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 16 - Forks: 1

BorealisAI/flora-opt

This is the official repository for the paper "Flora: Low-Rank Adapters Are Secretly Gradient Compressors" in ICML 2024.

Language: Python - Size: 672 KB - Last synced at: 1 day ago - Pushed at: about 1 year ago - Stars: 104 - Forks: 5

Jiacheng-Zhu-AIML/AsymmetryLoRA

Preprint: Asymmetry in Low-Rank Adapters of Foundation Models

Language: Python - Size: 51.8 KB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 35 - Forks: 3

MaharshPatelX/HumanTouch

Transform AI-generated text into natural, human-like writing using DoRA fine-tuned Qwen models. Features interactive training modes, Google Colab support, and comprehensive data processing pipeline for 32k context humanization.

Language: Python - Size: 50.8 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

AMfeta99/NLP_LLM

This repository is dedicated to small projects and some theoretical material that I used to get into NLP and LLM in a practical and efficient way.

Language: Jupyter Notebook - Size: 85.2 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

Facico/GOAT-PEFT

[ICML2025] Make LoRA Great Again: Boosting LoRA with Adaptive Singular Values and Mixture-of-Experts Optimization Alignment

Language: Python - Size: 39 MB - Last synced at: 3 months ago - Pushed at: 4 months ago - Stars: 100 - Forks: 11

gokhaneraslan/XTTS_V2

Training XTTS V2 and PEFT LORA Text-to-Speech (TTS)

Language: Python - Size: 275 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

arjunravi26/ai_psychologist

Fine tune llama 2 model with mental health counseling data from huggingface

Language: Jupyter Notebook - Size: 14.5 MB - Last synced at: about 1 month ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

abdussahid26/LLM-Post-training-Techniques

This repository contains implementations of LLM post-training techniques, including SFT, PEFT, RLHF, PPO, DPO, and more.

Language: Jupyter Notebook - Size: 317 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

kriskrisliu/PAT

[AAAI 2025] PAT: Pruning-Aware Tuning for Large Language Models

Language: Python - Size: 30.2 MB - Last synced at: 2 months ago - Pushed at: 7 months ago - Stars: 32 - Forks: 1

PRITHIVSAKTHIUR/GALLO-3XL

High Quality Image Generation Model - Powered with NVIDIA A100

Language: Python - Size: 11.2 MB - Last synced at: 18 days ago - Pushed at: about 1 year ago - Stars: 13 - Forks: 1

zyds/transformers-code

手把手带你实战 Huggingface Transformers 课程视频同步更新在B站与YouTube

Language: Jupyter Notebook - Size: 76.5 MB - Last synced at: 3 months ago - Pushed at: about 1 year ago - Stars: 2,903 - Forks: 398

Ajairajv/Fine-Tuning-a-Generative-AI-Model

Fine-tunes FLAN-T5 for dialogue summarization using full fine-tuning and LoRA-based PEFT, evaluated with ROUGE metrics and human assessment.

Language: Jupyter Notebook - Size: 28.3 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0