Topic: "finetuning"
meta-llama/llama-cookbook
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama model family and using them on various provider services
Language: Jupyter Notebook - Size: 265 MB - Last synced at: 7 days ago - Pushed at: 9 days ago - Stars: 17,977 - Forks: 2,634
linkedin/Liger-Kernel
Efficient Triton Kernels for LLM Training
Language: Python - Size: 18.6 MB - Last synced at: 6 days ago - Pushed at: 6 days ago - Stars: 5,772 - Forks: 419
h2oai/h2o-llmstudio
H2O LLM Studio - a framework and no-code GUI for fine-tuning LLMs. Documentation: https://docs.h2o.ai/h2o-llmstudio/
Language: Python - Size: 54.4 MB - Last synced at: 4 days ago - Pushed at: about 1 month ago - Stars: 4,702 - Forks: 499
microsoft/FLAML
A fast library for AutoML and tuning. Join our Discord: https://discord.gg/Cppx2vSPVP.
Language: Jupyter Notebook - Size: 209 MB - Last synced at: about 2 months ago - Pushed at: 2 months ago - Stars: 4,206 - Forks: 544
Dataherald/dataherald
Interact with your SQL database, Natural Language to SQL using LLMs
Language: Python - Size: 4.34 MB - Last synced at: 5 months ago - Pushed at: over 1 year ago - Stars: 3,508 - Forks: 254
eosphoros-ai/Awesome-Text2SQL
Curated tutorials and resources for Large Language Models, Text2SQL, Text2DSL、Text2API、Text2Vis and more.
Size: 187 KB - Last synced at: 13 days ago - Pushed at: about 1 month ago - Stars: 3,201 - Forks: 224
LazyAGI/LazyLLM
Easiest and laziest way for building multi-agent LLMs applications.
Language: Python - Size: 11.1 MB - Last synced at: 2 days ago - Pushed at: 2 days ago - Stars: 3,100 - Forks: 289
learnables/learn2learn
A PyTorch Library for Meta-learning Research
Language: Python - Size: 9.52 MB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 2,837 - Forks: 365
stochasticai/xTuring
Build, personalize and control your own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6
Language: Python - Size: 18.3 MB - Last synced at: 5 days ago - Pushed at: 7 days ago - Stars: 2,660 - Forks: 207
SocialAI-tianji/Tianji
制作懂人情世故的大语言模型 | 涵盖提示词工程、RAG、Agent、LLM微调教程
Language: Python - Size: 8.3 MB - Last synced at: 10 days ago - Pushed at: 6 months ago - Stars: 1,568 - Forks: 127
jina-ai/finetuner 📦
:dart: Task-oriented embedding tuning for BERT, CLIP, etc.
Language: Python - Size: 71.5 MB - Last synced at: 13 days ago - Pushed at: over 1 year ago - Stars: 1,505 - Forks: 70
adithya-s-k/AI-Engineering.academy
Mastering Applied AI, One Concept at a Time
Language: Jupyter Notebook - Size: 96.9 MB - Last synced at: 28 days ago - Pushed at: 28 days ago - Stars: 1,069 - Forks: 118
georgian-io/LLM-Finetuning-Toolkit
Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.
Language: Python - Size: 32.7 MB - Last synced at: 5 days ago - Pushed at: about 1 year ago - Stars: 858 - Forks: 104
data-prep-kit/data-prep-kit
Open source project for data preparation for GenAI applications
Language: HTML - Size: 237 MB - Last synced at: 9 days ago - Pushed at: 9 days ago - Stars: 832 - Forks: 222
daswer123/xtts-webui
Webui for using XTTS and for finetuning it
Language: Python - Size: 2.76 MB - Last synced at: 5 months ago - Pushed at: 10 months ago - Stars: 808 - Forks: 158
minosvasilias/godot-dodo
Finetuning large language models for GDScript generation.
Language: Python - Size: 8.01 MB - Last synced at: 1 day ago - Pushed at: over 2 years ago - Stars: 556 - Forks: 26
junxia97/awesome-pretrain-on-molecules
[IJCAI 2023 survey track]A curated list of resources for chemical pre-trained models
Size: 565 KB - Last synced at: 7 days ago - Pushed at: over 2 years ago - Stars: 532 - Forks: 59
dvgodoy/FineTuningLLMs
Official repository of my book "A Hands-On Guide to Fine-Tuning LLMs with PyTorch and Hugging Face"
Language: Jupyter Notebook - Size: 7.81 MB - Last synced at: 24 days ago - Pushed at: 28 days ago - Stars: 527 - Forks: 70
helixml/helix
♾️ Helix is a private GenAI stack for building AI agents with declarative pipelines, knowledge (RAG), API bindings, and first-class testing.
Language: Go - Size: 104 MB - Last synced at: 5 days ago - Pushed at: 5 days ago - Stars: 522 - Forks: 58
kaito-project/aikit
🏗️ Fine-tune, build, and deploy open-source LLMs easily!
Language: Go - Size: 5.32 MB - Last synced at: about 10 hours ago - Pushed at: 3 days ago - Stars: 486 - Forks: 49
xing61/xiaoyi-robot
优质稳定的OpenAI的API接口-For企业和开发者。OpenAI的api proxy,支持ChatGPT的API调用,支持openai的API接口,支持:gpt-4,gpt-3.5。不需要openai Key, 不需要买openai的账号,不需要美元的银行卡,通通不用的,直接调用就行,稳定好用!!智增增
Language: PHP - Size: 384 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 465 - Forks: 35
Xirider/finetune-gpt2xl
Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed
Language: Python - Size: 5.44 MB - Last synced at: almost 2 years ago - Pushed at: over 2 years ago - Stars: 408 - Forks: 70
microsoft/AzureML-BERT
End-to-End recipes for pre-training and fine-tuning BERT using Azure Machine Learning Service
Language: Jupyter Notebook - Size: 314 KB - Last synced at: 13 days ago - Pushed at: over 2 years ago - Stars: 400 - Forks: 126
baidubce/bce-qianfan-sdk
Provide best practices for LMOps, as well as elegant and convenient access to the features of the Qianfan MaaS Platform. (提供大模型工具链最佳实践,以及优雅且便捷地访问千帆大模型平台)
Language: Jupyter Notebook - Size: 75.7 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 372 - Forks: 58
zjysteven/lmms-finetune
A minimal codebase for finetuning large multimodal models, supporting llava-1.5/1.6, llava-interleave, llava-next-video, llava-onevision, llama-3.2-vision, qwen-vl, qwen2-vl, phi3-v etc.
Language: Python - Size: 12.9 MB - Last synced at: 6 days ago - Pushed at: 6 days ago - Stars: 345 - Forks: 39
LHRLAB/ChatKBQA
[ACL 2024] Official resources of "ChatKBQA: A Generate-then-Retrieve Framework for Knowledge Base Question Answering with Fine-tuned Large Language Models".
Language: Python - Size: 18.5 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 316 - Forks: 27
ServiceNow/TapeAgents
TapeAgents is a framework that facilitates all stages of the LLM Agent development lifecycle
Language: Python - Size: 188 MB - Last synced at: 24 days ago - Pushed at: 24 days ago - Stars: 297 - Forks: 37
JosefAlbers/Phi-3-Vision-MLX
Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon
Language: Jupyter Notebook - Size: 5.95 MB - Last synced at: 2 months ago - Pushed at: about 1 year ago - Stars: 273 - Forks: 22
promptslab/LLMtuner
FineTune LLMs in few lines of code (Text2Text, Text2Speech, Speech2Text)
Language: Python - Size: 591 KB - Last synced at: 3 months ago - Pushed at: almost 2 years ago - Stars: 240 - Forks: 15
gyunggyung/KoGPT2-FineTuning
🔥 Korean GPT-2, KoGPT2 FineTuning cased. 한국어 가사 데이터 학습 🔥
Language: Python - Size: 24.6 MB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 228 - Forks: 56
git-disl/awesome_LLM-harmful-fine-tuning-papers
A survey on harmful fine-tuning attack for large language model
Size: 3.77 MB - Last synced at: 5 days ago - Pushed at: 5 days ago - Stars: 215 - Forks: 6
babycommando/neuralgraffiti
Live-bending a foundation model’s output at neural network level.
Language: Jupyter Notebook - Size: 31.3 KB - Last synced at: 7 months ago - Pushed at: 7 months ago - Stars: 212 - Forks: 16
rasbt/dora-from-scratch
LoRA and DoRA from Scratch Implementations
Language: Jupyter Notebook - Size: 41 KB - Last synced at: 7 months ago - Pushed at: over 1 year ago - Stars: 199 - Forks: 15
git-cloner/llama2-lora-fine-tuning
llama2 finetuning with deepspeed and lora
Language: Python - Size: 22.6 MB - Last synced at: 3 days ago - Pushed at: over 2 years ago - Stars: 176 - Forks: 14
codelion/ellora
Enhancing LLMs with LoRA
Language: Jupyter Notebook - Size: 2.19 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 169 - Forks: 13
Snap-gen/Snapgen
🏗️ Build, fine-tune, and run generative models locally!
Language: Go - Size: 3.56 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 159 - Forks: 42
woctezuma/finetune-detr
Fine-tune Facebook's DETR (DEtection TRansformer) on Colaboratory.
Language: Jupyter Notebook - Size: 79.5 MB - Last synced at: 3 months ago - Pushed at: over 2 years ago - Stars: 150 - Forks: 24
NVIDIA-NeMo/Automodel
Pytorch DTensor native training library for LLMs/VLMs with OOTB Hugging Face support
Language: Python - Size: 8.05 MB - Last synced at: 4 days ago - Pushed at: 4 days ago - Stars: 141 - Forks: 17
git-cloner/llama-lora-fine-tuning
llama fine-tuning with lora
Language: Python - Size: 109 KB - Last synced at: 5 days ago - Pushed at: over 1 year ago - Stars: 140 - Forks: 15
kuutsav/llm-toys
Small finetuned LLMs for a diverse set of useful tasks
Language: Python - Size: 72.6 MB - Last synced at: about 1 month ago - Pushed at: over 2 years ago - Stars: 128 - Forks: 6
altengineer/awesome-ai-repositories
A curated list of open source repositories for AI Engineers
Size: 178 KB - Last synced at: 7 days ago - Pushed at: 8 months ago - Stars: 117 - Forks: 21
kyegomez/Lets-Verify-Step-by-Step
"Improving Mathematical Reasoning with Process Supervision" by OPENAI
Language: Python - Size: 52.7 KB - Last synced at: 5 days ago - Pushed at: 13 days ago - Stars: 111 - Forks: 11
Trainy-ai/llm-atc 📦
Fine-tuning and serving LLMs on any cloud
Language: Python - Size: 1.71 MB - Last synced at: 4 days ago - Pushed at: almost 2 years ago - Stars: 89 - Forks: 2
yifanzhang-pro/AutoMathText
Official implementation of ACL 2025 Findings paper "Autonomous Data Selection with Zero-shot Generative Classifiers for Mathematical Texts" (As Huggingface Daily Papers: https://huggingface.co/papers/2402.07625)
Language: Python - Size: 1.84 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 84 - Forks: 5
Oqura-ai/local-datagen-cli
synthetic dataset generation workflow using local file resources for finetuning llms.
Language: Python - Size: 2.82 MB - Last synced at: 8 days ago - Pushed at: 24 days ago - Stars: 80 - Forks: 8
kamalkraj/e5-mistral-7b-instruct
Finetune mistral-7b-instruct for sentence embeddings
Language: Python - Size: 34.2 KB - Last synced at: 5 months ago - Pushed at: over 1 year ago - Stars: 80 - Forks: 18
924973292/MambaPro
【AAAI2025】MambaPro: Multi-Modal Object Re-Identification with Mamba Aggregation and Synergistic Prompt
Language: Python - Size: 24.2 MB - Last synced at: about 2 months ago - Pushed at: 6 months ago - Stars: 77 - Forks: 3
GURPREETKAURJETHRA/Generative-AI-LLM-Projects
Gen AI Large Language Model Projects
Language: Jupyter Notebook - Size: 23 MB - Last synced at: 3 days ago - Pushed at: over 1 year ago - Stars: 77 - Forks: 22
LennartPurucker/finetune_tabpfn_v2
Code for finetuning TabPFN on one downstream tabular dataset.
Language: Python - Size: 67.4 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 76 - Forks: 14
Azure-Samples/azureai-foundry-finetuning-raft
A recipe that will walk you through using either Meta Llama 3.1 405B or OpenAI GPT-4o deployed on Azure AI to generate a synthetic dataset using UC Berkeley's Gorilla project RAFT method.
Language: Jupyter Notebook - Size: 41.2 MB - Last synced at: 17 days ago - Pushed at: 4 months ago - Stars: 74 - Forks: 26
Baijiong-Lin/LoRA-Torch
PyTorch Reimplementation of LoRA (featuring with supporting nn.MultiheadAttention in OpenCLIP)
Language: Python - Size: 60.5 KB - Last synced at: 20 days ago - Pushed at: 5 months ago - Stars: 72 - Forks: 7
LongxingTan/open-retrievals
All-in-One: Text Embedding, Retrieval, Reranking and RAG in Transformers
Language: Python - Size: 1.42 MB - Last synced at: 26 days ago - Pushed at: 3 months ago - Stars: 69 - Forks: 13
goodreasonai/praetor-data
Praetor is a lightweight finetuning data and prompt management tool
Language: Python - Size: 6.89 MB - Last synced at: 4 months ago - Pushed at: 12 months ago - Stars: 67 - Forks: 0
speediedan/finetuning-scheduler
A PyTorch Lightning extension that accelerates and enhances foundation model experimentation with flexible fine-tuning schedules.
Language: Python - Size: 2.74 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 65 - Forks: 6
zou-group/sirius
SiriuS: Self-improving Multi-agent Systems via Bootstrapped Reasoning
Language: Python - Size: 70.3 KB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 60 - Forks: 5
kyegomez/Finetuning-Suite
Finetune any model on HF in less than 30 seconds
Language: Jupyter Notebook - Size: 95.4 MB - Last synced at: 6 months ago - Pushed at: 7 months ago - Stars: 58 - Forks: 7
unit-mesh/unit-gen
UnitGen 是一个用于生成微调代码的数据框架 —— 直接从你的代码库中生成微调数据:代码补全、测试生成、文档生成等。UnitGen is a code fine-tuning data framework that generates data from your existing codebase.
Language: Kotlin - Size: 1.26 MB - Last synced at: 6 days ago - Pushed at: over 1 year ago - Stars: 58 - Forks: 11
sayedmohamedscu/Vision-language-models-VLM
vision language models finetuning notebooks & use cases (Medgemma - paligemma - florence .....)
Language: Jupyter Notebook - Size: 16.5 MB - Last synced at: 27 days ago - Pushed at: 27 days ago - Stars: 52 - Forks: 10
linhlpv/awesome-offline-to-online-RL-papers
A list of Offline to Online RL papers (continually updated)
Size: 16.6 KB - Last synced at: 7 days ago - Pushed at: about 1 year ago - Stars: 52 - Forks: 0
MohamedSebaie/Fight_Detection_From_Surveillance_Cameras-PyTorch_Project
Fight Detection From Surveillance Cameras by fine-tuning a PyTorch Pretrained Model
Language: Jupyter Notebook - Size: 208 MB - Last synced at: about 1 month ago - Pushed at: over 3 years ago - Stars: 51 - Forks: 14
deshwalmahesh/PHUDGE
Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute, relative and much more. It contains a list of all the available tool, methods, repo, code etc to detect hallucination, LLM evaluation, grading and much more.
Language: Jupyter Notebook - Size: 13.1 MB - Last synced at: 5 months ago - Pushed at: over 1 year ago - Stars: 49 - Forks: 7
chuangchuangtan/LLaVA-NeXT-Image-Llama3-Lora
LLaVA-NeXT-Image-Llama3-Lora, Modified from https://github.com/arielnlee/LLaVA-1.6-ft
Language: Python - Size: 11.6 MB - Last synced at: 4 days ago - Pushed at: over 1 year ago - Stars: 44 - Forks: 4
adithya-s-k/CompanionLLM
CompanionLLM - A framework to finetune LLMs to be your own sentient conversational companion
Language: Jupyter Notebook - Size: 40.1 MB - Last synced at: 4 days ago - Pushed at: almost 2 years ago - Stars: 44 - Forks: 5
poloclub/Fine-tuning-LLMs
Finetune Llama 2 on Colab for free on your own data: step-by-step tutorial
Language: Jupyter Notebook - Size: 9.12 MB - Last synced at: 5 months ago - Pushed at: over 1 year ago - Stars: 43 - Forks: 22
HenryNdubuaku/super-lazy-autograd
Hand-derived memory-efficient super lazy PyTorch VJPs for training LLMs on laptop, all using one op (bundled scaled matmuls).
Language: Python - Size: 1.32 MB - Last synced at: 6 days ago - Pushed at: 7 months ago - Stars: 41 - Forks: 0
conneroisu/Text-Dataset-Aid-Plugin
This is a obsidian plugin to help with the creation of personal jsonl datasets for text generation models.
Language: TypeScript - Size: 157 KB - Last synced at: 4 months ago - Pushed at: almost 2 years ago - Stars: 38 - Forks: 3
git-cloner/Llama2-chinese
Llama2 chinese finetuning
Language: Python - Size: 65.4 KB - Last synced at: over 1 year ago - Pushed at: over 2 years ago - Stars: 37 - Forks: 8
CogitoNTNU/TutorAI
TutorAI is a RAG system capable of assisting with learning academic subjects and using the curriculum and citing it. The project revolves around building an application that ingests a textbook in most formats and facilitates efficient learning of the course material.
Language: Python - Size: 20.8 MB - Last synced at: 6 days ago - Pushed at: 6 days ago - Stars: 36 - Forks: 14
MaxiDonkey/DelphiGemini
The Gemini API wrapper for Delphi utilizes advanced models developed by Google to provide robust capabilities, including interactive chat, text embeddings, code generation, image and video prompting, audio analysis and transcription, fine-tuning, caching, and integration with Google Search.
Language: Pascal - Size: 216 KB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 36 - Forks: 4
machinelearningnuremberg/QuickTune
[ICLR2024] Quick-Tune: Quickly Learning Which Pretrained Model to Finetune and How
Language: Python - Size: 5.86 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 33 - Forks: 4
avocardio/Zicklein
Finetuning instruct-LLaMA on german datasets.
Language: Python - Size: 9.78 MB - Last synced at: 5 days ago - Pushed at: about 2 years ago - Stars: 33 - Forks: 5
VatsaDev/nanoChatGPT
nanogpt turned into a chat model
Language: Python - Size: 266 MB - Last synced at: almost 2 years ago - Pushed at: about 2 years ago - Stars: 33 - Forks: 5
omerbsezer/Fast-LLM-Agent-MCP
This repo covers LLM, Agents concepts both theoretically and practically: LLMs, RAG, Fine Tuning, Agents, Tools, MCP, AWS Strands Agents, Google Agent Development Kit, ADK, Reference Documents, etc.
Language: Python - Size: 65.3 MB - Last synced at: about 2 months ago - Pushed at: 3 months ago - Stars: 30 - Forks: 8
Cre4T3Tiv3/unsloth-llama3-alpaca-lora
Advanced 4-bit QLoRA fine-tuning pipeline for LLaMA 3 8B with production-grade optimization. Memory-efficient training on consumer GPUs for instruction-following specialization. Demonstrates cutting-edge parameter-efficient fine-tuning with Unsloth integration.
Language: Jupyter Notebook - Size: 2.11 MB - Last synced at: 12 days ago - Pushed at: 4 months ago - Stars: 30 - Forks: 0
ssbuild/chatglm_rlhf
chatglm_rlhf_finetuning
Language: Python - Size: 149 KB - Last synced at: 5 days ago - Pushed at: about 2 years ago - Stars: 30 - Forks: 1
paulocoutinhox/mini-llm
Simple and lightweight tool to fine-tune GPT models (like GPT-2 and GPT-Neo) using your own data — built with Python and Transformers. Adapt powerful language models to your domain with ease.
Language: Python - Size: 89.8 KB - Last synced at: about 1 month ago - Pushed at: 6 months ago - Stars: 28 - Forks: 1
francoislanc/midistral
LLM finetuned for generating symbolic music
Language: Python - Size: 1.52 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 28 - Forks: 2
microsoft/Build25-LAB329
Fine-Tune End-to-End Distillation Models with Azure AI Foundry Models and Foundry Local
Language: Jupyter Notebook - Size: 1.4 MB - Last synced at: 13 days ago - Pushed at: 3 months ago - Stars: 27 - Forks: 12
neph1/finetrainers-ui
Gradio UI for training video models using finetrainers
Language: Python - Size: 103 KB - Last synced at: 7 months ago - Pushed at: 7 months ago - Stars: 27 - Forks: 2
dannylee1020/openpo
Building synthetic data for preference tuning
Language: Python - Size: 10.7 MB - Last synced at: 23 days ago - Pushed at: 10 months ago - Stars: 27 - Forks: 0
computational-cell-analytics/peft-sam
Parameter Efficient Fine-Tuning of Segment Anything Model
Language: Python - Size: 416 KB - Last synced at: 25 days ago - Pushed at: 25 days ago - Stars: 24 - Forks: 2
MaxiDonkey/DelphiMistralAI
DelphiMistralAI wrapper brings Mistral’s text-vision-audio models and agentic Conversations to Delphi, with chat, embeddings, Codestral codegen, fine-tuning, batching, moderation, async/await helpers and live request monitoring.
Language: Pascal - Size: 1.79 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 23 - Forks: 5
Shaurya-Sethi/transqlate
End-to-end natural language to SQL system: schema-aware model fine-tuning, retrieval-augmented prompting, and production-grade CLI, powered by a custom fine-tuned Phi-4 Mini.
Language: Python - Size: 1.7 MB - Last synced at: 2 months ago - Pushed at: 3 months ago - Stars: 23 - Forks: 1
chainstacklabs/web3-ai-trading-agent
Build an Autonomous Web3 AI Trading Agent (BASE + Uniswap V4 example)
Language: Python - Size: 1.11 MB - Last synced at: 2 months ago - Pushed at: 4 months ago - Stars: 23 - Forks: 5
shaheennabi/Production-Ready-Instruction-Finetuning-of-Meta-Llama-3.2-3B-Instruct-Project
Instruction Fine-Tuning of Meta Llama 3.2-3B Instruct on Kannada Conversations. Tailoring the model to follow specific instructions in Kannada, enhancing its ability to generate relevant, context-aware responses based on conversational inputs. Using the Kannada Instruct dataset for fine-tuning! Happy Finetuning 🎋
Language: Jupyter Notebook - Size: 692 KB - Last synced at: 8 days ago - Pushed at: 9 months ago - Stars: 23 - Forks: 6
meaningalignment/dft
Democratic Fine-tuning with a Moral Graph
Language: TypeScript - Size: 10 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 22 - Forks: 9
zhaoyl18/SEIKO
SEIKO is a novel reinforcement learning method to efficiently fine-tune diffusion models in an online setting. Our methods outperform all baselines (PPO, classifier-based guidance, direct reward backpropagation) for fine-tuning Stable Diffusion.
Language: Python - Size: 3.85 MB - Last synced at: 4 months ago - Pushed at: over 1 year ago - Stars: 22 - Forks: 0
adithya-s-k/Indic-llm
A open-source framework designed to adapt pre-trained Language Models (LLMs), such as Llama, Mistral, and Mixtral, to a wide array of domains and languages.
Language: Python - Size: 171 KB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 21 - Forks: 1
SIC98/GPT2-python-code-generator
GPT2 finetuning with transformers 🤗
Language: Jupyter Notebook - Size: 185 KB - Last synced at: over 2 years ago - Pushed at: over 4 years ago - Stars: 21 - Forks: 2
IBM/AutoVP
[ICLR24] AutoVP: An Automated Visual Prompting Framework and Benchmark
Language: Python - Size: 579 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 20 - Forks: 2
Hemanthkumar2112/Reward-Modeling-RLHF-Finetune-and-RAG
Gemma2(9B), Llama3-8B-Finetune-and-RAG, code base for sample, implemented in Kaggle platform
Language: Jupyter Notebook - Size: 18.6 MB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 20 - Forks: 6
lzzcd001/nabla-gfn
Official Implementation of Nabla-GFlowNet (ICLR 2025)
Language: Python - Size: 4.27 MB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 19 - Forks: 0
ThomasRochefortB/open-agentinstruct
An open-source recreation of the AgentInstruct agentic workflow for synthetic data generation
Language: Python - Size: 372 KB - Last synced at: 2 months ago - Pushed at: 7 months ago - Stars: 19 - Forks: 3
git-disl/Booster
This is the official code for the paper "Booster: Tackling Harmful Fine-tuning for Large Language Models via Attenuating Harmful Perturbation" (ICLR2025).
Language: Shell - Size: 293 KB - Last synced at: 8 months ago - Pushed at: 8 months ago - Stars: 19 - Forks: 0
git-disl/Vaccine
This is the official code for the paper "Vaccine: Perturbation-aware Alignment for Large Language Models" (NeurIPS2024)
Language: Shell - Size: 730 KB - Last synced at: 12 months ago - Pushed at: 12 months ago - Stars: 19 - Forks: 0
HomoScriptor-Project/HomoScriptor
Fuel innovation and advance language models with HomoScriptor: A vibrant, community-driven dataset for fine-tuning large language models.
Size: 63.5 KB - Last synced at: 11 months ago - Pushed at: about 2 years ago - Stars: 18 - Forks: 3
Raumberg/myllm
Multi-node distributed LLM training framework
Language: Python - Size: 1.66 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 17 - Forks: 1
itspranavajay/Merge-Diffusion-Tool
Merge Diffusion Tool is an open-source solution for merging LoRA models, integrating LoRA into checkpoints, and blending Flux And Stable Diffusion models (SD1.5, SD2, SD3, SDXL). Optimize your AI workflows with ease.
Language: Python - Size: 24.4 KB - Last synced at: 7 months ago - Pushed at: about 1 year ago - Stars: 17 - Forks: 0
ruimalheiro/training-custom-llama
Llama-style transformer in PyTorch with multi-node / multi-GPU training. Includes pretraining, SFT, DPO, LoRA, and knowledge distillation. Scripts for dataset mixing and training from scratch.
Language: Python - Size: 1.32 MB - Last synced at: 12 days ago - Pushed at: 12 days ago - Stars: 16 - Forks: 3
utahnlp/structured_tuning_srl
Implementation of our ACL 2020 paper: Structured Tuning for Semantic Role Labeling
Language: Python - Size: 802 KB - Last synced at: 11 months ago - Pushed at: over 1 year ago - Stars: 16 - Forks: 2