An open API service providing repository metadata for many open source software ecosystems.

Topic: "finetuning"

meta-llama/llama-cookbook

Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama model family and using them on various provider services

Language: Jupyter Notebook - Size: 265 MB - Last synced at: 7 days ago - Pushed at: 9 days ago - Stars: 17,977 - Forks: 2,634

linkedin/Liger-Kernel

Efficient Triton Kernels for LLM Training

Language: Python - Size: 18.6 MB - Last synced at: 6 days ago - Pushed at: 6 days ago - Stars: 5,772 - Forks: 419

h2oai/h2o-llmstudio

H2O LLM Studio - a framework and no-code GUI for fine-tuning LLMs. Documentation: https://docs.h2o.ai/h2o-llmstudio/

Language: Python - Size: 54.4 MB - Last synced at: 4 days ago - Pushed at: about 1 month ago - Stars: 4,702 - Forks: 499

microsoft/FLAML

A fast library for AutoML and tuning. Join our Discord: https://discord.gg/Cppx2vSPVP.

Language: Jupyter Notebook - Size: 209 MB - Last synced at: about 2 months ago - Pushed at: 2 months ago - Stars: 4,206 - Forks: 544

Dataherald/dataherald

Interact with your SQL database, Natural Language to SQL using LLMs

Language: Python - Size: 4.34 MB - Last synced at: 5 months ago - Pushed at: over 1 year ago - Stars: 3,508 - Forks: 254

eosphoros-ai/Awesome-Text2SQL

Curated tutorials and resources for Large Language Models, Text2SQL, Text2DSL、Text2API、Text2Vis and more.

Size: 187 KB - Last synced at: 13 days ago - Pushed at: about 1 month ago - Stars: 3,201 - Forks: 224

LazyAGI/LazyLLM

Easiest and laziest way for building multi-agent LLMs applications.

Language: Python - Size: 11.1 MB - Last synced at: 2 days ago - Pushed at: 2 days ago - Stars: 3,100 - Forks: 289

learnables/learn2learn

A PyTorch Library for Meta-learning Research

Language: Python - Size: 9.52 MB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 2,837 - Forks: 365

stochasticai/xTuring

Build, personalize and control your own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6

Language: Python - Size: 18.3 MB - Last synced at: 5 days ago - Pushed at: 7 days ago - Stars: 2,660 - Forks: 207

SocialAI-tianji/Tianji

制作懂人情世故的大语言模型 | 涵盖提示词工程、RAG、Agent、LLM微调教程

Language: Python - Size: 8.3 MB - Last synced at: 10 days ago - Pushed at: 6 months ago - Stars: 1,568 - Forks: 127

jina-ai/finetuner 📦

:dart: Task-oriented embedding tuning for BERT, CLIP, etc.

Language: Python - Size: 71.5 MB - Last synced at: 13 days ago - Pushed at: over 1 year ago - Stars: 1,505 - Forks: 70

adithya-s-k/AI-Engineering.academy

Mastering Applied AI, One Concept at a Time

Language: Jupyter Notebook - Size: 96.9 MB - Last synced at: 28 days ago - Pushed at: 28 days ago - Stars: 1,069 - Forks: 118

georgian-io/LLM-Finetuning-Toolkit

Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.

Language: Python - Size: 32.7 MB - Last synced at: 5 days ago - Pushed at: about 1 year ago - Stars: 858 - Forks: 104

data-prep-kit/data-prep-kit

Open source project for data preparation for GenAI applications

Language: HTML - Size: 237 MB - Last synced at: 9 days ago - Pushed at: 9 days ago - Stars: 832 - Forks: 222

daswer123/xtts-webui

Webui for using XTTS and for finetuning it

Language: Python - Size: 2.76 MB - Last synced at: 5 months ago - Pushed at: 10 months ago - Stars: 808 - Forks: 158

minosvasilias/godot-dodo

Finetuning large language models for GDScript generation.

Language: Python - Size: 8.01 MB - Last synced at: 1 day ago - Pushed at: over 2 years ago - Stars: 556 - Forks: 26

junxia97/awesome-pretrain-on-molecules

[IJCAI 2023 survey track]A curated list of resources for chemical pre-trained models

Size: 565 KB - Last synced at: 7 days ago - Pushed at: over 2 years ago - Stars: 532 - Forks: 59

dvgodoy/FineTuningLLMs

Official repository of my book "A Hands-On Guide to Fine-Tuning LLMs with PyTorch and Hugging Face"

Language: Jupyter Notebook - Size: 7.81 MB - Last synced at: 24 days ago - Pushed at: 28 days ago - Stars: 527 - Forks: 70

helixml/helix

♾️ Helix is a private GenAI stack for building AI agents with declarative pipelines, knowledge (RAG), API bindings, and first-class testing.

Language: Go - Size: 104 MB - Last synced at: 5 days ago - Pushed at: 5 days ago - Stars: 522 - Forks: 58

kaito-project/aikit

🏗️ Fine-tune, build, and deploy open-source LLMs easily!

Language: Go - Size: 5.32 MB - Last synced at: about 10 hours ago - Pushed at: 3 days ago - Stars: 486 - Forks: 49

xing61/xiaoyi-robot

优质稳定的OpenAI的API接口-For企业和开发者。OpenAI的api proxy,支持ChatGPT的API调用,支持openai的API接口,支持:gpt-4,gpt-3.5。不需要openai Key, 不需要买openai的账号,不需要美元的银行卡,通通不用的,直接调用就行,稳定好用!!智增增

Language: PHP - Size: 384 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 465 - Forks: 35

Xirider/finetune-gpt2xl

Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed

Language: Python - Size: 5.44 MB - Last synced at: almost 2 years ago - Pushed at: over 2 years ago - Stars: 408 - Forks: 70

microsoft/AzureML-BERT

End-to-End recipes for pre-training and fine-tuning BERT using Azure Machine Learning Service

Language: Jupyter Notebook - Size: 314 KB - Last synced at: 13 days ago - Pushed at: over 2 years ago - Stars: 400 - Forks: 126

baidubce/bce-qianfan-sdk

Provide best practices for LMOps, as well as elegant and convenient access to the features of the Qianfan MaaS Platform. (提供大模型工具链最佳实践,以及优雅且便捷地访问千帆大模型平台)

Language: Jupyter Notebook - Size: 75.7 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 372 - Forks: 58

zjysteven/lmms-finetune

A minimal codebase for finetuning large multimodal models, supporting llava-1.5/1.6, llava-interleave, llava-next-video, llava-onevision, llama-3.2-vision, qwen-vl, qwen2-vl, phi3-v etc.

Language: Python - Size: 12.9 MB - Last synced at: 6 days ago - Pushed at: 6 days ago - Stars: 345 - Forks: 39

LHRLAB/ChatKBQA

[ACL 2024] Official resources of "ChatKBQA: A Generate-then-Retrieve Framework for Knowledge Base Question Answering with Fine-tuned Large Language Models".

Language: Python - Size: 18.5 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 316 - Forks: 27

ServiceNow/TapeAgents

TapeAgents is a framework that facilitates all stages of the LLM Agent development lifecycle

Language: Python - Size: 188 MB - Last synced at: 24 days ago - Pushed at: 24 days ago - Stars: 297 - Forks: 37

JosefAlbers/Phi-3-Vision-MLX

Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon

Language: Jupyter Notebook - Size: 5.95 MB - Last synced at: 2 months ago - Pushed at: about 1 year ago - Stars: 273 - Forks: 22

promptslab/LLMtuner

FineTune LLMs in few lines of code (Text2Text, Text2Speech, Speech2Text)

Language: Python - Size: 591 KB - Last synced at: 3 months ago - Pushed at: almost 2 years ago - Stars: 240 - Forks: 15

gyunggyung/KoGPT2-FineTuning

🔥 Korean GPT-2, KoGPT2 FineTuning cased. 한국어 가사 데이터 학습 🔥

Language: Python - Size: 24.6 MB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 228 - Forks: 56

git-disl/awesome_LLM-harmful-fine-tuning-papers

A survey on harmful fine-tuning attack for large language model

Size: 3.77 MB - Last synced at: 5 days ago - Pushed at: 5 days ago - Stars: 215 - Forks: 6

babycommando/neuralgraffiti

Live-bending a foundation model’s output at neural network level.

Language: Jupyter Notebook - Size: 31.3 KB - Last synced at: 7 months ago - Pushed at: 7 months ago - Stars: 212 - Forks: 16

rasbt/dora-from-scratch

LoRA and DoRA from Scratch Implementations

Language: Jupyter Notebook - Size: 41 KB - Last synced at: 7 months ago - Pushed at: over 1 year ago - Stars: 199 - Forks: 15

git-cloner/llama2-lora-fine-tuning

llama2 finetuning with deepspeed and lora

Language: Python - Size: 22.6 MB - Last synced at: 3 days ago - Pushed at: over 2 years ago - Stars: 176 - Forks: 14

codelion/ellora

Enhancing LLMs with LoRA

Language: Jupyter Notebook - Size: 2.19 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 169 - Forks: 13

Snap-gen/Snapgen

🏗️ Build, fine-tune, and run generative models locally!

Language: Go - Size: 3.56 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 159 - Forks: 42

woctezuma/finetune-detr

Fine-tune Facebook's DETR (DEtection TRansformer) on Colaboratory.

Language: Jupyter Notebook - Size: 79.5 MB - Last synced at: 3 months ago - Pushed at: over 2 years ago - Stars: 150 - Forks: 24

NVIDIA-NeMo/Automodel

Pytorch DTensor native training library for LLMs/VLMs with OOTB Hugging Face support

Language: Python - Size: 8.05 MB - Last synced at: 4 days ago - Pushed at: 4 days ago - Stars: 141 - Forks: 17

git-cloner/llama-lora-fine-tuning

llama fine-tuning with lora

Language: Python - Size: 109 KB - Last synced at: 5 days ago - Pushed at: over 1 year ago - Stars: 140 - Forks: 15

kuutsav/llm-toys

Small finetuned LLMs for a diverse set of useful tasks

Language: Python - Size: 72.6 MB - Last synced at: about 1 month ago - Pushed at: over 2 years ago - Stars: 128 - Forks: 6

altengineer/awesome-ai-repositories

A curated list of open source repositories for AI Engineers

Size: 178 KB - Last synced at: 7 days ago - Pushed at: 8 months ago - Stars: 117 - Forks: 21

kyegomez/Lets-Verify-Step-by-Step

"Improving Mathematical Reasoning with Process Supervision" by OPENAI

Language: Python - Size: 52.7 KB - Last synced at: 5 days ago - Pushed at: 13 days ago - Stars: 111 - Forks: 11

Trainy-ai/llm-atc 📦

Fine-tuning and serving LLMs on any cloud

Language: Python - Size: 1.71 MB - Last synced at: 4 days ago - Pushed at: almost 2 years ago - Stars: 89 - Forks: 2

yifanzhang-pro/AutoMathText

Official implementation of ACL 2025 Findings paper "Autonomous Data Selection with Zero-shot Generative Classifiers for Mathematical Texts" (As Huggingface Daily Papers: https://huggingface.co/papers/2402.07625)

Language: Python - Size: 1.84 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 84 - Forks: 5

Oqura-ai/local-datagen-cli

synthetic dataset generation workflow using local file resources for finetuning llms.

Language: Python - Size: 2.82 MB - Last synced at: 8 days ago - Pushed at: 24 days ago - Stars: 80 - Forks: 8

kamalkraj/e5-mistral-7b-instruct

Finetune mistral-7b-instruct for sentence embeddings

Language: Python - Size: 34.2 KB - Last synced at: 5 months ago - Pushed at: over 1 year ago - Stars: 80 - Forks: 18

924973292/MambaPro

【AAAI2025】MambaPro: Multi-Modal Object Re-Identification with Mamba Aggregation and Synergistic Prompt

Language: Python - Size: 24.2 MB - Last synced at: about 2 months ago - Pushed at: 6 months ago - Stars: 77 - Forks: 3

GURPREETKAURJETHRA/Generative-AI-LLM-Projects

Gen AI Large Language Model Projects

Language: Jupyter Notebook - Size: 23 MB - Last synced at: 3 days ago - Pushed at: over 1 year ago - Stars: 77 - Forks: 22

LennartPurucker/finetune_tabpfn_v2

Code for finetuning TabPFN on one downstream tabular dataset.

Language: Python - Size: 67.4 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 76 - Forks: 14

Azure-Samples/azureai-foundry-finetuning-raft

A recipe that will walk you through using either Meta Llama 3.1 405B or OpenAI GPT-4o deployed on Azure AI to generate a synthetic dataset using UC Berkeley's Gorilla project RAFT method.

Language: Jupyter Notebook - Size: 41.2 MB - Last synced at: 17 days ago - Pushed at: 4 months ago - Stars: 74 - Forks: 26

Baijiong-Lin/LoRA-Torch

PyTorch Reimplementation of LoRA (featuring with supporting nn.MultiheadAttention in OpenCLIP)

Language: Python - Size: 60.5 KB - Last synced at: 20 days ago - Pushed at: 5 months ago - Stars: 72 - Forks: 7

LongxingTan/open-retrievals

All-in-One: Text Embedding, Retrieval, Reranking and RAG in Transformers

Language: Python - Size: 1.42 MB - Last synced at: 26 days ago - Pushed at: 3 months ago - Stars: 69 - Forks: 13

goodreasonai/praetor-data

Praetor is a lightweight finetuning data and prompt management tool

Language: Python - Size: 6.89 MB - Last synced at: 4 months ago - Pushed at: 12 months ago - Stars: 67 - Forks: 0

speediedan/finetuning-scheduler

A PyTorch Lightning extension that accelerates and enhances foundation model experimentation with flexible fine-tuning schedules.

Language: Python - Size: 2.74 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 65 - Forks: 6

zou-group/sirius

SiriuS: Self-improving Multi-agent Systems via Bootstrapped Reasoning

Language: Python - Size: 70.3 KB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 60 - Forks: 5

kyegomez/Finetuning-Suite

Finetune any model on HF in less than 30 seconds

Language: Jupyter Notebook - Size: 95.4 MB - Last synced at: 6 months ago - Pushed at: 7 months ago - Stars: 58 - Forks: 7

unit-mesh/unit-gen

UnitGen 是一个用于生成微调代码的数据框架 —— 直接从你的代码库中生成微调数据:代码补全、测试生成、文档生成等。UnitGen is a code fine-tuning data framework that generates data from your existing codebase.

Language: Kotlin - Size: 1.26 MB - Last synced at: 6 days ago - Pushed at: over 1 year ago - Stars: 58 - Forks: 11

sayedmohamedscu/Vision-language-models-VLM

vision language models finetuning notebooks & use cases (Medgemma - paligemma - florence .....)

Language: Jupyter Notebook - Size: 16.5 MB - Last synced at: 27 days ago - Pushed at: 27 days ago - Stars: 52 - Forks: 10

linhlpv/awesome-offline-to-online-RL-papers

A list of Offline to Online RL papers (continually updated)

Size: 16.6 KB - Last synced at: 7 days ago - Pushed at: about 1 year ago - Stars: 52 - Forks: 0

MohamedSebaie/Fight_Detection_From_Surveillance_Cameras-PyTorch_Project

Fight Detection From Surveillance Cameras by fine-tuning a PyTorch Pretrained Model

Language: Jupyter Notebook - Size: 208 MB - Last synced at: about 1 month ago - Pushed at: over 3 years ago - Stars: 51 - Forks: 14

deshwalmahesh/PHUDGE

Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute, relative and much more. It contains a list of all the available tool, methods, repo, code etc to detect hallucination, LLM evaluation, grading and much more.

Language: Jupyter Notebook - Size: 13.1 MB - Last synced at: 5 months ago - Pushed at: over 1 year ago - Stars: 49 - Forks: 7

chuangchuangtan/LLaVA-NeXT-Image-Llama3-Lora

LLaVA-NeXT-Image-Llama3-Lora, Modified from https://github.com/arielnlee/LLaVA-1.6-ft

Language: Python - Size: 11.6 MB - Last synced at: 4 days ago - Pushed at: over 1 year ago - Stars: 44 - Forks: 4

adithya-s-k/CompanionLLM

CompanionLLM - A framework to finetune LLMs to be your own sentient conversational companion

Language: Jupyter Notebook - Size: 40.1 MB - Last synced at: 4 days ago - Pushed at: almost 2 years ago - Stars: 44 - Forks: 5

poloclub/Fine-tuning-LLMs

Finetune Llama 2 on Colab for free on your own data: step-by-step tutorial

Language: Jupyter Notebook - Size: 9.12 MB - Last synced at: 5 months ago - Pushed at: over 1 year ago - Stars: 43 - Forks: 22

HenryNdubuaku/super-lazy-autograd

Hand-derived memory-efficient super lazy PyTorch VJPs for training LLMs on laptop, all using one op (bundled scaled matmuls).

Language: Python - Size: 1.32 MB - Last synced at: 6 days ago - Pushed at: 7 months ago - Stars: 41 - Forks: 0

conneroisu/Text-Dataset-Aid-Plugin

This is a obsidian plugin to help with the creation of personal jsonl datasets for text generation models.

Language: TypeScript - Size: 157 KB - Last synced at: 4 months ago - Pushed at: almost 2 years ago - Stars: 38 - Forks: 3

git-cloner/Llama2-chinese

Llama2 chinese finetuning

Language: Python - Size: 65.4 KB - Last synced at: over 1 year ago - Pushed at: over 2 years ago - Stars: 37 - Forks: 8

CogitoNTNU/TutorAI

TutorAI is a RAG system capable of assisting with learning academic subjects and using the curriculum and citing it. The project revolves around building an application that ingests a textbook in most formats and facilitates efficient learning of the course material.

Language: Python - Size: 20.8 MB - Last synced at: 6 days ago - Pushed at: 6 days ago - Stars: 36 - Forks: 14

MaxiDonkey/DelphiGemini

The Gemini API wrapper for Delphi utilizes advanced models developed by Google to provide robust capabilities, including interactive chat, text embeddings, code generation, image and video prompting, audio analysis and transcription, fine-tuning, caching, and integration with Google Search.

Language: Pascal - Size: 216 KB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 36 - Forks: 4

machinelearningnuremberg/QuickTune

[ICLR2024] Quick-Tune: Quickly Learning Which Pretrained Model to Finetune and How

Language: Python - Size: 5.86 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 33 - Forks: 4

avocardio/Zicklein

Finetuning instruct-LLaMA on german datasets.

Language: Python - Size: 9.78 MB - Last synced at: 5 days ago - Pushed at: about 2 years ago - Stars: 33 - Forks: 5

VatsaDev/nanoChatGPT

nanogpt turned into a chat model

Language: Python - Size: 266 MB - Last synced at: almost 2 years ago - Pushed at: about 2 years ago - Stars: 33 - Forks: 5

omerbsezer/Fast-LLM-Agent-MCP

This repo covers LLM, Agents concepts both theoretically and practically: LLMs, RAG, Fine Tuning, Agents, Tools, MCP, AWS Strands Agents, Google Agent Development Kit, ADK, Reference Documents, etc.

Language: Python - Size: 65.3 MB - Last synced at: about 2 months ago - Pushed at: 3 months ago - Stars: 30 - Forks: 8

Cre4T3Tiv3/unsloth-llama3-alpaca-lora

Advanced 4-bit QLoRA fine-tuning pipeline for LLaMA 3 8B with production-grade optimization. Memory-efficient training on consumer GPUs for instruction-following specialization. Demonstrates cutting-edge parameter-efficient fine-tuning with Unsloth integration.

Language: Jupyter Notebook - Size: 2.11 MB - Last synced at: 12 days ago - Pushed at: 4 months ago - Stars: 30 - Forks: 0

ssbuild/chatglm_rlhf

chatglm_rlhf_finetuning

Language: Python - Size: 149 KB - Last synced at: 5 days ago - Pushed at: about 2 years ago - Stars: 30 - Forks: 1

paulocoutinhox/mini-llm

Simple and lightweight tool to fine-tune GPT models (like GPT-2 and GPT-Neo) using your own data — built with Python and Transformers. Adapt powerful language models to your domain with ease.

Language: Python - Size: 89.8 KB - Last synced at: about 1 month ago - Pushed at: 6 months ago - Stars: 28 - Forks: 1

francoislanc/midistral

LLM finetuned for generating symbolic music

Language: Python - Size: 1.52 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 28 - Forks: 2

microsoft/Build25-LAB329

Fine-Tune End-to-End Distillation Models with Azure AI Foundry Models and Foundry Local

Language: Jupyter Notebook - Size: 1.4 MB - Last synced at: 13 days ago - Pushed at: 3 months ago - Stars: 27 - Forks: 12

neph1/finetrainers-ui

Gradio UI for training video models using finetrainers

Language: Python - Size: 103 KB - Last synced at: 7 months ago - Pushed at: 7 months ago - Stars: 27 - Forks: 2

dannylee1020/openpo

Building synthetic data for preference tuning

Language: Python - Size: 10.7 MB - Last synced at: 23 days ago - Pushed at: 10 months ago - Stars: 27 - Forks: 0

computational-cell-analytics/peft-sam

Parameter Efficient Fine-Tuning of Segment Anything Model

Language: Python - Size: 416 KB - Last synced at: 25 days ago - Pushed at: 25 days ago - Stars: 24 - Forks: 2

MaxiDonkey/DelphiMistralAI

DelphiMistralAI wrapper brings Mistral’s text-vision-audio models and agentic Conversations to Delphi, with chat, embeddings, Codestral codegen, fine-tuning, batching, moderation, async/await helpers and live request monitoring.

Language: Pascal - Size: 1.79 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 23 - Forks: 5

Shaurya-Sethi/transqlate

End-to-end natural language to SQL system: schema-aware model fine-tuning, retrieval-augmented prompting, and production-grade CLI, powered by a custom fine-tuned Phi-4 Mini.

Language: Python - Size: 1.7 MB - Last synced at: 2 months ago - Pushed at: 3 months ago - Stars: 23 - Forks: 1

chainstacklabs/web3-ai-trading-agent

Build an Autonomous Web3 AI Trading Agent (BASE + Uniswap V4 example)

Language: Python - Size: 1.11 MB - Last synced at: 2 months ago - Pushed at: 4 months ago - Stars: 23 - Forks: 5

shaheennabi/Production-Ready-Instruction-Finetuning-of-Meta-Llama-3.2-3B-Instruct-Project

Instruction Fine-Tuning of Meta Llama 3.2-3B Instruct on Kannada Conversations. Tailoring the model to follow specific instructions in Kannada, enhancing its ability to generate relevant, context-aware responses based on conversational inputs. Using the Kannada Instruct dataset for fine-tuning! Happy Finetuning 🎋

Language: Jupyter Notebook - Size: 692 KB - Last synced at: 8 days ago - Pushed at: 9 months ago - Stars: 23 - Forks: 6

meaningalignment/dft

Democratic Fine-tuning with a Moral Graph

Language: TypeScript - Size: 10 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 22 - Forks: 9

zhaoyl18/SEIKO

SEIKO is a novel reinforcement learning method to efficiently fine-tune diffusion models in an online setting. Our methods outperform all baselines (PPO, classifier-based guidance, direct reward backpropagation) for fine-tuning Stable Diffusion.

Language: Python - Size: 3.85 MB - Last synced at: 4 months ago - Pushed at: over 1 year ago - Stars: 22 - Forks: 0

adithya-s-k/Indic-llm

A open-source framework designed to adapt pre-trained Language Models (LLMs), such as Llama, Mistral, and Mixtral, to a wide array of domains and languages.

Language: Python - Size: 171 KB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 21 - Forks: 1

SIC98/GPT2-python-code-generator

GPT2 finetuning with transformers 🤗

Language: Jupyter Notebook - Size: 185 KB - Last synced at: over 2 years ago - Pushed at: over 4 years ago - Stars: 21 - Forks: 2

IBM/AutoVP

[ICLR24] AutoVP: An Automated Visual Prompting Framework and Benchmark

Language: Python - Size: 579 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 20 - Forks: 2

Hemanthkumar2112/Reward-Modeling-RLHF-Finetune-and-RAG

Gemma2(9B), Llama3-8B-Finetune-and-RAG, code base for sample, implemented in Kaggle platform

Language: Jupyter Notebook - Size: 18.6 MB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 20 - Forks: 6

lzzcd001/nabla-gfn

Official Implementation of Nabla-GFlowNet (ICLR 2025)

Language: Python - Size: 4.27 MB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 19 - Forks: 0

ThomasRochefortB/open-agentinstruct

An open-source recreation of the AgentInstruct agentic workflow for synthetic data generation

Language: Python - Size: 372 KB - Last synced at: 2 months ago - Pushed at: 7 months ago - Stars: 19 - Forks: 3

git-disl/Booster

This is the official code for the paper "Booster: Tackling Harmful Fine-tuning for Large Language Models via Attenuating Harmful Perturbation" (ICLR2025).

Language: Shell - Size: 293 KB - Last synced at: 8 months ago - Pushed at: 8 months ago - Stars: 19 - Forks: 0

git-disl/Vaccine

This is the official code for the paper "Vaccine: Perturbation-aware Alignment for Large Language Models" (NeurIPS2024)

Language: Shell - Size: 730 KB - Last synced at: 12 months ago - Pushed at: 12 months ago - Stars: 19 - Forks: 0

HomoScriptor-Project/HomoScriptor

Fuel innovation and advance language models with HomoScriptor: A vibrant, community-driven dataset for fine-tuning large language models.

Size: 63.5 KB - Last synced at: 11 months ago - Pushed at: about 2 years ago - Stars: 18 - Forks: 3

Raumberg/myllm

Multi-node distributed LLM training framework

Language: Python - Size: 1.66 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 17 - Forks: 1

itspranavajay/Merge-Diffusion-Tool

Merge Diffusion Tool is an open-source solution for merging LoRA models, integrating LoRA into checkpoints, and blending Flux And Stable Diffusion models (SD1.5, SD2, SD3, SDXL). Optimize your AI workflows with ease.

Language: Python - Size: 24.4 KB - Last synced at: 7 months ago - Pushed at: about 1 year ago - Stars: 17 - Forks: 0

ruimalheiro/training-custom-llama

Llama-style transformer in PyTorch with multi-node / multi-GPU training. Includes pretraining, SFT, DPO, LoRA, and knowledge distillation. Scripts for dataset mixing and training from scratch.

Language: Python - Size: 1.32 MB - Last synced at: 12 days ago - Pushed at: 12 days ago - Stars: 16 - Forks: 3

utahnlp/structured_tuning_srl

Implementation of our ACL 2020 paper: Structured Tuning for Semantic Role Labeling

Language: Python - Size: 802 KB - Last synced at: 11 months ago - Pushed at: over 1 year ago - Stars: 16 - Forks: 2