Topic: "llm-finetuning"
xingyaoww/code-act
Official Repo for ICML 2024 paper "Executable Code Actions Elicit Better LLM Agents" by Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhang, Yunzhu Li, Hao Peng, Heng Ji.
Language: Python - Size: 24.1 MB - Last synced at: 9 days ago - Pushed at: 11 months ago - Stars: 1,055 - Forks: 86

rohan-paul/LLM-FineTuning-Large-Language-Models
LLM (Large Language Model) FineTuning
Language: Jupyter Notebook - Size: 11.3 MB - Last synced at: 2 days ago - Pushed at: 20 days ago - Stars: 528 - Forks: 127

LLM-Tuning-Safety/LLMs-Finetuning-Safety
We jailbreak GPT-3.5 Turbo’s safety guardrails by fine-tuning it on only 10 adversarially designed examples, at a cost of less than $0.20 via OpenAI’s APIs.
Language: Python - Size: 23.2 MB - Last synced at: 5 months ago - Pushed at: about 1 year ago - Stars: 245 - Forks: 29

pdaicode/awesome-LLMs-finetuning
Collection of resources for finetuning Large Language Models (LLMs).
Size: 103 KB - Last synced at: 12 days ago - Pushed at: 3 months ago - Stars: 67 - Forks: 8

sugarcane-ai/sugarcane-ai
npm like package ecosystem for Prompts 🤖
Language: TypeScript - Size: 11.5 MB - Last synced at: 4 days ago - Pushed at: 3 months ago - Stars: 49 - Forks: 14

ROIM1998/APT
[ICML'24 Oral] APT: Adaptive Pruning and Tuning Pretrained Language Models for Efficient Training and Inference
Language: Python - Size: 4.08 MB - Last synced at: 2 days ago - Pushed at: 11 months ago - Stars: 38 - Forks: 1

user-wu/Chinese-Llama3-Quick-Start
中文llama3大模型快速上手,通用中文语言大模型finetune教程,基于Meta-llama3实现。
Language: Python - Size: 108 KB - Last synced at: 2 days ago - Pushed at: 10 months ago - Stars: 20 - Forks: 1

ShinoharaHare/LLM-Training
A distributed training framework for large language models powered by Lightning.
Language: Python - Size: 281 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 19 - Forks: 4

LucknowAI/Lucknow-LLM
Collecting data for Building Lucknow's first LLM
Language: Jupyter Notebook - Size: 773 KB - Last synced at: 2 days ago - Pushed at: 11 months ago - Stars: 17 - Forks: 27

HenryCai11/LLM-Self-Control
The official repo of paper "Self-Control of LLM Behaviors by Compressing Suffix Gradient into Prefix Controller"
Language: Jupyter Notebook - Size: 50 MB - Last synced at: 8 months ago - Pushed at: 8 months ago - Stars: 14 - Forks: 1

BY571/DistRL-LLM
Distributed Reinforcement Learning for LLM Fine-Tuning with multi-GPU utilization
Language: Python - Size: 374 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 9 - Forks: 0

Prismadic/tractor-beam
high-efficiency text & file scraper with smart tracking, client/server networking for building language model datasets fast
Language: Python - Size: 7.27 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 7 - Forks: 0

gmongaras/Wizard_QLoRA_Finetuning
Finetuning Some Wizard Models With QLoRA
Language: Python - Size: 76.2 KB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 7 - Forks: 9

truefoundry/llm-finetune
LLM Finetuning with Axolotl with decent defaults + Optional TrueFoundry Experiment Tracking Extension
Language: Python - Size: 1.44 MB - Last synced at: 3 days ago - Pushed at: 3 days ago - Stars: 6 - Forks: 0

bits-bytes-nn/bedrock-model-finetuner
A helper library for fine-tuning Amazon Bedrock models. This toolkit assists in generating Q&A datasets from documents and streamlines the LLM fine-tuning process.
Language: Python - Size: 84 KB - Last synced at: 19 days ago - Pushed at: 6 months ago - Stars: 5 - Forks: 0

sekerlipencere/zynpdata-zynp_ai-teknofest--acikhack2024tddi
Türkiye'nin en büyük teknoloji forumu Technopat'tan kazınmış, tamamen Türkçe ve açık kaynaklı en büyük veri setidir. 3 milyon konu ve 21 milyon yanıt içeren 7GB'lık bu veri seti, Türkçe NLP ve LLM projeleri için kapsamlı bir kaynak sağlar.#Acikhack2024TDDİ
Language: Python - Size: 30.3 MB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 5 - Forks: 1

yihedeng9/STIC
Enhancing Large Vision Language Models with Self-Training on Image Comprehension.
Language: Python - Size: 5.98 MB - Last synced at: 11 months ago - Pushed at: 11 months ago - Stars: 5 - Forks: 0

ethicalabs-ai/FlowerTune-Qwen2.5-Coder-0.5B-Instruct
FlowerTune LLM on Coding Dataset
Language: Python - Size: 530 KB - Last synced at: 17 days ago - Pushed at: 2 months ago - Stars: 3 - Forks: 0

mxagar/generative_ai_udacity
My personal notes, code and projects of the Udacity Generative AI Nanodegree.
Language: Jupyter Notebook - Size: 20.3 MB - Last synced at: 15 days ago - Pushed at: 3 months ago - Stars: 3 - Forks: 2

shankezh/LLMScratchs
This repository contains a comprehensive implementation of the full lifecycle of Large Language Model (LLM大模型) development and application, covering Pretraining (预训练), Supervised Fine-Tuning (SFT全参), Direct Preference Optimization (DPO偏好优化), Inference(推理) and Deployment. It supports frameworks such as PyTorch, HuggingFace, and DeepSpeed.
Language: Python - Size: 1.31 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 3 - Forks: 0

1rsh/qa-genie
This is a package for generating questions and answers from unstructured data to be used for NLP tasks.
Language: Python - Size: 68.4 KB - Last synced at: 11 days ago - Pushed at: about 1 year ago - Stars: 3 - Forks: 1

skyscrappers/DivyaDrishti
AI-Enhanced Learning and Interaction Assistant for Visually Impaired people at IIIT Delhi
Language: Jupyter Notebook - Size: 576 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 2 - Forks: 0

getflexai/flex_ai
simplifies fine-tuning and inference for 60+ open-source LLMs through a single API
Language: Python - Size: 1.05 MB - Last synced at: 17 days ago - Pushed at: 5 months ago - Stars: 2 - Forks: 0

gOsuzu/Efficient-LLM-Few-Examples-Supervised-Fine-Tuning
This is a final porject repository for Goergia Tech CS7643.
Language: Jupyter Notebook - Size: 12.1 MB - Last synced at: 12 months ago - Pushed at: 12 months ago - Stars: 2 - Forks: 0

AnveshakR/Reddit-Sports-Sentiment-Analysis
Natural Language Processing Class Project - Spring '23. Analysing and Generating Sports Fans Responses from Reddit Sport Subreddits
Language: Jupyter Notebook - Size: 32.6 MB - Last synced at: 3 months ago - Pushed at: over 1 year ago - Stars: 2 - Forks: 0

MDalamin5/CSE499-Project-Adaptive-Tutoring-System-AI-Based
This project is an AI-powered algebra tutor using the Phi-3 Mini model. It provides personalized learning through interactive chat, adapting to the student's level and offering detailed step-by-step solutions. Built with Streamlit for an engaging educational experience.
Language: Python - Size: 8.76 MB - Last synced at: 3 days ago - Pushed at: 3 days ago - Stars: 1 - Forks: 0

gjyotk/Menstrual-Health-Education-bot
Large Language Model based Menstrual Health Education solution
Language: Jupyter Notebook - Size: 437 KB - Last synced at: 11 days ago - Pushed at: 11 days ago - Stars: 1 - Forks: 0

mrs83/FlowerTune-Qwen2.5-7B-Instruct-Medical
FlowerTune LLM on Medical Dataset
Language: Python - Size: 1.13 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 1 - Forks: 0

s-araromi/My-SageMaker-IT-Domain-Expert-Project
This project demonstrates how to fine-tune a large language model (LLM) for the IT domain using Amazon SageMaker, creating an "IT Domain Expert" model.
Language: Jupyter Notebook - Size: 535 KB - Last synced at: 23 days ago - Pushed at: 6 months ago - Stars: 1 - Forks: 0

anurima-saha/Text-Classification
Wine reviews used to determine the type of wine training on imbalanced data using classification algorithms like SVM, Naive Bayes and Random Forest Classifier. Neural Network (CNN, RNN and LSTM) and LLM models (DistilBERT and RoBERTa) were also used followed by error analysis using SHAP.
Language: HTML - Size: 3.05 MB - Last synced at: about 2 months ago - Pushed at: 9 months ago - Stars: 1 - Forks: 0

eternalflame02/Single-Node-Finetuning-of-Tiny-LLama-using-Intel-Xeon-SPR
The project was undertaken as part of the Intel Unnati Industrial Training program for the year 2024. The primary objective of this project aligns with Problem Statement PS-04: Introduction to GenAI LLM Inference on CPUs and subsequent LLM Model Finetuning for the development of a Custom Chatbot.
Language: Jupyter Notebook - Size: 12.2 MB - Last synced at: 22 days ago - Pushed at: 9 months ago - Stars: 1 - Forks: 1

Prismadic/hygiene
A payload compression toolkit that makes it easy to create ideal data structures for LLMs; from training data to chain payloads.
Language: Python - Size: 6.38 MB - Last synced at: 1 day ago - Pushed at: about 1 year ago - Stars: 1 - Forks: 0

yuchengml/Adaptation-Tuning-PEFT
Comparison of different adaptation methods on PEFT for fine-tuning downstream tasks or benchmarks.
Language: Python - Size: 130 KB - Last synced at: 11 months ago - Pushed at: about 1 year ago - Stars: 1 - Forks: 0

medxiaorudan/LLM_NER_MultiNERD
LLM fine-tuned for Named Entity Recognition in MultiNERD Dataset
Language: Jupyter Notebook - Size: 2.19 MB - Last synced at: about 1 month ago - Pushed at: about 1 year ago - Stars: 1 - Forks: 0

MaximeRobeyns/llm_examples
A collection of examples for training or fine-tuning LLMs.
Language: Python - Size: 82 KB - Last synced at: 8 days ago - Pushed at: over 1 year ago - Stars: 1 - Forks: 1

Dhanush-R-git/MH-Analysis
The MHRoberta is Mental Health Roberta model. The pretrained Roberta transformer based model fine-tunned on Mental Health dataset by adopting PEFT method.
Language: Jupyter Notebook - Size: 3.42 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 0 - Forks: 0

zhangzhangco/iso-standard-expert-model
A specialized fine-tuning project that transforms DeepSeek 7B into an ISO standards expert, capable of providing professional guidance on ISO document drafting according to ISO/IEC Directives, Part 2. Features efficient LoRA training, GGUF export for mobile deployment, and comprehensive evaluation workflows.
Language: Python - Size: 114 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

BasharatWali/TableTalk_CodeGen
Language: Jupyter Notebook - Size: 3.3 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

ethicalabs-ai/FlowerTune-xLSTM-7b-NLP
FlowerTune LLM on NLP Dataset
Language: Python - Size: 278 KB - Last synced at: 17 days ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

ethicalabs-ai/FlowerTune-SmolLM2-1.7B-Instruct-Finance
FlowerTune LLM on Finance Dataset (sentiment analysis)
Language: Python - Size: 287 KB - Last synced at: 17 days ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

abhash-rai/E-Commerce-LLM-Based-Recommendation-System
This project develops an e-commerce recommendation system using FastAPI, SQLModel for database management, and Qdrant for storing product embeddings. Google FLAN-T5 is fine-tuned on training data generated from project database itself to generate personalized recommendations and adapt over time.
Language: Jupyter Notebook - Size: 377 KB - Last synced at: 25 days ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

mrs83/FlowerTune-phi-4-NLP
FlowerTune LLM on NLP Dataset
Language: Python - Size: 192 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

MohamedAlaouiMhamdi/Finetuning_LLM
This project demonstrates the fine-tuning of the Mistral-7B language model using QLoRA (Quantized Low-Rank Adaptation) to optimize the model for dialogue summarization and customer support tasks.
Language: Jupyter Notebook - Size: 0 Bytes - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 0 - Forks: 0

mrs83/FlowerTune-Mistral-7B-Instruct-v0.3-Medical
FlowerTune LLM on Medical Dataset
Language: Python - Size: 1.13 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 0 - Forks: 0

tinoSibz/gen-ai-intensive
Applied techniques and workflows for leveraging Generative AI, particularly Large Language Models (LLMs), in practical and domain-specific scenarios.
Language: Jupyter Notebook - Size: 353 KB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 0 - Forks: 0

alecruces/GPT-Truth
Enter the realm of truth detection with GPT-Truth - fine-tuning GPT-3.5 for unparalleled accuracy in identifying deceptive opinions
Language: Jupyter Notebook - Size: 2.08 MB - Last synced at: 7 days ago - Pushed at: 5 months ago - Stars: 0 - Forks: 0

dpoulopoulos/bert-qa-finetuning
Fine-tuning BERT on the SQuAD dataset for Question-Answering tasks
Language: Jupyter Notebook - Size: 3.24 MB - Last synced at: 7 months ago - Pushed at: 7 months ago - Stars: 0 - Forks: 0

ShahDishank/gemma-finetuned
Gemma-2b-it LLM has been finetuned on a dataset of Python codes, enabling it to proficiently learn Python syntax and assist in debugging tasks, offering valuable guidance to programmers.
Language: Jupyter Notebook - Size: 35.2 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

0xnu/fine_tune_llm_docker 📦
Fine-tune large language models (LLMs) using the Hugging Face Transformers library.
Language: Python - Size: 9.77 KB - Last synced at: 19 days ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

elahehaghaarabi/language_model_grant_classifier
A language model is fine-tuned using domain data to identify pre-defined groups of documents
Language: Jupyter Notebook - Size: 43.9 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

prince-css/semrep
Factuality check of the SemRep Predications
Language: Jupyter Notebook - Size: 4.2 MB - Last synced at: 2 days ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

nabojyoti/LLM-Finetuning-Playground
Language: Jupyter Notebook - Size: 197 KB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0

QuyAnh2005/neurips-llm-challenge
A winner of NeurIPS LLM 2023 Competition
Language: Jupyter Notebook - Size: 6.44 MB - Last synced at: 3 days ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 1

dishadas168/reddit-chatgpt
Streamlit application for Reddit posts powered by OpenAI, Pinecone and Langchain
Language: Python - Size: 4.89 MB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0
