GitHub / CAI991108 / Machine-Learning-and-Language-Model
This project explores GPT-2 and Llama models through pre-training, fine-tuning, and Chain-of-Thought (CoT) prompting. It includes memory-efficient optimizations (SGD, LoRA, BAdam) and evaluations on math datasets (GSM8K, NumGLUE, StimulEq, SVAMP).
Stars: 1
Forks: 0
Open issues: 0
License: mit
Language: Python
Size: 54.5 MB
Dependencies parsed at: Pending
Created at: 4 months ago
Updated at: 4 months ago
Pushed at: 4 months ago
Last synced at: 3 months ago
Topics: chainofthought, finetune-llm, gpt2, llama, llm, llm-inference, pretrained-language-model