GitHub / 2003HARSH / Transformer-based-Decoder-only-Language-Model
This repository explores building a character-level transformer decoder in PyTorch, similar to GPT while focusing more on understanding individual components. My goal is to gain deep transformer knowledge and see if character-level learning improves handling of unseen words. The code allows for hyperparameter tuning and experiment customization.
Stars: 0
Forks: 0
Open issues: 0
License: mit
Language: Jupyter Notebook
Size: 74.3 MB
Dependencies parsed at: Pending
Created at: 11 months ago
Updated at: 7 months ago
Pushed at: 8 months ago
Last synced at: 2 months ago
Topics: from-scratch-in-python, gpt, language-model, pytorch, transformers