GitHub topics: rotary-positional-embedding
andreyYaky/mingpt
GPTs trained with shakespeare dataset.
Language: Python - Size: 56.6 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0
xycoord/Language-Modelling
Implementations and Experiments: Transformers, RoPE, KV cache, SAEs, Tokenisers
Language: Python - Size: 1.52 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0
gerdis/pytorch_video_classification
Transformer-Based Video Classification in PyTorch using RoPE
Language: Jupyter Notebook - Size: 37.1 KB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 0 - Forks: 0
liaoyanqing666/Decoder-only-transformer_Time_Series_Prediction
使用Decoder-only的Transformer进行时序预测,包含SwiGLU和RoPE(Rotary Positional Embedding),Time series prediction using Decoder-only Transformer, Including SwiGLU and RoPE(Rotary Positional Embedding)
Language: Python - Size: 31.3 KB - Last synced at: 8 months ago - Pushed at: almost 2 years ago - Stars: 10 - Forks: 1
lucadellalib/llama3
A single-file implementation of LLaMA 3, with support for jitting, KV caching and prompting
Language: Python - Size: 26.4 KB - Last synced at: about 1 month ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0
prajeshshrestha/Llama-2.0-architecture-and-inference-from-scratch-with-PyTorch
Language: Python - Size: 24.4 KB - Last synced at: 8 months ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0