GitHub topics: transformer-architecture
shaharoded/Transform-EMR
This model is a decoder transformer based model aiming to model events predictions from EMR records as a sequential text generation problem. This project is a part of my thesis research.
Language: Jupyter Notebook - Size: 10.6 MB - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 0 - Forks: 0

hyuki875/Transformers
The Transformers repository provides a comprehensive implementation of the Transformer architecture, a groundbreaking model that has revolutionized both Natural Language Processing (NLP) and Computer Vision tasks. Introduced in the seminal paper "Attention is All You Need" by Vaswani et al.
Size: 1.95 KB - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 1 - Forks: 0

pribaditzy/Memorized-Q-A-Simulator
🤖 A toy Transformer Q&A model simulator demonstrating core concepts of large language models through memorized Q&A pairs. Educational demo with interactive web interface.
Size: 1000 Bytes - Last synced at: 2 days ago - Pushed at: 2 days ago - Stars: 0 - Forks: 0

takara-ai/SwarmFormer
A pytorch implementation of SwarmFormer for text classification.
Language: Python - Size: 72.3 KB - Last synced at: 5 days ago - Pushed at: 5 days ago - Stars: 15 - Forks: 4

tairov/llama2.mojo
Inference Llama 2 in one file of pure 🔥
Language: Mojo - Size: 2.61 MB - Last synced at: 3 days ago - Pushed at: about 1 year ago - Stars: 2,115 - Forks: 139

GabMartino/TransformerForDummies
Annotated implementation of vanilla Transformers to guide through all the ambiguities.
Language: Python - Size: 7.39 MB - Last synced at: 6 days ago - Pushed at: 6 days ago - Stars: 10 - Forks: 0

vilari-mickopf/mmwave-gesture-recognition
Basic Gesture Recognition Using mmWave Sensor - TI AWR1642
Language: Python - Size: 1.71 MB - Last synced at: 6 days ago - Pushed at: 10 months ago - Stars: 135 - Forks: 23

versoindustries/HighNoonLLM
HighNoon LLM uses Hierarchical Spatial Neural Memory (HSMN) to process language like humans, organizing text into a tree for efficiency. It cuts computing needs by 78x, excelling in summarization, coding, and Q&A, while running locally for privacy.
Language: Python - Size: 7.9 MB - Last synced at: 7 days ago - Pushed at: 7 days ago - Stars: 12 - Forks: 0

nlpodyssey/spago
Self-contained Machine Learning and Natural Language Processing library in Go
Language: Go - Size: 19.5 MB - Last synced at: 7 days ago - Pushed at: 3 months ago - Stars: 1,806 - Forks: 87

kyegomez/MultiModalMamba
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest AI framework ever.
Language: Python - Size: 2.2 MB - Last synced at: about 11 hours ago - Pushed at: 19 days ago - Stars: 452 - Forks: 25

Yunika-Bajracharya/Extractive-Nepali-QA
Extractive Nepali Question Answering System | Browser Extension & Web Application
Language: Jupyter Notebook - Size: 18.6 MB - Last synced at: 8 days ago - Pushed at: 8 days ago - Stars: 3 - Forks: 0

mohitsarawgi/Generative-AI
A curated, structured repository covering Generative AI from Python basics to Transformers, LangChain, and LLM-powered apps. Perfect for anyone learning Gen AI hands-on.
Language: Jupyter Notebook - Size: 9.96 MB - Last synced at: 8 days ago - Pushed at: 9 days ago - Stars: 0 - Forks: 0

genieincodebottle/generative-ai
Comprehensive resources on Generative AI, including a detailed roadmap, projects, use cases, interview preparation, and coding preparation.
Language: Jupyter Notebook - Size: 52.7 MB - Last synced at: 9 days ago - Pushed at: 9 days ago - Stars: 1,021 - Forks: 267

DoQuantum/r1.7-planck-pioneers
Integrating QC techniques into Sparse Attention for Transformers
Language: Python - Size: 7.81 KB - Last synced at: 10 days ago - Pushed at: 10 days ago - Stars: 0 - Forks: 0

kowshik14/FakeNewsDetection-TweetGuard
TweetGuard combines Transformer and Bi-LSTM architectures to detect fake news on Twitter. Using the 'TruthSeeker' dataset and BERTweet tokenization, it outperforms traditional models, setting a new standard for detecting misinformation across multiple datasets.
Language: Jupyter Notebook - Size: 11.3 MB - Last synced at: 10 days ago - Pushed at: 10 days ago - Stars: 1 - Forks: 0

dame-cell/Triformer
Transformers components but in Triton
Language: Python - Size: 2.09 MB - Last synced at: 7 days ago - Pushed at: about 2 months ago - Stars: 34 - Forks: 0

snailpt/MSCFormer
Multi-Scale Convolutional Transformer Network for Motor Imagery Brain-Computer Interface
Language: Python - Size: 300 KB - Last synced at: 12 days ago - Pushed at: 12 days ago - Stars: 93 - Forks: 0

cmhungsteve/Awesome-Transformer-Attention
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
Size: 5.65 MB - Last synced at: 12 days ago - Pushed at: 11 months ago - Stars: 4,884 - Forks: 494

LipSync-Edusync/multispeaker-tts
Transfer Learning for Multispeaker TTS: Implementation of the NeurIPS 2018 paper "Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis" (Jia et al.). Synthesizes speech for both seen and unseen speakers using a pre-trained speaker encoder and Tacotron 2.
Language: Python - Size: 2.92 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 1 - Forks: 1

TobiasBN1005/From-thought-to-movement-
A Transformer-based neural network that decodes movement intentions in real time from EEG, EMG, and IMU signals. It classifies intended actions and sends control signals to external actuators, such as robotic arms or electrical muscle stimulation systems.
Language: C - Size: 41.4 MB - Last synced at: 14 days ago - Pushed at: 14 days ago - Stars: 3 - Forks: 0

songqiang321/Awesome-AI-Papers
This repository is used to collect papers and code in the field of AI.
Size: 4.58 MB - Last synced at: 12 days ago - Pushed at: 23 days ago - Stars: 66 - Forks: 5

Glodanale/PIMGT
Physics-Informed Meta Graph Transformer for Travel State Estimation via Traffic Density
Language: Jupyter Notebook - Size: 250 MB - Last synced at: 15 days ago - Pushed at: 15 days ago - Stars: 1 - Forks: 0

JericN/rice-disease-detection
Pipeline for fine-tuning and evaluating state-of-the-art vision transformer and hybrid models for image classification.
Language: Jupyter Notebook - Size: 36 MB - Last synced at: 18 days ago - Pushed at: 18 days ago - Stars: 1 - Forks: 0

Binny-Shukla/HN-Transformer
Language: Jupyter Notebook - Size: 11.7 KB - Last synced at: 19 days ago - Pushed at: 19 days ago - Stars: 1 - Forks: 0

jshuadvd/LongRoPE
Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper
Language: Python - Size: 562 KB - Last synced at: 6 days ago - Pushed at: 11 months ago - Stars: 137 - Forks: 14

berniwal/swin-transformer-pytorch
Implementation of the Swin Transformer in PyTorch.
Language: Python - Size: 201 KB - Last synced at: 19 days ago - Pushed at: about 4 years ago - Stars: 831 - Forks: 131

WalkingBugFix/Building-a-miniature-GPT
A short code guide to building a super miniature GPT model, providing detailed understanding of model architecture and workflow composition. The model and components were constructed with references from the 2017 Attention is all you need paper.
Language: Jupyter Notebook - Size: 605 KB - Last synced at: 21 days ago - Pushed at: 21 days ago - Stars: 1 - Forks: 0

awslabs/sockeye
Sequence-to-sequence framework with a focus on Neural Machine Translation based on PyTorch
Language: Python - Size: 9.83 MB - Last synced at: 3 days ago - Pushed at: 8 months ago - Stars: 1,218 - Forks: 323

cjerzak/LinkOrgs-software
LinkOrgs: An R package for linking linking records on organizations using half a billion open-collaborated records from LinkedIn
Language: HTML - Size: 178 MB - Last synced at: 23 days ago - Pushed at: 23 days ago - Stars: 12 - Forks: 1

kyegomez/Algorithm-Of-Thoughts
My implementation of "Algorithm of Thoughts: Enhancing Exploration of Ideas in Large Language Models"
Language: Python - Size: 294 KB - Last synced at: 21 days ago - Pushed at: over 1 year ago - Stars: 98 - Forks: 15

sgrvinod/a-PyTorch-Tutorial-to-Transformers
Attention Is All You Need | a PyTorch Tutorial to Transformers
Language: Python - Size: 27.5 MB - Last synced at: 18 days ago - Pushed at: over 1 year ago - Stars: 313 - Forks: 49

hkproj/transformer-from-scratch-notes
Notes about "Attention is all you need" video (https://www.youtube.com/watch?v=bCz4OMemCcA)
Size: 1.32 MB - Last synced at: 24 days ago - Pushed at: about 2 years ago - Stars: 281 - Forks: 62

wgcban/ChangeFormer
[IGARSS'22]: A Transformer-Based Siamese Network for Change Detection
Language: Python - Size: 13.8 MB - Last synced at: 25 days ago - Pushed at: over 1 year ago - Stars: 486 - Forks: 66

scaomath/eit-transformer
[ICLR 2023] Interplay between the Attention and Electrical Impedance Tomography
Language: Python - Size: 31.3 KB - Last synced at: 6 days ago - Pushed at: over 2 years ago - Stars: 34 - Forks: 5

kyegomez/GPT3
An implementation of the base GPT-3 Model architecture from the paper by OPENAI "Language Models are Few-Shot Learners"
Language: Python - Size: 236 KB - Last synced at: 17 days ago - Pushed at: 12 months ago - Stars: 19 - Forks: 3

omron-sinicx/crystalframer
The official code respository for "Rethinking the role of frames for SE(3)-invariant crystal structure modeling" (ICLR 2025)
Language: Python - Size: 9.74 MB - Last synced at: 13 days ago - Pushed at: about 2 months ago - Stars: 7 - Forks: 1

zhongkaifu/Seq2SeqSharp
Seq2SeqSharp is a tensor based fast & flexible deep neural network framework written by .NET (C#). It has many highlighted features, such as automatic differentiation, different network types (Transformer, LSTM, BiLSTM and so on), multi-GPUs supported, cross-platforms (Windows, Linux, x86, x64, ARM), multimodal model for text and images and so on.
Language: C# - Size: 432 MB - Last synced at: about 22 hours ago - Pushed at: 2 months ago - Stars: 204 - Forks: 42

SverreNystad/transformer
An implementation of the Transformer architecture from Attention is all you need: https://arxiv.org/abs/1706.03762
Language: Python - Size: 121 KB - Last synced at: 7 days ago - Pushed at: about 1 month ago - Stars: 2 - Forks: 0

onlybooks/llm
LLM을 활용한 실전 AI 애플리케이션 개발
Language: Jupyter Notebook - Size: 4.47 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 162 - Forks: 124

Plachtaa/VALL-E-X
An open source implementation of Microsoft's VALL-E X zero-shot TTS model. Demo is available in https://plachtaa.github.io/vallex/
Language: Python - Size: 56.5 MB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 7,865 - Forks: 787

PengBoXiangShang/multigraph_transformer
IEEE TNNLS 2021, transformer, multi-graph transformer, graph, graph classification, sketch recognition, sketch classification, free-hand sketch, official code of the paper "Multi-Graph Transformer for Free-Hand Sketch Recognition"
Language: Python - Size: 2.36 MB - Last synced at: about 1 month ago - Pushed at: about 4 years ago - Stars: 298 - Forks: 32

wjf5203/SeqFormer
SeqFormer: Sequential Transformer for Video Instance Segmentation (ECCV 2022 Oral)
Language: Python - Size: 16.3 MB - Last synced at: about 1 month ago - Pushed at: almost 3 years ago - Stars: 347 - Forks: 30

abdur75648/Deep-Learning-Specialization-Coursera
This repo contains the updated version of all the assignments/labs (done by me) of Deep Learning Specialization on Coursera by Andrew Ng. It includes building various deep learning models from scratch and implementing them for object detection, facial recognition, autonomous driving, neural machine translation, trigger word detection, etc.
Language: Jupyter Notebook - Size: 161 MB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 447 - Forks: 373

Ma-Lab-Berkeley/CRATE
Code for CRATE (Coding RAte reduction TransformEr).
Language: Python - Size: 55.8 MB - Last synced at: about 1 month ago - Pushed at: 8 months ago - Stars: 1,226 - Forks: 96

falseywinchnet/ConvexGPT
Mostly Convex ICNN based Large Language Model
Language: Jupyter Notebook - Size: 1.83 MB - Last synced at: about 18 hours ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

google-research/maxvit 📦
[ECCV 2022] Official repository for "MaxViT: Multi-Axis Vision Transformer". SOTA foundation models for classification, detection, segmentation, image quality, and generative modeling...
Language: Jupyter Notebook - Size: 2.6 MB - Last synced at: about 2 months ago - Pushed at: about 2 years ago - Stars: 472 - Forks: 33

M-Taghizadeh/Persian_Question_Answering_Voice2Voice_AI
This repository hosts BonyadAI, a Persian question answering AI Model. We developed an initial web crawler and scraper to gather the dataset. The second phase involved building a machine learning model based on word embeddings and NLP techniques. This AI model operates end-to-end, receiving user voice input and providing responses in Persian voice.
Language: Jupyter Notebook - Size: 89.4 MB - Last synced at: about 2 months ago - Pushed at: 12 months ago - Stars: 5 - Forks: 3

codewithdark-git/Transformers
The Transformers repository provides a comprehensive implementation of the Transformer architecture, a groundbreaking model that has revolutionized both Natural Language Processing (NLP) and Computer Vision tasks. Introduced in the seminal paper "Attention is All You Need" by Vaswani et al.
Language: Jupyter Notebook - Size: 2.09 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

narges-d/Course
Language: Jupyter Notebook - Size: 13.4 MB - Last synced at: 10 days ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

Yramklass/BCI-simulation
Machine learning framework that simulates basic brain-computer interfaces for thought-driven prosthetic limb control.
Language: Python - Size: 254 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

VSainteuf/pytorch-psetae
PyTorch implementation of the model presented in "Satellite Image Time Series Classification with Pixel-Set Encoders and Temporal Self-Attention"
Language: Python - Size: 1.98 MB - Last synced at: about 2 months ago - Pushed at: over 3 years ago - Stars: 198 - Forks: 43

fastnlp/CPT
CPT: A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation
Language: Python - Size: 1.33 MB - Last synced at: about 1 month ago - Pushed at: over 2 years ago - Stars: 487 - Forks: 72

e-hossam96/arabic-nano-gpt
Arabic Nano GPT Trained on Arabic Wikipedia Dataset from Wikimedia
Language: Jupyter Notebook - Size: 1.65 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 2 - Forks: 0

prasannaghimiree/PRATYASHA-A-deep-learning-approach-for-Visual-Impairment
An complete flutter based app for blind people incorporating Nepali currency detection, OCR and Environment Captioning for enhancing day to day activities of visually impaired person.
Language: Jupyter Notebook - Size: 10.8 MB - Last synced at: about 1 month ago - Pushed at: 4 months ago - Stars: 2 - Forks: 0

frankaging/Multimodal-Transformer
Attention Based Multi-modal Emotion Recognition; Stanford Emotional Narratives Dataset
Language: Python - Size: 458 MB - Last synced at: about 2 months ago - Pushed at: almost 6 years ago - Stars: 18 - Forks: 1

ZixuanKe/PyContinual
PyContinual (An Easy and Extendible Framework for Continual Learning)
Language: Python - Size: 3 MB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 314 - Forks: 65

joeynmt/joeynmt
Minimalist NMT for educational purposes
Language: Python - Size: 37.8 MB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 692 - Forks: 215

quanghuy0497/Transformers4Vision
A summarization of Transformer-based architectures for CV tasks, including image classification, object detection, segmentation, and Few-shot Learning. Keep updated frequently.
Size: 13.6 MB - Last synced at: about 2 months ago - Pushed at: about 3 years ago - Stars: 109 - Forks: 18

linwhitehat/ET-BERT
The repository of ET-BERT, a network traffic classification model on encrypted traffic. The work has been accepted as The Web Conference (WWW) 2022 accepted paper.
Language: Python - Size: 13.3 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 458 - Forks: 92

jmcheon/Transformers
Language: Jupyter Notebook - Size: 25.8 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

leehyeonbeen/TimeSeriesSeq2Seq
Sequence-to-sequence model implementations including RNN, CNN, Attention, and Transformers using PyTorch
Language: Python - Size: 82 KB - Last synced at: 11 days ago - Pushed at: almost 2 years ago - Stars: 18 - Forks: 1

CambrianTech/sentinel-ai
Sentinel-AI is a neuroscience-inspired transformer that dynamically prunes and regrows attention heads. Guided by controller feedback and entropy-based pruning, it self-optimizes through biologically informed cycles—compressing, adapting, and evolving its architecture over time.
Language: Python - Size: 8.92 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 1 - Forks: 0

topazape/ViT-Pytorch
Vision Transformer in Pytorch
Language: Python - Size: 4.62 MB - Last synced at: 6 days ago - Pushed at: over 2 years ago - Stars: 8 - Forks: 1

CiscoDevNet/g2p_seq2seq_pytorch
Grapheme to phoneme model for PyTorch
Language: Python - Size: 1.48 MB - Last synced at: about 1 month ago - Pushed at: almost 3 years ago - Stars: 41 - Forks: 11

UIC-Liu-Lab/ContinualLM
An Extensible Continual Learning Framework Focused on Language Models (LMs)
Language: Python - Size: 696 KB - Last synced at: 3 months ago - Pushed at: over 1 year ago - Stars: 272 - Forks: 21

cuiziteng/Illumination-Adaptive-Transformer
🌕 [BMVC 2022] You Only Need 90K Parameters to Adapt Light: A Light Weight Transformer for Image Enhancement and Exposure Correction. SOTA for low light enhancement, 0.004 seconds try this for pre-processing.
Language: Python - Size: 29.4 MB - Last synced at: 3 months ago - Pushed at: over 1 year ago - Stars: 513 - Forks: 49

Cyanosite/Facial-Attribute-Recognition
Facial Attribute Recognition using the Transformer architecture, 91% on CelebA
Language: Jupyter Notebook - Size: 27.4 MB - Last synced at: 3 months ago - Pushed at: about 1 year ago - Stars: 3 - Forks: 0

google-research/human-scene-transformer
Human Scene Transformer: A framework for trajectory prediction and wrappers for reframing the JRDB dataset for the prediction task.
Language: Python - Size: 2.18 MB - Last synced at: 3 months ago - Pushed at: 11 months ago - Stars: 62 - Forks: 10

prakhar21/TextAugmentation-GPT2
Fine-tuned pre-trained GPT2 for custom topic specific text generation. Such system can be used for Text Augmentation.
Language: Python - Size: 655 KB - Last synced at: about 2 months ago - Pushed at: almost 2 years ago - Stars: 193 - Forks: 43

NishantkSingh0/Mail-Classification
Automate mail classification using custom transformer's Encoder architecture
Language: Jupyter Notebook - Size: 21.5 KB - Last synced at: 3 months ago - Pushed at: 12 months ago - Stars: 3 - Forks: 0

NishantkSingh0/Movie-review-classification
Used Transformer's Encoder to classify movie reviews. From scratch
Language: Jupyter Notebook - Size: 8.79 KB - Last synced at: 3 months ago - Pushed at: 11 months ago - Stars: 3 - Forks: 0

NishantkSingh0/Vision-Transformer
Train vision transformer architecture for classifying Fashion-MNIST.
Language: Jupyter Notebook - Size: 14.6 KB - Last synced at: 3 months ago - Pushed at: 10 months ago - Stars: 3 - Forks: 0

NishantkSingh0/Generative-Language-Model
Trained Decoder only model on large BookCorpus Dataset. First time!
Language: Jupyter Notebook - Size: 48.8 KB - Last synced at: 3 months ago - Pushed at: 10 months ago - Stars: 3 - Forks: 0

NishantkSingh0/News_Classification
Used custom parallel transformer's Encoder architecture to train model on News title & News description separately.
Language: Jupyter Notebook - Size: 6.05 MB - Last synced at: 3 months ago - Pushed at: 10 months ago - Stars: 4 - Forks: 0

erfanzar/OST-OpenSourceTransformers
OST Collection: An AI-powered suite of models that predict the next word matches with remarkable accuracy (Text Generative Models). OST Collection is based on a novel approach to work as a full and intelligent NLP Model.
Language: Jupyter Notebook - Size: 134 MB - Last synced at: 2 months ago - Pushed at: over 1 year ago - Stars: 15 - Forks: 1

kyegomez/Tiktokx
Tiktok is an advanced multimedia recommender system that fuses the generative modality-aware collaborative self-augmentation and contrastive cross-modality dependency encoding to achieve superior performance compared to existing state-of-the-art multi-model recommenders.
Language: Python - Size: 229 KB - Last synced at: about 2 months ago - Pushed at: almost 2 years ago - Stars: 12 - Forks: 0

RohanMenon/LipShiFT
This repo contains code to reproduce results for LipShiFT.
Language: Python - Size: 43 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

mohd-faizy/06P_Sentiment-Analysis-With-Deep-Learning-Using-BERT
Finetuning BERT in PyTorch for sentiment analysis.
Language: Jupyter Notebook - Size: 20.6 MB - Last synced at: 3 months ago - Pushed at: over 2 years ago - Stars: 22 - Forks: 9

tobna/TaylorShift
This repository contains the code for the paper "TaylorShift: Shifting the Complexity of Self-Attention from Squared to Linear (and Back) using Taylor-Softmax"
Language: Python - Size: 98.6 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 8 - Forks: 0

labteral/ernie
Simple State-of-the-Art BERT-Based Sentence Classification with Keras / TensorFlow 2. Built with HuggingFace's Transformers.
Language: Python - Size: 326 KB - Last synced at: about 1 month ago - Pushed at: about 1 year ago - Stars: 200 - Forks: 31

REFLOXO/NLP-Translator
Language Translation Tool built using Python and CustomTkinter. It supports both auto-translate and manual translate modes, allowing users to translate text between multiple languages.
Size: 1000 Bytes - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

alessoh/Neural-Symbolic-Superintelligence-scaling
Neural-Symbolic-Superintelligence scaling
Size: 5.86 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

dingo-actual/om
An LLM architecture utilizing a recurrent structure and multi-layer memory
Language: Python - Size: 1.04 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 12 - Forks: 0

sine2pi/Echo
Whisper like ASR transformer model. Experimental.
Language: Python - Size: 410 KB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 0 - Forks: 0

ES7/Transformer-from-Scratch
In this repository, I have explained the working of the Transformer architecture, provided the code for building it from scratch, and demonstrated how to train it.
Language: Python - Size: 17.6 KB - Last synced at: 3 months ago - Pushed at: about 1 year ago - Stars: 9 - Forks: 3

heisenberg141/Pointcloud-Segmentation
This repository contains sensor fusion between a lidar and camera, semantic segmentation on point clouds and ICP registration of multiple point clouds.
Language: Python - Size: 118 MB - Last synced at: 3 months ago - Pushed at: about 2 years ago - Stars: 9 - Forks: 1

omron-sinicx/crystalformer
The official code respository for "Crystalformer: Infinitely Connected Attention for Periodic Structure Encoding" (ICLR 2024)
Language: Python - Size: 3.15 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 14 - Forks: 0

shuddha2021/Memorized-Q-A-Simulator
🤖 A toy Transformer Q&A model simulator demonstrating core concepts of large language models through memorized Q&A pairs. Educational demo with interactive web interface.
Language: Python - Size: 27.3 KB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 0 - Forks: 0

skl0726/AI-Paper-Review
Al (especially computer vision) paper review and model code implementation by PyTorch
Language: Jupyter Notebook - Size: 5.92 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 1 - Forks: 0

zouguojian/Traffic-speed-prediction
Using to predict the highway traffic speed
Language: Python - Size: 646 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 30 - Forks: 3

bniladridas/rag-transformer
Searches through a diverse dataset of machine learning, science fiction, and cosmic content.
Language: Python - Size: 898 KB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 0 - Forks: 0

abhaskumarsinha/MinimalGPT
MinimalGPT is a concise, adaptable, and streamlined code framework that encompasses the essential components necessary for the construction, training, inference, and fine-tuning of the GPT model. This framework is implemented exclusively using Keras and TensorFlow, ensuring compatibility and coherence within the broader deep learning ecosystem.
Language: Python - Size: 320 KB - Last synced at: 3 months ago - Pushed at: about 1 year ago - Stars: 23 - Forks: 6

MuzzammilShah/GPT-TransformerModel-1
A step-by-step implementation of a GPT-style language model on a combined harry potter novels dataset, inspired by Andrej Karpathy’s lecture.
Language: Jupyter Notebook - Size: 6.57 MB - Last synced at: 24 days ago - Pushed at: 4 months ago - Stars: 1 - Forks: 0

flavioschenker/functaformer
FunctaFormer is a novel, domain-agnostic super-resolution method that leverages Implicit Neural Representations (INRs) to upscale diverse data types, including images, audio, videos, 3D shapes, manifold data, and LiDAR scans.
Language: Python - Size: 37.1 KB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 0 - Forks: 0

OguzhanCOG/EnhanceFX
EnhanceFX™ is a novel, state-of-the-art hybrid super-resolution model architecture.
Size: 3.71 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 2 - Forks: 0

mahshid1378/SwinTransformerPytorch
Implementation of the Swin Transformer in PyTorch. and use Article: arxiv.org/pdf/2103.14030.pdf
Language: Python - Size: 185 KB - Last synced at: 23 days ago - Pushed at: 4 months ago - Stars: 0 - Forks: 0

odarasimi/ml-architectures
architectures in code
Language: Python - Size: 9.77 KB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 0 - Forks: 0

Sciform/transformer-attention-is-all-you-need-implementation
Implementation of the transformer model form the NeuRIPS-paper 2017 "Attention is all you need!"
Language: Python - Size: 32.6 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 1 - Forks: 0

willGuimont/transformers
Flexible transformer implementation for research
Language: Python - Size: 107 KB - Last synced at: 2 months ago - Pushed at: 4 months ago - Stars: 5 - Forks: 1

rubenpjove/tabularT-OS-fingerprinting
This repository employs two different architectures of Tabular Transformer models for Operating System fingerprinting from three different datasets.
Language: Jupyter Notebook - Size: 3.98 MB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 0 - Forks: 0
