An open API service providing repository metadata for many open source software ecosystems.

GitHub topics: attention-mechanisms

pprp/awesome-attention-mechanism-in-cv

Awesome List of Attention Modules and Plug&Play Modules in Computer Vision

Language: Python - Size: 3.25 MB - Last synced at: 10 minutes ago - Pushed at: about 2 years ago - Stars: 1,185 - Forks: 170

lucidrains/muse-maskgit-pytorch

Implementation of Muse: Text-to-Image Generation via Masked Generative Transformers, in Pytorch

Language: Python - Size: 285 KB - Last synced at: about 18 hours ago - Pushed at: about 1 year ago - Stars: 893 - Forks: 83

zalkklop/LVSM

Official code for "LVSM: A Large View Synthesis Model with Minimal 3D Inductive Bias" (ICLR 2025 Oral)

Language: Python - Size: 193 KB - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 0 - Forks: 0

lucidrains/mmdit

Implementation of a single layer of the MMDiT, proposed in Stable Diffusion 3, in Pytorch

Language: Python - Size: 156 KB - Last synced at: about 18 hours ago - Pushed at: 4 months ago - Stars: 354 - Forks: 9

lucidrains/make-a-video-pytorch

Implementation of Make-A-Video, new SOTA text to video generator from Meta AI, in Pytorch

Language: Python - Size: 227 KB - Last synced at: 1 day ago - Pushed at: about 1 year ago - Stars: 1,969 - Forks: 186

changzy00/pytorch-attention

πŸ¦–Pytorch implementation of popular Attention Mechanisms, Vision Transformers, MLP-Like models and CNNs.πŸ”₯πŸ”₯πŸ”₯

Language: Python - Size: 3.5 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 469 - Forks: 47

lucidrains/BS-RoFormer

Implementation of Band Split Roformer, SOTA Attention network for music source separation out of ByteDance AI Labs

Language: Python - Size: 186 KB - Last synced at: 2 days ago - Pushed at: 4 months ago - Stars: 537 - Forks: 18

lucidrains/equiformer-pytorch

Implementation of the Equiformer, SE3/E3 equivariant attention network that reaches new SOTA, and adopted for use by EquiFold for protein folding

Language: Python - Size: 17.5 MB - Last synced at: about 20 hours ago - Pushed at: 5 months ago - Stars: 270 - Forks: 27

lucidrains/iTransformer

Unofficial implementation of iTransformer - SOTA Time Series Forecasting using Attention networks, out of Tsinghua / Ant group

Language: Python - Size: 220 KB - Last synced at: 1 day ago - Pushed at: 5 months ago - Stars: 492 - Forks: 38

lucidrains/local-attention

An implementation of local windowed attention for language modeling

Language: Python - Size: 34.1 MB - Last synced at: 2 days ago - Pushed at: 4 months ago - Stars: 449 - Forks: 48

lucidrains/magvit2-pytorch

Implementation of MagViT2 Tokenizer in Pytorch

Language: Python - Size: 1.87 MB - Last synced at: 2 days ago - Pushed at: 4 months ago - Stars: 602 - Forks: 33

lucidrains/MEGABYTE-pytorch

Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Pytorch

Language: Python - Size: 34.5 MB - Last synced at: 1 day ago - Pushed at: 5 months ago - Stars: 643 - Forks: 55

lucidrains/meshgpt-pytorch

Implementation of MeshGPT, SOTA Mesh generation using Attention, in Pytorch

Language: Python - Size: 1.15 MB - Last synced at: 1 day ago - Pushed at: 5 months ago - Stars: 823 - Forks: 67

lucidrains/musiclm-pytorch

Implementation of MusicLM, Google's new SOTA model for music generation using attention networks, in Pytorch

Language: Python - Size: 196 KB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 3,249 - Forks: 264

lucidrains/PaLM-rlhf-pytorch

Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM

Language: Python - Size: 34.3 MB - Last synced at: 2 days ago - Pushed at: 23 days ago - Stars: 7,806 - Forks: 677

lucidrains/phenaki-pytorch

Implementation of Phenaki Video, which uses Mask GIT to produce text guided videos of up to 2 minutes in length, in Pytorch

Language: Python - Size: 266 KB - Last synced at: about 17 hours ago - Pushed at: 10 months ago - Stars: 772 - Forks: 81

lucidrains/recurrent-memory-transformer-pytorch

Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch

Language: Python - Size: 34.3 MB - Last synced at: 1 day ago - Pushed at: 4 months ago - Stars: 407 - Forks: 17

lucidrains/q-transformer

Implementation of Q-Transformer, Scalable Offline Reinforcement Learning via Autoregressive Q-Functions, out of Google Deepmind

Language: Python - Size: 1.56 MB - Last synced at: 1 day ago - Pushed at: 3 months ago - Stars: 375 - Forks: 20

lucidrains/taylor-series-linear-attention

Explorations into the recently proposed Taylor Series Linear Attention

Language: Python - Size: 779 KB - Last synced at: 1 day ago - Pushed at: 9 months ago - Stars: 100 - Forks: 3

lucidrains/toolformer-pytorch

Implementation of Toolformer, Language Models That Can Use Tools, by MetaAI

Language: Python - Size: 161 KB - Last synced at: 1 day ago - Pushed at: 10 months ago - Stars: 2,031 - Forks: 127

JulesBelveze/time-series-autoencoder

PyTorch Dual-Attention LSTM-Autoencoder For Multivariate Time Series

Language: Python - Size: 343 KB - Last synced at: 1 day ago - Pushed at: 2 months ago - Stars: 675 - Forks: 67

cmhungsteve/Awesome-Transformer-Attention

An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites

Size: 5.65 MB - Last synced at: 2 days ago - Pushed at: 10 months ago - Stars: 4,860 - Forks: 493

kyegomez/MambaTransformer

Integrating Mamba/SSMs with Transformer for Enhanced Long Context and High-Quality Sequence Modeling

Language: Python - Size: 2.23 MB - Last synced at: about 1 hour ago - Pushed at: about 1 month ago - Stars: 193 - Forks: 16

lucidrains/alphafold3-pytorch

Implementation of Alphafold 3 from Google Deepmind in Pytorch

Language: Python - Size: 13.6 MB - Last synced at: 3 days ago - Pushed at: 4 months ago - Stars: 1,426 - Forks: 185

lucidrains/audiolm-pytorch

Implementation of AudioLM, a SOTA Language Modeling Approach to Audio Generation out of Google Research, in Pytorch

Language: Python - Size: 571 KB - Last synced at: 3 days ago - Pushed at: 4 months ago - Stars: 2,537 - Forks: 275

kyegomez/MambaFormer

Implementation of MambaFormer in Pytorch ++ Zeta from the paper: "Can Mamba Learn How to Learn? A Comparative Study on In-Context Learning Tasks"

Language: Python - Size: 2.16 MB - Last synced at: about 9 hours ago - Pushed at: 25 days ago - Stars: 21 - Forks: 1

erogol/BlaGPT

Experimental playground for benchmarking language model (LM) architectures, layers, and tricks on smaller datasets. Designed for flexible experimentation and exploration.

Language: Python - Size: 856 KB - Last synced at: 5 days ago - Pushed at: 8 days ago - Stars: 31 - Forks: 3

SaifPunjwani/weight-of-thought

Weight-of-Thought Reasoning: Exploring Neural Network Weights for Enhanced LLM Reasoning

Language: Python - Size: 2.93 KB - Last synced at: 8 days ago - Pushed at: 8 days ago - Stars: 0 - Forks: 0

lucidrains/flash-attention-jax

Implementation of Flash Attention in Jax

Language: Python - Size: 181 KB - Last synced at: 8 days ago - Pushed at: about 1 year ago - Stars: 208 - Forks: 24

lucidrains/robotic-transformer-pytorch

Implementation of RT1 (Robotic Transformer) in Pytorch

Language: Python - Size: 161 KB - Last synced at: 9 days ago - Pushed at: 7 months ago - Stars: 421 - Forks: 33

earthai-tech/fusionlab-learn

fusionlab-learn: Igniting Next-Gen Temporal Fusion Architectures

Language: Python - Size: 3.57 MB - Last synced at: about 2 hours ago - Pushed at: about 2 hours ago - Stars: 1 - Forks: 0

kyegomez/LongNet

Implementation of plug in and play Attention from "LongNet: Scaling Transformers to 1,000,000,000 Tokens"

Language: Python - Size: 40.3 MB - Last synced at: about 9 hours ago - Pushed at: over 1 year ago - Stars: 704 - Forks: 63

landskape-ai/triplet-attention

Official PyTorch Implementation for "Rotate to Attend: Convolutional Triplet Attention Module." [WACV 2021]

Language: Jupyter Notebook - Size: 9.73 MB - Last synced at: 8 days ago - Pushed at: 3 months ago - Stars: 422 - Forks: 50

kyegomez/Jamba

PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"

Language: Python - Size: 2.17 MB - Last synced at: 9 days ago - Pushed at: about 1 month ago - Stars: 168 - Forks: 11

lucidrains/simple-hierarchical-transformer

Experiments around a simple idea for inducing multiple hierarchical predictive model within a GPT

Language: Python - Size: 34.1 MB - Last synced at: 10 days ago - Pushed at: 9 months ago - Stars: 212 - Forks: 13

jshuadvd/LongRoPE

Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper

Language: Python - Size: 562 KB - Last synced at: 3 days ago - Pushed at: 10 months ago - Stars: 136 - Forks: 14

DonorNearBy/YOLO12-UnderWater

YOLOv12 Underwater Object Detection is an open-source suite for underwater object detection, built on YOLOv12. It offers an end-to-end pipeline with GPU-accelerated training, customizable data augmentations, real-time inference via Gradio, and support for model export (ONNX & PyTorch).

Language: Python - Size: 52.1 MB - Last synced at: 16 days ago - Pushed at: 16 days ago - Stars: 0 - Forks: 0

selfcontrol7/Korean_Voice_Phishing_Detection

All codes implemented on Korean voice phishing detection papers

Language: Jupyter Notebook - Size: 135 MB - Last synced at: 17 days ago - Pushed at: 17 days ago - Stars: 14 - Forks: 8

tinh2044/YOLO12-UnderWater

YOLOv12 Underwater Object Detection is an open-source suite for underwater object detection, built on YOLOv12. It offers an end-to-end pipeline with GPU-accelerated training, customizable data augmentations, real-time inference via Gradio, and support for model export (ONNX & PyTorch).

Language: Python - Size: 52.1 MB - Last synced at: 12 days ago - Pushed at: 18 days ago - Stars: 0 - Forks: 0

lucidrains/transformer-directed-evolution

Explorations into whether a transformer with RL can direct a genetic algorithm to converge faster

Language: Python - Size: 32.2 KB - Last synced at: 21 days ago - Pushed at: 21 days ago - Stars: 67 - Forks: 0

lucidrains/coconut-pytorch

Implementation of πŸ₯₯ Coconut, Chain of Continuous Thought, in Pytorch

Language: Python - Size: 187 KB - Last synced at: 13 days ago - Pushed at: 5 months ago - Stars: 165 - Forks: 9

lucidrains/agent-attention-pytorch

Implementation of Agent Attention in Pytorch

Language: Python - Size: 516 KB - Last synced at: 5 days ago - Pushed at: 10 months ago - Stars: 89 - Forks: 5

lucidrains/flash-cosine-sim-attention

Implementation of fused cosine similarity attention in the same style as Flash Attention

Language: Cuda - Size: 34.4 MB - Last synced at: 9 days ago - Pushed at: over 2 years ago - Stars: 213 - Forks: 12

lucidrains/block-recurrent-transformer-pytorch

Implementation of Block Recurrent Transformer - Pytorch

Language: Python - Size: 34.2 MB - Last synced at: 9 days ago - Pushed at: 9 months ago - Stars: 217 - Forks: 20

lucidrains/recurrent-interface-network-pytorch

Implementation of Recurrent Interface Network (RIN), for highly efficient generation of images and video without cascading networks, in Pytorch

Language: Python - Size: 731 KB - Last synced at: about 1 month ago - Pushed at: about 1 year ago - Stars: 203 - Forks: 15

kyegomez/MobileVLM

Implementation of the LDP module block in PyTorch and Zeta from the paper: "MobileVLM: A Fast, Strong and Open Vision Language Assistant for Mobile Devices"

Language: Python - Size: 2.17 MB - Last synced at: 9 days ago - Pushed at: about 1 year ago - Stars: 16 - Forks: 0

cbaziotis/neat-vision

Neat (Neural Attention) Vision, is a visualization tool for the attention mechanisms of deep-learning models for Natural Language Processing (NLP) tasks. (framework-agnostic)

Language: Vue - Size: 25.4 MB - Last synced at: about 1 month ago - Pushed at: about 7 years ago - Stars: 250 - Forks: 24

lucidrains/complex-valued-transformer

Implementation of the transformer proposed in "Building Blocks for a Complex-Valued Transformer Architecture"

Language: Python - Size: 34.4 MB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 73 - Forks: 6

lucidrains/Mega-pytorch

Implementation of Mega, the Single-head Attention with Multi-headed EMA architecture that currently holds SOTA on Long Range Arena

Language: Python - Size: 34.2 MB - Last synced at: 2 days ago - Pushed at: over 1 year ago - Stars: 203 - Forks: 11

lucidrains/medical-chatgpt

Implementation of ChatGPT, but tailored towards primary care medicine, with the reward being able to collect patient histories in a thorough and efficient manner and come up with a reasonable differential diagnosis

Language: Python - Size: 27.3 KB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 315 - Forks: 30

lucidrains/mixture-of-attention

Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts

Language: Python - Size: 34.1 MB - Last synced at: 6 days ago - Pushed at: 7 months ago - Stars: 118 - Forks: 4

lucidrains/diffusion-policy

Implementation of Diffusion Policy, Toyota Research's supposed breakthrough in leveraging DDPMs for learning policies for real-world Robotics

Language: Python - Size: 1.02 MB - Last synced at: about 1 month ago - Pushed at: 10 months ago - Stars: 106 - Forks: 2

kyegomez/FlashMHA

An simple pytorch implementation of Flash MultiHead Attention

Language: Jupyter Notebook - Size: 85 KB - Last synced at: 9 days ago - Pushed at: over 1 year ago - Stars: 21 - Forks: 2

kyegomez/SparseAttention

Pytorch Implementation of the sparse attention from the paper: "Generating Long Sequences with Sparse Transformers"

Language: Python - Size: 2.16 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 77 - Forks: 2

kyegomez/MGQA

The open source implementation of the multi grouped query attention by the paper "GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints"

Language: Python - Size: 248 KB - Last synced at: 9 days ago - Pushed at: over 1 year ago - Stars: 14 - Forks: 1

lucidrains/transframer-pytorch

Implementation of Transframer, Deepmind's U-net + Transformer architecture for up to 30 seconds video generation, in Pytorch

Language: Python - Size: 159 KB - Last synced at: 16 days ago - Pushed at: over 2 years ago - Stars: 70 - Forks: 6

monk1337/Various-Attention-mechanisms

This repository contain various types of attention mechanism like Bahdanau , Soft attention , Additive Attention , Hierarchical Attention etc in Pytorch, Tensorflow, Keras

Language: Python - Size: 643 KB - Last synced at: about 1 month ago - Pushed at: over 3 years ago - Stars: 125 - Forks: 25

lucidrains/CALM-pytorch

Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composition", out of Google Deepmind

Language: Python - Size: 939 KB - Last synced at: about 1 month ago - Pushed at: 8 months ago - Stars: 174 - Forks: 11

lucidrains/zorro-pytorch

Implementation of Zorro, Masked Multimodal Transformer, in Pytorch

Language: Python - Size: 197 KB - Last synced at: 10 days ago - Pushed at: over 1 year ago - Stars: 97 - Forks: 6

JoanaR/multi-mode-CNN-pytorch

A PyTorch implementation of the Multi-Mode CNN to reconstruct Chlorophyll-a time series in the global ocean from oceanic and atmospheric physical drivers

Language: Jupyter Notebook - Size: 8.97 MB - Last synced at: about 1 month ago - Pushed at: almost 2 years ago - Stars: 8 - Forks: 2

lucidrains/CoLT5-attention

Implementation of the conditionally routed attention in the CoLT5 architecture, in Pytorch

Language: Python - Size: 187 KB - Last synced at: 7 days ago - Pushed at: 8 months ago - Stars: 226 - Forks: 13

kyegomez/Hedgehog

Implementation of the model "Hedgehog" from the paper: "The Hedgehog & the Porcupine: Expressive Linear Attentions with Softmax Mimicry"

Language: Python - Size: 2.16 MB - Last synced at: 9 days ago - Pushed at: about 1 year ago - Stars: 13 - Forks: 0

lucidrains/pause-transformer

Yet another random morning idea to be quickly tried and architecture shared if it works; to allow the transformer to pause for any amount of time on any token

Language: Python - Size: 659 KB - Last synced at: 16 days ago - Pushed at: over 1 year ago - Stars: 53 - Forks: 1

kyegomez/PaLM2-VAdapter

Implementation of "PaLM2-VAdapter:" from the multi-modal model paper: "PaLM2-VAdapter: Progressively Aligned Language Model Makes a Strong Vision-language Adapter"

Language: Python - Size: 2.17 MB - Last synced at: 9 days ago - Pushed at: 6 months ago - Stars: 17 - Forks: 0

kyegomez/ShallowFF

Zeta implemantion of "Rethinking Attention: Exploring Shallow Feed-Forward Neural Networks as an Alternative to Attention Layers in Transformers"

Language: Python - Size: 36.2 MB - Last synced at: 9 days ago - Pushed at: 11 days ago - Stars: 10 - Forks: 1

lucidrains/autoregressive-linear-attention-cuda

CUDA implementation of autoregressive linear attention, with all the latest research findings

Language: Python - Size: 5.86 KB - Last synced at: 16 days ago - Pushed at: almost 2 years ago - Stars: 44 - Forks: 3

lucidrains/kalman-filtering-attention

Implementation of the Kalman Filtering Attention proposed in "Kalman Filtering Attention for User Behavior Modeling in CTR Prediction"

Size: 4.88 KB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 57 - Forks: 4

johnsmithm/multi-heads-attention-image-classification

Multi heads attention for image classification

Language: Python - Size: 2.93 KB - Last synced at: about 1 month ago - Pushed at: about 7 years ago - Stars: 81 - Forks: 34

lucidrains/flash-genomics-model

My own attempt at a long context genomics model, leveraging recent advances in long context attention modeling (Flash Attention + other hierarchical methods)

Language: Python - Size: 12.7 KB - Last synced at: 16 days ago - Pushed at: almost 2 years ago - Stars: 52 - Forks: 5

lucidrains/equiformer-diffusion

Implementation of Denoising Diffusion for protein design, but using the new Equiformer (successor to SE3 Transformers) with some additional improvements

Size: 2.93 KB - Last synced at: about 2 months ago - Pushed at: over 2 years ago - Stars: 56 - Forks: 4

PraveenKumar-Rajendran/Udacity-Natural-Language-Processing-Engineer-Nanodegree

Projects Implemented for the Udacity Natural Language Processing Engineer Nanodegree Program

Size: 1.42 MB - Last synced at: about 2 months ago - Pushed at: 9 months ago - Stars: 0 - Forks: 0

KB9/BraccioVisualAttention

An active vision system which builds a 3D environment map autonomously using visual attention mechanisms.

Language: Python - Size: 102 MB - Last synced at: 5 months ago - Pushed at: about 7 years ago - Stars: 1 - Forks: 0

dynastes-team/dynastes

A collection of layers, ops, utilities and more for TensorFlow 2.0 high-level API Keras

Language: Python - Size: 658 KB - Last synced at: 25 days ago - Pushed at: almost 5 years ago - Stars: 10 - Forks: 1

programmer290399/pyqna

A simple python package for question answering.

Language: Python - Size: 3.95 MB - Last synced at: 9 months ago - Pushed at: about 2 years ago - Stars: 11 - Forks: 5

lucidrains/coordinate-descent-attention

Implementation of an Attention layer where each head can attend to more than just one token, using coordinate descent to pick topk

Language: Python - Size: 34.1 MB - Last synced at: 16 days ago - Pushed at: almost 2 years ago - Stars: 46 - Forks: 2

kyegomez/KosmosG

My implementation of the model KosmosG from "KOSMOS-G: Generating Images in Context with Multimodal Large Language Models"

Language: Python - Size: 2.79 MB - Last synced at: 9 days ago - Pushed at: 6 months ago - Stars: 14 - Forks: 1

kyegomez/CELESTIAL-1

Omni-Modality Processing, Understanding, and Generation

Language: Python - Size: 2.49 MB - Last synced at: 9 days ago - Pushed at: about 1 year ago - Stars: 7 - Forks: 1

DIAGNijmegen/prostateMR_3D-CAD-csPCa πŸ“¦

Hierarchical probabilistic 3D U-Net, with attention mechanisms (β€”π˜ˆπ˜΅π˜΅π˜¦π˜―π˜΅π˜ͺ𝘰𝘯 𝘜-π˜•π˜¦π˜΅, π˜šπ˜Œπ˜™π˜¦π˜΄π˜•π˜¦π˜΅) and a nested decoder structure with deep supervision (β€”π˜œπ˜•π˜¦π˜΅++). Built in TensorFlow 2.5. Configured for voxel-level clinically significant prostate cancer detection in multi-channel 3D bpMRI scans.

Language: Python - Size: 21.2 MB - Last synced at: 11 months ago - Pushed at: over 3 years ago - Stars: 39 - Forks: 6

ssghost/var-attn Fork of harvardnlp/var-attn

Language: Python - Size: 93.2 MB - Last synced at: about 1 year ago - Pushed at: almost 7 years ago - Stars: 0 - Forks: 0

pouyasattari/Automatic-Generative-Code-with-Neural-Machine-Translation-for-data-security-purpose

Transformers, including the T5 and MarianMT, enabled effective understanding and generating complex programming codes. Consequently, they can help us in Data Security field. Let's see how!

Language: Jupyter Notebook - Size: 130 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 1 - Forks: 0

soran-ghaderi/make-a-video

"Make-A-Video", new SOTA text to video by Meta-FAIR - Tensorflow

Language: Python - Size: 705 KB - Last synced at: about 1 month ago - Pushed at: over 2 years ago - Stars: 14 - Forks: 2

arnavdantuluri/long-context-transformers

A repository to get train transformers to access longer context for causal language models, most of these methods are still in testing. Try them out if you'd like but please lmk your results so we don't duplicate work :)

Language: Python - Size: 188 KB - Last synced at: about 1 year ago - Pushed at: almost 2 years ago - Stars: 5 - Forks: 2

KevinAtsou/formerslab

A simple a set of Transformers building blocks that can be used to build language models

Language: Python - Size: 27.3 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0

GiantPandaCV/yolov3-point

Learning YOLOv3 from scratch δ»Žι›ΆεΌ€ε§‹ε­¦δΉ YOLOv3代码

Language: Jupyter Notebook - Size: 94.6 MB - Last synced at: over 1 year ago - Pushed at: about 3 years ago - Stars: 203 - Forks: 53

vene/sparse-structured-attention

Sparse and structured neural attention mechanisms

Language: Python - Size: 102 KB - Last synced at: over 1 year ago - Pushed at: over 4 years ago - Stars: 215 - Forks: 36

super-m-a-n/covid19-vaccine-tweets-sentiment-analysis

Deep learning methods for sentiment analysis classification of covid-19 vaccination tweets

Language: Jupyter Notebook - Size: 3.89 MB - Last synced at: almost 2 years ago - Pushed at: almost 2 years ago - Stars: 0 - Forks: 0

umapornp/PAC-MAN

πŸ‘» The PyTorch implementation for the IEEE Access paper: "PAC-MAN: Multi-Relation Network in Social Community for Personalized Hashtag Recommendation".

Language: Python - Size: 318 KB - Last synced at: almost 2 years ago - Pushed at: almost 2 years ago - Stars: 1 - Forks: 0

umapornp/ARERec

✨ The implementation for the IEEE Access paper: "ARERec: Attentive Local Interaction Model for Sequential Recommendation".

Language: Python - Size: 7 MB - Last synced at: about 2 years ago - Pushed at: about 2 years ago - Stars: 0 - Forks: 0

elbuco1/AttentionMechanismsTrajectoryPrediction

In this repository, one can find the code for my master's thesis project. The main goal of the project was to study and improve attention mechanisms for trajectory prediction of moving agents.

Language: Python - Size: 13.8 MB - Last synced at: about 2 years ago - Pushed at: over 5 years ago - Stars: 41 - Forks: 11

veqtor/veqtor_keras πŸ“¦

Collection of various of my custom TensorFlow-Keras 2.0+ layers, utils and such

Language: Python - Size: 14.6 KB - Last synced at: about 2 years ago - Pushed at: over 5 years ago - Stars: 0 - Forks: 0

acadTags/Automated-Social-Annotation

Joint Multi-label Attention Network (JMAN)

Language: Python - Size: 59.1 MB - Last synced at: about 2 years ago - Pushed at: over 4 years ago - Stars: 11 - Forks: 3

subho406/Sequence-to-Sequence-and-Attention-from-scratch-using-Tensorflow

Sequence to Sequence and attention from scratch using Tensorflow

Language: Jupyter Notebook - Size: 33.2 KB - Last synced at: about 2 years ago - Pushed at: over 7 years ago - Stars: 29 - Forks: 16

bulatkh/image_captioning

Master Project on Image Captioning using Supervised Deep Learning Methods

Language: Jupyter Notebook - Size: 17.4 MB - Last synced at: about 2 years ago - Pushed at: over 2 years ago - Stars: 3 - Forks: 2

ssea-lab/DL4ETI

Computer-aided diagnosis in histopathological images of the Endometrium

Language: Python - Size: 38.7 MB - Last synced at: about 2 years ago - Pushed at: almost 5 years ago - Stars: 16 - Forks: 4

Related Keywords
attention-mechanisms 94 deep-learning 58 artificial-intelligence 50 transformers 35 attention-is-all-you-need 15 attention-mechanism 12 attention 12 ai 7 pytorch 7 tensorflow 7 machine-learning 7 computer-vision 7 natural-language-processing 6 transformer 5 keras 5 ml 5 multimodal 5 gpt4 5 object-detection 4 denoising-diffusion 4 self-attention 3 time-series 3 video-generation 3 linear-attention 3 robotics 3 long-context 3 transformer-models 3 neural-networks 3 recurrent-neural-networks 3 gpt 3 llm 3 long-context-attention 3 multi-modal 3 text-to-video 3 mixed-precision-training 2 attention-lstm 2 model-export 2 long-context-transformers 2 feedforward 2 onnx 2 real-time-inference 2 vision-transformer 2 neural-nets 2 transformer-architecture 2 training-pipeline 2 underwater-imaging 2 reasoning 2 underwater-object-detection 2 deeplearning 2 text-to-image 2 sequence-to-sequence 2 python 2 nlp 2 artificial-neural-networks 2 tokenization 2 ssm 2 brackish-water 2 reinforcement-learning 2 data-augmentation 2 memory 2 recurrence 2 attention-model 2 gpu-acceleration 2 recommendation-system 2 recommender-system 2 bahdanau-attention 2 convolutional-neural-networks 2 gradio 2 machine-translation 1 language-modeling 1 nlp-projects 1 pos-tagging 1 hidden-markov-models 1 mlps 1 speech-recognition 1 text-processing 1 udacity-nanodegree 1 udacity-nanodegree-projects 1 endometrium 1 active-vision 1 autonomous-systems 1 braccio-arduino 1 convolutional-neural-network 1 ffns 1 routing 1 efficient-attention 1 seasonal-interannual-timescale 1 satellite-data 1 remote-sensing 1 regression 1 phytoplankton 1 physical-drivers 1 oceanography 1 ocean-color 1 marine-science 1 interpretability 1 global-scale 1 earth-observation 1 api-calling 1 automatic-speech-recognition 1