An open API service providing repository metadata for many open source software ecosystems.

Topic: "transformer-architecture"

kyegomez/LongVit

A simplistic pytorch implementation of LongVit using my previous implementation of LongNet as a foundation.

Language: Shell - Size: 2.15 MB - Last synced at: 9 days ago - Pushed at: 6 months ago - Stars: 7 - Forks: 0

GregorKobsik/Octree-Transformer

Octree Transformer: Autoregressive 3D Shape Generation on Hierarchically Structured Sequences - CVPRW: StruCo3D, 2023

Language: Python - Size: 671 KB - Last synced at: 10 months ago - Pushed at: 10 months ago - Stars: 7 - Forks: 1

ArikReuter/TNTM

This repository contains the code for the Transformer-Representation Neural Topic Model (TNTM) based on the paper "Probabilistic Topic Modelling with Transformer Representations" by Arik Reuter, Anton Thielmann, Christoph Weisser, Benjamin Säfken and Thomas Kneib

Language: Jupyter Notebook - Size: 31.8 MB - Last synced at: 10 months ago - Pushed at: 10 months ago - Stars: 7 - Forks: 1

SachinKalsi/annotated-research-papers

This repository is a comprehensive collection of research papers, annotations, and concise summaries in the field of Natural Language Processing (NLP). It focuses on machine learning and deep learning techniques, providing valuable resources for NLP enthusiasts and researchers.

Size: 7.26 MB - Last synced at: 6 days ago - Pushed at: over 1 year ago - Stars: 7 - Forks: 2

AspirinCode/drug-likeness_space

Explore drug-like space with deep generative models

Language: Python - Size: 46.1 MB - Last synced at: 2 months ago - Pushed at: about 2 years ago - Stars: 7 - Forks: 0

wojtekcz/language2motion

The goal of this project is to create multi-modal implementation of Transformer architecture in Swift.

Language: Jupyter Notebook - Size: 11.1 MB - Last synced at: almost 2 years ago - Pushed at: over 3 years ago - Stars: 7 - Forks: 4

hrithickcodes/transformer-tf

This repository contains the code for the paper "Attention Is All You Need" i.e The Transformer.

Language: Jupyter Notebook - Size: 30.4 MB - Last synced at: about 1 year ago - Pushed at: over 2 years ago - Stars: 6 - Forks: 1

expellialbus/English-to-Turkish-Translator

An implementation of Encoder-Decoder architecture built with Transformers to translate English sentences to their Turkish equivalents.

Language: Python - Size: 4.96 MB - Last synced at: about 1 year ago - Pushed at: about 3 years ago - Stars: 6 - Forks: 1

ccaven/music-transformer

Deep learning transformer model that generates unique music sequences.

Language: Python - Size: 6.24 MB - Last synced at: about 1 year ago - Pushed at: over 3 years ago - Stars: 6 - Forks: 0

LaurentVeyssier/Abstractive-Text-Summarization-model-in-Keras

Abstractive Text Summarization using Transformer model

Language: Jupyter Notebook - Size: 12.3 MB - Last synced at: about 1 year ago - Pushed at: almost 5 years ago - Stars: 6 - Forks: 5

omron-sinicx/crystalframer

The official code respository for "Rethinking the role of frames for SE(3)-invariant crystal structure modeling" (ICLR 2025)

Language: Python - Size: 3.82 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 5 - Forks: 0

willGuimont/transformers

Flexible transformer implementation for research

Language: Python - Size: 107 KB - Last synced at: about 1 month ago - Pushed at: 3 months ago - Stars: 5 - Forks: 1

M-Taghizadeh/Persian_Question_Answering_Voice2Voice_AI

This repository hosts BonyadAI, a Persian question answering AI Model. We developed an initial web crawler and scraper to gather the dataset. The second phase involved building a machine learning model based on word embeddings and NLP techniques. This AI model operates end-to-end, receiving user voice input and providing responses in Persian voice.

Language: Jupyter Notebook - Size: 89.4 MB - Last synced at: 5 days ago - Pushed at: 10 months ago - Stars: 5 - Forks: 3

jeromepatel/EyeSeg-OpenEDS-2021

3D Eye Segmentation Track 1 OpenEDS 2021, Point Transformer and other CNN based models Implementation

Language: Python - Size: 1.29 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 5 - Forks: 0

ginofft/flax-BLT

FLAX Implementation of Biodirectional Layout Transformer

Language: Python - Size: 11.3 MB - Last synced at: about 1 year ago - Pushed at: over 1 year ago - Stars: 5 - Forks: 0

cheedinmar/SocialMedia-App

Full Stack Social Media App with Nuxt 3, TailwindCSS , Prisma, Auth using JWT.

Language: Vue - Size: 5.28 MB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 5 - Forks: 0

abhaskumarsinha/Keras-implementation-of-Transformer-Architecture

This repository presents a Python-based implementation of the Transformer architecture on Keras TensorFlow library, as proposed by Vaswani et al. in their 2017 paper "Attention is all you need."

Language: Jupyter Notebook - Size: 223 KB - Last synced at: about 1 month ago - Pushed at: over 2 years ago - Stars: 5 - Forks: 3

henryhungle/MM_DST

Code for the paper Multimodal Dialogue State Tracking (NAACL22)

Language: Python - Size: 1.24 MB - Last synced at: about 2 years ago - Pushed at: almost 3 years ago - Stars: 5 - Forks: 1

NewJerseyStyle/anime-translator 📦

Applying deep learning to translate animation and re-generate audio.

Language: Python - Size: 36.1 KB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 4 - Forks: 1

pyserve/Real-Time-Video-Captioning

Visio Text is a real-time video captioning project that leverages the capabilities of artificial intelligence to provide dynamic text captions for videos.

Language: Jupyter Notebook - Size: 67.1 MB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 4 - Forks: 2

NishantkSingh0/News_Classification

Used custom parallel transformer's Encoder architecture to train model on News title & News description separately.

Language: Jupyter Notebook - Size: 6.05 MB - Last synced at: about 1 month ago - Pushed at: 9 months ago - Stars: 4 - Forks: 0

TheCrouchingDragon/clinical_trials_LLM

Match synthetic patients to clinicaltrials.gov restAPI data with ChatGPT4. Course project for Johns Hopkins University EN.705.651.8VL : Large Language Models: Theory and Practice

Language: Jupyter Notebook - Size: 13 MB - Last synced at: 14 days ago - Pushed at: 9 months ago - Stars: 4 - Forks: 1

DebeshJha/TransRUPNet

TransRUPNet for Improved Out-of-Distribution Generalization in Polyp Segmentation (IEEE EMBC)

Language: Python - Size: 827 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 4 - Forks: 0

phuijse/irregular_ts_transformers

Transformer architectures for irregulary sampled time series classification

Language: Jupyter Notebook - Size: 230 KB - Last synced at: 2 months ago - Pushed at: about 1 year ago - Stars: 4 - Forks: 0

YunghuiHsu/ebird_project

Extraction of deep features/representation of birds by deep learning algorithms.

Language: Jupyter Notebook - Size: 8.42 MB - Last synced at: over 1 year ago - Pushed at: about 2 years ago - Stars: 4 - Forks: 1

naokishibuya/simple_transformer

A Transformer Implementation that is easy to understand and customizable.

Language: Python - Size: 46.9 KB - Last synced at: about 2 years ago - Pushed at: over 2 years ago - Stars: 4 - Forks: 2

Merterm/COSMic

Public repo for the paper: "COSMic: A Coherence-Aware Generation Metric for Image Descriptions" by Mert İnan, Piyush Sharma, Baber Khalid, Radu Soricut, Matthew Stone, Malihe Alikhani

Language: Python - Size: 396 KB - Last synced at: over 1 year ago - Pushed at: about 3 years ago - Stars: 4 - Forks: 0

Yunika-Bajracharya/Extractive-Nepali-QA

Extractive Nepali Question Answering System | Browser Extension & Web Application

Language: Jupyter Notebook - Size: 18.6 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 3 - Forks: 0

BOUALILILila/hybridseq2seq

Hyperbolic geometry for representing structural information to improve the systematic generalization of sequence-to-sequence models

Language: Python - Size: 92.8 KB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 3 - Forks: 0

ranfysvalle02/ai-self-attention

This repository provides a basic implementation of self-attention. The code demonstrates how attention mechanisms work in predicting the next word in a sequence. It's a basic implementation that demonstrates the core concept of attention but lacks the complexity of more advanced models like Transformers.

Language: Python - Size: 263 KB - Last synced at: 25 days ago - Pushed at: 8 months ago - Stars: 3 - Forks: 0

NishantkSingh0/Vision-Transformer

Train vision transformer architecture for classifying Fashion-MNIST.

Language: Jupyter Notebook - Size: 14.6 KB - Last synced at: about 1 month ago - Pushed at: 8 months ago - Stars: 3 - Forks: 0

Digital-Defiance/nlp-metaformer

An ablation study on the transformer network for Natural Language Processing

Language: Rust - Size: 25.8 MB - Last synced at: 8 months ago - Pushed at: 8 months ago - Stars: 3 - Forks: 0

NishantkSingh0/Generative-Language-Model

Trained Decoder only model on large BookCorpus Dataset. First time!

Language: Jupyter Notebook - Size: 48.8 KB - Last synced at: about 1 month ago - Pushed at: 8 months ago - Stars: 3 - Forks: 0

NishantkSingh0/Movie-review-classification

Used Transformer's Encoder to classify movie reviews. From scratch

Language: Jupyter Notebook - Size: 8.79 KB - Last synced at: about 1 month ago - Pushed at: 10 months ago - Stars: 3 - Forks: 0

NishantkSingh0/Mail-Classification

Automate mail classification using custom transformer's Encoder architecture

Language: Jupyter Notebook - Size: 21.5 KB - Last synced at: about 1 month ago - Pushed at: 10 months ago - Stars: 3 - Forks: 0

Cyanosite/Facial-Attribute-Recognition

Facial Attribute Recognition using the Transformer architecture, 91% on CelebA

Language: Jupyter Notebook - Size: 27.4 MB - Last synced at: about 1 month ago - Pushed at: about 1 year ago - Stars: 3 - Forks: 0

garyb9/pytorch-transformers

Transformers architecture code playground repository in python using PyTorch.

Language: Python - Size: 1.53 MB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 3 - Forks: 2

speediedan/deep_classiflie

Deep Classiflie is a framework for developing ML models that bolster fact-checking efficiency. As a POC, the initial alpha release of Deep Classiflie generates/analyzes a model that continuously classifies a single individual's statements (Donald Trump) using a single ground truth labeling source (The Washington Post). For statements the model deems most likely to be labeled falsehoods, the @DeepClassiflie twitter bot tweets out a statement analysis and model interpretation "report"

Language: Python - Size: 204 MB - Last synced at: about 1 month ago - Pushed at: over 1 year ago - Stars: 3 - Forks: 0

Fuzzytariy/CMF-DGCN

A Chinese Sentiment Analysis Model based on Transmembrane State Attention for Modal Fusion and Multimodal Dynamic Gradient Regulation.

Language: Python - Size: 4.04 MB - Last synced at: about 2 years ago - Pushed at: about 2 years ago - Stars: 3 - Forks: 0

daspartho/simple-transformer

transformer from scratch

Language: Python - Size: 8.79 KB - Last synced at: about 1 month ago - Pushed at: about 2 years ago - Stars: 3 - Forks: 0

LRY89757/Transformer-models-in-CV

some basic transformer models about computer vision

Language: Python - Size: 650 KB - Last synced at: about 2 years ago - Pushed at: about 3 years ago - Stars: 3 - Forks: 0

arnavsshah/Meeting-Summarization

Leveraging extractive summarization techniques for abstractive summarization.

Language: Jupyter Notebook - Size: 5.03 MB - Last synced at: about 2 years ago - Pushed at: over 3 years ago - Stars: 3 - Forks: 0

adam-mehdi/V2Iformer

A new attention-based architecture for synthesizing images from videos or 3D-images

Language: Python - Size: 331 KB - Last synced at: about 2 years ago - Pushed at: over 3 years ago - Stars: 3 - Forks: 0

CalogeroZarbo/bioshield

BioShield is intended to be an AI system that predicts anti-viral molecules and search among already FDA approved drugs the most similar to be used on humans while the novel predicted molecule is tested and approved.

Language: Jupyter Notebook - Size: 1.85 MB - Last synced at: over 1 year ago - Pushed at: about 5 years ago - Stars: 3 - Forks: 0

e-hossam96/arabic-nano-gpt

Arabic Nano GPT Trained on Arabic Wikipedia Dataset from Wikimedia

Language: Jupyter Notebook - Size: 1.65 MB - Last synced at: 19 days ago - Pushed at: 19 days ago - Stars: 2 - Forks: 0

TobiasBN1005/From-thought-to-movement-

A Transformer-based neural network that decodes movement intentions in real time from EEG, EMG, and IMU signals. It classifies intended actions and sends control signals to external actuators, such as robotic arms or electrical muscle stimulation systems.

Language: C - Size: 40.8 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 2 - Forks: 0

prasannaghimiree/PRATYASHA-A-deep-learning-approach-for-Visual-Impairment

An complete flutter based app for blind people incorporating Nepali currency detection, OCR and Environment Captioning for enhancing day to day activities of visually impaired person.

Language: Jupyter Notebook - Size: 10.8 MB - Last synced at: 1 day ago - Pushed at: 2 months ago - Stars: 2 - Forks: 0

OguzhanCOG/EnhanceFX

EnhanceFX™ is a novel, state-of-the-art hybrid super-resolution model architecture.

Size: 3.71 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 2 - Forks: 0

ihaeyong/Soft-TF

Soft-Transformers For Continual Learning

Size: 2.93 KB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 2 - Forks: 0

zhahoi/swinGAN

This repository uses modules from Swin Transformer to build Transformer-based Generative Adversarial Networks (GANs) models.

Language: Python - Size: 48.8 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 2 - Forks: 0

mytechnotalent/kgpt

A custom GPT based on [Zero To Hero](https://karpathy.ai/zero-to-hero.html) utilizing tiktoken with the intent to augment AI Transformer-model education and reverse engineer GPT models from scratch.

Language: Python - Size: 1.17 MB - Last synced at: almost 2 years ago - Pushed at: almost 2 years ago - Stars: 2 - Forks: 0

tedtedtedtedtedted/Solve-Rubiks-Cube-Via-Transformer

Applying regular transformer and decision transformer on solving the Rubik's cube. A paper is also written to document the results

Language: Python - Size: 107 MB - Last synced at: almost 2 years ago - Pushed at: about 2 years ago - Stars: 2 - Forks: 0

RFLeijenaar/NLP-Explanations-for-NLI

Fine-tuning and evaluation of a language model for explanation generation of natural language inference. Fine-tuning scripts for a pre-trained T5 model supporting both full model fine-tuning as well as LoRA are included in this repository.

Language: Jupyter Notebook - Size: 1.46 MB - Last synced at: 5 months ago - Pushed at: about 2 years ago - Stars: 2 - Forks: 0

jyanqa/case-2022-causual-event

Codebase of the CASE@EMNLP 2022 paper: Causality Detection using Multiple Annotation Decision.

Language: Python - Size: 1.22 MB - Last synced at: almost 2 years ago - Pushed at: over 2 years ago - Stars: 2 - Forks: 0

PrasunDatta/Multimodal-Brain-Tumor-Segmenatation-BraTS-2020

This research work basically highlights my undergrad thesis works. In my thesis, I have worked on the BraTS 2020 dataset. My total journey of thesis from building various models to writing paper is presented here.

Language: Jupyter Notebook - Size: 32.5 MB - Last synced at: about 1 year ago - Pushed at: over 2 years ago - Stars: 2 - Forks: 0

LaurentVeyssier/TRAX_transformer_abstractive_summarization_model

Use Trax and Transformer architecture for generative summarizer deep learning model

Language: Jupyter Notebook - Size: 1.02 MB - Last synced at: about 1 year ago - Pushed at: about 4 years ago - Stars: 2 - Forks: 1

hyuki875/Transformers

The Transformers repository provides a comprehensive implementation of the Transformer architecture, a groundbreaking model that has revolutionized both Natural Language Processing (NLP) and Computer Vision tasks. Introduced in the seminal paper "Attention is All You Need" by Vaswani et al.

Size: 1.95 KB - Last synced at: 4 days ago - Pushed at: 4 days ago - Stars: 1 - Forks: 0

CambrianTech/sentinel-ai

Sentinel-AI is a neuroscience-inspired transformer that dynamically prunes and regrows attention heads. Guided by controller feedback and entropy-based pruning, it self-optimizes through biologically informed cycles—compressing, adapting, and evolving its architecture over time.

Language: Python - Size: 8.92 MB - Last synced at: 26 days ago - Pushed at: 26 days ago - Stars: 1 - Forks: 0

skl0726/AI-Paper-Review

Al (especially computer vision) paper review and model code implementation by PyTorch

Language: Jupyter Notebook - Size: 5.92 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

Sciform/transformer-attention-is-all-you-need-implementation

Implementation of the transformer model form the NeuRIPS-paper 2017 "Attention is all you need!"

Language: Python - Size: 32.6 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

ShahirAnsari/TelegramChatbot-Using-Hugging-face-Transformers

This project was build to create a Telegram bot that uses hugging Face Transformers to process text and reply.

Language: Python - Size: 6.84 KB - Last synced at: 2 months ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

Ctrl408/ViT-implementations

Implementation of Vision Transformers (ViT) with a token merging mechanism

Language: Python - Size: 15.6 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 1 - Forks: 0

RistoAle97/yati

Yet Another Transformer Implementation

Language: Python - Size: 779 KB - Last synced at: 4 months ago - Pushed at: 4 months ago - Stars: 1 - Forks: 0

HenryZNNUE/transformer_gpu

Experimental Implementation of Transformer (With Smolgen) on NVIDIA GPU with the help of LibTorch

Language: C++ - Size: 18.6 KB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 1 - Forks: 0

PrateekJannu/Vision-GPT

Coding a Multi-Modal vision model like GPT-4o from scratch, inspired by @hkproj and PaliGemma

Language: Python - Size: 591 KB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 1 - Forks: 0

minhnhat2001vt/End-2-End-Question-Answering

Language: Jupyter Notebook - Size: 69.3 KB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 1 - Forks: 0

jhaayush2004/My-Transformer

Code implementation of Transformer Model in "Attention is All You Need" in PyTorch.

Language: Jupyter Notebook - Size: 5.25 MB - Last synced at: about 1 month ago - Pushed at: 12 months ago - Stars: 1 - Forks: 0

gustavecortal/transformer

Slides from my NLP course on the transformer architecture

Size: 1.41 MB - Last synced at: 12 months ago - Pushed at: 12 months ago - Stars: 1 - Forks: 0

RuochenT/transformer_hybrid

This study aims to investigate the effectiveness of three Transformers (BERT, RoBERTa, XLNet) in handling data sparsity and cold start problems in the recommender system. We present a Transformer-based hybrid recommender system that predicts missing ratings and ex- tracts semantic embeddings from user reviews to mitigate the issues.

Language: Jupyter Notebook - Size: 3.27 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 1 - Forks: 0

ayan-cs/bangla-ocr-transformer

Implementation of the paper 'Towards Full page Offline Bangla Handwritten Text Recognition using Image-to-Sequence Architecture'. For details, please read the README section.

Language: Python - Size: 9.2 MB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 1 - Forks: 0

aju22/Transformer-Distribution-Visualizer

A Research Toolkit that allows users to select a transformer model, specify a decoding strategy to generate text, which are then displayed in an interactive fashion to understand model probabilities.

Language: HTML - Size: 3.67 MB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 1 - Forks: 0

lotuspalm/transformer_PokerOrder

悟透Transformer--AI是如何学习扑克牌排序的

Size: 5.86 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 1 - Forks: 0

matchild/DeepBrownianMotion

A deep learning classification tool for anomalous diffusion trajectories.

Language: Jupyter Notebook - Size: 4.03 MB - Last synced at: almost 2 years ago - Pushed at: almost 2 years ago - Stars: 1 - Forks: 0

SpydazWebAI-NLP/BasicNeuralNetWork2023

A Basic Multi layered Neural Network, With Attention Masking Features

Language: Visual Basic .NET - Size: 36.1 KB - Last synced at: almost 2 years ago - Pushed at: almost 2 years ago - Stars: 1 - Forks: 1

JelinR/Plasma_Evolution_Analysis

Analyzing Plasma Evolution (Time Series) with Transformers to detect anomalies

Language: Jupyter Notebook - Size: 0 Bytes - Last synced at: almost 2 years ago - Pushed at: almost 2 years ago - Stars: 1 - Forks: 0

IvanBongiorni/shakespeare-gpt

How to build a custom GPT for text generation, based on TensorFlow 2.x and Maximal. Trained on the Shakespeare corpus.

Language: Python - Size: 23.4 KB - Last synced at: about 2 months ago - Pushed at: almost 2 years ago - Stars: 1 - Forks: 1

SatishAdhikari/Generative-AI-with-LLMs

Generative AI with Large Language Models (LLMs) - how generative AI works, and how to deploy it in real-world applications. Coursera link -> https://www.coursera.org/learn/generative-ai-with-llms

Language: Jupyter Notebook - Size: 9.4 MB - Last synced at: almost 2 years ago - Pushed at: almost 2 years ago - Stars: 1 - Forks: 2

thaisaraujom/about-gpt-and-transformer

This repository provides a comprehensive and intuitive explanation of the architecture of OpenAI's GPT (Generative Pretrained Transformer) models.

Size: 6.84 KB - Last synced at: 12 months ago - Pushed at: almost 2 years ago - Stars: 1 - Forks: 0

ThomasRochefortB/resovit-pytorch

Implementation of a variable resolution image pipeline for training Vision Transformers in PyTorch without padding or resizing.

Language: Jupyter Notebook - Size: 81.1 KB - Last synced at: about 2 years ago - Pushed at: about 2 years ago - Stars: 1 - Forks: 0

blaze7451/Project-Haruki-Murakami2

Haruki Murakami style text generation project using transformers.

Language: Jupyter Notebook - Size: 1.9 MB - Last synced at: over 1 year ago - Pushed at: over 2 years ago - Stars: 1 - Forks: 0

paper-and-code/learn-transformer-by-solving

Transformerを書く勉強会 Learn Transformer by Solving

Language: Python - Size: 47.9 KB - Last synced at: 2 days ago - Pushed at: over 2 years ago - Stars: 1 - Forks: 0

markpesic/Vit

Vit implementation in pytorch

Language: Python - Size: 49.8 KB - Last synced at: over 1 year ago - Pushed at: over 2 years ago - Stars: 1 - Forks: 0

adam-mehdi/FastTimeSformer

Attention-based video classifier running on accelerated attention approximations

Language: Python - Size: 98.6 KB - Last synced at: about 2 years ago - Pushed at: almost 3 years ago - Stars: 1 - Forks: 0

dariodellamura/NLG-The-impact-of-data-quality-on-automatic-text-generation-from-RDF-data

The work presented was developed during the internship, as researchers in the field of Natural Language Generation, at the Insid&s Lab laboratory in Milan-Bicocca. The work carried out deals with the creation of a framework for the correct assessment of the impact of the quality of the input datasets on the quality of the text generated by the NLG models, specifically: Creation of the "Concept-Based" and "Entity-Based" versions of the WebNLG dataset; Evaluation of the quality of the datasets created; Training of LSTM and Transformer models using the OpenNMT tool; Natural language text generation by LSTM and Transformer models; Evaluation of the quality of the text generated by the NLG models; Final analysis.

Language: Jupyter Notebook - Size: 1.22 MB - Last synced at: about 2 years ago - Pushed at: almost 3 years ago - Stars: 1 - Forks: 0

robertocarlosmedina/attention-transformer-translator-1

Sequence to Sequence Transformer implementation in order to train a model to translate over Cap-verdian criole to English.

Language: Python - Size: 5.5 MB - Last synced at: about 2 years ago - Pushed at: almost 3 years ago - Stars: 1 - Forks: 0

brianhill11/ViViT-tensorflow

Unofficial Tensorflow implementation of the ViViT model architecture

Language: Python - Size: 19.5 KB - Last synced at: about 2 years ago - Pushed at: over 3 years ago - Stars: 1 - Forks: 0

yashsmehta/transformers-playground

Repo for playing around with transformer architectures and trying out simple experiments

Language: Python - Size: 1.95 KB - Last synced at: about 2 years ago - Pushed at: almost 4 years ago - Stars: 1 - Forks: 0

speediedan/deep_classiflie_db

Deep_classiflie_db is the backend data system for managing Deep Classiflie metadata, analyzing Deep Classiflie intermediate datasets and orchestrating Deep Classiflie model training pipelines. Deep_classiflie_db includes data scraping modules for the initial model data sources. Deep Classiflie depends upon deep_classiflie_db for much of its analytical and dataset generation functionality but the data system is currently maintained as a separate repository here to maximize architectural flexibility. Depending on how Deep Classiflie evolves (e.g. as it supports distributed data stores etc.), it may make more sense to integrate deep_classiflie_db back into deep_classiflie. Currently, deep_classiflie_db releases are synchronized to deep_classiflie releases. To learn more, visit deepclassiflie.org.

Language: Jupyter Notebook - Size: 14.3 MB - Last synced at: about 1 month ago - Pushed at: over 4 years ago - Stars: 1 - Forks: 0

patricebechard/transformer

Implementation of the transformer architecture (Vaswani et. al. 2017) in PyTorch.

Language: TeX - Size: 22.7 MB - Last synced at: 11 months ago - Pushed at: over 6 years ago - Stars: 1 - Forks: 0

pribaditzy/Memorized-Q-A-Simulator

🤖 A toy Transformer Q&A model simulator demonstrating core concepts of large language models through memorized Q&A pairs. Educational demo with interactive web interface.

Size: 1000 Bytes - Last synced at: 4 days ago - Pushed at: 4 days ago - Stars: 0 - Forks: 0

codewithdark-git/Transformers

The Transformers repository provides a comprehensive implementation of the Transformer architecture, a groundbreaking model that has revolutionized both Natural Language Processing (NLP) and Computer Vision tasks. Introduced in the seminal paper "Attention is All You Need" by Vaswani et al.

Language: Jupyter Notebook - Size: 2.09 MB - Last synced at: 11 days ago - Pushed at: 11 days ago - Stars: 0 - Forks: 0

narges-d/Course

Language: Jupyter Notebook - Size: 1.95 KB - Last synced at: 12 days ago - Pushed at: 12 days ago - Stars: 0 - Forks: 0

Yramklass/BCI-simulation

Machine learning framework that simulates basic brain-computer interfaces for thought-driven prosthetic limb control.

Language: Python - Size: 254 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 0 - Forks: 0

jmcheon/Transformers

Language: Jupyter Notebook - Size: 25.8 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

RohanMenon/LipShiFT

This repo contains code to reproduce results for LipShiFT.

Language: Python - Size: 43 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

REFLOXO/NLP-Translator

Language Translation Tool built using Python and CustomTkinter. It supports both auto-translate and manual translate modes, allowing users to translate text between multiple languages.

Size: 1000 Bytes - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

alessoh/Neural-Symbolic-Superintelligence-scaling

Neural-Symbolic-Superintelligence scaling

Size: 5.86 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

sine2pi/Echo

Whisper like ASR transformer model. Experimental.

Language: Python - Size: 410 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

shuddha2021/Memorized-Q-A-Simulator

🤖 A toy Transformer Q&A model simulator demonstrating core concepts of large language models through memorized Q&A pairs. Educational demo with interactive web interface.

Language: Python - Size: 27.3 KB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

bniladridas/rag-transformer

Searches through a diverse dataset of machine learning, science fiction, and cosmic content.

Language: Python - Size: 898 KB - Last synced at: 2 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0

Related Topics
deep-learning 68 transformer 60 pytorch 57 machine-learning 51 nlp 47 attention-mechanism 36 transformers 33 natural-language-processing 30 transformer-models 28 python 27 artificial-intelligence 22 tensorflow 22 llm 18 computer-vision 18 ai 18 nlp-machine-learning 15 vision-transformer 14 gpt 13 self-attention 13 neural-network 13 attention-is-all-you-need 12 language-model 12 large-language-models 9 keras 9 python3 9 deep-neural-networks 9 transformer-encoder 9 lstm 9 numpy 8 attention-model 8 tensorflow2 8 ml 8 bert 8 neural-networks 7 generative-model 7 pytorch-implementation 7 machine-translation 7 fine-tuning 7 llms 6 transformer-pytorch 6 transformers-models 6 seq2seq 6 bert-model 6 huggingface 6 pandas 6 deeplearning 5 classification 5 translation 5 gpt-2 5 neural-machine-translation 5 vit 5 generative-ai 5 question-answering 5 attention 5 rnn 5 lstm-neural-networks 5 sentiment-analysis 4 huggingface-transformers 4 multimodal 4 reinforcement-learning 4 openai 4 chatbot 4 word-embeddings 4 encoder-decoder-architecture 4 time-series 4 transfer-learning 4 text-classification 4 encoder-decoder 4 machine-learning-algorithms 4 t5-model 3 language-modeling 3 web-interface 3 cnn 3 roberta 3 glove-embeddings 3 graph-neural-networks 3 multimodal-learning 3 resnet 3 multimodal-deep-learning 3 sequence-to-sequence-models 3 text-summarization 3 self-supervised-learning 3 bag-of-words 3 transformer-tensorflow2 3 seq2seq-model 3 time-series-forecasting 3 gpt-4 3 llm-training 3 graph-neural-network 3 convolutional-neural-networks 3 multi-head-attention 3 keras-tensorflow 3 data-visualization 3 albert 3 llama 3 continual-learning 3 image-classification 3 research-project 3 natural-language-generation 3 chatgpt 3