GitHub / hkproj / pytorch-transformer-distributed
Distributed training (multi-node) of a Transformer model
JSON API: http://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/hkproj%2Fpytorch-transformer-distributed
PURL: pkg:github/hkproj/pytorch-transformer-distributed
Stars: 66
Forks: 29
Open issues: 0
License: None
Language: Python
Size: 4.03 MB
Dependencies parsed at: Pending
Created at: over 1 year ago
Updated at: about 2 months ago
Pushed at: about 1 year ago
Last synced at: about 2 months ago
Topics: collective-communication, data-parallelism, deep-learning, distributed-data-parallel, distributed-training, gradient-accumulation, machine-learning, model-parallelism, pytorch, tutorial