models
gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
About
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
GitHub Stats
Stars7.4k
Forks1.1k
Watchers0
Open Issues0
Details
LanguagePython
LicenseApache-2.0
Deploymentboth
StatusActive
RepositoryEleutherAI/gpt-neox
Last push2/3/2026