models

gpt-neox

An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries

About

An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries

GitHub Stats

Stars7.4k
Forks1.1k
Watchers0
Open Issues0

Details

LanguagePython
LicenseApache-2.0
Deploymentboth
StatusActive
Last push2/3/2026