Decoder-only language models by EleutherAI, designed for high-throughput training and capable zero-shot performance on a range of natural language tasks.
/gptj
directory within ModelZoo. Here’s how it’s organized:
/configs
: Contains YAML configuration files for GPT-J./continuous_pretraining/configs
: Contains configs for continuous pretraining of GPT-J.model.py
: The implementation of the GPT-Neox model.gptneox/model.py
: The implementation of the GPT-J model.gpt2_model.py
.Configuration | Description |
---|---|
params_gptj_6B.yaml | Standard 6B parameter GPT-J model. |
params_gptj_6B_muP.yaml | GPT-J model configured with μ-parameterization for scaling. |
params_gptj_6B_TRC2.yaml | GPT-J 6B model for continued pretraining on TRC2 datasets. |
params_gpt_neox_20B.yaml | GPT-Neox model with 20B parameters. |