Multilingual decoder-only language model with ALiBi positional embeddings, designed to generalize across 46 natural and 13 programming languages.
bloom
directory.
configs/
: Contains YAML configuration files for various BLOOM variants.model.py
: Shared model code reused from GPT-2, modified to support ALiBi embeddings.gpt2_model.py
.Configuration | Description |
---|---|
params_bloom_7b.yaml | BLOOM-7B model with hidden_size=4096 , num_hidden_layers=30 , and num_heads=32 . Uses ALiBi position embeddings. |
model.position_embedding_type: alibi
model.alibi_trainable_slopes: false
(recommended, based on the ALiBi paper’s findings)