BigCodeEvalHarness
callbacks.
winogrande
as part of the pre-training run. To do this, you will need to augment the configuration with the EleutherEvalHarness
callback as such:
winogrande
.
eval_frequency
specified as part of the trainer’s loop (YAML) or in the TrainingLoop
object (Python) also controls the frequency of downstream validation; i.e., for your example above, validation on EEH task winogrande
will be run every 1K steps.
tasks
argument to configure downstream validation for more EEH tasks. Note that only a single generative EEH task may be specified per callback.
humaneval
, please augment the YAML configuration file with the the [BigCodeEvalHarness
](cerebras.modelzoo.trainer.extensions.bigcode.BigCodeEvalHarness “cerebras.modelzoo.trainer.extensions.bigcode.BigCodeEvalHarness”) callback as such:
BigCodeEvalHarness
callback object and pass it to the Trainer’s constructor as follows. Note that the BCEH arguments are passed to the callback via the BigCodeCLIArgs
object, comprising the list of supported BCEH command line arguments.
humaneval
.
BigCodeEvalHarness
callback to run downstream validation for more BCEH tasks.
BigCodeEvalHarness
callbacks.
Let’s augment the full YAML configuration file to run downstream validation on EEH tasks hellaswag
, gsm8k
and winogrande
, and BCEH task mbpp
with the callbacks as follows:
BigCodeEvalHarness
objects, respectively.
./pretrain_downstream_llama_8b.yaml
. Then, to run pre-training using the training script that comes packaged as part of ModelZoo, you can run the code below on the command line.
./pretrain_downstream_llama_8b.py
. Then, to run pre-training, all there is to do is to execute that python script.