A framework for training large language models with custom loss functions and model architectures. This project provides a flexible and extensible training pipeline that supports various model types, datasets, and training configurations.
- Clone the repository:
git clone https://github.com/efrick2002/highly-rewarding.git- Run the setup script:
cd highly-rewarding
source setup.shNote: the setup script installs
uv. The environment in activated withsource .venv/bin/activate.
train.py: Main training scriptmodeling.py: Model architecture definitions and registrylosses.py: Custom loss function implementationsdataset.py: Dataset handling and data loadingutils.py: Utility functionsmodel_type_registry.py: Model type registration system
wandb login- Configure your training parameters in a YAML config file
- Run the training script:
deepspeed --num_gpus=8 train.py -c configs/bt_debug.yamlThe training configuration should be specified in a YAML file.
See configs for examples.
Contributions are welcome! Please feel free to submit a Pull Request.