Skip to content

efrick2002/highly-rewarding

Repository files navigation

Highly Rewarding

A framework for training large language models with custom loss functions and model architectures. This project provides a flexible and extensible training pipeline that supports various model types, datasets, and training configurations.

Setup

  1. Clone the repository:
git clone https://github.com/efrick2002/highly-rewarding.git
  1. Run the setup script:
cd highly-rewarding
source setup.sh

Note: the setup script installs uv. The environment in activated with source .venv/bin/activate.

Project Structure

  • train.py: Main training script
  • modeling.py: Model architecture definitions and registry
  • losses.py: Custom loss function implementations
  • dataset.py: Dataset handling and data loading
  • utils.py: Utility functions
  • model_type_registry.py: Model type registration system

Usage

  1. wandb login
  2. Configure your training parameters in a YAML config file
  3. Run the training script:
deepspeed --num_gpus=8 train.py -c configs/bt_debug.yaml

Configuration

The training configuration should be specified in a YAML file.

See configs for examples.

Contributing

Contributions are welcome! Please feel free to submit a Pull Request.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors