18th November 2023
nanoDPO v0.1 Release, a pioneering implementation of Direct Preference Optimization (DPO) for time series data, inspired by "Direct Preference Optimization: Your Language Model is Secretly a Reward Model," the cutting-edge DPO approach in language model fine-tuning.
Key Features:
- Causal Transformer and LSTM Integration: Incorporating Causal Transformer and LSTM models to handle time series data effectively.
- DPO Algorithm Implementation: Direct Preference Optimization for nuanced understanding and prediction of time series trends.
- DPO and Multi-Class Trainers: Two distinct training models catering to different time series analysis requirements.
- Customizable Training Configurations: Enhanced flexibility with adjustable learning rates, batch sizes, and model specifications.
- Robust performance metrics including accuracy and loss visualizations.
- Compatibility with popular machine learning tools like PyTorch and wandb.
Documentation:
For more information, visit the GitHub README and the detailed Documentation.