We are excited to announce the initial release of nanoPPO, version 0.13! This release lays the foundation for reinforcement learning practitioners, providing a lightweight and efficient implementation of the Proximal Policy Optimization (PPO) algorithm.
**Highlights:**
- **PPO Implementation:** Besides supporting discrete action spaces in v0.1, now supporting continuous action spaces in v0.13 for a wide range of applications.
- **Ease of Use:** Simple API to get started with PPO training quickly.
- **Examples Included:** Contains examples to help users understand how to train agents on various environments.
- **Custom Environments:** We create two environments: PointMass1D and PointMass2D for easy testing of the PPO agent training.
- **Test Suite:** Initial test suite to ensure code quality and functionality.
**Installation:**
You can install nanoPPO via PyPI:
bash
pip install nanoPPO
Or clone the repository and install from source:
bash
git clone https://github.com/jamesliu/nanoPPO.git
cd nanoPPO
pip install .
**Support & Contribution:**
We welcome feedback, issues, and contributions. Please refer to our [contribution guidelines](https://github.com/jamesliu/nanoPPO/blob/main/CONTRIBUTING.md) for more details.
Thank you for your interest in nanoPPO, and we look forward to hearing your feedback and seeing what you build with it!