Reinforced-lib

Latest version: v1.1.4

Safety actively analyzes 688007 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 1 of 3

1.1.4

- Add experimental masked MAB agent.

1.1.3

- Add epsilon decay in e-greedy MAB.

1.1.2

- Update dependencies.
- Fix error with action space size of one.

1.1.1

Improvements:
- Update documentation.
- Add reference to the SoftwareX paper.

Fix:
- Normal Thompson sampling allows the `lam` parameter to be zero.
- Bernoulli Thompson sampling is stationary by default.
- Update the default value of the `decay` parameter the in `ra-sim` example.

1.1.0

Major API changes:
- Migrate from `haiku` (deprecated) to `flax` as the base naural network library.
- Update agent names to match literature:
- QLearning (deep Q-learning) -> DQN,
- DQN (deep double Q-learning) -> DDQN.
- Move particle filter from `agents` to `utils`.
- New behavior of loggers - all declared loggers get values from all sources.

New functionalities:
- Add Weights & Biases logger.

Other important changes:
- Fix updates with empty replay buffer.
- Fix logging of arrays to TensorBoard.
- Minor improvements in documentation.
- Rewrite Gymnasium integration example in documentation.
- Improve the CCOD example to better reflect the original implementation.

1.0.4

Improvements:
- Update documentation.
- Enable the use of 64-bit JAX.

Page 1 of 3

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.