-------------------------------------------------------------------------------
major:
- New Algorithm: Randomized Popularity (Popularity)
- This bandit uses a probabilistically weighted selection based on mean arm rewards
- Currently, it assumes that the rewards are non-negative
-------------------------------------------------------------------------------