- All inputs (not just nonbundle activities) are used to train new bundles.
- **Increases** in element activities are used to learn sequences, rather than absolute values.
- Instead of choosing one goal every time step, it is now permissible not to choose a goal. Sequences are chosen as goals probabilistically based on their expected value.
Additional context
I just hit a roadblock in the development of Becca 8. The approach I'm using estimates reward for state-action pairs (similar to a traditional value function). I realized I need to revert back to the approach in Becca 7 where I used state-action-state tuples as my basis. This allows for state predictions, world modeling and flexible multi-step planning. Becca 8 taught me several things about how to best represent and combine inputs. Now I need to move on to Becca 9 and integrate the best of both 7 and 8. I'll release the current state as Becca 0.8.2 and move on.