Additions:
- Transformer modules - `Transformer`, `TransformerEncoder`, and `TransformerDecoder`.
- Attention modules - `Attention`, `MultiheadAttention`.
- `PositionalEncoding` module.
- Softmax as a (splittable) aggregation function.
- Concatenation as an aggregation function.
- Reshape transformation function.
- Slice transformation function.
- Learning rate decay as an optimizer argument.
- Batching support.
- Progress bar for tracking dataset-building progress.
- Special arithmetic predicates for addition (`add`), subtraction (`sub`), and modulo (`mod`).
Changes:
- Concatenation can now be parametrized with a dimension.
- Refactored MLP module - the current version is more compact.
- Changed the behavior of `Variable`/`V` factory - only the first letter is capitalized.
- Performance improvements.