* bump ONNX protobuf file and support external tensors format * new "skinny" kernels for avx2/fma f32 multiplication (positive impact on low, non 1 batch size for DNN-heavy loads)
0.16.7
* Softmax is now an operator in core, coming with a direct quantized implementation * new TypedFact constructor API ( f32::fact(&[1, 4, 12]), f32::fact(shape!(Symbol::from('N'), 12))) * fixes and optimisation of re-quantization pipeline * fixes around symbols in NNEF/OPL
0.16.6
* Various changes around quantization support (qi32 appearance)
0.16.5
* Intel optimisation are back * Range is now more flexible, should unlock some BERT models with symbolic dimensions.
0.16.4
* some optimisations in depthwise convolutions * various bugfixes * [Breaking] Fixed nnef "tile" operator definition ("repeats" is plural). As a consequence models using "tile" serialized with tract with prior versions can not be loaded anymore (and vice-versa).
0.16.3
* [Breaking] tract-opl models Scan syntax changed a bit. Models exported by <0.16.2 are loadable in >=0.16.2, but not the other way around. * Optimisation in deconv