Fixed - Dropout setting not properly passed in many attentions [facebookresearch/xformers123]
0.0.6
Fixed - Fix self attention optimization not being triggered, broken residual path [facebookresearch/xformers119] - Improve speed by not using contiguous Tensors when not needed [facebookresearch/xformers119]
Fixed - Homogenizing the masks, additive or bool [facebookresearch/xformers79][facebookresearch/xformers85][facebookresearch/xformers86] - Fix causality flag not being respected [facebookresearch/xformers103] - Enabling FusedLayerNorm by default in the factory if Triton is available - Fixing Favor with fp16 - Fixing Favor trainability
Added - Fused dropout/bias/activation layer [facebookresearch/xformers58] - Fused layernorm used by default in the factory [facebookresearch/xformers92]