Documentation
* docs: Update contributor table (1071)
update contributor table ([`778d7a3`](https://github.com/embeddings-benchmark/mteb/commit/778d7a3bf85b2023cc8ba9b2c35a810dcfa5e924))
Fix
* fix: restructure test suite to avoid to many external dependencies (1070)
* fix: restructure test suite to avoid to many external dependencies
The intention behind this is to make tests faster and more robust. Especially to avoid false positives where the test works, but fails due to interaction with HF, the OS (CLI commands).
- Restructure test suite to make it easier to find specific tests and add tests in the future. e.g. added all evaluators test to a folder
- splits test_mteb into multiple segments, one for testing SentenceTransformers, one for testing evaluation workflow, one for testing integration with datasets
- avoid using encoders when testing evaluation workflow (one extra interaction with hf which can fail and it is slow)
- Added MockEncoders for testing torch, numpy and torch bf16 return types (see FP16 embeddings fail with BitextMining 1064)
- Added MockTask for all task types to avoid downloading too many datasets and to allow for more diverse without making the test suite too slow
The slowest test are now the mteb_rerank tests along with the integration tests. orionw was hoping we might be able to to replace the dataset with a dummy dataset. However the slowest atm. is 20s
This PR also added a few minor fixes (surprisingly few actually) for handling torch bf16 which was not handled and a minor fix to AbstaskClusteringFast (isaac-chung)
* fix: Remove precision for more readable scores and lower memory footprint (1075)
* fix: reduce precision of results for readability
* Added example cases
* Removed &34;dev&34; as a default split from flores
* remove flores results
* fixes errors due to merge ([`636018b`](https://github.com/embeddings-benchmark/mteb/commit/636018be6ec04a2051881abd0068ce018e485be5))
Refactor
* refactor: update TaskMetadata (1076)
* update TaskMetadata.py
* update _add_stats.grit
* update _add_stats.grit
* update 415 files
* update _add_stats.grit and TaskMetadata.py
* update _add_stats.grit
* update 86 files
* update 6 files update 341 files
* delete 1 file and update 416 files
* misc. ([`57c1c12`](https://github.com/embeddings-benchmark/mteb/commit/57c1c12727b6300465cb647f2e55c9af0049d759))
Unknown
* Update points table ([`cff1ae6`](https://github.com/embeddings-benchmark/mteb/commit/cff1ae6c2515b8aa97b15da7d72c47b824c7d94f))
* Update points table ([`86eb18c`](https://github.com/embeddings-benchmark/mteb/commit/86eb18c1c8a2d2c04c0a2752fd9302726119100f))
* Change refresh description (1074)
Update adding_a_model.md ([`d98dcd8`](https://github.com/embeddings-benchmark/mteb/commit/d98dcd8958ecaeda9b7f97f340412be42f6f9efb))