Fix
* fix: ensure that results from parallel datasets are formatted correctly (974)
* fix: ensure that results from parallel datasets are formatted corrected.
Additionally updated a few results.
* add pytest coverage
* remove unfinished results file
* Add test for multilingual subset loader
* removed upper bound on numpy
* sped up tests
* add trust remote code for new datasets
---------
Co-authored-by: Isaac Chung <chungisaac1217gmail.com> ([`6004ec7`](https://github.com/embeddings-benchmark/mteb/commit/6004ec7b6e99afb2d31a41784ac0b3d4a6ded935))
Unknown
* Update points table ([`ae27b5c`](https://github.com/embeddings-benchmark/mteb/commit/ae27b5c16bd2655663be28f86657569e049d7ea4))
* LLM2Vec models (926)
* adding llm2vec model loader
* fix merge
* update import error
Co-authored-by: Kenneth Enevoldsen <kennethcenevoldsengmail.com>
* llm2vec use instructions
* prompt_name separate out of kwargs
* format
* scores
* use flash attention if available
* fixed bug for retieval
* user can provide instructions for LLM2Vec
* fix type error
* making code py 3.8 and 3.9 compatible
* proper storing results
* type combination compliant with py 3.8, 3.9
* add points
* add semicolon
* updated scores
* format
---------
Co-authored-by: Kenneth Enevoldsen <kennethcenevoldsengmail.com> ([`a6c784c`](https://github.com/embeddings-benchmark/mteb/commit/a6c784c1835ce11a9b1a974f8a6d43439d165805))