Feature
* feat: Use prompts instead of encode_corpus and encode_queries (1278)
* add prompt per task type
* fix prompt
* upd test
* lint
* fix test
* fix DeprecatedSummarizationEvaluator
* fix prompts
* add test
* lint
* logger info
* use task type only in model_encode
* lint
* update interface
* add prompt types to docs
* fix test
* mock tasks
* mock task registry
* remove last task_type
* fix tests
* lint
* fix test
* fix
* use wrapper and new prompts
* fix tests
* lint
* fix test
* remove conftest
* validate task to prompt_name
* override model prompts
* task to prompt name optional
* fix tests
* fix models
* remove task_to_prompt_name
* remove from mteb __init__
* update docs
* load existing model prompts if model_prompts is None
* fix
* lint
* change wrapper loader
* add wrapper class
* lint
* add wrapper file
* update logging
* upd logging
* refactor reranking
* lint
* remove prints ([`2a61821`](https://github.com/embeddings-benchmark/mteb/commit/2a61821d9294eb5b0cb053e1e676c199f23be12b))
Unknown
* Leaderboard (1235)
* Add leaderboard dev
* Renamed MTEBResults to TaskResult
* Moved model and model meta loading utilities into overview.py
* Added get_model_metas to retrieve filtered metadata for models
* Restructured results object and made it into a class instead of a dict
* Added utilities for filtering models on BenchmarkResults objects
* Added to_table utility function to BenchmarkResults
* Added serialization utilities to BenchmarkResults
* Attempted fixing tests
* Added get_model_metas to __init__
* Added get_benchmarks to __init__ and made it return all benchmarks by default
* Added get_benchmarks to __init__
* Made tasks hashable
* Added task filtering based on task objects on BenchmarkResults
* Added BenchmarkResults to __init__
* Added additional arguments to get_scores on two classes
* Made get_scores smarter on BenchmarkResult
* Added basic multilingual benchmark
* Modified benchmark to be able to easily access results
* Added useful properties and filtering functions to BenchmarkResults
* Added minimal functioning example
* Added smarter table, task-list updating and tried fixing dropdown scrolling
* Made restrict_results into a private function
Co-authored-by: Kenneth Enevoldsen <kennethcenevoldsengmail.com>
* Removed old leaderboard scripts
* Hardcoded max and min model size
* Removed redundant utils file
* Ran linting
* added leaderboard dependencies as optional
* Fixed union type error on Python 3.9
* Removed references to Dict in task aggregation
* Fixed name errors in _restrict_task_results
* Fixed _restrict_task_results
* Made hf_subsets={&39;default&39;} when the task is monolingual in _restric_task_results
* Task dropdown now gets filtered based on the other criteria
* Ran linting again
* Introduced hotfix for reranking test
* Added BenchmarkResults to __all__ in __init__
* Fixed validate_and_filter_scores method, and replaced _restric_task_results with it
---------
Co-authored-by: Kenneth Enevoldsen <kennethcenevoldsengmail.com> ([`094f922`](https://github.com/embeddings-benchmark/mteb/commit/094f9225973268fedae55b322eb9af7bb0ae2110))