Sentence-transformers

Latest version: v4.0.1

Safety actively analyzes 723650 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 10 of 24

6.87

4.0.1

pip install sentence-transformers[onnx-gpu]==4.0.1
pip install sentence-transformers[onnx]==4.0.1
pip install sentence-transformers[openvino]==4.0.1


> [!TIP]
> My [Training and Finetuning Reranker Models with Sentence Transformers v4](https://huggingface.co/blog/train-reranker) blogpost is an excellent place to learn 1) why finetuning rerankers makes sense and 2) how you can do it, too!

Reranker (Cross Encoder) training refactor (3222)
The v4.0 release centers around this huge modernization of the training approach for `CrossEncoder` models, following v3.0 which introduced the same for `SentenceTransformer` models. Whereas training before v4.0 used to be all about `InputExample`, `DataLoader` and `model.fit`, the new training approach relies on 5 components. You can learn more about these components in our [Training and Finetuning Embedding Models with Sentence Transformers v4](https://huggingface.co/blog/train-reranker) blogpost. Additionally, you can read the new [Training Overview](https://sbert.net/docs/cross_encoder/training_overview.html), check out the [Training Examples](https://sbert.net/docs/cross_encoder/training/examples.html), or read this summary:

1. [Dataset](https://sbert.net/docs/sentence_transformer/training_overview.html#dataset)
A training [`Dataset`](https://huggingface.co/docs/datasets/main/en/package_reference/main_classes#datasets.Dataset) or [`DatasetDict`](https://huggingface.co/docs/datasets/main/en/package_reference/main_classes#datasets.DatasetDict). This class is much more suited for sharing & efficient modifications than lists/DataLoaders of `InputExample` instances. A `Dataset` can contain multiple text columns that will be fed in order to the corresponding loss function. So, if the loss expects (anchor, positive, negative) triplets, then your dataset should also have 3 columns. The names of these columns are irrelevant. If there is a "label" or "score" column, it is treated separately, and used as the labels during training.
A `DatasetDict` can be used to train with multiple datasets at once, e.g.:
python
DatasetDict({
natural_questions: Dataset({
features: ['anchor', 'positive'],
num_rows: 392702
})
gooaq: Dataset({
features: ['anchor', 'positive', 'negative'],
num_rows: 549367
})
stsb: Dataset({
features: ['sentence1', 'sentence2', 'label'],
num_rows: 5749
})
})

When a `DatasetDict` is used, the `loss` parameter to the `CrossEncoderTrainer` must also be a dictionary with these dataset keys, e.g.:
python
{
'natural_questions': CachedMultipleNegativesRankingLoss(...),
'gooaq': CachedMultipleNegativesRankingLoss(...),
'stsb': BinaryCrossEntropyLoss(...),
}

2. [Loss Function](https://sbert.net/docs/cross_encoder/training_overview.html#loss-function)
A loss function, or a dictionary of loss functions like described above.
3. [Training Arguments](https://sbert.net/docs/cross_encoder/training_overview.html#training-arguments)
A CrossEncoderTrainingArguments instance, subclass of a [TrainingArguments](https://huggingface.co/docs/transformers/main/en/main_classes/trainer#transformers.TrainingArguments) instance. This powerful class controls the specific details of the training.
4. [Evaluator](https://sbert.net/docs/cross_encoder/training_overview.html#evaluator)
An optional [`SentenceEvaluator`](https://sbert.net/docs/package_reference/evaluation.html) instance. Unlike before, models can now be evaluated both on an evaluation dataset with some loss function and/or a `SentenceEvaluator` instance.
5. [Trainer](https://sbert.net/docs/sentence_transformer/training_overview.html#trainer)
The new `CrossEncoderTrainer` instance based on the `transformers` `Trainer`. This instance can be initialized with a CrossEncoder model, a CrossEncoderTrainingArguments class, a SentenceEvaluator, a training and evaluation Dataset/DatasetDict and a loss function/dict of loss functions. Most of these parameters are optional. Once provided, all you have to do is call `trainer.train()`.

Some of the major features that are now implemented include:
* MultiGPU Training (Data Parallelism (DP) and Distributed Data Parallelism (DDP))
* bf16 training support
* Loss logging
* Evaluation datasets + evaluation loss
* Improved callback support (built-in via Weights and Biases, TensorBoard, CodeCarbon, etc., as well as custom callbacks)
* Gradient checkpointing
* Gradient accumulation
* Improved model card generation
* Warmup ratio
* Pushing to the Hugging Face Hub on every model checkpoint
* Resuming from a training checkpoint
* Hyperparameter Optimization

This script is a minimal example (no evaluator, no training arguments) of training [`mpnet-base`](https://huggingface.co/microsoft/mpnet-base) on a part of the [`sentence-transformers/hotpotqa` dataset](https://huggingface.co/datasets/sentence-transformers/hotpotqa) using [`BinaryCrossEntropyLoss`](https://sbert.net/docs/package_reference/cross_encoder/losses.html#binarycrossentropyloss):

python
from datasets import load_dataset
from sentence_transformers import CrossEncoder, CrossEncoderTrainer
from sentence_transformers.cross_encoder.losses import BinaryCrossEntropyLoss

1. Define the model. Either from scratch of by loading a pre-trained model
model = CrossEncoder("microsoft/mpnet-base")

2. Load a dataset to finetune on
dataset = load_dataset("sentence-transformers/hotpotqa", "triplet", split="train")

def triplet_to_labeled_pair(batch):
anchors = batch["anchor"]
positives = batch["positive"]
negatives = batch["negative"]
return {
"sentence_A": anchors * 2,
"sentence_B": positives + negatives,
"labels": [1] * len(positives) + [0] * len(negatives),
}

dataset = dataset.map(triplet_to_labeled_pair, batched=True, remove_columns=dataset.column_names)
train_dataset = dataset.select(range(10_000))
eval_dataset = dataset.select(range(10_000, 11_000))

3. Define a loss function
loss = BinaryCrossEntropyLoss(model)

4. Create a trainer & train
trainer = CrossEncoderTrainer(
model=model,
train_dataset=train_dataset,
eval_dataset=eval_dataset,
loss=loss,
)
trainer.train()

5. Save the trained model
model.save_pretrained("models/mpnet-base-hotpotqa")
model.push_to_hub("mpnet-base-hotpotqa")


Additionally, trained models now automatically produce extensive model cards. Each of the following models were trained using some script from the [Training Examples](https://sbert.net/docs/cross_encoder/training/examples.html), and the model cards were not edited manually whatsoever:
* [tomaarsen/reranker-MiniLM-L12-gooaq-bce](https://huggingface.co/tomaarsen/reranker-MiniLM-L12-gooaq-bce)
* [tomaarsen/reranker-msmarco-MiniLM-L12-H384-uncased-lambdaloss](https://huggingface.co/tomaarsen/reranker-msmarco-MiniLM-L12-H384-uncased-lambdaloss)
* [tomaarsen/reranker-distilroberta-base-nli](https://huggingface.co/tomaarsen/reranker-distilroberta-base-nli)

Prior to the Sentence Transformer v4 release, all reranker models would be trained using the [`CrossEncoder.fit`](https://sbert.net/docs/package_reference/cross_encoder/cross_encoder.html#sentence_transformers.cross_encoder.CrossEncoder.fit) method. Rather than deprecating this method, starting from v4.0, this method will use the [`CrossEncoderTrainer`](https://sbert.net/docs/package_reference/cross_encoder/trainer.html#sentence_transformers.cross_encoder.trainer.CrossEncoderTrainer) behind the scenes. This means that your old training code should still work, and should even be upgraded with the new features such as multi-gpu training, loss logging, etc. That said, the new training approach is much more powerful, so it is **recommended** to write new training scripts using the new approach.

To help you out, all of the Cross Encoder (a.k.a. reranker) training scripts were updated to use the new Trainer-based approach.

Is finetuning worth it?

Finetuning reranker models on your data is very valuable. Consider for example these 2 models that I finetuned on 100k samples from the GooAQ dataset in 30 minutes and 1 hour, respectively. After finetuning, my models heavily outperformed general-purpose reranker models, even though GooAQ is a very generic dataset/domain!

* [tomaarsen/reranker-ModernBERT-base-gooaq-bce](https://huggingface.co/tomaarsen/reranker-ModernBERT-base-gooaq-bce)
* [tomaarsen/reranker-ModernBERT-large-gooaq-bce](https://huggingface.co/tomaarsen/reranker-ModernBERT-large-gooaq-bce)

![Model size vs NDCG for Rerankers on GooAQ](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/train-reranker/reranker_gooaq_model_size_ndcg.png)

Read my [Training and Finetuning Reranker Models with Sentence Transformers v4](https://huggingface.co/blog/train-reranker) blogpost for many more details on these models and how they were trained.

Resources:

* How to **use** Cross Encoder models? [Cross Encoder > Usage](https://sbert.net/docs/cross_encoder/usage/usage.html)
* What Cross Encoder **models** can I use? [Cross Encoder > Pretrained Models](https://sbert.net/docs/cross_encoder/pretrained_models.html)
* How do I **train/finetune** a Cross Encoder model? [Cross Encoder > Training Overview](https://sbert.net/docs/cross_encoder/training_overview.html)

Refactor Stats
* Code:
* New Trainer, Training Arguments, Data Collator, Model Card generation + template, with backwards compatibility
* [11 new losses](https://sbert.net/docs/package_reference/cross_encoder/losses.html)
* [1 new, 3 refactored, 6 deprecated evaluators](https://sbert.net/docs/package_reference/cross_encoder/evaluation.html)
* Tests:
* [84 tests for CrossEncoder Loading, inference, training, etc.](https://github.com/UKPLab/sentence-transformers/tree/master/tests/cross_encoder)
* Docs:
* All new [Training Overview](https://sbert.net/docs/cross_encoder/training_overview.html), [Loss Overview](https://sbert.net/docs/cross_encoder/loss_overview.html), [API Reference](https://sbert.net/docs/package_reference/cross_encoder/index.html) docs
* [5 new, 1 refactored training examples docs pages](https://sbert.net/docs/cross_encoder/training/examples.html)
* [13 new, 6 refactored training scripts](https://github.com/UKPLab/sentence-transformers/tree/master/examples/cross_encoder/training)
* Migration guide ([2.x -> 3.x](https://sbert.net/docs/migration_guide.html#migrating-from-v2-x-to-v3-x), [3.x -> 4.x](https://sbert.net/docs/migration_guide.html#migrating-from-v3-x-to-v4-x))

Small Features
* Introduce `show_progress_bar` for the `InformationRetrievalEvaluator` (3227)
* Replace `SubsetRandomSampler` with `RandomSampler` in the default batch sampler, should result in reduced memory usage and increased training speed! (3261)
* Allow resuming from checkpoint when training with the deprecated `SentenceTransformer.fit` (3269)
* Allow truncation and setting `model.max_seq_length` for CLIP models (2969)

Bug Fixes
* Fixed `MatryoshkaLoss` with `n_dims_per_step` and an unsorted `matryoshka_dims` crashing (3203)
* Fixed `GISTEmbedLoss` failing with some base models whose tokenizers don't have the `vocab` attribute (3219, 3226)
* Fixed support of `Asym`-based `SentenceTransformer` models (3220, 3244)
* Fixed some evaluator outputs not being converted to a Python float, i.e. staying as `numpy` or `torch` (3277)

Examples
* Improved TSDAE examples (3263, 3265)

Note
The `v4.0.0` version did not include the `model_card_template.md` in the package, this has been resolved in `v4.0.1` via ba1260d58c8804e97989e5af06ac90a0f4be8594.

What's Changed
* fix MatryoshkaLoss bug: sort sampled dimension indices to maintain descending dimension order by emapco in https://github.com/UKPLab/sentence-transformers/pull/3203
* [`docs`] Resolve broken URL due to weird & behaviour in pretrained ST models by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3213
* Update Evaluation Script for Reranking by milistu in https://github.com/UKPLab/sentence-transformers/pull/3198
* [`docs`] Update incorrect name: pairwise_similarity -> similarity_pairwise by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3224
* [`fix`] Use .get_vocab() instead of .vocab for checking tokenizer vocabulary by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3226
* [`feat`] Add progress bar support for corpus in IR Evaluator by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3227
* Update CoSENTLoss.py documentation by johneckberg in https://github.com/UKPLab/sentence-transformers/pull/3230
* NoDuplicatesDataLoader Compatability with Asymmetric models by OsamaS99 in https://github.com/UKPLab/sentence-transformers/pull/3220
* [`fix`] Fix Syntax issue; move 'as fIn' to after the if-else in `STSDataReader` by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3235
* Model Card Compatability & BinaryClassificationEvaluator with Asymmetric Models by OsamaS99 in https://github.com/UKPLab/sentence-transformers/pull/3244
* [fix] Changed value error for missing model into FileNotFoundError by PhorstenkampFuzzy in https://github.com/UKPLab/sentence-transformers/pull/3238
* Add check for hpu and wrap_in_hpu_graph availability. by vshekhawat-hlab in https://github.com/UKPLab/sentence-transformers/pull/3249
* Replacing SubsetRandomSampler by RandomSampler in BATCH_SAMPLER by NohTow in https://github.com/UKPLab/sentence-transformers/pull/3261
* Fix: Reorder dataset columns for DenoisingAutoEncoderLoss in TSADE examples by HuangBugWei in https://github.com/UKPLab/sentence-transformers/pull/3263
* Update to fit_mixin.fit to allow fine tuning to resume from a checkpoint by NRamirez01 in https://github.com/UKPLab/sentence-transformers/pull/3269
* Fix: dynamic noise addition during training in TSADE examples by HuangBugWei in https://github.com/UKPLab/sentence-transformers/pull/3265
* [`typing`] Fix the type hints in CGISTEmbedLoss by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3272
* typing: fix typing on encode by stephantul in https://github.com/UKPLab/sentence-transformers/pull/3270
* feat: add 'Path' parameter for ModelCard template by sam-hey in https://github.com/UKPLab/sentence-transformers/pull/3253
* Always convert the evaluation metrics to float, also without a 'name' by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3277
* Add truncation to CLIP model by MrLoh in https://github.com/UKPLab/sentence-transformers/pull/2969
* [`v4`] CrossEncoder Training refactor - MultiGPU, loss logging, bf16, etc. by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3222
* Bump jinja2 from 3.1.5 to 3.1.6 in /docs by dependabot in https://github.com/UKPLab/sentence-transformers/pull/3282
* Update the core README in preparation for the v4.0 release by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3283
* Make minor updates to docs by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3285
* Add the .htaccess to git, automatically include it in builds by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3286
* Update main description by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3287

New Contributors
* emapco made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3203
* OsamaS99 made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3220
* PhorstenkampFuzzy made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3238
* vshekhawat-hlab made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3249
* NohTow made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3261
* HuangBugWei made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3263
* NRamirez01 made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3269
* stephantul made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3270
* sam-hey made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3253
* MrLoh made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/2969

A special shoutout to milistu for contributing the LambdaLoss & ListNetLoss and yjoonjang for contributing the ListMLELoss, PListMLELoss, and RankNetLoss. Much appreciated, you really helped improve this release!

**Full Changelog**: https://github.com/UKPLab/sentence-transformers/compare/v3.4.1...v4.0.1

3.4.1

pip install sentence-transformers[onnx-gpu]==3.4.1
pip install sentence-transformers[onnx]==3.4.1
pip install sentence-transformers[openvino]==3.4.1


Full Model2Vec integration
This release introduces support to load an efficient [Model2Vec](https://huggingface.co/models?library=model2vec) embedding model directly in Sentence Transformers:
python
from sentence_transformers import SentenceTransformer

Download from the 🤗 Hub
model = SentenceTransformer(
"minishlab/potion-base-8M",
device="cpu",
)

Run inference
sentences = [
'Gadofosveset-enhanced MR angiography of carotid arteries: does steady-state imaging improve accuracy of first-pass imaging?',
'To evaluate the diagnostic accuracy of gadofosveset-enhanced magnetic resonance (MR) angiography in the assessment of carotid artery stenosis, with digital subtraction angiography (DSA) as the reference standard, and to determine the value of reading first-pass, steady-state, and "combined" (first-pass plus steady-state) MR angiograms.',
'In a longitudinal study we investigated in vivo alterations of CVO during neuroinflammation, applying Gadofluorine M- (Gf) enhanced magnetic resonance imaging (MRI) in experimental autoimmune encephalomyelitis, an animal model of multiple sclerosis. SJL/J mice were monitored by Gadopentate dimeglumine- (Gd-DTPA) and Gf-enhanced MRI after adoptive transfer of proteolipid-protein-specific T cells. Mean Gf intensity ratios were calculated individually for different CVO and correlated to the clinical disease course. Subsequently, the tissue distribution of fluorescence-labeled Gf as well as the extent of cellular inflammation was assessed in corresponding histological slices.',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
[3, 1024]

Get the similarity scores for the embeddings
similarities = model.similarity(embeddings[0], embeddings[1:])
print(similarities)

3.4.0

pip install sentence-transformers[onnx-gpu]==3.4.0
pip install sentence-transformers[onnx]==3.4.0
pip install sentence-transformers[openvino]==3.4.0


Matryoshka & Cached loss compatibility (3068, 3107)
It is now possible to combine the strong Cached losses ([CachedMultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedmultiplenegativesrankingloss), [CachedGISTEmbedLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedgistembedloss), [CachedMultipleNegativesSymmetricRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedmultiplenegativessymmetricrankingloss)) with the Matryoshka loss modifier:

python
from sentence_transformers import SentenceTransformer, SentenceTransformerTrainer, losses
from datasets import Dataset

model = SentenceTransformer("microsoft/mpnet-base")
train_dataset = Dataset.from_dict({
"anchor": ["It's nice weather outside today.", "He drove to work."],
"positive": ["It's so sunny.", "He took the car to the office."],
})
loss = losses.CachedMultipleNegativesRankingLoss(model, mini_batch_size=16)
loss = losses.MatryoshkaLoss(model, loss, [768, 512, 256, 128, 64])

trainer = SentenceTransformerTrainer(
model=model,
train_dataset=train_dataset,
loss=loss,
)
trainer.train()


See for example [tomaarsen/mpnet-base-gooaq-cmnrl-mrl](https://huggingface.co/tomaarsen/mpnet-base-gooaq-cmnrl-mrl) which was trained with [CachedMultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedmultiplenegativesrankingloss) (CMNRL) with the Matryoshka loss modifier (MRL).

Resolve memory leak when Model and Trainer are reinitialized (3144)
Due to a circular dependency in the `SentenceTransformerTrainer` -> `SentenceTransformer` -> `SentenceTransformerModelCardData` -> `SentenceTransformerTrainer`, deleting the trainer and model still doesn't clear them up via garbage disposal. I've moved a lot of components around, and now `SentenceTransformerModelCardData` does not need to store the `SentenceTransformerTrainer`, breaking the cycle.

We ran the seed optimization script (which frequently creates and deletes models and trainers):
* **Before**: Approximate highest recorded VRAM:

16332MiB / 24576MiB

* **After**: Approximate highest recorded VRAM:

8222MiB / 24576MiB


Small Features
* Add Matthews Correlation Coefficient to the BinaryClassificationEvaluator in 3051.
* Add a triplet `margin` parameter to the TripletEvaluator in 2862.
* Put dataset information in the automatically generated model card in "expanding sections" blocks if there are many datasets in 3088.
* Add multi-GPU (and CPU multi-process) support for [`mine_hard_negatives`](https://sbert.net/docs/package_reference/util.html#sentence_transformers.util.mine_hard_negatives) in 2967.

Notable Bug Fixes
* Subsequent batches were identical when using the `no_duplicates` Batch Sampler (3069). This has been resolved in 3073
* The old-style `model.fit()` training with `write_csv` on an evaluator would crash (3062). This has been resolved in 3066.
* The output types of some evaluators were `np.float` instead of `float` (3075). This has been resolved in 3076 and 3096.
* It was not possible to specify a `revision` or `cache_dir` when loading a PEFT Adapter model (3061). This has been resolved in 3079 and 3174.
* The CrossEncoder was lazily placed on the incorrect device, did not respond to `model.to` (3078). This has been resolved in 3104.
* If a model used a custom module with custom kwargs, those `kwargs` keys were not saved in `modules.json` correctly, e.g. relevant for [jina-embeddings-v3](https://huggingface.co/jinaai/jina-embeddings-v3) (#3111). This has been resolved in 3112.
* `HfArgumentParser(SentenceTransformerTrainingArguments)` would crash due to `prompts` typing (3090). This has been resolved in 3178.

Example Updates
* Update the quantization script in 3070.
* Update the seed optimization script in 3092.
* Update the TSDAE scripts in 3137.
* Add PEFT Adapter script in 3180.

Documentation Updates
* Add PEFT Adapter documentation in 3180.
* Add links to [backend-export](https://huggingface.co/spaces/sentence-transformers/backend-export) in [Speeding up Inference](https://sbert.net/docs/sentence_transformer/usage/efficiency.html).

All Changes
* [`training`] Pass `steps`/`epoch`/`output_path` to Evaluator during training by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3066
* [`examples`] Update the quantization script by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3070
* [`fix`] Fix different batches per epoch in NoDuplicatesBatchSampler by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3073
* [`docs`] Add links to backend-export in Speeding up Inference by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3071
* add MCC to BinaryClassificationEvaluator by JINO-ROHIT in https://github.com/UKPLab/sentence-transformers/pull/3051
* support cached losses in combination with matryoshka loss by Marcel256 in https://github.com/UKPLab/sentence-transformers/pull/3068
* align model_card_templates.py with code by amitport in https://github.com/UKPLab/sentence-transformers/pull/3081
* converting np float result to float in binary classification evaluator by JINO-ROHIT in https://github.com/UKPLab/sentence-transformers/pull/3076
* Add triplet margin for distance functions in TripletEvaluator by zivicmilos in https://github.com/UKPLab/sentence-transformers/pull/2862
* [`model_card`] Keep the model card readable even with many datasets by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3088
* [`docs`] Add NanoBEIR to the Training Overview evaluators by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3089
* [fix] revision of the adapter model can now be specified. by pesuchin in https://github.com/UKPLab/sentence-transformers/pull/3079
* [`docs`] Update from Sphinx==3.5.4 to 8.1.3, recommonmark -> myst-parser by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3099
* normalize to float in NanoBEIREvaluator, InformationRetrievalEvaluator, MSEEvaluator by JINO-ROHIT in https://github.com/UKPLab/sentence-transformers/pull/3096
* [`docs`] List 'prompts' as a key training argument by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3101
* revert float type cast manually in BinaryClassificationEvaluator by JINO-ROHIT in https://github.com/UKPLab/sentence-transformers/pull/3102
* update train_sts_seed_optimization with SentenceTransformerTrainer by JINO-ROHIT in https://github.com/UKPLab/sentence-transformers/pull/3092
* Fix cross encoder device issue by susnato in https://github.com/UKPLab/sentence-transformers/pull/3104
* [`enhancement`] Make MultipleNegativesRankingLoss easier to understand by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3100
* [`fix`] Fix breaking change in PyLate when loading modules by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3110
* multi-GPU support for mine_hard_negatives by alperctnkaya in https://github.com/UKPLab/sentence-transformers/pull/2967
* raises error when dataset is an empty list in NanoBEIREvaluator by JINO-ROHIT in https://github.com/UKPLab/sentence-transformers/pull/3122
* Added a note to the documentation stating that the similarity method does not support embeddings other than non-quantized ones. by pesuchin in https://github.com/UKPLab/sentence-transformers/pull/3131
* [`typo`] Add missing space between sentences in error message by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3125
* raises ValueError when num_label !=1 when using Crossencoder.rank() by JINO-ROHIT in https://github.com/UKPLab/sentence-transformers/pull/3126
* fix backward pass for cached losses by Marcel256 in https://github.com/UKPLab/sentence-transformers/pull/3114
* Adding evaluation checks to prevent Transformer ValueError by stsfaroz in https://github.com/UKPLab/sentence-transformers/pull/3105
* [typo] Fix incorrect spelling for "corpus" by ignasgr in https://github.com/UKPLab/sentence-transformers/pull/3154
* [`fix`] Save custom module `kwargs` if specified by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3112
* [`memory`] Avoid storing trainer in ModelCardCallback and SentenceTransformerModelCardData by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3144
* Suport for embedded representation by Radu1999 in https://github.com/UKPLab/sentence-transformers/pull/3156
* [DRAFT] tests for nanobeir evaluator by JINO-ROHIT in https://github.com/UKPLab/sentence-transformers/pull/3127
* Update TSDAE examples with SentenceTransformerTrainer by JINO-ROHIT in https://github.com/UKPLab/sentence-transformers/pull/3137
* [`docs`] Update the Static Embedding example snippet by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3177
* fix: propagate cache dir to find adapter by lauralehoczki11 in https://github.com/UKPLab/sentence-transformers/pull/3174
* [`fix`] Use HfArgumentParser-compatible typing for prompts by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3178
* testcases for community detection by JINO-ROHIT in https://github.com/UKPLab/sentence-transformers/pull/3163
* [`docs`] Add PEFT documentation + training example by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3180
* [`tests`] Make TripletEvaluator test more consistent by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3183
* [`deprecation`] Clarify that datasets and readers are deprecated since v3 by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3184
* [docs] Update the documentation surrounding Matryoshka + Cached losses by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3190

New Contributors
* JINO-ROHIT made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3051
* Marcel256 made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3068
* amitport made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3081
* zivicmilos made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/2862
* susnato made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3104
* alperctnkaya made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/2967
* stsfaroz made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3105
* ignasgr made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3154
* Radu1999 made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3156
* lauralehoczki11 made their first contribution in https://github.com/UKPLab/sentence-transformers/pull/3174

An explicit thanks to JINO-ROHIT who has made a large amount of contributions in this release.

**Full Changelog**: https://github.com/UKPLab/sentence-transformers/compare/v3.3.1...v3.4.0

3.3.1

pip install sentence-transformers[onnx-gpu]==3.3.1
pip install sentence-transformers[onnx]==3.3.1
pip install sentence-transformers[openvino]==3.3.1


Details
If you're loading model under this scenario:
* Your model is hosted on Hugging Face.
* Your model is private.
* You haven't set the `HF_TOKEN` environment variable via `huggingface-cli login` or some other approach.
* You're passing the `token` argument to `SentenceTransformer` to load the model.

Then you may have encountered a crash in v3.3.0. This should be resolved now.

All Changes
* [`docs`] Fix the prompt link to the training script by tomaarsen in https://github.com/UKPLab/sentence-transformers/pull/3060
* [Fix] Resolve loading private Transformer model in version 3.3.0 by pesuchin in https://github.com/UKPLab/sentence-transformers/pull/3058

**Full Changelog**: https://github.com/UKPLab/sentence-transformers/compare/v3.3.0...v3.3.1

3.3.0

pip install sentence-transformers[onnx-gpu]==3.3.0
pip install sentence-transformers[onnx]==3.3.0
pip install sentence-transformers[openvino]==3.3.0


OpenVINO int8 static quantization (https://github.com/UKPLab/sentence-transformers/pull/3025)
We introduce int8 static quantization using [OpenVINO](https://github.com/openvinotoolkit/openvino), a highly performant solution that outperforms all other current backends by a mile, at a minimal loss in performance. Here are the updated benchmarks:

<p align="center">
<img src="https://github.com/user-attachments/assets/96f96da5-b65e-4293-8b67-d47430aa5fae" width="50%" />
</p>

Quantizing directly to the Hugging Face Hub

python
from sentence_transformers import SentenceTransformer, export_static_quantized_openvino_model

1. Load a model with the OpenVINO backend
model = SentenceTransformer("all-MiniLM-L6-v2", backend="openvino")

2. Quantize the model to int8, push the model to https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2
as a pull request:
export_static_quantized_openvino_model(
model,
quantization_config=None,
model_name_or_path="sentence-transformers/all-MiniLM-L6-v2",
push_to_hub=True,
create_pr=True,
)

You can immediately use the model, even before it's merged, by using the `revision` argument:
python
from sentence_transformers import SentenceTransformer

pull_request_nr = 2 TODO: Update this to the number of your pull request
model = SentenceTransformer(
"all-MiniLM-L6-v2",
backend="openvino",
model_kwargs={"file_name": "openvino_model_qint8_quantized.xml"},
revision=f"refs/pr/{pull_request_nr}"
)

And once it's merged:
python
from sentence_transformers import SentenceTransformer

model = SentenceTransformer(
"all-MiniLM-L6-v2",
backend="openvino",
model_kwargs={"file_name": "openvino/openvino_model_qint8_quantized.xml"},
)


Quantizing locally
You can also quantize a model and save it locally:
python
from sentence_transformers import SentenceTransformer, export_static_quantized_openvino_model
from optimum.intel import OVQuantizationConfig

model = SentenceTransformer("all-mpnet-base-v2", backend="openvino")
model.save_pretrained("path/to/all-mpnet-base-v2-local")
quantization_config = OVQuantizationConfig() <- You can update settings here
export_static_quantized_openvino_model(model, quantization_config, "path/to/all-mpnet-base-v2-local")

And after quantizing, you can load it like so:
python
from sentence_transformers import SentenceTransformer

model = SentenceTransformer(
"path/to/all-mpnet-base-v2-local",
backend="openvino",
model_kwargs={"file_name": "openvino_model_qint8_quantized.xml"},
)


All [original Sentence Transformer models](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) already have these new ` openvino_model_qint8_quantized.xml` files, so you can load them without exporting directly! I would recommend making pull requests for [other models on Hugging Face](https://huggingface.co/models?library=sentence-transformers) that you'd like to see quantized.

Learn more about how to Speed up Inference in the documentation: https://sbert.net/docs/sentence_transformer/usage/efficiency.html

Training with Prompts (https://github.com/UKPLab/sentence-transformers/pull/2964)
Many modern embedding models are trained with “instructions” or “prompts” following the [INSTRUCTOR paper](https://arxiv.org/abs/2212.09741). These prompts are strings, prefixed to each text to be embedded, allowing the model to distinguish between different types of text.

For example, the [mixedbread-ai/mxbai-embed-large-v1](https://huggingface.co/mixedbread-ai/mxbai-embed-large-v1) model was trained with Represent this sentence for searching relevant passages: as the prompt for all queries. This prompt is stored in the [model configuration](https://huggingface.co/mixedbread-ai/mxbai-embed-large-v1/blob/main/config_sentence_transformers.json) under the prompt name "query", so users can specify that prompt_name in model.encode:

python
from sentence_transformers import SentenceTransformer

model = SentenceTransformer("mixedbread-ai/mxbai-embed-large-v1")
query_embedding = model.encode("What are Pandas?", prompt_name="query")
or
query_embedding = model.encode("What are Pandas?", prompt="Represent this sentence for searching relevant passages: ")
document_embeddings = model.encode([
"Pandas is a software library written for the Python programming language for data manipulation and analysis.",
"Pandas are a species of bear native to South Central China. They are also known as the giant panda or simply panda.",
"Koala bears are not actually bears, they are marsupials native to Australia.",
])
similarity = model.similarity(query_embedding, document_embeddings)
print(similarity)

Page 10 of 24

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.