Technical University of Denmark
Browse

Sentence BERT models trained on STS benchmark

software
posted on 2023-09-28, 06:46 authored by Beatrix Miranda Ginn NielsenBeatrix Miranda Ginn Nielsen
<h3>Sentence BERT models trained on STS benchmark </h3> <p><br></p> <p>See sentence transformers (https://www.sbert.net/) for training and use of sentence BERT models in general.</p> <p>The STS benchmark dataset was fetched from https://sbert.net/datasets/stsbenchmark.tsv.gz</p> <p><br></p> <p>Models follow the naming scheme: </p> <p>sts_bert_[base model name]_[distance function]_[placement of most dissimilar]_z_[z score normalisation]_n_[normalisation of embeddings]_c_[centering of embeddings]_seed[random seed] </p> <p>Example: sts_bert_distilroberta-base_cos_dist_ORTHOGONAL_z_False_n_False_c_False_seed1</p> <p><br></p> <h4>Base model name</h4> <p>Three different base models were used:</p> <p>- distilroberta-base (https://huggingface.co/distilroberta-base)</p> <p>- microsoft-MiniLM-L12-H384-uncased (https://huggingface.co/microsoft/MiniLM-L12-H384-uncased)</p> <p>- microsoft-mpnet-base (https://huggingface.co/microsoft/mpnet-base)</p> <h4><br></h4> <h4>Distance function</h4> <p>Three different distance functions were used. </p> <p>- Cosine similarity (cos)</p> <p>- Cosine distance (cos_dist)</p> <p>- Euclidean distance (euclidean)</p> <h4><br></h4> <h4>Placement of most dissimilar</h4> <p>This describes how we want the embeddings of two texts which are completetly dissimilar to ideally be placed. </p> <p>In all these models we have chosen ORTHOGONAL, so the two embeddings of dissimilar texts should have 90 degrees between them. </p> <p>Another choice could have been OPPOSITE where the embeddings should have 180 degrees between them when they are completely dissimilar. </p> <h4><br></h4> <h4>Z-score normalisation</h4> <p>Whether z-score normalisation of the embeddings was performed during training. True/False.</p> <h4><br></h4> <h4>Normalisation of embeddings</h4> <p>Whether normalisation of embeddings was performed during training. True/False.</p> <h4><br></h4> <h4>Centering of embeddings</h4> <p>Whether centering of the embeddings was performed during training. True/False.</p> <h4><br></h4> <h4>Random seed</h4> <p>What random seed was used for training. </p> <p><br></p> <p><br></p> <p><br></p>

Funding

Danish Pioneer Centre for AI, DNRF grant number P1

History

Related Materials

ORCID for corresponding depositor

Usage metrics

    DTU Compute

    Categories

    Licence

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC