Technical University of Denmark
Browse

Pretrained Sentence BERT models Yahoo Answers small embeddings

dataset
posted on 2023-05-03, 10:57 authored by Beatrix Miranda Ginn NielsenBeatrix Miranda Ginn Nielsen
<p> Embeddings on 10% of the Yahoo Answers dataset using pretrained Sentence BERT models.<br> </p> <p>Yahoo Answers dataset: https://www.kaggle.com/datasets/yacharki/yahoo-answers-10-categories-for-nlp-csv</p> <p>Indexes used can be found in the code repository.</p> <p><br></p> <p>Pretrained models used:</p> <p> all-distilroberta-v1: https://huggingface.co/sentence-transformers/all-distilroberta-v1</p> <p> all-MiniLM-L12-v2: https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2</p> <p> all-mpnet-base-v2: https://huggingface.co/sentence-transformers/all-mpnet-base-v2</p> <p> multi-qa-distilbert-cos-v1: https://huggingface.co/sentence-transformers/multi-qa-distilbert-cos-v1</p> <p><br></p> <p><br> <br> <br> <br> <br> </p>

Funding

Danish Pioneer Centre for AI, DNRF grant number P1

History

Related Materials

ORCID for corresponding depositor

Usage metrics

    DTU Compute

    Categories

    Licence

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC