Pretrained Sentence BERT models Yahoo Answers small embeddings
dataset
posted on 2023-05-03, 10:57 authored by Beatrix Miranda Ginn NielsenBeatrix Miranda Ginn Nielsen<p> Embeddings on 10% of the Yahoo Answers dataset using pretrained Sentence BERT models.<br>
</p>
<p>Yahoo Answers dataset: https://www.kaggle.com/datasets/yacharki/yahoo-answers-10-categories-for-nlp-csv</p>
<p>Indexes used can be found in the code repository.</p>
<p><br></p>
<p>Pretrained models used:</p>
<p> all-distilroberta-v1: https://huggingface.co/sentence-transformers/all-distilroberta-v1</p>
<p> all-MiniLM-L12-v2: https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2</p>
<p> all-mpnet-base-v2: https://huggingface.co/sentence-transformers/all-mpnet-base-v2</p>
<p> multi-qa-distilbert-cos-v1: https://huggingface.co/sentence-transformers/multi-qa-distilbert-cos-v1</p>
<p><br></p>
<p><br>
<br>
<br>
<br>
<br>
</p>
Funding
Danish Pioneer Centre for AI, DNRF grant number P1
History
Related Materials
- 1.
ORCID for corresponding depositor
Usage metrics
Categories
Keywords
Licence
Exports
RefWorksRefWorks
BibTeXBibTeX
Ref. managerRef. manager
EndnoteEndnote
DataCiteDataCite
NLMNLM
DCDC


