language: []
library_name:sentence-transformerstags:-sentence-transformers-sentence-similarity-feature-extraction-generated_from_trainer-dataset_size:1115700-loss:MatryoshkaLoss-loss:MultipleNegativesRankingLossbase_model:UBC-NLP/serengeti-E250datasets: []
metrics:-pearson_cosine-spearman_cosine-pearson_manhattan-spearman_manhattan-pearson_euclidean-spearman_euclidean-pearson_dot-spearman_dot-pearson_max-spearman_maxwidget:-source_sentence:Ndegemwenyemdomomrefukatikatiyandege.sentences:-Panyaanayekimbiajuuyagurudumu.-Mtuanashindanakatikamashindanoyambio.-Ndegeanayeruka.-source_sentence:>- Msichana mchanga mwenye nywele nyeusi anakabili kamera na kushikilia mfuko wa karatasi wakati amevaa shati la machungwa na mabawa ya kipepeo yenye rangi nyingi.sentences:-Mwanamkemzeeanakataakupigwapicha.-mtuakilanamvulanamdogokwenyekijiachajiji-Msichanamchangaanakabilikamera.-source_sentence:>- Wanawake na watoto wameketi nje katika kivuli wakati kikundi cha watoto wadogo wameketi ndani katika kivuli.sentences:-Mwanamkenawatotonakukaachini.-Mwanamkehuyoanakimbia.-Watuwanasafirikwabaiskeli.-source_sentence:>- Mtoto mdogo anaruka mikononi mwa mwanamke aliyevalia suti nyeusi ya kuogelea akiwa kwenye dimbwi.sentences:->- Mtoto akiruka mikononi mwa mwanamke aliyevalia suti ya kuogelea kwenye dimbwi.-Someoneisholdingorangesandwalking-Mamanabintiwakinunuaviatu.-source_sentence:>- Mwanamume na mwanamke wachanga waliovaa mikoba wanaweka au kuondoa kitu kutoka kwenye mti mweupe wa zamani, huku watu wengine wamesimama au wameketi nyuma.sentences:-taihuruka-mwanamumenamwanamkewenyemikoba-Wanaumewawiliwameketikaribunamwanamke.pipeline_tag:sentence-similaritymodel-index:-name:SentenceTransformerbasedonUBC-NLP/serengeti-E250results:-task:type:semantic-similarityname:SemanticSimilaritydataset:name:ststest768type:sts-test-768metrics:-type:pearson_cosinevalue:0.7084016023985643name:PearsonCosine-type:spearman_cosinevalue:0.7080643276583263name:SpearmanCosine-type:pearson_manhattanvalue:0.7163851544290831name:PearsonManhattan-type:spearman_manhattanvalue:0.7066259909380899name:SpearmanManhattan-type:pearson_euclideanvalue:0.716171309296757name:PearsonEuclidean-type:spearman_euclideanvalue:0.7064427148038006name:SpearmanEuclidean-type:pearson_dotvalue:0.38463559218643695name:PearsonDot-type:spearman_dotvalue:0.3566836293112297name:SpearmanDot-type:pearson_maxvalue:0.7163851544290831name:PearsonMax-type:spearman_maxvalue:0.7080643276583263name:SpearmanMax-task:type:semantic-similarityname:SemanticSimilaritydataset:name:ststest512type:sts-test-512metrics:-type:pearson_cosinevalue:0.7059523092716506name:PearsonCosine-type:spearman_cosinevalue:0.7046582726338858name:SpearmanCosine-type:pearson_manhattanvalue:0.714245009590492name:PearsonManhattan-type:spearman_manhattanvalue:0.7048777976859945name:SpearmanManhattan-type:pearson_euclideanvalue:0.7150194670982656name:PearsonEuclidean-type:spearman_euclideanvalue:0.7055458365374757name:SpearmanEuclidean-type:pearson_dotvalue:0.3855295554891442name:PearsonDot-type:spearman_dotvalue:0.3585966097040326name:SpearmanDot-type:pearson_maxvalue:0.7150194670982656name:PearsonMax-type:spearman_maxvalue:0.7055458365374757name:SpearmanMax-task:type:semantic-similarityname:SemanticSimilaritydataset:name:ststest256type:sts-test-256metrics:-type:pearson_cosinevalue:0.7069259070512649name:PearsonCosine-type:spearman_cosinevalue:0.7072103115498357name:SpearmanCosine-type:pearson_manhattanvalue:0.7151518946293685name:PearsonManhattan-type:spearman_manhattanvalue:0.7050845216566457name:SpearmanManhattan-type:pearson_euclideanvalue:0.7154956682724514name:PearsonEuclidean-type:spearman_euclideanvalue:0.70486417475867name:SpearmanEuclidean-type:pearson_dotvalue:0.37291132473389677name:PearsonDot-type:spearman_dotvalue:0.3480769113927452name:SpearmanDot-type:pearson_maxvalue:0.7154956682724514name:PearsonMax-type:spearman_maxvalue:0.7072103115498357name:SpearmanMax-task:type:semantic-similarityname:SemanticSimilaritydataset:name:ststest128type:sts-test-128metrics:-type:pearson_cosinevalue:0.7022542784280805name:PearsonCosine-type:spearman_cosinevalue:0.7062378358777478name:SpearmanCosine-type:pearson_manhattanvalue:0.711575484251127name:PearsonManhattan-type:spearman_manhattanvalue:0.701312903814612name:SpearmanManhattan-type:pearson_euclideanvalue:0.7125043324593673name:PearsonEuclidean-type:spearman_euclideanvalue:0.7011154675785318name:SpearmanEuclidean-type:pearson_dotvalue:0.34394993785114003name:PearsonDot-type:spearman_dotvalue:0.31686351995727197name:SpearmanDot-type:pearson_maxvalue:0.7125043324593673name:PearsonMax-type:spearman_maxvalue:0.7062378358777478name:SpearmanMax-task:type:semantic-similarityname:SemanticSimilaritydataset:name:ststest64type:sts-test-64metrics:-type:pearson_cosinevalue:0.6950172826546709name:PearsonCosine-type:spearman_cosinevalue:0.6993973161633343name:SpearmanCosine-type:pearson_manhattanvalue:0.7059726901866531name:PearsonManhattan-type:spearman_manhattanvalue:0.6938542774412633name:SpearmanManhattan-type:pearson_euclideanvalue:0.7066346687971139name:PearsonEuclidean-type:spearman_euclideanvalue:0.6949014564343952name:SpearmanEuclidean-type:pearson_dotvalue:0.30982738809482646name:PearsonDot-type:spearman_dotvalue:0.2855406388879541name:SpearmanDot-type:pearson_maxvalue:0.7066346687971139name:PearsonMax-type:spearman_maxvalue:0.6993973161633343name:SpearmanMax
SentenceTransformer based on UBC-NLP/serengeti-E250
This is a sentence-transformers model finetuned from UBC-NLP/serengeti-E250 on the Mollel/swahili-n_li-triplet-swh-eng dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("Mollel/MultiLinguSwahili-MultiLinguSwahili-serengeti-E250-nli-matryoshka-nli-matryoshka")
# Run inference
sentences = [
'Mwanamume na mwanamke wachanga waliovaa mikoba wanaweka au kuondoa kitu kutoka kwenye mti mweupe wa zamani, huku watu wengine wamesimama au wameketi nyuma.',
'mwanamume na mwanamke wenye mikoba',
'tai huruka',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 768]# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
Approximate statistics based on the first 1000 samples:
anchor
positive
negative
type
string
string
string
details
min: 5 tokens
mean: 18.07 tokens
max: 53 tokens
min: 4 tokens
mean: 9.45 tokens
max: 33 tokens
min: 4 tokens
mean: 10.27 tokens
max: 29 tokens
Samples:
anchor
positive
negative
Two women are embracing while holding to go packages.
Two woman are holding packages.
The men are fighting outside a deli.
Wanawake wawili wanakumbatiana huku wakishikilia vifurushi vya kwenda.
Wanawake wawili wanashikilia vifurushi.
Wanaume hao wanapigana nje ya duka la vyakula vitamu.
Two young children in blue jerseys, one with the number 9 and one with the number 2 are standing on wooden steps in a bathroom and washing their hands in a sink.
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
MatryoshkaLoss
@misc{kusupati2024matryoshka,
title={Matryoshka Representation Learning},
author={Aditya Kusupati and Gantavya Bhatt and Aniket Rege and Matthew Wallingford and Aditya Sinha and Vivek Ramanujan and William Howard-Snyder and Kaifeng Chen and Sham Kakade and Prateek Jain and Ali Farhadi},
year={2024},
eprint={2205.13147},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
MultipleNegativesRankingLoss
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}