{MODEL_NAME}
This is a sentence-transformers model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
Usage (Sentence-Transformers)
Using this model becomes easy when you have sentence-transformers installed:
pip install -U sentence-transformers
Then you can use the model like this:
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('{MODEL_NAME}')
embeddings = model.encode(sentences)
print(embeddings)
Usage (HuggingFace Transformers)
Without sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
from transformers import AutoTokenizer, AutoModel
import torch
#Mean Pooling - Take attention mask into account for correct averaging
def mean_pooling(model_output, attention_mask):
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
# Sentences we want sentence embeddings for
sentences = ['This is an example sentence', 'Each sentence is converted']
# Load model from HuggingFace Hub
tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}')
model = AutoModel.from_pretrained('{MODEL_NAME}')
# Tokenize sentences
encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
# Compute token embeddings
with torch.no_grad():
model_output = model(**encoded_input)
# Perform pooling. In this case, mean pooling.
sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
print("Sentence embeddings:")
print(sentence_embeddings)
Evaluation Results
Model | id_raw_acc | vn_raw_acc | br_raw_acc | th_raw_acc | my_raw_acc | ph_raw_acc | sg_raw_acc | avg |
---|---|---|---|---|---|---|---|---|
thtang/SetFit_ALL_200M_itr5 | 74.24% | 64.04% | 58.98% | 67.24% | 70.77% | 70.63% | 70.58% | 68.07% |
('ViT-B-16-SigLIP-i18n-256', 'webli') | 69.38% | 57.92% | 47.40% | 56.40% | 65.20% | 65.72% | 65.12% | 61.02% |
('xlm-roberta-base-ViT-B-32', 'laion5b_s13b_b90k') | 66.23% | 54.05% | 49.26% | 55.39% | 65.61% | 66.11% | 66.72% | 60.48% |
('xlm-roberta-large-ViT-H-14', 'frozen_laion5b_s13b_b90k') | 66.05% | 52.77% | 46.46% | 53.44% | 62.70% | 64.40% | 64.24% | 58.58% |
('ViT-L-14', 'commonpool_xl_s13b_b90k') | 65.48% | 53.80% | 46.61% | 51.00% | 62.01% | 64.37% | 63.94% | 58.17% |
('ViT-L-14', 'commonpool_xl_clip_s13b_b90k') | 66.73% | 49.82% | 45.25% | 38.32% | 63.64% | 66.17% | 65.29% | 56.46% |
('ViT-B-16', 'commonpool_l_s1b_b8k') | 62.14% | 49.25% | 45.20% | 39.47% | 61.15% | 63.03% | 62.63% | 54.69% |
('ViT-bigG-14-CLIPA', 'datacomp1b') | 69.21% | 44.39% | 48.25% | 20.54% | 62.83% | 68.15% | 66.48% | 54.26% |
('ViT-bigG-14-CLIPA-336', 'datacomp1b') | 69.17% | 44.22% | 48.06% | 20.48% | 62.79% | 67.74% | 66.63% | 54.15% |
('ViT-H-14-CLIPA-336', 'datacomp1b') | 68.03% | 42.79% | 47.52% | 20.82% | 62.38% | 67.06% | 66.92% | 53.65% |
('ViT-H-14-CLIPA', 'datacomp1b') | 68.18% | 42.82% | 47.33% | 20.68% | 62.31% | 67.26% | 66.56% | 53.59% |
('ViT-B-16', 'commonpool_l_clip_s1b_b8k') | 63.68% | 42.24% | 44.87% | 28.59% | 62.04% | 65.18% | 64.97% | 53.08% |
('ViT-B-32-256', 'datacomp_s34b_b86k') | 65.44% | 38.94% | 43.57% | 25.11% | 62.39% | 65.82% | 64.94% | 52.32% |
('ViT-L-14-CLIPA-336', 'datacomp1b') | 66.99% | 38.69% | 45.25% | 20.36% | 61.47% | 66.78% | 65.56% | 52.16% |
('ViT-L-14-CLIPA', 'datacomp1b') | 66.86% | 38.34% | 45.21% | 20.18% | 61.51% | 66.71% | 65.41% | 52.03% |
('ViT-H-14-CLIPA-336', 'laion2b') | 64.62% | 35.52% | 44.73% | 21.27% | 61.01% | 67.12% | 65.76% | 51.43% |
('ViT-B-32', 'datacomp_xl_s13b_b90k') | 64.57% | 37.26% | 42.06% | 22.61% | 61.96% | 65.59% | 64.63% | 51.24% |
('ViT-L-14', 'datacomp_xl_s13b_b90k') | 64.37% | 37.78% | 40.65% | 22.89% | 60.72% | 65.26% | 64.30% | 50.85% |
('EVA02-E-14-plus', 'laion2b_s9b_b144k') | 63.51% | 31.79% | 42.52% | 23.71% | 60.74% | 64.74% | 63.97% | 50.14% |
('ViT-H-14-quickgelu', 'metaclip_fullcc') | 59.75% | 34.61% | 43.12% | 22.69% | 60.61% | 65.47% | 64.58% | 50.12% |
('ViT-B-16', 'datacomp_xl_s13b_b90k') | 63.15% | 36.19% | 39.81% | 22.39% | 60.66% | 63.96% | 63.31% | 49.92% |
('ViT-bigG-14', 'laion2b_s39b_b160k') | 63.03% | 31.52% | 41.20% | 23.65% | 60.52% | 65.11% | 63.99% | 49.86% |
('ViT-B-16', 'commonpool_l_basic_s1b_b8k') | 62.56% | 36.99% | 40.87% | 22.16% | 59.57% | 63.56% | 63.06% | 49.82% |
intfloat/multilingual-e5-large | 52.99% | 42.00% | 33.92% | 47.69% | 55.82% | 57.76% | 58.16% | 49.76% |
intfloat/multilingual-e5-base | 52.06% | 43.21% | 34.17% | 47.41% | 55.28% | 57.38% | 57.45% | 49.57% |
('ViT-B-16', 'commonpool_l_image_s1b_b8k') | 61.48% | 36.08% | 40.87% | 22.62% | 59.17% | 63.47% | 62.80% | 49.50% |
('convnext_large_d', 'laion2b_s26b_b102k_augreg') | 61.61% | 29.78% | 39.92% | 23.49% | 60.93% | 65.69% | 64.60% | 49.43% |
('EVA01-g-14-plus', 'merged2b_s11b_b114k') | 62.34% | 30.29% | 39.02% | 22.80% | 60.83% | 65.19% | 63.49% | 49.14% |
('convnext_large_d_320', 'laion2b_s29b_b131k_ft') | 61.18% | 29.24% | 39.09% | 23.23% | 60.65% | 65.64% | 64.12% | 49.02% |
('ViT-B-32', 'laion2b_s34b_b79k') | 61.21% | 29.82% | 37.51% | 24.49% | 60.21% | 65.28% | 64.08% | 48.94% |
('convnext_large_d_320', 'laion2b_s29b_b131k_ft_soup') | 60.91% | 29.28% | 38.97% | 22.61% | 60.78% | 65.76% | 63.84% | 48.88% |
('convnext_xxlarge', 'laion2b_s34b_b82k_augreg_soup') | 61.55% | 30.17% | 38.85% | 22.30% | 60.28% | 64.83% | 63.22% | 48.74% |
('ViT-B-32', 'laion2b_e16') | 61.44% | 28.15% | 38.05% | 24.49% | 59.93% | 65.14% | 63.87% | 48.72% |
('ViT-B-16', 'datacomp_l_s1b_b8k') | 61.33% | 29.35% | 38.67% | 23.31% | 60.29% | 64.42% | 63.64% | 48.72% |
('ViT-H-14', 'laion2b_s32b_b79k') | 61.45% | 29.19% | 38.91% | 22.64% | 60.56% | 64.86% | 63.30% | 48.70% |
('EVA02-E-14', 'laion2b_s4b_b115k') | 61.63% | 29.60% | 38.57% | 22.89% | 60.22% | 64.83% | 63.18% | 48.70% |
('convnext_xxlarge', 'laion2b_s34b_b82k_augreg_rewind') | 61.24% | 30.22% | 39.04% | 22.40% | 60.02% | 64.75% | 62.99% | 48.67% |
('ViT-B-32-quickgelu', 'metaclip_fullcc') | 58.26% | 29.70% | 38.99% | 23.24% | 60.07% | 65.67% | 64.30% | 48.60% |
('convnext_xxlarge', 'laion2b_s34b_b82k_augreg') | 60.94% | 29.90% | 39.49% | 22.08% | 60.10% | 64.50% | 63.15% | 48.59% |
('ViT-g-14', 'laion2b_s12b_b42k') | 61.46% | 27.70% | 38.23% | 22.46% | 60.65% | 65.68% | 63.87% | 48.58% |
('ViT-g-14', 'laion2b_s34b_b88k') | 60.83% | 29.56% | 39.37% | 21.63% | 59.87% | 64.68% | 63.30% | 48.46% |
('ViT-L-14-quickgelu', 'metaclip_fullcc') | 56.99% | 31.07% | 40.45% | 23.13% | 59.21% | 64.77% | 63.50% | 48.45% |
intfloat/multilingual-e5-small | 49.50% | 42.68% | 30.96% | 47.42% | 54.44% | 56.44% | 57.04% | 48.35% |
('ViT-B-16-quickgelu', 'metaclip_fullcc') | 58.00% | 28.59% | 37.68% | 23.22% | 59.42% | 65.03% | 64.10% | 48.01% |
('ViT-L-14', 'laion2b_s32b_b82k') | 60.18% | 28.09% | 36.28% | 23.70% | 59.89% | 64.86% | 63.01% | 48.00% |
('ViT-B-32-quickgelu', 'laion400m_e32') | 59.74% | 25.92% | 36.98% | 25.19% | 59.67% | 64.79% | 63.68% | 48.00% |
('ViT-B-32-quickgelu', 'laion400m_e31') | 59.86% | 25.92% | 36.84% | 25.20% | 59.56% | 64.76% | 63.79% | 47.99% |
('convnext_base_w', 'laion2b_s13b_b82k_augreg') | 60.97% | 27.03% | 36.75% | 22.90% | 59.70% | 64.78% | 63.46% | 47.94% |
('ViT-L-14', 'laion400m_e32') | 60.01% | 24.45% | 37.24% | 23.95% | 59.17% | 65.02% | 63.78% | 47.66% |
('EVA01-g-14', 'laion400m_s11b_b41k') | 60.51% | 25.96% | 36.17% | 23.69% | 59.57% | 64.40% | 63.22% | 47.64% |
('ViT-B-16-plus-240', 'laion400m_e32') | 59.84% | 25.29% | 36.80% | 23.73% | 59.31% | 64.99% | 63.43% | 47.63% |
('ViT-B-16-plus-240', 'laion400m_e31') | 59.69% | 25.22% | 36.79% | 23.69% | 59.44% | 64.92% | 63.53% | 47.61% |
('ViT-B-16', 'laion2b_s34b_b88k') | 59.82% | 27.45% | 35.12% | 24.41% | 59.39% | 64.37% | 62.66% | 47.60% |
('ViT-L-14', 'laion400m_e31') | 59.91% | 24.26% | 37.53% | 23.84% | 59.08% | 64.90% | 63.64% | 47.60% |
('ViT-L-16-SigLIP-256', 'webli') | 65.54% | 20.39% | 44.65% | 15.18% | 60.10% | 64.64% | 62.44% | 47.56% |
('roberta-ViT-B-32', 'laion2b_s12b_b32k') | 59.70% | 25.15% | 39.81% | 17.10% | 59.95% | 65.81% | 65.00% | 47.50% |
('ViT-L-14', 'commonpool_xl_laion_s13b_b90k') | 58.13% | 26.95% | 34.93% | 23.34% | 59.05% | 64.51% | 63.63% | 47.22% |
('ViT-B-16-SigLIP', 'webli') | 64.31% | 19.87% | 44.78% | 14.87% | 58.38% | 65.16% | 62.44% | 47.12% |
('ViT-B-16-SigLIP-256', 'webli') | 64.24% | 20.94% | 44.15% | 15.35% | 58.22% | 64.41% | 62.43% | 47.10% |
('ViT-B-16-SigLIP-384', 'webli') | 64.36% | 20.06% | 44.41% | 15.11% | 58.03% | 64.68% | 62.10% | 46.96% |
('ViT-L-16-SigLIP-384', 'webli') | 64.49% | 20.17% | 44.01% | 14.80% | 58.89% | 64.92% | 61.39% | 46.95% |
('ViT-B-32', 'laion400m_e31') | 59.06% | 26.66% | 35.69% | 23.68% | 58.00% | 62.82% | 62.68% | 46.94% |
('ViT-B-16-SigLIP-512', 'webli') | 64.28% | 19.61% | 44.17% | 15.09% | 57.71% | 64.83% | 62.44% | 46.88% |
('convnext_base_w_320', 'laion_aesthetic_s13b_b82k_augreg') | 57.60% | 26.52% | 35.01% | 24.43% | 57.05% | 64.54% | 62.74% | 46.84% |
('ViT-B-16', 'commonpool_l_text_s1b_b8k') | 59.57% | 28.15% | 37.37% | 20.89% | 57.54% | 62.68% | 61.63% | 46.83% |
('ViT-B-32', 'laion400m_e32') | 59.05% | 26.62% | 35.44% | 23.54% | 58.00% | 62.74% | 62.27% | 46.81% |
('convnext_base_w', 'laion2b_s13b_b82k') | 58.65% | 26.97% | 34.80% | 23.26% | 58.31% | 63.39% | 61.56% | 46.71% |
sentence-transformers/gtr-t5-xxl | 59.93% | 24.82% | 40.79% | 17.23% | 58.41% | 64.00% | 61.57% | 46.68% |
('ViT-B-16', 'laion400m_e32') | 59.01% | 24.34% | 35.07% | 21.84% | 59.04% | 64.58% | 62.73% | 46.66% |
('ViT-B-16', 'laion400m_e31') | 58.94% | 24.20% | 34.92% | 21.58% | 59.11% | 64.77% | 63.09% | 46.66% |
('convnext_base', 'laion400m_s13b_b51k') | 58.44% | 24.99% | 34.05% | 23.99% | 58.33% | 63.79% | 62.59% | 46.60% |
('EVA02-L-14-336', 'merged2b_s6b_b61k') | 59.54% | 23.19% | 34.54% | 22.36% | 59.24% | 63.90% | 63.40% | 46.60% |
('coca_ViT-B-32', 'laion2b_s13b_b90k') | 58.70% | 27.10% | 33.22% | 24.13% | 57.53% | 63.56% | 61.87% | 46.59% |
('EVA02-L-14', 'merged2b_s4b_b131k') | 59.64% | 23.18% | 34.62% | 22.55% | 59.11% | 63.86% | 63.10% | 46.58% |
thenlper/gte-large | 55.10% | 28.16% | 33.96% | 18.73% | 59.50% | 65.19% | 63.52% | 46.31% |
('ViT-L-14-quickgelu', 'metaclip_400m') | 54.32% | 25.87% | 34.30% | 23.41% | 58.50% | 64.48% | 63.24% | 46.30% |
('coca_ViT-L-14', 'laion2b_s13b_b90k') | 57.92% | 25.78% | 33.97% | 24.17% | 57.64% | 63.08% | 61.55% | 46.30% |
('coca_ViT-L-14', 'mscoco_finetuned_laion2b_s13b_b90k') | 58.07% | 25.32% | 34.18% | 24.60% | 57.77% | 62.80% | 61.28% | 46.29% |
('ViT-B-32-quickgelu', 'metaclip_400m') | 55.85% | 27.37% | 31.91% | 21.76% | 58.64% | 64.69% | 63.11% | 46.19% |
sentence-transformers/paraphrase-multilingual-mpnet-base-v2 | 49.03% | 32.58% | 32.82% | 38.43% | 55.30% | 57.36% | 57.34% | 46.12% |
('convnext_base_w', 'laion_aesthetic_s13b_b82k') | 57.39% | 25.68% | 33.71% | 23.82% | 56.64% | 63.22% | 62.22% | 46.10% |
('ViT-B-32', 'commonpool_m_clip_s128m_b4k') | 56.09% | 26.70% | 38.25% | 22.79% | 56.52% | 61.26% | 61.05% | 46.09% |
('convnext_base_w_320', 'laion_aesthetic_s13b_b82k') | 56.96% | 25.60% | 33.77% | 24.64% | 56.32% | 63.33% | 61.87% | 46.07% |
('ViT-B-16', 'commonpool_l_laion_s1b_b8k') | 56.37% | 25.70% | 31.07% | 23.18% | 58.65% | 63.93% | 63.49% | 46.06% |
('ViT-B-16-quickgelu', 'metaclip_400m') | 55.90% | 25.88% | 32.67% | 21.57% | 58.65% | 64.48% | 63.04% | 46.03% |
intfloat/e5-large | 55.45% | 28.54% | 36.69% | 18.15% | 57.78% | 62.92% | 61.83% | 45.91% |
('EVA02-B-16', 'merged2b_s8b_b131k') | 58.08% | 24.45% | 31.80% | 22.36% | 58.45% | 63.25% | 62.44% | 45.83% |
sentence-transformers/LaBSE | 50.30% | 32.82% | 33.15% | 39.79% | 54.95% | 53.71% | 55.06% | 45.68% |
thenlper/gte-base | 55.46% | 27.88% | 32.77% | 17.20% | 58.09% | 63.68% | 62.03% | 45.30% |
intfloat/e5-large-v2 | 55.10% | 28.06% | 35.95% | 17.16% | 57.16% | 61.21% | 60.84% | 45.07% |
('ViT-SO400M-14-SigLIP', 'webli') | 60.18% | 29.39% | 38.90% | 13.73% | 52.79% | 59.15% | 56.81% | 44.42% |
('ViT-B-32', 'commonpool_m_s128m_b4k') | 50.30% | 32.12% | 37.08% | 23.02% | 53.63% | 57.64% | 56.91% | 44.39% |
sentence-transformers/sentence-t5-xxl | 50.98% | 18.38% | 36.37% | 16.91% | 59.25% | 64.82% | 63.75% | 44.35% |
infgrad/stella-base-en-v2 | 52.42% | 26.24% | 30.61% | 18.81% | 56.84% | 63.03% | 61.67% | 44.23% |
('RN50x4', 'openai') | 56.39% | 25.77% | 29.99% | 21.48% | 55.31% | 61.02% | 59.42% | 44.20% |
('RN50x16', 'openai') | 56.58% | 25.09% | 29.77% | 21.03% | 54.81% | 61.28% | 58.47% | 43.86% |
('RN101-quickgelu', 'openai') | 56.57% | 25.83% | 29.66% | 21.09% | 54.50% | 60.18% | 58.74% | 43.80% |
('RN101', 'openai') | 56.57% | 25.83% | 29.66% | 21.09% | 54.50% | 60.18% | 58.74% | 43.80% |
llmrails/ember-v1 | 50.85% | 24.76% | 31.02% | 17.20% | 57.62% | 63.06% | 62.04% | 43.79% |
sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2 | 44.88% | 28.32% | 29.45% | 36.40% | 53.97% | 56.87% | 56.14% | 43.72% |
BAAI/bge-large-en-v1.5 | 49.81% | 25.55% | 30.68% | 17.41% | 56.89% | 62.87% | 61.72% | 43.56% |
('RN50x64', 'openai') | 55.34% | 22.19% | 30.63% | 20.79% | 55.18% | 60.93% | 59.45% | 43.50% |
('nllb-clip-large', 'v1') | 48.84% | 23.45% | 33.92% | 32.38% | 53.67% | 55.36% | 56.76% | 43.48% |
BAAI/bge-base-en-v1.5 | 51.73% | 24.30% | 31.51% | 17.53% | 56.21% | 62.37% | 60.25% | 43.42% |
intfloat/e5-small | 51.31% | 27.36% | 32.05% | 16.66% | 55.15% | 60.39% | 59.06% | 43.14% |
BAAI/bge-small-en-v1.5 | 51.37% | 25.16% | 29.99% | 16.13% | 56.17% | 61.69% | 61.01% | 43.07% |
('ViT-L-14', 'openai') | 54.57% | 21.44% | 30.13% | 19.50% | 54.99% | 60.94% | 59.59% | 43.02% |
('ViT-L-14-336', 'openai') | 54.12% | 21.52% | 30.63% | 19.47% | 55.41% | 60.77% | 58.87% | 42.97% |
intfloat/e5-small-v2 | 51.41% | 26.82% | 33.04% | 16.30% | 54.97% | 58.66% | 58.68% | 42.84% |
('ViT-SO400M-14-SigLIP-384', 'webli') | 62.68% | 15.00% | 32.38% | 7.32% | 56.65% | 64.12% | 61.49% | 42.81% |
('RN50-quickgelu', 'openai') | 53.15% | 24.79% | 29.57% | 20.84% | 53.15% | 59.19% | 57.59% | 42.61% |
('RN50', 'openai') | 53.15% | 24.79% | 29.57% | 20.84% | 53.15% | 59.19% | 57.59% | 42.61% |
('ViT-B-16', 'openai') | 53.31% | 22.22% | 27.96% | 21.22% | 53.68% | 59.47% | 58.45% | 42.33% |
('ViT-B-32', 'openai') | 52.93% | 23.44% | 28.70% | 20.78% | 52.96% | 59.38% | 57.93% | 42.30% |
('ViT-B-32-quickgelu', 'openai') | 52.93% | 23.44% | 28.70% | 20.78% | 52.96% | 59.38% | 57.93% | 42.30% |
sentence-transformers/all-MiniLM-L6-v2 | 50.80% | 25.76% | 27.04% | 15.81% | 54.63% | 60.07% | 59.68% | 41.97% |
('ViT-B-32', 'commonpool_m_basic_s128m_b4k') | 52.54% | 22.67% | 30.25% | 16.17% | 53.22% | 59.40% | 58.31% | 41.80% |
sentence-transformers/all-MiniLM-L12-v2 | 48.98% | 24.05% | 25.74% | 16.41% | 54.51% | 60.38% | 58.90% | 41.28% |
('ViT-B-32', 'commonpool_m_image_s128m_b4k') | 51.93% | 20.40% | 29.44% | 16.53% | 53.16% | 58.71% | 58.17% | 41.19% |
sentence-transformers/clip-ViT-B-32-multilingual-v1 | 44.45% | 27.34% | 28.00% | 28.25% | 50.30% | 54.05% | 53.39% | 40.82% |
sentence-transformers/distiluse-base-multilingual-cased-v2 | 43.51% | 23.86% | 28.41% | 26.90% | 53.14% | 53.54% | 54.38% | 40.53% |
('ViT-B-32', 'datacomp_m_s128m_b4k') | 51.60% | 19.45% | 26.58% | 16.46% | 52.54% | 59.03% | 58.03% | 40.53% |
('ViT-B-32', 'commonpool_m_text_s128m_b4k') | 50.38% | 20.31% | 27.01% | 16.00% | 52.61% | 58.82% | 58.10% | 40.46% |
sentence-transformers/all-mpnet-base-v2 | 46.97% | 23.15% | 24.75% | 16.31% | 52.66% | 59.07% | 57.75% | 40.09% |
('nllb-clip-base', 'v1') | 42.72% | 23.90% | 29.29% | 33.96% | 48.33% | 49.09% | 51.21% | 39.79% |
sentence-transformers/paraphrase-mpnet-base-v2 | 46.00% | 20.45% | 26.92% | 14.75% | 52.89% | 58.71% | 58.20% | 39.70% |
sentence-transformers/all-distilroberta-v1 | 46.74% | 22.34% | 24.06% | 17.59% | 51.49% | 57.54% | 56.45% | 39.46% |
sentence-transformers/paraphrase-MiniLM-L6-v2 | 44.92% | 23.59% | 26.12% | 14.23% | 51.84% | 57.14% | 56.03% | 39.12% |
('ViT-B-32', 'commonpool_m_laion_s128m_b4k') | 42.94% | 19.21% | 19.70% | 17.26% | 50.84% | 57.59% | 56.06% | 37.66% |
('RN50-quickgelu', 'cc12m') | 40.71% | 18.10% | 16.78% | 16.23% | 45.55% | 52.89% | 50.77% | 34.43% |
('RN50', 'cc12m') | 39.76% | 17.32% | 16.15% | 15.76% | 44.25% | 52.46% | 49.18% | 33.55% |
('RN101', 'yfcc15m') | 33.79% | 18.04% | 16.05% | 11.10% | 37.62% | 43.50% | 42.45% | 28.94% |
('RN101-quickgelu', 'yfcc15m') | 32.79% | 16.89% | 14.45% | 11.56% | 37.77% | 42.86% | 41.93% | 28.32% |
('ViT-B-32', 'commonpool_s_clip_s13m_b4k') | 33.80% | 13.26% | 18.82% | 12.42% | 37.36% | 42.09% | 40.39% | 28.31% |
('RN50', 'yfcc15m') | 31.81% | 15.87% | 14.88% | 8.99% | 37.42% | 42.06% | 41.19% | 27.46% |
('RN50-quickgelu', 'yfcc15m') | 31.57% | 15.90% | 14.44% | 8.99% | 36.81% | 41.81% | 41.20% | 27.24% |
('ViT-B-32', 'commonpool_s_s13m_b4k') | 29.42% | 12.57% | 16.82% | 11.00% | 32.42% | 36.77% | 35.48% | 24.93% |
('ViT-B-32', 'commonpool_s_text_s13m_b4k') | 28.02% | 10.61% | 12.49% | 9.85% | 31.18% | 37.10% | 34.85% | 23.44% |
('ViT-B-32', 'commonpool_s_basic_s13m_b4k') | 27.87% | 10.72% | 12.67% | 8.16% | 30.11% | 36.13% | 32.68% | 22.62% |
('coca_ViT-B-32', 'mscoco_finetuned_laion2b_s13b_b90k') | 12.60% | 7.91% | 5.11% | 9.96% | 17.15% | 20.67% | 20.32% | 13.39% |
('ViT-B-32', 'commonpool_s_image_s13m_b4k') | 15.20% | 5.59% | 5.91% | 4.63% | 16.80% | 20.74% | 18.78% | 12.52% |
('ViT-B-32', 'datacomp_s_s13m_b4k') | 15.20% | 5.59% | 5.91% | 4.63% | 16.80% | 20.74% | 18.78% | 12.52% |
('ViT-B-32', 'commonpool_s_laion_s13m_b4k') | 11.72% | 5.12% | 4.05% | 4.23% | 14.33% | 18.82% | 16.44% | 10.67% |
Training
The model was trained with the parameters:
DataLoader:
torch.utils.data.dataloader.DataLoader
of length 1468721 with parameters:
{'batch_size': 160, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
Loss:
sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss
Parameters of the fit()-Method:
{
"epochs": 1,
"evaluation_steps": 0,
"evaluator": "NoneType",
"max_grad_norm": 1,
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
"optimizer_params": {
"lr": 2e-05
},
"scheduler": "WarmupLinear",
"steps_per_epoch": null,
"warmup_steps": 100,
"weight_decay": 0.01
}
Full Model Architecture
SentenceTransformer(
(0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
(1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
)
Citing & Authors
- Downloads last month
- 11
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.