Sentence Similarity
sentence-transformers
PyTorch
TensorFlow
Rust
Safetensors
Transformers
English
bert
feature-extraction
Inference Endpoints
5 papers

Has all-MiniLM-L6-v2 ever changed it's architecture? I have two different models

#49
by rcland12 - opened

I have two models of all-MiniLM-L6-v2 in ONNX format that I have converted to TensorRT format and been using for a while.
One model has the inputs:

  • input_ids
  • attention_mask
    Outputs:
  • token_embeddings
  • sentence_embedding

The other model has inputs:

  • input_ids
  • token_type_ids
  • attention_mask
    Outputs:
  • last_hidden_state

Can someone explain this, or has anyone run into this?

Sentence Transformers org

Hello!

No, it hasn't changed its architecture. However, you can load this model in 2 ways: via Sentence Transformers where your inputs are input_ids & attention_mask and the outputs are token_embeddings & sentence_embedding, and via transformers where your inputs are input_ids, token_type_ids, attention_mask and the output is last_hidden_state. See also the model card for a bit more info on this.

  • Tom Aarsen

Sign up or log in to comment