Has all-MiniLM-L6-v2 ever changed it's architecture? I have two different models
#49
by
rcland12
- opened
I have two models of all-MiniLM-L6-v2 in ONNX format that I have converted to TensorRT format and been using for a while.
One model has the inputs:
- input_ids
- attention_mask
Outputs: - token_embeddings
- sentence_embedding
The other model has inputs:
- input_ids
- token_type_ids
- attention_mask
Outputs: - last_hidden_state
Can someone explain this, or has anyone run into this?
Hello!
No, it hasn't changed its architecture. However, you can load this model in 2 ways: via Sentence Transformers where your inputs are input_ids & attention_mask and the outputs are token_embeddings
& sentence_embedding
, and via transformers
where your inputs are input_ids
, token_type_ids
, attention_mask
and the output is last_hidden_state
. See also the model card for a bit more info on this.
- Tom Aarsen