license: mit | |
pipeline_tag: sentence-similarity | |
datasets: | |
- dell-research-harvard/headlines-semantic-similarity | |
- dell-research-harvard/AmericanStories | |
tags: | |
- sentence-transformers | |
- feature-extraction | |
- sentence-similarity | |
- transformers | |
language: | |
- en | |
base_model: "StoriesLM/StoriesLM-v1-1963" | |
# RepresentLM-v1 | |
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. | |
The model is trained on the [HEADLINES](https://huggingface.co/datasets/dell-research-harvard/headlines-semantic-similarity) semantic similarity dataset, using the [StoriesLM-v1-1963](https://huggingface.co/StoriesLM/StoriesLM-v1-1963) model as a base. | |
## Usage | |
First install the [sentence-transformers](https://www.SBERT.net) package: | |
``` | |
pip install -U sentence-transformers | |
``` | |
The model can then be used to encode language sequences: | |
```python | |
from sentence_transformers import SentenceTransformer | |
sequences = ["This is an example sequence", "Each sequence is embedded"] | |
model = SentenceTransformer('RepresentLM/RepresentLM-v1') | |
embeddings = model.encode(sequences) | |
print(embeddings) | |
``` | |