Research interests

Relation distilled BERT: lexical relation embedding model driven by pre-trained language models.

Team members 2

Organization Card
About org cards

RelBERT is a high-quality semantic representative embedding of word pairs powered by pre-trained language model. Install relbert via pip,

   
      pip install relbert
   

and play with RelBERT models.

   
      from relbert import RelBERT
      model = RelBERT('relbert/relbert-roberta-large')
      vector = model.get_embedding(['Tokyo', 'Japan'])  # shape of (1024, )
   

See more information bellow.