license: apache-2.0 library_name: transformers
PEG: Towards Robust Text Retrieval with Progressive Learning
Model Details
We propose the PEG model (a Progressively Learned Textual Embedding), which progressively adjusts the weights of samples contributing to the loss within an extremely large batch, based on the difficulty levels of negative samples. we have amassed an extensive collection of over 110 million data, spanning a wide range of fields such as general knowledge, finance, tourism, medicine, and more.
Our technical report is available at Paper
Usage (HuggingFace Transformers)
Install transformers:
pip install transformers
Then load model and predict:
from transformers import AutoModel, AutoTokenizer
import torch
# Load model from HuggingFace Hub
tokenizer = AutoTokenizer.from_pretrained('TownsWu/PEG')
model = AutoModel.from_pretrained('TownsWu/PEG')
sentences = ['如何更换花呗绑定银行卡', '花呗更改绑定银行卡']
# Tokenize sentences
inputs = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
# Compute token embeddings
with torch.no_grad():
last_hidden_state = model(**inputs, return_dict=True).last_hidden_state
embeddings = last_hidden_state[:, 0]
print("embeddings:")
print(embeddings)
Contact
If you have any question or suggestion related to this project, feel free to open an issue or pull request. You also can email Tong Wu(townswu@tencent.com).
Citation
If you find our work helpful for your research, please consider citing the following BibTeX entry:
@article{wu2023towards,
title={Towards Robust Text Retrieval with Progressive Learning},
author={Wu, Tong and Qin, Yulei and Zhang, Enwei and Xu, Zihan and Gao, Yuting and Li, Ke and Sun, Xing},
journal={arXiv preprint arXiv:2311.11691},
year={2023}
}
- Downloads last month
- 127
Inference API (serverless) does not yet support transformers models for this pipeline type.
Spaces using TownsWu/PEG 2
Evaluation results
- map on MTEB CMedQAv1test set self-reported84.091
- mrr on MTEB CMedQAv1test set self-reported86.629
- map on MTEB CMedQAv2test set self-reported86.558
- mrr on MTEB CMedQAv2test set self-reported89.433
- map_at_1 on MTEB CmedqaRetrievalself-reported26.101
- map_at_10 on MTEB CmedqaRetrievalself-reported38.239
- map_at_100 on MTEB CmedqaRetrievalself-reported40.083
- map_at_1000 on MTEB CmedqaRetrievalself-reported40.205
- map_at_3 on MTEB CmedqaRetrievalself-reported34.386
- map_at_5 on MTEB CmedqaRetrievalself-reported36.426