File size: 1,650 Bytes
0fa331b
 
4f42d19
0fa331b
ccb2d20
4f42d19
 
1aa10ca
 
 
 
ddba5e8
5538fa7
4f42d19
 
 
 
 
 
 
 
 
 
 
 
5538fa7
4f42d19
 
 
 
be5acba
 
 
 
 
 
 
 
 
 
 
 
 
4f42d19
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
---
license: apache-2.0
pipeline_tag: text-classification
---


## Description:
Passage Re-Ranker a encoder based model that takes a search query, and a passage, and calculates the relevancy of the passage to the query. This is used in conjunction with sentence transformers to re-rank the passages matched by the sentence transformer, there-by improving relevance of Information Retrieval processes.

The Model is Finetuned using MS-Marco, and tested using Science QA datasets.

The Model is an integral part of `Neural Search` Information Retreival process used by the Science Discovery Engine, Along with the finetuned sentence transformer (https://huggingface.co/nasa-impact/nasa-smd-ibm-st-v2).

## Evaluation:

Model Evaluation on msmarco dev set, and NASA Science Questions:

![image/png](https://cdn-uploads.huggingface.co/production/uploads/61099e5d86580d4580767226/jJnEkMijBvnTSN_cb_lDn.png)


## Intended uses & limitations

Both query and passagehave to fit in 512 Tokens. The intended use is to rerank the first dozens of embedding search results.

## How to use

```
from transformers import AutoTokenizer, AutoModelForSequenceClassification
tokenizer = AutoTokenizer.from_pretrained("nasa-impact/nasa-smd-ibm-ranker")
model = AutoModelForSequenceClassification.from_pretrained("nasa-impact/nasa-smd-ibm-ranker")
```


## Cite this Model
```
@misc {nasa-impact_2024,
	author       = { {NASA-IMPACT} },
	title        = { nasa-smd-ibm-ranker (Revision 4f42d19) },
	year         = 2024,
	url          = { https://huggingface.co/nasa-impact/nasa-smd-ibm-ranker },
	doi          = { 10.57967/hf/1849 },
	publisher    = { Hugging Face }
}
```