soumyasanyal commited on
Commit
40891f0
1 Parent(s): 3571969

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -2,12 +2,14 @@
2
  language:
3
  - en
4
  ---
5
- # entailment-verifier-xxl
6
 
7
  ## Model description
8
- Entailment-verifier-xxl is based on [flan-t5-xxl model](https://huggingface.co/google/flan-t5-xxl) and finetuned with a ranking objective (rank the most supported hypothesis from a given pair of hypotheses for a given premise). Please refer to our paper [Are Machines Better at Complex Reasoning? Unveiling Human-Machine Inference Gaps in Entailment Verification](https://arxiv.org/abs/2402.03686) for more detals.
9
 
10
- It is built to verify whether a given premise entails (or supports) a hypothesis or not. It works for both NLI-style of datasets and CoT rationales.
 
 
11
 
12
  ## Usage
13
 
@@ -29,7 +31,7 @@ def get_score(model, tokenizer, input_ids):
29
  return scores
30
 
31
  tokenizer = AutoTokenizer.from_pretrained('google/flan-t5-xxl')
32
- model = AutoModelForSeq2SeqLM.from_pretrained('soumyasanyal/entailment-verifier-xxl')
33
 
34
  premise = "A fossil fuel is a kind of natural resource. Coal is a kind of fossil fuel."
35
  hypothesis = "Coal is a kind of natural resource."
 
2
  language:
3
  - en
4
  ---
5
+ # nli-entailment-verifier-xxl
6
 
7
  ## Model description
8
+ **nli-entailment-verifier-xxl** is based on [flan-t5-xxl model](https://huggingface.co/google/flan-t5-xxl) and finetuned with a ranking objective (rank the most supported hypothesis from a given pair of hypotheses for a given premise). Please refer to our paper [Are Machines Better at Complex Reasoning? Unveiling Human-Machine Inference Gaps in Entailment Verification](https://arxiv.org/abs/2402.03686) for more detals.
9
 
10
+ It is built to verify whether a given premise supports a hypothesis or not. It works for both NLI-style datasets and CoT rationales. This model is specifically trained to handle multi-sentence premises (similar to what we expect in CoT rationales and other modern LLM use cases).
11
+
12
+ **Note**: You can use 4-bit/8-bit [quantization](https://huggingface.co/docs/bitsandbytes/main/en/index) to reduce GPU memory usage.
13
 
14
  ## Usage
15
 
 
31
  return scores
32
 
33
  tokenizer = AutoTokenizer.from_pretrained('google/flan-t5-xxl')
34
+ model = AutoModelForSeq2SeqLM.from_pretrained('soumyasanyal/nli-entailment-verifier-xxl')
35
 
36
  premise = "A fossil fuel is a kind of natural resource. Coal is a kind of fossil fuel."
37
  hypothesis = "Coal is a kind of natural resource."