mlburnham commited on
Commit
3568d6f
1 Parent(s): 9ff7d93

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -13
README.md CHANGED
@@ -25,6 +25,16 @@ The model was trained using [Moritz Laurer's](https://huggingface.co/MoritzLaure
25
 
26
  The PolNLI dataset contains documents from social media, news outlets, congressional bills, court case summaries, and more. Classification should work well across a broad set of document sources and subjects, but for best performance refer to the recommendations section.
27
 
 
 
 
 
 
 
 
 
 
 
28
  # Using NLI Classifiers
29
  NLI classifiers work by pairing a document (AKA the 'premise') with a 'hypothesis', and determining if the hypothesis is true given the content of the document. The hypothesis is supplied by the user and can be thought of as a simplified version of a prompt for an LLM. It's best to keep hypotheses short and simple with a single classification criteria. If a more complicated hypothesis is necessary, consider few-shot training.
30
  For more detailed reading on using NLI classifiers see:
@@ -58,16 +68,4 @@ Evaluation is primarily conducted on the PolNLI test set. No hypotheses in the P
58
  <img src="https://cdn-uploads.huggingface.co/production/uploads/64d0341901931c60161f2a06/kDZXC-OsZtmTCxq0HsJMg.png" width="750" height="500" />
59
 
60
  ### Few-shot performance vs. and Electra classifier trained on 2,000 documents:
61
- <img src="https://cdn-uploads.huggingface.co/production/uploads/64d0341901931c60161f2a06/WpQ_ZofMJMFPraCK_a3Fb.png" width="750" height="500" />
62
-
63
-
64
- # Citation
65
- If you use this model or the PolNLI data set please cite:
66
- ```
67
- @article{burnham2024debate,
68
- title={Political DEBATE: Efficient Zero-shot and Few-shot Classifiers for Political Text},
69
- author={Burnham, Michael, Kayla Kahn, Rachel Peng, Ryan Wang},
70
- journal={working manuscript},
71
- year={2024}
72
- }
73
- ```
 
25
 
26
  The PolNLI dataset contains documents from social media, news outlets, congressional bills, court case summaries, and more. Classification should work well across a broad set of document sources and subjects, but for best performance refer to the recommendations section.
27
 
28
+ If you use this model or the PolNLI data set please cite:
29
+ ```
30
+ @article{burnham2024debate,
31
+ title={Political DEBATE: Efficient Zero-shot and Few-shot Classifiers for Political Text},
32
+ author={Burnham, Michael, Kayla Kahn, Rachel Peng, Ryan Wang},
33
+ journal={working manuscript},
34
+ year={2024}
35
+ }
36
+ ```
37
+
38
  # Using NLI Classifiers
39
  NLI classifiers work by pairing a document (AKA the 'premise') with a 'hypothesis', and determining if the hypothesis is true given the content of the document. The hypothesis is supplied by the user and can be thought of as a simplified version of a prompt for an LLM. It's best to keep hypotheses short and simple with a single classification criteria. If a more complicated hypothesis is necessary, consider few-shot training.
40
  For more detailed reading on using NLI classifiers see:
 
68
  <img src="https://cdn-uploads.huggingface.co/production/uploads/64d0341901931c60161f2a06/kDZXC-OsZtmTCxq0HsJMg.png" width="750" height="500" />
69
 
70
  ### Few-shot performance vs. and Electra classifier trained on 2,000 documents:
71
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/64d0341901931c60161f2a06/WpQ_ZofMJMFPraCK_a3Fb.png" width="750" height="500" />