Update README.md
Browse files
README.md
CHANGED
@@ -25,6 +25,16 @@ The model was trained using [Moritz Laurer's](https://huggingface.co/MoritzLaure
|
|
25 |
|
26 |
The PolNLI dataset contains documents from social media, news outlets, congressional bills, court case summaries, and more. Classification should work well across a broad set of document sources and subjects, but for best performance refer to the recommendations section.
|
27 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
28 |
# Using NLI Classifiers
|
29 |
NLI classifiers work by pairing a document (AKA the 'premise') with a 'hypothesis', and determining if the hypothesis is true given the content of the document. The hypothesis is supplied by the user and can be thought of as a simplified version of a prompt for an LLM. It's best to keep hypotheses short and simple with a single classification criteria. If a more complicated hypothesis is necessary, consider few-shot training.
|
30 |
For more detailed reading on using NLI classifiers see:
|
@@ -58,16 +68,4 @@ Evaluation is primarily conducted on the PolNLI test set. No hypotheses in the P
|
|
58 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/64d0341901931c60161f2a06/kDZXC-OsZtmTCxq0HsJMg.png" width="750" height="500" />
|
59 |
|
60 |
### Few-shot performance vs. and Electra classifier trained on 2,000 documents:
|
61 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/64d0341901931c60161f2a06/WpQ_ZofMJMFPraCK_a3Fb.png" width="750" height="500" />
|
62 |
-
|
63 |
-
|
64 |
-
# Citation
|
65 |
-
If you use this model or the PolNLI data set please cite:
|
66 |
-
```
|
67 |
-
@article{burnham2024debate,
|
68 |
-
title={Political DEBATE: Efficient Zero-shot and Few-shot Classifiers for Political Text},
|
69 |
-
author={Burnham, Michael, Kayla Kahn, Rachel Peng, Ryan Wang},
|
70 |
-
journal={working manuscript},
|
71 |
-
year={2024}
|
72 |
-
}
|
73 |
-
```
|
|
|
25 |
|
26 |
The PolNLI dataset contains documents from social media, news outlets, congressional bills, court case summaries, and more. Classification should work well across a broad set of document sources and subjects, but for best performance refer to the recommendations section.
|
27 |
|
28 |
+
If you use this model or the PolNLI data set please cite:
|
29 |
+
```
|
30 |
+
@article{burnham2024debate,
|
31 |
+
title={Political DEBATE: Efficient Zero-shot and Few-shot Classifiers for Political Text},
|
32 |
+
author={Burnham, Michael, Kayla Kahn, Rachel Peng, Ryan Wang},
|
33 |
+
journal={working manuscript},
|
34 |
+
year={2024}
|
35 |
+
}
|
36 |
+
```
|
37 |
+
|
38 |
# Using NLI Classifiers
|
39 |
NLI classifiers work by pairing a document (AKA the 'premise') with a 'hypothesis', and determining if the hypothesis is true given the content of the document. The hypothesis is supplied by the user and can be thought of as a simplified version of a prompt for an LLM. It's best to keep hypotheses short and simple with a single classification criteria. If a more complicated hypothesis is necessary, consider few-shot training.
|
40 |
For more detailed reading on using NLI classifiers see:
|
|
|
68 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/64d0341901931c60161f2a06/kDZXC-OsZtmTCxq0HsJMg.png" width="750" height="500" />
|
69 |
|
70 |
### Few-shot performance vs. and Electra classifier trained on 2,000 documents:
|
71 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/64d0341901931c60161f2a06/WpQ_ZofMJMFPraCK_a3Fb.png" width="750" height="500" />
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|