Transformer Encoder for Social Science (TESS)
TESS is a deep neural network model intended for social science related NLP tasks. The model is developed by Haosen Ge, In Young Park, Xuancheng Qian, and Grace Zeng.
We demonstrate in two validation tests that TESS outperforms BERT and RoBERTa by 16.7% on average, especially when the number of training samples is limited (<1,000 training instances). The results display the superiority of TESS on social science text processing tasks.
|Preferential Trade Agreements||ToTA|
|Congressional Bills||Kornilova and Eidelman (2019)|
|Firms' Annual Reports||Loughran and McDonald (2016)|
|U.S. Court Opinions||Caselaw Access Project|
The model is trained on 4 NVIDIA A100 GPUs for 120K steps.
- Downloads last month