jiazhengli commited on
Commit
3e82df1
1 Parent(s): 14c037e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -1
README.md CHANGED
@@ -19,5 +19,30 @@ widget:
19
 
20
  # Model Card for long-t5-tglobal-large-AERA
21
 
22
- This repository provides a fine-tuned version of long-t5-tglobal-large, using our proposed [AERA](https://github.com/lijiazheng99/aera/tree/main) framework presented in the paper: [Distilling ChatGPT for Explainable Automated Student Answer Assessment](https://aclanthology.org/2023.findings-emnlp.399.pdf).
23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  # Model Card for long-t5-tglobal-large-AERA
21
 
22
+ This repository provides a fine-tuned version of long-t5-tglobal-large, using our proposed [AERA](https://github.com/lijiazheng99/aera/tree/main) framework presented in the paper: [Distilling ChatGPT for Explainable Automated Student Answer Assessment](https://aclanthology.org/2023.findings-emnlp.399.pdf).
23
 
24
+ ## Citation Information
25
+
26
+ ```bibtex
27
+ @inproceedings{li-etal-2023-distilling,
28
+ title = "Distilling {C}hat{GPT} for Explainable Automated Student Answer Assessment",
29
+ author = "Li, Jiazheng and
30
+ Gui, Lin and
31
+ Zhou, Yuxiang and
32
+ West, David and
33
+ Aloisi, Cesare and
34
+ He, Yulan",
35
+ editor = "Bouamor, Houda and
36
+ Pino, Juan and
37
+ Bali, Kalika",
38
+ booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2023",
39
+ month = dec,
40
+ year = "2023",
41
+ address = "Singapore",
42
+ publisher = "Association for Computational Linguistics",
43
+ url = "https://aclanthology.org/2023.findings-emnlp.399",
44
+ doi = "10.18653/v1/2023.findings-emnlp.399",
45
+ pages = "6007--6026",
46
+ abstract = "Providing explainable and faithful feedback is crucial for automated student answer assessment. In this paper, we introduce a novel framework that explores using ChatGPT, a cutting-edge large language model, for the concurrent tasks of student answer scoring and rationale generation. We identify the appropriate instructions by prompting ChatGPT with different templates to collect the rationales, where inconsistent rationales are refined to align with marking standards. The refined ChatGPT outputs enable us to fine-tune a smaller language model that simultaneously assesses student answers and provides rationales. Extensive experiments on the benchmark dataset show that the proposed method improves the overall QWK score by 11{\%} compared to ChatGPT. Furthermore, our thorough analysis and human evaluation demonstrate that the rationales generated by our proposed method are comparable to those of ChatGPT. Our approach provides a viable solution to achieve explainable automated assessment in education",
47
+ }
48
+ ```