Spaces:
Runtime error
Runtime error
A newer version of the Gradio SDK is available:
5.4.0
metadata
title: CTC_Eval
datasets:
- null
tags:
- evaluate
- metric
description: >-
This repo contains code of an automatic evaluation metric described in the
paper Compression, Transduction, and Creation: A Unified Framework for
Evaluating Natural Language Generation
sdk: gradio
sdk_version: 3.0.2
app_file: app.py
pinned: false
Metric Card for CTC_Eval
Metric Description
- Previous work on NLG evaluation has typically focused on a single task and developed individual evaluation metrics based on specific intuitions.
- In this work, we propose a unifying perspective based on the nature of information change in NLG tasks, including compression (e.g., summarization), transduction (e.g., text rewriting), and creation (e.g., dialog).
- A common concept underlying the three broad categories is information alignment, which we define as the extent to which the information in one generation component is grounded in another.
- We adopt contextualized language models to measure information alignment.
How to Use
Example:
>>> ctc_score = evaluate.load("yzha/ctc_eval")
>>> results = ctc_score.compute(references=['hello world'], predictions='hi world')
>>> print(results)
{'ctc_score': 0.5211202502250671}
Inputs
- input_field
references
: The document contains all the informationpredictions
: NLG model generated text
Output Values
The CTC Score.
Citation
@inproceedings{deng2021compression, title={Compression, Transduction, and Creation: A Unified Framework for Evaluating Natural Language Generation}, author={Deng, Mingkai and Tan, Bowen and Liu, Zhengzhong and Xing, Eric and Hu, Zhiting}, booktitle={Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing}, pages={7580--7605}, year={2021} }