yzha commited on
Commit
630c62a
1 Parent(s): 2c2f56c
Files changed (1) hide show
  1. README.md +24 -23
README.md CHANGED
@@ -5,7 +5,8 @@ datasets:
5
  tags:
6
  - evaluate
7
  - metric
8
- description: "TODO: add a description here"
 
9
  sdk: gradio
10
  sdk_version: 3.0.2
11
  app_file: app.py
@@ -14,37 +15,37 @@ pinned: false
14
 
15
  # Metric Card for CTC_Eval
16
 
17
- ***Module Card Instructions:*** *Fill out the following subsections. Feel free to take a look at existing metric cards if you'd like examples.*
18
-
19
  ## Metric Description
20
- *Give a brief overview of this metric, including what task(s) it is usually used for, if any.*
 
 
 
21
 
22
  ## How to Use
23
- *Give general statement of how to use the metric*
 
24
 
25
- *Provide simplest possible example for using the metric*
 
 
 
 
26
 
27
  ### Inputs
28
- *List all input arguments in the format below*
29
- - **input_field** *(type): Definition of input, with explanation if necessary. State any default value(s).*
 
30
 
31
  ### Output Values
32
 
33
- *Explain what this metric outputs and provide an example of what the metric output looks like. Modules should return a dictionary with one or multiple key-value pairs, e.g. {"bleu" : 6.02}*
34
-
35
- *State the range of possible values that the metric's output can take, as well as what in that range is considered good. For example: "This metric can take on any value between 0 and 100, inclusive. Higher scores are better."*
36
-
37
- #### Values from Popular Papers
38
- *Give examples, preferrably with links to leaderboards or publications, to papers that have reported this metric, along with the values they have reported.*
39
-
40
- ### Examples
41
- *Give code examples of the metric being used. Try to include examples that clear up any potential ambiguity left from the metric description above. If possible, provide a range of examples that show both typical and atypical results, as well as examples where a variety of input parameters are passed.*
42
-
43
- ## Limitations and Bias
44
- *Note any known limitations or biases that the metric has, with links and references if possible.*
45
 
46
  ## Citation
47
- *Cite the source where this metric was introduced.*
 
 
 
 
 
 
48
 
49
- ## Further References
50
- *Add any useful further references.*
 
5
  tags:
6
  - evaluate
7
  - metric
8
+ description: "This repo contains code of an automatic evaluation metric described in the paper
9
+ Compression, Transduction, and Creation: A Unified Framework for Evaluating Natural Language Generation"
10
  sdk: gradio
11
  sdk_version: 3.0.2
12
  app_file: app.py
 
15
 
16
  # Metric Card for CTC_Eval
17
 
 
 
18
  ## Metric Description
19
+ * Previous work on NLG evaluation has typically focused on a single task and developed individual evaluation metrics based on specific intuitions.
20
+ * In this work, we propose a unifying perspective based on the nature of information change in NLG tasks, including compression (e.g., summarization), transduction (e.g., text rewriting), and creation (e.g., dialog).
21
+ * A common concept underlying the three broad categories is information alignment, which we define as the extent to which the information in one generation component is grounded in another.
22
+ * We adopt contextualized language models to measure information alignment.
23
 
24
  ## How to Use
25
+ Example:
26
+ ```python
27
 
28
+ >>> ctc_score = evaluate.load("yzha/ctc_eval")
29
+ >>> results = ctc_score.compute(references=['hello world'], predictions='hi world')
30
+ >>> print(results)
31
+ {'ctc_score': 0.5211202502250671}
32
+ ```
33
 
34
  ### Inputs
35
+ - **input_field**
36
+ - `references`: The document contains all the information
37
+ - `predictions`: NLG model generated text
38
 
39
  ### Output Values
40
 
41
+ The CTC Score.
 
 
 
 
 
 
 
 
 
 
 
42
 
43
  ## Citation
44
+ @inproceedings{deng2021compression,
45
+ title={Compression, Transduction, and Creation: A Unified Framework for Evaluating Natural Language Generation},
46
+ author={Deng, Mingkai and Tan, Bowen and Liu, Zhengzhong and Xing, Eric and Hu, Zhiting},
47
+ booktitle={Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing},
48
+ pages={7580--7605},
49
+ year={2021}
50
+ }
51