ctc_eval / README.md
yzha's picture
update
630c62a
---
title: CTC_Eval
datasets:
-
tags:
- evaluate
- metric
description: "This repo contains code of an automatic evaluation metric described in the paper
Compression, Transduction, and Creation: A Unified Framework for Evaluating Natural Language Generation"
sdk: gradio
sdk_version: 3.0.2
app_file: app.py
pinned: false
---
# Metric Card for CTC_Eval
## Metric Description
* Previous work on NLG evaluation has typically focused on a single task and developed individual evaluation metrics based on specific intuitions.
* In this work, we propose a unifying perspective based on the nature of information change in NLG tasks, including compression (e.g., summarization), transduction (e.g., text rewriting), and creation (e.g., dialog).
* A common concept underlying the three broad categories is information alignment, which we define as the extent to which the information in one generation component is grounded in another.
* We adopt contextualized language models to measure information alignment.
## How to Use
Example:
```python
>>> ctc_score = evaluate.load("yzha/ctc_eval")
>>> results = ctc_score.compute(references=['hello world'], predictions='hi world')
>>> print(results)
{'ctc_score': 0.5211202502250671}
```
### Inputs
- **input_field**
- `references`: The document contains all the information
- `predictions`: NLG model generated text
### Output Values
The CTC Score.
## Citation
@inproceedings{deng2021compression,
title={Compression, Transduction, and Creation: A Unified Framework for Evaluating Natural Language Generation},
author={Deng, Mingkai and Tan, Bowen and Liu, Zhengzhong and Xing, Eric and Hu, Zhiting},
booktitle={Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing},
pages={7580--7605},
year={2021}
}