winograd_th / README.md
pakphum's picture
Update README.md
7ca1dd2 verified
---
language:
- th
license: cc-by-4.0
task_categories:
- multiple-choice
dataset_info:
features:
- name: qn
dtype: int64
- name: label
dtype: int64
- name: pronoun
dtype: string
- name: quote
dtype: string
- name: source
dtype: string
- name: text
dtype: string
- name: options
sequence: string
- name: pronoun_loc
dtype: int64
- name: quote_loc
dtype: int64
splits:
- name: test
num_bytes: 108414
num_examples: 285
download_size: 36552
dataset_size: 108414
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
size_categories:
- n<1K
---
# A collection of Thai Winograd Schemas
We present a collection of Winograd Schemas in the Thai language. These schemas are adapted from the original set of English Winograd Schemas proposed by [Levesque et al.](https://commonsensereasoning.org/2011/papers/Levesque.pdf), which was based on [Ernest Davis's collection](https://cs.nyu.edu/~davise/papers/WinogradSchemas/WSCollection.xml).
## Dataset Translation
Two professional translators, who were native Thai speakers fluent in English and had experience translating from English to Thai, were hired. In a pilot translation phase, one native speaker translated the first 85 sentences. Based on a qualitative analysis of these 85 sentences, guidelines were provided for a second native speaker to translate the remaining 200 sentences. In total, 285 sentences were translated from English to Thai. Translation guidelines were provided, instructing them to adapt names and contexts to suit the Thai language while preserving the ambiguity and nuances of the original schema. The translators were also asked to mark any translated names and translations they were unsure about, so that the validator in the next step could pay extra attention to those instances. For example, in \autoref{fig:example}, the names Paul and George were changed to Mana and Piti, respectively, adapting the names to better suit the Thai context while preserving the essence of the original content.
## Dataset Validation
The translated Winograd Schemas were reviewed with three native Thai speakers, and final adjustments were made to ensure clarity. A validator was provided with the translations and tasked with identifying any potential issues. They were instructed to pay closer attention to the text marked by the translators, which included changes to names and translations that the translators were unsure about. Based on the validator’s feedback, final adjustments to the translations were made, and typographical errors were corrected. Furthermore, the dataset will be made publicly available\footnote{A link to the dataset will be added in the de-anonymized version.}, inviting other native Thai speakers to verify it and suggest any necessary adjustments.
# Dataset Structure
### Data Instances
Each instance contains a text passage with a designated pronoun and two possible answers indicating which entity in
the passage the pronoun represents. An example instance looks like the following:
```python
{
'qn': 0,
'label': 0,
'pronoun': 'พวกเขา',
'quote': 'พวกเขากลัวความรุนแรง',
'source': '(Winograd 1972)',
'text': 'สมาชิกสภาเทศบาลเมืองปฏิเสธใบอนุญาตผู้ชุมนุมเพราะพวกเขากลัวความรุนแรง'
'options': ('สมาชิกสภาเทศบาลเมือง', 'ผู้ชุมนุม'),
'pronoun_loc': 48,
'quote_loc': 48
}
```
### Data Fields
- `qn` (int): The number of the question based on [winograd_wsc285](https://huggingface.co/datasets/winograd_wsc)
- `label` (int): The index of the correct option in the `options` field
- `pronoun` (str): The pronoun in the sequence to be resolved
- `quote` (str): The substr with the key action or context surrounding the pronoun
- `source` (str): A description of the source who contributed the example
- `text` (str): The text sequence
- `options` (tuple[str]): The two entity options that the pronoun may be referring to
- `pronoun_loc` (int): The starting position of the pronoun in the sequence
- `quote_loc` (int): The starting position of the quote in the sequence
### Data Splits
Only a test split is included.
# Evaluation
## Model Accuracy in English and Thai
| Model | Accuracy (English) | Accuracy (Thai) |
|-----------------|--------------------|-----------------|
| Typhoon | 64.56% | 54.04% |
| Claude-3-Haiku | 62.81% | 51.58% |
| Claude-3-Sonnet | 80.70% | 63.51% |
| Claude-3-Opus | 92.63% | 77.54% |
| GPT-3.5 | 71.93% | 51.23% |
| GPT-4 | 93.68% | 72.28% |
| Human | 90% | -% |
**Table 1: Accuracy vs. Model in English and Thai**
## Acknowledgement
We extend our gratitude to Chanikarn Inthongpan and Korakoch Rienmek, who translated the schemas into the Thai language. We would also like to thank Sakrapee Namsak and Chonnawit Khumchoo for validating the translated Thai dataset.
Dataset Curated and Maintained by Phakphum Artkaew
Any comments or concerns can be reached at pa2497@nyu.edu