f-coref / README.md
shon711's picture
Update README.md
e91dfff
|
raw
history blame
2.32 kB
metadata
language:
  - en
tags:
  - fast
  - coreference-resolution
license: mit
datasets:
  - multi_news
  - ontonotes
metrics:
  - CoNLL
task_categories:
  - coreference-resolution
model-index:
  - name: biu-nlp/f-coref
    results:
      - task:
          type: coreference-resolution
          name: coreference-resolution
        dataset:
          name: ontonotes
          type: coreference
        metrics:
          - name: Avg. F1
            type: CoNLL
            value: 78.5

F-Coref: Fast, Accurate and Easy to Use Coreference Resolution

F-Coref allows to process 2.8K OntoNotes documents in 25 seconds on a V100 GPU (compared to 6 minutes for the LingMess model, and to 12 minutes of the popular AllenNLP coreference model) with only a modest drop in accuracy. The fast speed is achieved through a combination of distillation of a compact model from the LingMess model, and an efficient batching implementation using a technique we call leftover

Please check the official repository for more details and updates.

Experiments

Model Runtime Memory
Joshi et al. (2020) 12:06 27.4
Otmazgin et al. (2022) 06:43 4.6
+ Batching 06:00 6.6
Kirstain et al. (2021) 04:37 4.4
Dobrovolskii (2021) 03:49 3.5
F-Coref 00:45 3.3
+ Batching 00:35 4.5
+ Leftovers batching 00:25 4.0
The inference time(Min:Sec) and memory(GiB) for each model on 2.8K documents. Average of 3 runs. Hardware, NVIDIA Tesla V100 SXM2.

Citation

@inproceedings{Otmazgin2022FcorefFA,
  title={F-coref: Fast, Accurate and Easy to Use Coreference Resolution},
  author={Shon Otmazgin and Arie Cattan and Yoav Goldberg},
  booktitle={AACL},
  year={2022}
}

F-coref: Fast, Accurate and Easy to Use Coreference Resolution (Otmazgin et al., AACL-IJCNLP 2022)