biu-nlp/f-coref

transformersentransformerspytorchrobertafastcoreference-resolutionenmit
325.8K

F-Coref: Fast, Accurate and Easy to Use Coreference Resolution

F-Coref allows to process 2.8K OntoNotes documents in 25 seconds on a V100 GPU (compared to 6 minutes for the LingMess model, and to 12 minutes of the popular AllenNLP coreference model) with only a modest drop in accuracy. The fast speed is achieved through a combination of distillation of a compact model from the LingMess model, and an efficient batching implementation using a technique we call leftover

Please check the official repository for more details and updates.

Experiments

ModelRuntimeMemory
Joshi et al. (2020)12:0627.4
Otmazgin et al. (2022)06:434.6
+ Batching06:006.6
Kirstain et al. (2021)04:374.4
Dobrovolskii (2021)03:493.5
F-Coref00:453.3
+ Batching00:354.5
+ Leftovers batching00:254.0
The inference time(Min:Sec) and memory(GiB) for each model on 2.8K documents. Average of 3 runs. Hardware, NVIDIA Tesla V100 SXM2.

Citation

@inproceedings{Otmazgin2022FcorefFA,
  title={F-coref: Fast, Accurate and Easy to Use Coreference Resolution},
  author={Shon Otmazgin and Arie Cattan and Yoav Goldberg},
  booktitle={AACL},
  year={2022}
}

F-coref: Fast, Accurate and Easy to Use Coreference Resolution (Otmazgin et al., AACL-IJCNLP 2022)

DEPLOY IN 60 SECONDS

Run f-coref on Runcrate

Deploy on H100, A100, or RTX GPUs. Pay only for what you use. No setup required.