Skip to content

bltlab/conll-sharp

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 

Repository files navigation

CoNLL#

This is the repository for the LREC-COLING 2024 paper CoNLL#: Fine-grained Error Analysis and a Corrected Test Set for CoNLL-03 English.

We are in the process of releasing the data.

Contact

Please email lignos at brandeis dot edu with any questions.

Citation

@inproceedings{rueda-etal-2024-conll-fine,
    title = "{C}o{NLL}{\#}: Fine-grained Error Analysis and a Corrected Test Set for {C}o{NLL}-03 {E}nglish",
    author = "Rueda, Andrew  and
      Alvarez-Mellado, Elena  and
      Lignos, Constantine",
    editor = "Calzolari, Nicoletta  and
      Kan, Min-Yen  and
      Hoste, Veronique  and
      Lenci, Alessandro  and
      Sakti, Sakriani  and
      Xue, Nianwen",
    booktitle = "Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)",
    month = may,
    year = "2024",
    address = "Torino, Italy",
    publisher = "ELRA and ICCL",
    url = "https://aclanthology.org/2024.lrec-main.330",
    pages = "3718--3728",
    abstract = "Modern named entity recognition systems have steadily improved performance in the age of larger and more powerful neural models. However, over the past several years, the state-of-the-art has seemingly hit another plateau on the benchmark CoNLL-03 English dataset. In this paper, we perform a deep dive into the test outputs of the highest-performing NER models, conducting a fine-grained evaluation of their performance by introducing new document-level annotations on the test set. We go beyond F1 scores by categorizing errors in order to interpret the true state of the art for NER and guide future work. We review previous attempts at correcting the various flaws of the test set and introduce CoNLL{\#}, a new corrected version of the test set that addresses its systematic and most prevalent errors, allowing for low-noise, interpretable error analysis.",
}

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published