This paper describes an undergoing experiment to compare two tagsets for Named Entities (NE) annotation. We compared Klue 2 tagset, developed by IBM Research, with HAREM tagset, developed for tagging the Portuguese corpora used in Second HAREM competition. From this report, we expected to evaluate our methodology for comparison and to survey the problems that arise from it.
@inproceedings{acl-news-2015, author = {Real, Livy and Rademaker, Alexandre}, title = {HAREM and Klue: how to compare two tagsets for named entities annotation}, booktitle = {Proceedings of 53rd Annual Meeting of the Association for Computational Linguistics and The 7th International Joint Conference on Natural Language Processing of Asian Federation of Natural Language Processing - Named Entities Workshop (NEWS 2015)}, year = {2015}, pdflink1 = {/files/acl-news-2015.pdf}, pdflink2 = {https://aclweb.org/anthology/W/W15/W15-3906.pdf}, month = jul, address = {Beijing, China} }