Skip to main content
. Author manuscript; available in PMC: 2018 Jan 5.
Published in final edited form as: CEUR Workshop Proc. 2016 Sep;1609:28–42.

Table 4.

System performance for plain entity recognition on the EMEA test corpus. Data shown in italic font presents runs that were submitted after the official deadline. The median and average are computed solely using the official runs. A * symbol indicates statistically significant difference of a run with the runs ranked before and after it, per student test.

Team TP FP FN Precision Recall F-measure
Erasmus-run3.unofficial* 1729 685 475 0.716 0.785 0.749
Erasmus-run2* 1732 1001 472 0.634 0.786 0.702
Erasmus-run1* 1757 1063 447 0.623 0.797 0.699
LITL-run1* 879 242 1325 0.784 0.399 0.529
LITL-run2 867 264 1337 0.767 0.393 0.520
SIBM-run1* 834 716 1370 0.538 0.378 0.444
SIBM-run2* 724 483 1480 0.600 0.329 0.425
BITEM-run1* 406 371 1798 0.523 0,184 0.272
UPF-run1* 512 3463 1835 0.129 0.218 0.162
UPF-run2.unofficial* 420 4025 1816 0.095 0.188 0.126

average 0.575 0.436 0.469
median 0.611 0.386 0.482