JPARK logo

This page provides additional details on JPARK, an ontological knowledge powered probabilistic approach for jointly revising multiple NLP entity annotations.

The proposed approach is fully implemented and evaluated in the following paper:

  • Joint Posterior Revision of NLP Annotations via Ontological Knowledge
    By Marco Rospocher and Francesco Corcoglioniti.
    In Proceedings of the 27th International Joint Conference on Artificial Intelligence and the 23rd European Conference on Artificial Intelligence, IJCAI-ECAI 2018, Stockholm, Sweden, July 13-19, 2018
    [bib] [pre-print/mirror]

JPARK has been evaluated on three reference datasets for Named Entity Recognition and Classification (NERC) and Entity Linking (EL):

  • AIDA CoNLL-YAGO: This dataset consists of 1,393 English news wire articles from Reuters, with 34,999 mentions hand-annotated with named entity types (PER, ORG, LOC, MISC) for the CONLL2003 shared task on named entity recognition, and later hand-annotated with the YAGO2 entities and corresponding Wikipedia page URLs. It is split in three parts: eng.train (946 docs), eng.testa (216 docs), eng.testb (231 docs).
  • MEANTIME: The NewsReader MEANTIME corpus consists of 480 news articles from Wikinews, in four languages. In our evaluation, we used only the English section and its 120 articles. The dataset, used as part of the SemEval 2015 task on TimeLine extraction, includes manual annotations for named entity types (only PER, ORG, LOC) and DBpedia entity links.
  • TAC-KBP: Developed for the TAC KBP 2011 Knowledge Base Population Track, this dataset consists of 2,231 English documents, including newswire articles and posts to blogs, newsgroups, and discussion fora. For each document, it is known that all the mentions of one or a few query entities can be linked to a certain Wikipedia page and to a specific NERC type (only PER, ORG, LOC), giving rise to a (partially) annotated gold standard for NERC and EL.

The following JPARK resources are made available:

  • TSV (~39MB) containing the model used in the experiments and trained on AIDA CoNLL-YAGO (eng.train). Its columns contain:
    1. a YAGO Class Set (classes in the set are space separated)
    2. the conditional probability — cf. eq. (6) in the paper — of having that class set given a NERC PER annotation
    3. the conditional probability — cf. eq. (6) in the paper — of having that class set given a NERC ORG annotation
    4. the conditional probability — cf. eq. (6) in the paper — of having that class set given a NERC LOC annotation
    5. the conditional probability — cf. eq. (6) in the paper — of having that class set given a NERC MISC annotation
    6. the prior probability of that class set estimated from the ontological background knowledge — cf. eq. (7) in the paper
    7. all the entities (space separated) having as types exactly the classes in that class set
  • PDF (~83KB) file containing all evaluation metrics computed for all measures, with and without using JPARK, by
    • micro-averaging, considering only mentions in the gold standard;
    • micro-averaging, considering all mentions returned by the system;
    • macro-averaging by document;
    • macro-averaging by NERC type.
  • ZIP (~985KB) package of the evaluation folder, containing:
    • the official TAC scorer;
    • commands for computing scores (and statistical significance) for all metrics and measures considered (cf. the paper for details on interpreting the values);
    • gold, standard, and PSL4EA annotations for all datasets (excluding TAC-KBP, under LDC copyright).

Back to top

Last Published: 2018/05/30.

Reflow Maven skin by Andrius Velykis.

Data and Knowledge Management tools