This page provides additional details on the use of Knowledge Extraction techniques, as implemented in PIKES, to improve the performances of Information Retrieval. The proposed approach is fully implemented in the following paper:

  • Knowledge Extraction for Information Retrieval
    By Francesco Corcoglioniti, Mauro Dragoni, Marco Rospocher, and Alessio Palmero Aprosio.
    In Proceedings of the 13th European Semantic Web Conference (ESWC2016), Anissaras, Crete, Greece, May 29-June 2, 2016
    [bib] [pre-print/mirror]

Here, we provide a brief overview of the approach, make available for downlaod all the code and data used in the evaluation (to allow reproducibility of our results), and provide all the reports and the additional material we produced as part of the evaluation.


The goal in Information Retrieval is to determine, for a given text query, the relevant documents in a text collection, ranking them according to their relevance degree for the query.

In our approach, named KE4IR (read: kee-fer) and implemented on top of PIKES and Apache Lucene, both queries and documents are processed to extract semantic terms pertaining to the following semantic layers:

  • URI layer, containing URIs of entities mentioned in the text, disambiguated against DBpedia;
  • TYPE layer, containing YAGO classes associated to noun phrases in the text, extracted via disambiguation to WordNet (plus WordNet to YAGO mappings) or imported from DBpedia after entity linking;
  • TIME layer, containing temporal values explicitly expressed in the text, or imported from DBpedia after entity linking;
  • FRAME layer, containing <frame type, participant> pairs where the frame results from knowledge extraction and the participant is a disambiguated entity.

We adopt a retrieval model inspired to the Vector Space Model (VSM). We represent both documents and queries as term vectors whose elements are the weight of textual and semantic terms, computed based on TF / IDF values opportunely extended for the use with semantic information, plus a weight that allows assigning different importances to distinct layers (in our experiments, we use a setup where textual and semantic information contribute equally). The similarity score between a document and a query is computed via the scalar product of their vectors, and is used to identify and rank relevant documents.


We built an evaluation infrastructure that implements the KE4IR approach and allows applying it on arbitrary documents and queries, measuring retrieval performances against gold relevance judgments.

The infrastructure requires in input the NLP annotations and the knowledge graphs extracted with PIKES for both documents and queries of the test dataset, plus an index containing background knowledge that will be injected into the knowledge graphs prior to their processing. For each document or query, textual terms are extracted starting from the tokenization in the NLP annotations, while semantic terms are extracted from the knowledge graph. Document terms are indexed in a Lucene inverted index. At search time, query terms are OR-ed in a Lucene query that locates documents containing at least one term. Then, matched documents are scored and ranked externally to Lucene (for ease of testing) according to KE4IR retrieval model. The resulting document ranking is compared with the gold relevance judgments to compute a comprehensive set of evaluation metrics, which are averaged along different queries. The evaluation infrastructure emits several CSV files:

  • aggregates.csv: aggregated results over all queries of the dataset;
  • query-NNN.csv: results on query NNN, for each feasible layer combination;
  • ranking-NNN.csv: document rankings per layer on query NNN;
  • settings-ABC.csv: results for a given layer combination ABC, on each query of the dataset.

We provide the evaluation infrastructure as a separate GitHub project. A precompiled version is available. Please refer to the documentation on the GitHub project for specific instructions on how to load a dataset and run the code.

Evaluation 1: WES Dataset

KE4IR has been validated on the ad-hoc IR task, consisting in performing a set of queries over a document collection for which the list of relevance judgments is available, comparing for each query the document rankings produced by KE4IR with the gold rankings. We adopted the document collection from, described in the paper:

Waitelonis, J., Exeler, C., Sack, H.: Linked Data enabled generalized vector space model to improve document retrieval

  • Linked Data enabled Generalized Vector Space Model to improve document retrieval
    By Jörg Waitelonis and Claudia Exeler and Harald Sack.
    In Proc. of NLP & DBpedia 2015 workshop in conjunction with 14th International Semantic Web Conference (ISWC2015), 2015, CEUR Workshop Proceedings
    [pdf] [slides]

The collection is composed of 331 documents from a web blog and 35 queries. The relevance of each document is expressed in a multi-value scale with scores going from 5 (the document contains exact information with respect to what the user is looking for) to 1 (the document is of no interest for the query). Document scored 2 to 5 are thus relevant, while document not scored or scored 1 are irrelevant. For NDCG, we directly used the score as is (range 2-5) for computing the metric.

The raw texts distributed in the dataset at misses some pieces of text (corresponding to links and other special markup in original HTML pages). We fixed this issue by downloading and converting again the original HTML texts. The obtained texts, together with the queries and the gold relevance judgements from, are all made available here for download using the data formats supported by our evaluation infrastructure (NAF for texts, TSV for relevance judgements); data in the NIF format, including manual entity linking annotations not used here, is available on the original website.

Below we provide all the input, output and intermediate data involved in the evaluation, divided in multiple files for convenience:

  • Dataset files
  • Knowledge extraction (PIKES) and enrichment results
  • Term extraction results
  • Evaluation results
    • zip package containing the CSV files produced by the evaluation infrastructure - (180 KB):
    • spreadsheet with performances aggregated over queries, using different layer combinations (data related to Table 3) - results_aggregates.ods (60 KB)
    • spreadsheet with performances (and top 10 results with their scores) of each query, for each significant layer combination (data related to Table 4) - results_queries.ods (292 KB)
    • spreadsheet with rankings returned by each query considering one semantic layer at a time, with scores obtained before applying layer weight w(l(t)); data not used in the paper but possibly useful to further compare layers’ performances - results_rankings.ods (212 KB)
    • CSV file with performances averaged over all queries, measured by varying the total weight assigned to semantic layers in 0.01 steps (data related to Figure 2) - results_semantic_weight_analysis.csv (576 KB)
    • TSV with retrieval performances using SOLR (slightly worse than our textual baseline) - solr_result.tsv (3 KB)
      If all you want is to reproduce our results as quickly as possible, we provide a zipped folder with the minimum required data (NLP annotations and enriched graphs for documents and queries, relevance judgments): (206 MB)


  • download and extract the ZIP file containing the data folder
  • download the precompiled binaries of the evaluation infrastructure
  • in a shell, execute ke4ir-eval -p /path/to/data/folder/
  • you should get an output similar to this one, ending with a table reporting the values of several metrics for different layer combinations


  • a Linux /Mac OS X / Unix-like machine
  • Java 8

Note: the data folder contains the relevance judgments, the NLP annotations of documents and queries, and the knowledge graphs of documents and queries already enriched with background knowledge, so to avoid shipping the (very large) background knowledge index. Document and query terms, the lucene index, and the CSV report files are not included as they are generated by running the command above (the spreadsheets were built manually from the CSV reports, while the CSV file with the semantic weight analysis was produced calling the command multiple times with some script hack).

Evaluation 2: Fernandez et al. Dataset

For this evaluation, we used the semantic search dataset proposed in:

  • Semantically enhanced Information Retrieval: an ontology-based approach
    By Miriam Fernandez, Ivan Cantador, Vanesa Lopez, David Vallet, Pablo Castells, Enrico Motta.
    Web Semantics: Science, Services and Agents on the World Wide Web, [S.l.], v. 9, n. 4, Jan. 2012. ISSN 1570-8268.
    [pdf] (more details also available on the dataset webpage)

Evaluation Results:

Evaluation 3: Trec 6-7-8-9-2001 Datasets

For this evaluation, we used the datasets adopted in the Trec 6, 7, 8, 9. and 2001 evaluation campaigns. More details can be found on the Trec web-site.

Evaluation Results:

Back to top

Last Published: 2017/01/19.

Reflow Maven skin by Andrius Velykis.

Data and Knowledge Management tools