• Víctor Rodríguez Doncel, Universidad Politécnica de Madrid, Spain
  • Mohamed Yahya, Bloomberg, USA
  • Javier D. Fernández, Vienna University of Economics and Business, Austria
  • Vinh Nguyen, Wright State University, USA
  • Lydia Pintscher, Wikimedia Berlin, Germany
  • Ricardo Usbeck, University Leipzig, Germany
  • Vanessa Lopez, IBM, Ireland
  • Thomas Pellissier, ENS Lyon, France
  • Lucie-Aimée Kaffee, University of Southampton, United Kingdom
  • Manolis Koubarakis, National and Kapodistrian University of Athens, Greece
  • Johann Stan, European Patent Office, The Hague, Netherlands

Detailed program (under construction)















July 3rd


Ricardo Usbeck


Víctor Rodríguez Doncel


Demo + poster session 


Demo + poster session


July 4th




Johann Stan


Javier D. Fernández


Javier D. Fernández (Hands on)

Welcome Aperitif

Wednesday, July 5th




 Vanessa Lopez


Manolis Koubarakis


Departure by bus

Boat trip


July 6th


Mohamed Yahya


Lydia Pintscher 


Vinh Nguyen


Vinh Nguyen
(Hands on)


July 7th

Registration Coffee

 T. Pellissier
L.-A. Kaffee


T. Pellissier
L.-A. Kaffee







  • Víctor Rodríguez Doncel
    Question-Answering for providing Technology Intelligence Services to Small and Medium Enterprises
    Benefits of Technology Intelligence (TI) services for SMEs include the assessment of the technologies in use, the identification of new technologies or technology-based products and services, the opportunities for technical collaboration or technology innovation, the recognition of competitors and providers, and the detection of potential and experienced customers for their portfolio of products and services.  In this talk, I will present the design of innovative TI services to provide solutions to technology-based small and medium enterprises (SMEs) and how the system is being implemented using the IBM Bluemix components and IBM Watson components.

  • Mohamed Yahya
    Question Answering in Bloomberg.
    Bloomberg is the world's leading financial data provider. The data, which comes in structured and unstructured flavors, is made available to our 400k subscribers through various functions in the "Bloomberg Terminal". In this three-part talk I present our ongoing effort to create a unified question answering (QA) interface within the Bloomberg Terminal to streamline users' access to financial knowledge.

  • Javier D. Fernández
    Managing Compressed Big Semantic Data: Theory and Practice
    The steady adoption of Linked Data in recent years has led to a significant increase in the volume of RDF datasets.  The potential of this Semantic Big Data is under-exploited when data management is based on traditional, human-readable RDF representations, which add unnecessary overheads when storing, exchanging and consuming RDF in the context of a large-scale and machine-understandable Semantic Web. This scenario calls for efficient and functional representation formats for RDF as an essential tool for RDF preservation, sharing, and management.
    In the first part of the session, after introducing the main challenges emerging in a Big Semantic Data scenario, we will present fundamental concepts of Compact Data Structures and RDF self-indexes. We will analyze how RDF can be effectively compressed by detecting and removing two different sources of syntactic redundancy. Then, we will introduce HDT, a compact data structure and binary serialization format that keeps big datasets compressed, saving space while maintaining search and browse operations without prior decompression. We will present the HDT deployment in projects such as Linked Data Fragments, which provides a uniform and lightweight interface to access RDF in the Web, indexing/reasoning systems like HDT-FoQ/Jena and LOD Laundromat, a project serving a crawl of a very big subset of the Linked Open Data Cloud.
    Finally, we will inspect the challenges of representing and querying evolving semantic data. In particular we will present different modeling strategies and compact indexes to cope with RDF versions, allowing cross-time queries to understand and analyse the history and evolution of dynamic datasets.

    In the hands-on session, students will gain experience on compressed semantic big data management, with a main focus on HDT and related applications. Materials will be composed of slides, code snippets, datasets and existing libraries in . All tutorial material will be available to anyone on Github and the project website
  • Vinh Nguyen
    Representing, Querying, and Reasoning with RDF Statements about Statements using Singleton Property Model
    Abstract: RDF statement about statement describes various contextual information about a statement such as time, location, certainty, and provenance. In the first session, I will present the singleton property model for representing and tracking the context of the RDF statements, how its graph pattern can be queried in SPARQL, and how the context for inferred triples can be derived.
    In the hands-on session, I will go through various examples of the existing knowledge bases such as Yago2S and BKR for demonstrating the use of the singleton property model in comparison with existing models such as named graph and reification. Students are welcome to bring examples from their own research projects for discussion. A tutorial for transforming existing knowledge bases to the singleton property model using the tool RDF-contextualizer will also be given.

  • Lydia Pintscher
    Wikidata: current state and challenges
    Wikidata is 4,5 years old now. As Wikimedia's structured data project it has made great strides over that time. It has attracted a large community and figured out ways to maintain a multilingual and multicultural knowledge base in the open. In this talk we'll look at the current state of the project, the challenges it still faces as well as some of the remaining open research questions.

  • Ricardo Usbeck
    Question Answering - Problems and the Search for Solutions
    Man-made AI is able to defeat the world-best Jeopardy players and thousands of research papers address Question Answering (QA), but recent QA benchmarks based on the Web of Data still show huge performance gaps. Using the Question Answering over Linked Data (QALD) challenge, we identified two main root causes: a) missing measurement of KPIs to drive research in the right direction b) missing technologies to solve particular types of questions generically.  Here, we are going to inspect recent takes on improving the measurement of QA systems and suggest requirements for future developments. The take-aways will include a list of reviewer-convincing list of techniques to develop and publish novel QA systems.

  • Vanessa Lopez
    Semantics and QA in Action: cognitive solutions for Integrated Care
    Cognitive technologies promise to have significant societal impact in domains where there is a need to transform multidisciplinary information into actionable services. But with most of information still unstructured, how do we demonstrate the impact of QA on the web of data? We argue that to know if we are answering the important questions and doing what it needs to be done to push the state of the art we need a story to tell. To do that we look at the potential use of semantic and cognitive QA in the health and social care sector. In this talk, we present a use case that require harvesting large amounts of data and discuss research challenges and future directions. Embodying cognitive approaches that combine semantics, NLP and learning remain as a significant challenge. In particular for knowledge acquisition, to find and combine meaningful pieces of knowledge from sources with evidence for answers, even without a complete representation of a question, and to facilitate intuitive human interaction, in which professionals interact in a natural way with the system and the systems reacts and adapts its understanding and knowledge to give better answers to questions.
  • Thomas Pellissier and Lucie Kaffee
    Practical Wikidata and librairies (hands-on)
    In the hands-on session for Wikidata, we will introduce various ways of working with Wikidata and getting the information you need. Whether that is by using the SPARQL endpoint and learning all the small tricks to handle it or understanding how to get the whole of Wikidata at once and looking for that one information that’s needed. Furthermore, we will show you how to understand Wikidata, and how to become part of its editor community. What happens if information is missing? Where does the information come from (provenance of data)? These and many more questions will be answered.

  • Manolis Koubarakis
    Writing European Project Proposals

  • Johann Stan
    Patent writing and evaluating. Overview of patents in the QA domain

    Knowledge in patenting represents a skill that can make a difference in academia or industry. In this talk, as a patent examiner I will elaborate on the patenting process with examples from information retrieval and question answering systems. Further, the main criteria for obtaining a software patent in Europe will be discussed.