Skip to content

Latest commit

 

History

History
285 lines (182 loc) · 10.2 KB

README.md

File metadata and controls

285 lines (182 loc) · 10.2 KB

SemanticBot

Build Status

SemanticBot is a semi-interactive ontology mapping tool. It provides an easy-to-use interface in order to semantize (i.e. to map ontologies on) OpenDataSoft datasets. Schema View

Publication

For more details, you can read the following demonstration paper:

Benjamin Moreau, Nicolas Terpolilli, Patricia Serrano-Alvarado. A Semi-Automatic Tool for Linked Data Integration. 18th International Semantic Web Conference (ISWC2019), Oct 2019, Auckland, New Zealand.

Glossary

DBPedia

DBpedia is a knowledge graph containing informations extracted from wikipedia . SemanticBot uses DBpedia to perform named entity recognition. In other word, to find class of entities (e.g., Italia is a country/PopulatedPLace.. or B.Obama is a President/Person..)

LOV (Linked Open Vocabularies)

LOV is an ontology search engine. It is used by the SemanticBot to find candidate ontologies for opendatasoft datasets.

Ontology (Vocabulary)

An ontology is a vocabulary defining the concepts and relationships used to describe an area of concern. It's composed of:

  • classes (e.g. Car, Building, Person, Disease, Source Code) to represent a concept.
  • properties (e.g. horsepower, financed By, date of birth, has Symptoms, Author) to represent relation between concepts.
  • rules (e.g. A person have a unique date of birth). Ontologies can be created for every area of concern and by everyone using RDF (Resource Description Framework), RDFS (RDF Schema) and OWL (Web Ontology Language).

RML & YARRRML

RML is a generic mapping language to describe multi-format to RDF transformations. YARRRML is a simple syntax to respresent RML rules.

Installation

Assuming you already have python 3.6, pip, yarn and docker-compose:

Clone the repository and go to the root folder: semantic-bot.

It is strongly recommended to create a new virtualenv.

Automatic installation

Run the installation script

./install.sh

Then, in file chatbot_app/local_settings.py update the secret key.

SECRET_KEY = "<SECRET_KEY>"

If you get errors, proceed to the manual installation.

Manual installation

install python dependencies with pip

pip install -r requirements.txt

Download ttl dumps of DBPedia types and labels at this address:

https://eu.ftp.opendatasoft.com/bmoreau/ttl_data_dumps.zip

and override /data_dumps

create a file chatbot_app/local_settings.py and add a secret key.

SECRET_KEY = "<SECRET_KEY>"

install js dependencies

yarn

run Elasticsearch (ES) with docker-compose

docker-compose up -d

Finally, when Elasticsearch is ready, load all the turtle files into ES with:

python -m scripts.bulk_load

Mapping serializer

Resulting mapping can be serialized using RML turtle syntax or YARRRML (default) syntax. To choose between the two serializer, add:

MAPPING_SERIALIZER = 'RML'

or

MAPPING_SERIALIZER = 'YARRRML'

in your chatbot_app/local_settings.py file.

Run the demo

Navigate to semantic-bot folder and execute the two commands:

yarn run build
python manage.py runserver

App should be running on http://127.0.0.1:8000/

API

SemanticBot is powered by an API exposed by this service:

Correspondances API

Correspondences are the semantic correspondences between a dataset's field and an ontology. Fields can be linked to classes (Car, Person, ...) or poroperties of class (Engine horsepower, full name, ...).

GET /api/{dataset-id}/correspondances/classes to retrieve field values/class correspondences (powered by DBpedia, Yago and LOV).

GET /api/{dataset-id}/correspondances/properties to retrieve field name/properties correspondences (powered by LOV).

GET /api/{dataset-id}/correspondances/ to retrieve both correspondences.

POST /api/{dataset-id}/correspondances/mapping to translate a set of correspondences into valid RML mapping file.

POST /api/{dataset-id}/correspondances/confirmed push confirmed correspondences after semantization.

POST /api/{dataset-id}/correspondances/awaiting push confirmed correspondences after semantization.

POST /api/{dataset-id}/correspondances/denied push denied correspondences after semantization.

Conversation API

Conversations API is used to translate possible correspondences into Human Readable questions.

POST /api/conversation/question/class to retrieve a question about a class/field correspondence.

POST /api/conversation/question/property to retrieve a question about a property/field correspondence.

POST /api/conversation/question/property-class to retrieve a question about a class/property correspondence.

GET /api/conversation/error/lov-unavailable to retrieve phrase when LOV is unavailable.

GET /api/conversation/error/no-classes to retrieve phrase when no class is found.

GET /api/conversation/greeting to retrieve welcome phrase.

GET /api/conversation/instructions to retrieve instructions to use the SemanticBot.

GET /api/conversation/answer/positive to retrieve response to positive user input.

GET /api/conversation/answer/neutral to retrieve response to neutral user input.

GET /api/conversation/answer/negative to retrieve response to negative user input.

GET /api/conversation/salutation to retrieve phrase to say goodbye to the user.

Named Entity Recognition API

GET /api/ner?q=[query]&lang=[language] returns the class of the term in the query.

How it works?

Global approach

Our approach works with 3 automatic steps:

  1. Entities Recognition
  2. Properties Recognition
  3. Mapping Generation

Between each step, semantic bot asks simple questions to the user in order to validate the semantization.

Global approach

1. Entities Recognition

The goal of this step is to identify which class correspond to each field of the dataset.

N-first values of each fields are searched in DBPedia and Yago. Classes of relevant resources are returned. For each field, class that occur the most will be search in LOV and proposed to the user. Class questions are made using rdfs:label of classes.

the following figure illustrate the entities recognition step.

Entities Recognition

2. Properties Recognition

The goal of this step is to identify which property correspond to each field of the dataset.

field name and datatype of each field are searched in LOV property api. Property questions are made using rdfs:label and of properties.

the following figure illustrate the Properties recognition step. Properties Recognition

3. Mapping Generation

Before serializing the RDF mapping, a lot of mapping rules can be inferred.

First, using rdfs:domain and rdfs:range of properties, we can infer new classes that correspond to dataset fields.

In the following example, class dbo:place is inferred as a corresponding class of field Birth Province because the rdfs:range of the property dbo:birthPlace is dbo:place.

Domain Range

Then, we add rdfs:label property for each type of resources in the RDF mapping. This is important to have a human-readable RDF dataset. Value of the rdfs:label is the field that was used to identify the class of the resource.

The following example show how rdfs:label properties are infered.

RDFS Label

Finally, we add new class and properties by saturating the RDF mapping using using subClassOf, equivalentClass, subPropertyOf, equivalentProperty rules.

The following example shows inferred classes using rdfs:subClassOf, owl:equivalentClass

Sub Class

The following example shows inferred properties using rdfs:subPropertyOf, owl:equivalentProperty

Sub Property

At the end, mapping is serialized and sent to the user.

Result

Logs

logs folder contains data about the usage of the SemanticBot:

  • Number of semantized dataset
  • Number of cancelled semantization
  • Number of failed semantization
  • Average time to semantize a dataset
  • Field/URI (class or property) associations

Analyse logs with:

python logs/log_analyser.py logs/chatbot_results.log

Fully Automatic Approach (Beta)

We also provide a fully automatic approach to generate an RDF mapping for datasets. The three automatic steps (i.e., entities recognition, properties recognition and, mapping generation) are executed sequentially. Between each step, class and property correspondences are confirmed by the SemanticBot without asking the user.

This approach can be useful for users that want to generate RDF mappings for a large set of datasets.

For better results, we recommend to check and improve the quality of each mapping after generation.

How To

Proceed to the installation described in Section Installation

In your chatbot_app/local_settings.py file (create one if it does not exist). Add the api_key for the domain of your dataset by adding the following line:

DATA_API_KEY = 'YOUR_API_KEY_HERE'

Run the script

Then you can run the script on your dataset with the following command:

python chatbot/automatic.py -D domain-id -d dataset-id -o output-file

where:

  • domain-id is the domain-id of the domain of your dataset
  • dataset-id is the dataset-id of your dataset
  • output-file is the filename of the resulting RDF mapping

working example:

python chatbot/automatic.py -D data -d roman-emperors@public -o roman-emperors@public.rml.yml
notes:

Chatbot icon designed by Oksana Latysheva, is licensed under CC BY 3.0.