structure learning with distributed parameter learning for...

37
Structure Learning with Distributed Parameter Learning for Probabilistic Ontologies Giuseppe Cota 1 Riccardo Zese 1 Elena Bellodi 1 Fabrizio Riguzzi 2 Evelina Lamma 1 Dipartimento di Ingegneria – University of Ferrara Dipartimento di Matematica e Informatica – University of Ferrara [giuseppe.cota,riccardo.zese,elena.bellodi,fabrizio.riguzzi, evelina.lamma]@unife.it September 23, 2015 Cota, Zese et al. (UNIFE) LEAP MR September 23, 2015 1 / 21

Upload: others

Post on 30-Jul-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning with Distributed Parameter Learningfor Probabilistic Ontologies

Giuseppe Cota1 Riccardo Zese1 Elena Bellodi1

Fabrizio Riguzzi2 Evelina Lamma1

Dipartimento di Ingegneria – University of Ferrara

Dipartimento di Matematica e Informatica – University of Ferrara[giuseppe.cota,riccardo.zese,elena.bellodi,fabrizio.riguzzi,

evelina.lamma]@unife.it

September 23, 2015

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 1 / 21

Page 2: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Outline

1 Probabilistic Description LogicsIntroductionDISPONTEProbabilistic Inductive Logic ProgrammingCorresponding Systems

2 EDGE: Learning DISPONTE Probabilities

3 Distributed Parameter Learning

4 Class Expression LearningDefinitionCELOE

5 Structure LearningPut it all together!LEAP: Learning DISPONTE theories

6 Distributed Structure Learning

7 Experiments and Results

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 2 / 21

Page 3: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics Introduction

Probabilistic Description Logics: Introduction

• Basic Concepts:• Ontology: a formal and explicit description of a domain of interest• Description Logics (DLs): family of logical formalism (decidable)

fragments of first order logic• Knowledge Base: set of axioms expressed by a DL

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 3 / 21

Page 4: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics DISPONTE

The semantics: DISPONTE

• In real world domains the information is often uncertain

• The Distribution Semantics underlies many probabilistic logicprogramming languages, such as PRISM, ICL, LPADs, ProbLog

• DISPONTE applies the Distribution Semantics of probabilisticlogic programming to description logics

• DISPONTE semantics for probabilistic SROIQ(D).• Probabilistic axioms: p :: E

e.g., p :: C v D represents the fact that we believe in the truth ofC v D with probability p.

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 4 / 21

Page 5: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics DISPONTE

Example

0.4 :: fluffy : Cat (1)

0.3 :: tom : Cat (2)

0.6 :: Cat v Pet (3)

∃hasAnimal .Pet v NatureLover (4)

(kevin, fluffy) : hasAnimal (5)

(kevin, tom) : hasAnimal (6)

• Q = kevin : NatureLover has two explanations:

{ (1), (3) }{ (2), (3) }

• P(Q) = 0.4× 0.6× (1− 0.3) + 0.3× 0.6 = 0.348

• P(Q) can be computed with the inference system BUNDLE

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 5 / 21

Page 6: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics Probabilistic Inductive Logic Programming

???

What does structure learning mean (in this context)?

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 6 / 21

Page 7: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics Probabilistic Inductive Logic Programming

Reasoning task

In the field of Probabilistic Inductive Logic Programming the reasoningtask is composed by three main issues:

1 Inference: we want to compute the probability of a query

2 Parameter Learning: given a structure (the logical formulas), wewant to know the parameters (weights) of the logical formulas

3 Structure Learning: we want to learn both the structure and theparameters

• We are here!

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Page 8: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics Probabilistic Inductive Logic Programming

Reasoning task

In the field of Probabilistic Inductive Logic Programming the reasoningtask is composed by three main issues:

1 Inference: we want to compute the probability of a query

2 Parameter Learning: given a structure (the logical formulas), wewant to know the parameters (weights) of the logical formulas

3 Structure Learning: we want to learn both the structure and theparameters

• We are here!

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Page 9: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics Probabilistic Inductive Logic Programming

Reasoning task

In the field of Probabilistic Inductive Logic Programming the reasoningtask is composed by three main issues:

1 Inference: we want to compute the probability of a query

2 Parameter Learning: given a structure (the logical formulas), wewant to know the parameters (weights) of the logical formulas

3 Structure Learning: we want to learn both the structure and theparameters

• We are here!

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Page 10: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics Probabilistic Inductive Logic Programming

Reasoning task

In the field of Probabilistic Inductive Logic Programming the reasoningtask is composed by three main issues:

1 Inference: we want to compute the probability of a query

2 Parameter Learning: given a structure (the logical formulas), wewant to know the parameters (weights) of the logical formulas

3 Structure Learning: we want to learn both the structure and theparameters

• We are here!

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Page 11: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics Probabilistic Inductive Logic Programming

Reasoning task

In the field of Probabilistic Inductive Logic Programming the reasoningtask is composed by three main issues:

1 Inference: we want to compute the probability of a query

2 Parameter Learning: given a structure (the logical formulas), wewant to know the parameters (weights) of the logical formulas

3 Structure Learning: we want to learn both the structure and theparameters • We are here!

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Page 12: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics Corresponding Systems

Systems and Algorithms

1 Inference• BUNDLE: a system for reasoning on DISPONTE KBs using Binary

Decision Diagrams (BDDs)

2 Parameter Learning• EDGE: an algorithm that learns the probabilistic parameters of a

DISPONTE KB• EDGEMR: distributed version of EDGE

3 Structure Learning• LEAP: structure learning algorithm for DISPONTE KBs• LEAPMR: distributed version of LEAP

• We are here!

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 8 / 21

Page 13: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Probabilistic Description Logics Corresponding Systems

Systems and Algorithms

1 Inference• BUNDLE: a system for reasoning on DISPONTE KBs using Binary

Decision Diagrams (BDDs)

2 Parameter Learning• EDGE: an algorithm that learns the probabilistic parameters of a

DISPONTE KB• EDGEMR: distributed version of EDGE

3 Structure Learning• LEAP: structure learning algorithm for DISPONTE KBs• LEAPMR: distributed version of LEAP • We are here!

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 8 / 21

Page 14: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

EDGE: Learning DISPONTE Probabilities

EDGE: Em over bDds for description loGics paramEterlearning

• Supervised parameter learning algorithm• Input: a DISPONTE theory, a set of examples that represent queries

• The queries are assertional axioms divided into:

positive examples information that we regard as true for which wewould like to get high probability;

negative examples information that we regard as false for which wewould like to get low probability.

• Compute the explanations of each example using BUNDLE• Enter the EM cycle

• Expectation• computes the expected counts by traversing twice the BDDs

• Maximization• computes maximum likelihood parameters from the expected counts

• The EM algorithm is guaranteed to find a local maximum of theprobability of the examples

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 9 / 21

Page 15: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Distributed Parameter Learning

EDGEMR: Distributed Parameter Learning by MapReduce

• Initialization: the data is replicated among all workers. Each workerparses its copy of the probabilistic knowledge base and stores it inmain memory.

• Query resolution: the examples are divided among all the workersEach worker generates its private subset of BDDs and keeps them inmemory. Two different scheduling techniques can be applied for thisoperation.

• Expectation-Maximization: during the Expectation step all theworkers traverse their BDDs and calculate their local set ofexpectations. Then the master gathers all the expectations from theslaves, aggregates them and performs Maximization

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 10 / 21

Page 16: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Distributed Parameter Learning

Scheduling TechniquesSingle-step scheduling

• N slaves, the master divides the total number of queries into N + 1chunks

• The master begins to compute its queries while, for each other chunkof queries, it sends the chunk to the corresponding slave.

• Then the master waits for the results from the slaves. When theslowest slave returns its results to the master, EDGEMR proceeds tothe EM cycle.

Master

Slave1 Slave2

q1 q2 q3 q4 q5 qn

Master

Slave1 Slave2

chunk1q1 q2 q3 q4 q5 qn

chunk2 chunk3

Master

Slave1 Slave2

qnq4

q1 q2 q3

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 11 / 21

Page 17: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Distributed Parameter Learning

Scheduling TechniquesDynamic scheduling

• Handling each query may require a different amount of time• Chunk dimension established by the user• At first each machine is assigned a chunk of query in order• Then if the master ends handling its chunk it just takes the next one,

instead, if a slave ends handling its chunk, it asks the master foranother one

• When all the queries are evaluated, EDGEMR starts the EM cycle

Master

Slave1 Slave2

q1 q2 q3 q4 q5 qn

Master

Slave1 Slave2

q1

q2 q3

q4 q5 qn

Master

Slave1 Slave2

q1

q3 q4q2

q5 qn

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 12 / 21

Page 18: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Class Expression Learning Definition

Class Expression Learning

Definition

• Given a knowledge base K• Given a Target class contained in K• Given RK(C ) a retrieval function of the reasoner that returns all the

instances of C

The class learning consists in finding an expression C such that:

RK(Target) = RK(C )

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 13 / 21

Page 19: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Class Expression Learning CELOE

CELOE: Class Expression Learning for OntologyEngineering

• Class Expression Learning for OntologyEngineering.

• Generates a set of Class Expressions Ci

ordered according to an heuristics.

• Provides a semi-automatic approach to addaxioms of the form Target ≡ Ci oTarget v Ci .

• Combination of a refinement operator anda search algorithm

• Top-down

• Inside DL-Learner project

Person Car Building . . .

. . .

. . .

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 14 / 21

Page 20: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning Put it all together!

Put it all together!

• Our recipe:• BUNDLE: for Inference

• EDGE: for parameter learning

• CELOE: for candidate axiomsgeneration

LEAP

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 15 / 21

Page 21: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning Put it all together!

Put it all together!

• Our recipe:• BUNDLE: for Inference

• EDGE: for parameter learning

• CELOE: for candidate axiomsgeneration

LEAP

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 15 / 21

Page 22: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning Put it all together!

Put it all together!

• Our recipe:• BUNDLE: for Inference

• EDGE: for parameter learning

• CELOE: for candidate axiomsgeneration

LEAP

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 15 / 21

Page 23: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning LEAP: Learning DISPONTE theories

LEAP: LEArning Probabilistic description logics

• Structure and parameter learner for probabilistic KBs underDISPONTE

• Exploits two algorithms:

CELOE (Class Expression Learning for Ontology Engineering)for learning the structure of a KB

EDGE for learning the parameters of a probabilistic KB

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 16 / 21

Page 24: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning LEAP: Learning DISPONTE theories

LEAP: algorithm

1 Input: a knowledge base K, a set of examples that represent queries

2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms

• For each axiom, the probability is initialized to the accuracy returnedby CELOE

3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C

1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the

new log-likelihood LL′

3 If LL′ > LL keep the axiom ax4 Else remove it

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21

Page 25: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning LEAP: Learning DISPONTE theories

LEAP: algorithm

1 Input: a knowledge base K, a set of examples that represent queries

2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms

• For each axiom, the probability is initialized to the accuracy returnedby CELOE

3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C

1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the

new log-likelihood LL′

3 If LL′ > LL keep the axiom ax4 Else remove it

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21

Page 26: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning LEAP: Learning DISPONTE theories

LEAP: algorithm

1 Input: a knowledge base K, a set of examples that represent queries

2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms

• For each axiom, the probability is initialized to the accuracy returnedby CELOE

3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C

1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the

new log-likelihood LL′

3 If LL′ > LL keep the axiom ax4 Else remove it

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21

Page 27: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning LEAP: Learning DISPONTE theories

LEAP: algorithm

1 Input: a knowledge base K, a set of examples that represent queries

2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms

• For each axiom, the probability is initialized to the accuracy returnedby CELOE

3 Run EDGE to compute the initial log-likelihood LL of K

4 For each ax ∈ C

1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the

new log-likelihood LL′

3 If LL′ > LL keep the axiom ax4 Else remove it

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21

Page 28: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning LEAP: Learning DISPONTE theories

LEAP: algorithm

1 Input: a knowledge base K, a set of examples that represent queries

2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms

• For each axiom, the probability is initialized to the accuracy returnedby CELOE

3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C

1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the

new log-likelihood LL′

3 If LL′ > LL keep the axiom ax4 Else remove it

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21

Page 29: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning LEAP: Learning DISPONTE theories

LEAP: algorithm

1 Input: a knowledge base K, a set of examples that represent queries

2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms

• For each axiom, the probability is initialized to the accuracy returnedby CELOE

3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C

1 Add the axiom ax to the knowledge base K

2 Run EDGE on the new KB to compute the new parameters and thenew log-likelihood LL′

3 If LL′ > LL keep the axiom ax4 Else remove it

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21

Page 30: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning LEAP: Learning DISPONTE theories

LEAP: algorithm

1 Input: a knowledge base K, a set of examples that represent queries

2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms

• For each axiom, the probability is initialized to the accuracy returnedby CELOE

3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C

1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the

new log-likelihood LL′

3 If LL′ > LL keep the axiom ax4 Else remove it

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21

Page 31: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning LEAP: Learning DISPONTE theories

LEAP: algorithm

1 Input: a knowledge base K, a set of examples that represent queries

2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms

• For each axiom, the probability is initialized to the accuracy returnedby CELOE

3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C

1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the

new log-likelihood LL′

3 If LL′ > LL keep the axiom ax

4 Else remove it

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21

Page 32: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Structure Learning LEAP: Learning DISPONTE theories

LEAP: algorithm

1 Input: a knowledge base K, a set of examples that represent queries

2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms

• For each axiom, the probability is initialized to the accuracy returnedby CELOE

3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C

1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the

new log-likelihood LL′

3 If LL′ > LL keep the axiom ax4 Else remove it

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21

Page 33: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Distributed Structure Learning

LEAPMR: Structure Learning with Distributed ParameterLearning

• Evolution of LEAP

• It adapts LEAP to perform EDGEMR

• It exploits:• CELOE for learning the structure of a KB• EDGEMR for learning the parameters of a probabilistic KB

• Same algorithm of LEAP, but it always uses EDGEMR instead ofEDGE

• The execution of CELOE is not distributed

• Future work• Distributing the scoring of the refinements generated during the

execution of CELOE

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 18 / 21

Page 34: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Experiments and Results

Experiments

• LEAPMR has been implemented in Java

• Cluster of 64-bit Linux machines with 8-cores Intel Haswell 2.40 GHzCPUs and 2 GB (max) Java memory allocated per node

• Dataset: Moral• 202 individuals and 4710 axioms (22 axioms are probabilistic)

• The experiments were performed on 1 (LEAP), 3, 5, 9 and 17 nodes

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 19 / 21

Page 35: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Experiments and Results

Results

• Speedup of LEAPMR relative to LEAP for Moral KB:

3 5 9 172

4

6

8

10

12

N. of Nodes

Speedup

• Speedup is significant but sublinear

• Sublinearity caused by MPI communication overhead

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 20 / 21

Page 36: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Experiments and Results

Thank you for listening!

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 21 / 21

Page 37: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21

Experiments and Results

Thank you for listening!

Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 21 / 21