structure learning with distributed parameter learning for...
TRANSCRIPT
![Page 1: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/1.jpg)
Structure Learning with Distributed Parameter Learningfor Probabilistic Ontologies
Giuseppe Cota1 Riccardo Zese1 Elena Bellodi1
Fabrizio Riguzzi2 Evelina Lamma1
Dipartimento di Ingegneria – University of Ferrara
Dipartimento di Matematica e Informatica – University of Ferrara[giuseppe.cota,riccardo.zese,elena.bellodi,fabrizio.riguzzi,
evelina.lamma]@unife.it
September 23, 2015
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 1 / 21
![Page 2: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/2.jpg)
Outline
1 Probabilistic Description LogicsIntroductionDISPONTEProbabilistic Inductive Logic ProgrammingCorresponding Systems
2 EDGE: Learning DISPONTE Probabilities
3 Distributed Parameter Learning
4 Class Expression LearningDefinitionCELOE
5 Structure LearningPut it all together!LEAP: Learning DISPONTE theories
6 Distributed Structure Learning
7 Experiments and Results
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 2 / 21
![Page 3: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/3.jpg)
Probabilistic Description Logics Introduction
Probabilistic Description Logics: Introduction
• Basic Concepts:• Ontology: a formal and explicit description of a domain of interest• Description Logics (DLs): family of logical formalism (decidable)
fragments of first order logic• Knowledge Base: set of axioms expressed by a DL
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 3 / 21
![Page 4: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/4.jpg)
Probabilistic Description Logics DISPONTE
The semantics: DISPONTE
• In real world domains the information is often uncertain
• The Distribution Semantics underlies many probabilistic logicprogramming languages, such as PRISM, ICL, LPADs, ProbLog
• DISPONTE applies the Distribution Semantics of probabilisticlogic programming to description logics
• DISPONTE semantics for probabilistic SROIQ(D).• Probabilistic axioms: p :: E
e.g., p :: C v D represents the fact that we believe in the truth ofC v D with probability p.
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 4 / 21
![Page 5: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/5.jpg)
Probabilistic Description Logics DISPONTE
Example
0.4 :: fluffy : Cat (1)
0.3 :: tom : Cat (2)
0.6 :: Cat v Pet (3)
∃hasAnimal .Pet v NatureLover (4)
(kevin, fluffy) : hasAnimal (5)
(kevin, tom) : hasAnimal (6)
• Q = kevin : NatureLover has two explanations:
{ (1), (3) }{ (2), (3) }
• P(Q) = 0.4× 0.6× (1− 0.3) + 0.3× 0.6 = 0.348
• P(Q) can be computed with the inference system BUNDLE
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 5 / 21
![Page 6: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/6.jpg)
Probabilistic Description Logics Probabilistic Inductive Logic Programming
???
What does structure learning mean (in this context)?
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 6 / 21
![Page 7: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/7.jpg)
Probabilistic Description Logics Probabilistic Inductive Logic Programming
Reasoning task
In the field of Probabilistic Inductive Logic Programming the reasoningtask is composed by three main issues:
1 Inference: we want to compute the probability of a query
2 Parameter Learning: given a structure (the logical formulas), wewant to know the parameters (weights) of the logical formulas
3 Structure Learning: we want to learn both the structure and theparameters
• We are here!
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21
![Page 8: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/8.jpg)
Probabilistic Description Logics Probabilistic Inductive Logic Programming
Reasoning task
In the field of Probabilistic Inductive Logic Programming the reasoningtask is composed by three main issues:
1 Inference: we want to compute the probability of a query
2 Parameter Learning: given a structure (the logical formulas), wewant to know the parameters (weights) of the logical formulas
3 Structure Learning: we want to learn both the structure and theparameters
• We are here!
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21
![Page 9: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/9.jpg)
Probabilistic Description Logics Probabilistic Inductive Logic Programming
Reasoning task
In the field of Probabilistic Inductive Logic Programming the reasoningtask is composed by three main issues:
1 Inference: we want to compute the probability of a query
2 Parameter Learning: given a structure (the logical formulas), wewant to know the parameters (weights) of the logical formulas
3 Structure Learning: we want to learn both the structure and theparameters
• We are here!
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21
![Page 10: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/10.jpg)
Probabilistic Description Logics Probabilistic Inductive Logic Programming
Reasoning task
In the field of Probabilistic Inductive Logic Programming the reasoningtask is composed by three main issues:
1 Inference: we want to compute the probability of a query
2 Parameter Learning: given a structure (the logical formulas), wewant to know the parameters (weights) of the logical formulas
3 Structure Learning: we want to learn both the structure and theparameters
• We are here!
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21
![Page 11: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/11.jpg)
Probabilistic Description Logics Probabilistic Inductive Logic Programming
Reasoning task
In the field of Probabilistic Inductive Logic Programming the reasoningtask is composed by three main issues:
1 Inference: we want to compute the probability of a query
2 Parameter Learning: given a structure (the logical formulas), wewant to know the parameters (weights) of the logical formulas
3 Structure Learning: we want to learn both the structure and theparameters • We are here!
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21
![Page 12: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/12.jpg)
Probabilistic Description Logics Corresponding Systems
Systems and Algorithms
1 Inference• BUNDLE: a system for reasoning on DISPONTE KBs using Binary
Decision Diagrams (BDDs)
2 Parameter Learning• EDGE: an algorithm that learns the probabilistic parameters of a
DISPONTE KB• EDGEMR: distributed version of EDGE
3 Structure Learning• LEAP: structure learning algorithm for DISPONTE KBs• LEAPMR: distributed version of LEAP
• We are here!
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 8 / 21
![Page 13: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/13.jpg)
Probabilistic Description Logics Corresponding Systems
Systems and Algorithms
1 Inference• BUNDLE: a system for reasoning on DISPONTE KBs using Binary
Decision Diagrams (BDDs)
2 Parameter Learning• EDGE: an algorithm that learns the probabilistic parameters of a
DISPONTE KB• EDGEMR: distributed version of EDGE
3 Structure Learning• LEAP: structure learning algorithm for DISPONTE KBs• LEAPMR: distributed version of LEAP • We are here!
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 8 / 21
![Page 14: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/14.jpg)
EDGE: Learning DISPONTE Probabilities
EDGE: Em over bDds for description loGics paramEterlearning
• Supervised parameter learning algorithm• Input: a DISPONTE theory, a set of examples that represent queries
• The queries are assertional axioms divided into:
positive examples information that we regard as true for which wewould like to get high probability;
negative examples information that we regard as false for which wewould like to get low probability.
• Compute the explanations of each example using BUNDLE• Enter the EM cycle
• Expectation• computes the expected counts by traversing twice the BDDs
• Maximization• computes maximum likelihood parameters from the expected counts
• The EM algorithm is guaranteed to find a local maximum of theprobability of the examples
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 9 / 21
![Page 15: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/15.jpg)
Distributed Parameter Learning
EDGEMR: Distributed Parameter Learning by MapReduce
• Initialization: the data is replicated among all workers. Each workerparses its copy of the probabilistic knowledge base and stores it inmain memory.
• Query resolution: the examples are divided among all the workersEach worker generates its private subset of BDDs and keeps them inmemory. Two different scheduling techniques can be applied for thisoperation.
• Expectation-Maximization: during the Expectation step all theworkers traverse their BDDs and calculate their local set ofexpectations. Then the master gathers all the expectations from theslaves, aggregates them and performs Maximization
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 10 / 21
![Page 16: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/16.jpg)
Distributed Parameter Learning
Scheduling TechniquesSingle-step scheduling
• N slaves, the master divides the total number of queries into N + 1chunks
• The master begins to compute its queries while, for each other chunkof queries, it sends the chunk to the corresponding slave.
• Then the master waits for the results from the slaves. When theslowest slave returns its results to the master, EDGEMR proceeds tothe EM cycle.
Master
Slave1 Slave2
q1 q2 q3 q4 q5 qn
Master
Slave1 Slave2
chunk1q1 q2 q3 q4 q5 qn
chunk2 chunk3
Master
Slave1 Slave2
qnq4
q1 q2 q3
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 11 / 21
![Page 17: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/17.jpg)
Distributed Parameter Learning
Scheduling TechniquesDynamic scheduling
• Handling each query may require a different amount of time• Chunk dimension established by the user• At first each machine is assigned a chunk of query in order• Then if the master ends handling its chunk it just takes the next one,
instead, if a slave ends handling its chunk, it asks the master foranother one
• When all the queries are evaluated, EDGEMR starts the EM cycle
Master
Slave1 Slave2
q1 q2 q3 q4 q5 qn
Master
Slave1 Slave2
q1
q2 q3
q4 q5 qn
Master
Slave1 Slave2
q1
q3 q4q2
q5 qn
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 12 / 21
![Page 18: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/18.jpg)
Class Expression Learning Definition
Class Expression Learning
Definition
• Given a knowledge base K• Given a Target class contained in K• Given RK(C ) a retrieval function of the reasoner that returns all the
instances of C
The class learning consists in finding an expression C such that:
RK(Target) = RK(C )
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 13 / 21
![Page 19: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/19.jpg)
Class Expression Learning CELOE
CELOE: Class Expression Learning for OntologyEngineering
• Class Expression Learning for OntologyEngineering.
• Generates a set of Class Expressions Ci
ordered according to an heuristics.
• Provides a semi-automatic approach to addaxioms of the form Target ≡ Ci oTarget v Ci .
• Combination of a refinement operator anda search algorithm
• Top-down
• Inside DL-Learner project
Person Car Building . . .
. . .
. . .
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 14 / 21
![Page 20: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/20.jpg)
Structure Learning Put it all together!
Put it all together!
• Our recipe:• BUNDLE: for Inference
• EDGE: for parameter learning
• CELOE: for candidate axiomsgeneration
⇓
LEAP
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 15 / 21
![Page 21: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/21.jpg)
Structure Learning Put it all together!
Put it all together!
• Our recipe:• BUNDLE: for Inference
• EDGE: for parameter learning
• CELOE: for candidate axiomsgeneration
⇓
LEAP
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 15 / 21
![Page 22: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/22.jpg)
Structure Learning Put it all together!
Put it all together!
• Our recipe:• BUNDLE: for Inference
• EDGE: for parameter learning
• CELOE: for candidate axiomsgeneration
⇓
LEAP
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 15 / 21
![Page 23: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/23.jpg)
Structure Learning LEAP: Learning DISPONTE theories
LEAP: LEArning Probabilistic description logics
• Structure and parameter learner for probabilistic KBs underDISPONTE
• Exploits two algorithms:
CELOE (Class Expression Learning for Ontology Engineering)for learning the structure of a KB
EDGE for learning the parameters of a probabilistic KB
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 16 / 21
![Page 24: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/24.jpg)
Structure Learning LEAP: Learning DISPONTE theories
LEAP: algorithm
1 Input: a knowledge base K, a set of examples that represent queries
2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms
• For each axiom, the probability is initialized to the accuracy returnedby CELOE
3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C
1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the
new log-likelihood LL′
3 If LL′ > LL keep the axiom ax4 Else remove it
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21
![Page 25: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/25.jpg)
Structure Learning LEAP: Learning DISPONTE theories
LEAP: algorithm
1 Input: a knowledge base K, a set of examples that represent queries
2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms
• For each axiom, the probability is initialized to the accuracy returnedby CELOE
3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C
1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the
new log-likelihood LL′
3 If LL′ > LL keep the axiom ax4 Else remove it
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21
![Page 26: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/26.jpg)
Structure Learning LEAP: Learning DISPONTE theories
LEAP: algorithm
1 Input: a knowledge base K, a set of examples that represent queries
2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms
• For each axiom, the probability is initialized to the accuracy returnedby CELOE
3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C
1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the
new log-likelihood LL′
3 If LL′ > LL keep the axiom ax4 Else remove it
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21
![Page 27: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/27.jpg)
Structure Learning LEAP: Learning DISPONTE theories
LEAP: algorithm
1 Input: a knowledge base K, a set of examples that represent queries
2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms
• For each axiom, the probability is initialized to the accuracy returnedby CELOE
3 Run EDGE to compute the initial log-likelihood LL of K
4 For each ax ∈ C
1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the
new log-likelihood LL′
3 If LL′ > LL keep the axiom ax4 Else remove it
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21
![Page 28: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/28.jpg)
Structure Learning LEAP: Learning DISPONTE theories
LEAP: algorithm
1 Input: a knowledge base K, a set of examples that represent queries
2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms
• For each axiom, the probability is initialized to the accuracy returnedby CELOE
3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C
1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the
new log-likelihood LL′
3 If LL′ > LL keep the axiom ax4 Else remove it
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21
![Page 29: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/29.jpg)
Structure Learning LEAP: Learning DISPONTE theories
LEAP: algorithm
1 Input: a knowledge base K, a set of examples that represent queries
2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms
• For each axiom, the probability is initialized to the accuracy returnedby CELOE
3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C
1 Add the axiom ax to the knowledge base K
2 Run EDGE on the new KB to compute the new parameters and thenew log-likelihood LL′
3 If LL′ > LL keep the axiom ax4 Else remove it
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21
![Page 30: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/30.jpg)
Structure Learning LEAP: Learning DISPONTE theories
LEAP: algorithm
1 Input: a knowledge base K, a set of examples that represent queries
2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms
• For each axiom, the probability is initialized to the accuracy returnedby CELOE
3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C
1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the
new log-likelihood LL′
3 If LL′ > LL keep the axiom ax4 Else remove it
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21
![Page 31: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/31.jpg)
Structure Learning LEAP: Learning DISPONTE theories
LEAP: algorithm
1 Input: a knowledge base K, a set of examples that represent queries
2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms
• For each axiom, the probability is initialized to the accuracy returnedby CELOE
3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C
1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the
new log-likelihood LL′
3 If LL′ > LL keep the axiom ax
4 Else remove it
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21
![Page 32: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/32.jpg)
Structure Learning LEAP: Learning DISPONTE theories
LEAP: algorithm
1 Input: a knowledge base K, a set of examples that represent queries
2 Run CELOE to obtain a set C of candidate probabilistic subclass-ofaxioms
• For each axiom, the probability is initialized to the accuracy returnedby CELOE
3 Run EDGE to compute the initial log-likelihood LL of K4 For each ax ∈ C
1 Add the axiom ax to the knowledge base K2 Run EDGE on the new KB to compute the new parameters and the
new log-likelihood LL′
3 If LL′ > LL keep the axiom ax4 Else remove it
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 17 / 21
![Page 33: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/33.jpg)
Distributed Structure Learning
LEAPMR: Structure Learning with Distributed ParameterLearning
• Evolution of LEAP
• It adapts LEAP to perform EDGEMR
• It exploits:• CELOE for learning the structure of a KB• EDGEMR for learning the parameters of a probabilistic KB
• Same algorithm of LEAP, but it always uses EDGEMR instead ofEDGE
• The execution of CELOE is not distributed
• Future work• Distributing the scoring of the refinements generated during the
execution of CELOE
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 18 / 21
![Page 34: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/34.jpg)
Experiments and Results
Experiments
• LEAPMR has been implemented in Java
• Cluster of 64-bit Linux machines with 8-cores Intel Haswell 2.40 GHzCPUs and 2 GB (max) Java memory allocated per node
• Dataset: Moral• 202 individuals and 4710 axioms (22 axioms are probabilistic)
• The experiments were performed on 1 (LEAP), 3, 5, 9 and 17 nodes
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 19 / 21
![Page 35: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/35.jpg)
Experiments and Results
Results
• Speedup of LEAPMR relative to LEAP for Moral KB:
3 5 9 172
4
6
8
10
12
N. of Nodes
Speedup
• Speedup is significant but sublinear
• Sublinearity caused by MPI communication overhead
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 20 / 21
![Page 36: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/36.jpg)
Experiments and Results
Thank you for listening!
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 21 / 21
![Page 37: Structure Learning with Distributed Parameter Learning for ...aixia2015.unife.it/wp-content/uploads/slides/DC/DC_2.pdf · Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 7 / 21](https://reader034.vdocuments.site/reader034/viewer/2022043002/5f800fd065f136004130a386/html5/thumbnails/37.jpg)
Experiments and Results
Thank you for listening!
Cota, Zese et al. (UNIFE) LEAPMR September 23, 2015 21 / 21