490 likes | 756 Vues
Markov Logic Networks. Speaker: Benedict Fehringer Seminar: Probabilistic Models for Information Extraction by Dr. Martin Theobald and Maximilian Dylla Based on Richards, M., and Domingos , P. (2006). Outline. Part 1: Why we need Markov Logic Networks (MLN’s) Markov Networks
E N D
Markov Logic Networks Speaker: Benedict Fehringer Seminar: Probabilistic Models for Information Extraction by Dr. Martin Theobald and Maximilian Dylla Based on Richards, M., and Domingos, P. (2006) Markov Logic Networks
Outline • Part 1: Why we need Markov Logic Networks (MLN’s) • Markov Networks • First-Order Logic • Conclusion and Motivation • Part 2: How the MLN’s work? • Part 3: Are they better than other methods? Markov Logic Networks
Outline • Part 1: Why we need Markov Logic Networks (MLN’s) • Markov Networks • First-Order Logic • Conclusion and Motivation • Part 2: How the MLN’s work? • Part 3: Are they better than other methods? Markov Logic Networks
Markov Networks • Set of variables: • The distribution is given by: with as normalization factor and as potential function Markov Logic Networks
Markov Networks • Representation as log-linear model: • In our case there will be only binary features: • Each feature corresponds to each possible state • The weight is equal to the log of the potential: Markov Logic Networks
Little example scenario There is a company witch has a Playstation and each company’s employee has the right to play with it. If two employees are friends then the probability is high (ω=3) that both play all day long or both do not. If someone plays all day long then the chance for her/him is high (ω=2) to get fired. Markov Logic Networks
Little example scenario There is a company witch has a Playstation and each company’s employee has the right to play with it. If two employees are friends then the probability is high (ω=3) that both play all day long or both do not. If someone plays all day long then the chance for her/him is high (ω=2) to get fired. Markov Logic Networks
Markov Networks Has playing Friend Plays Fired One possibility… Markov Logic Networks
Markov Networks Is Friend with Some playing Employee Plays Fired And another one. Markov Logic Networks
Little example scenario There is a company witch has a Playstation and each company’s employee has the right to play with it. If two employees are friends then the probability is high (ω=3) that both play all day long or both do not. If someone plays all day long then the chance for her/him is high (ω=2) to get fired. If an employee A can convince another employee B to play, depends on the lability of B. For a high lability of B there is a higher probability (ω=4) than for a low lability (ω=2) MarkovLogic Networks
Markov Networks Is labile Is Friend with Playing Person Plays Fired Could be. Markov Logic Networks
Markov Networks • Advantages: • Efficiently handling uncertainty • Tolerant against imperfection and contradictory knowledge • Disadvantages: • Very complex networks for a wide variety of knowledge • Difficult to incorporate a wide range of domain knowledge Markov Logic Networks
Outline • Part 1: Why we need Markov Logic Networks (MLN’s) • Markov Networks • First-Order Logic • Conclusion and Motivation • Part 2: How the MLN’s work? • Part 3: Are they better than other methods? Markov Logic Networks
First-Order Logic • Four types of symbols: Constants: concrete object in the domain (e.g., people: Anna, Bob) Variables: range over the objects in the domain Functions: Mapping from tuples of objects to objects (e.g., GrandpaOf) Predicates: relations among objects in the domain (e.g., Friends) or attributes of objects (e.g. Fired) • Term: Any expression representing an object in the domain. Consisting of a constant, a variable, or a function applied to a tuple of terms. • Atomic formula or atom: predicate applied to a tuple of terms • Logical connectives and quantifier: Markov Logic Networks
Translation in First-Order Logic If two employees are friends then the probability is high that both play all day long or both do not. In Clausal Form: If someone plays all day long then the chance for her/him is high to get fired. In Clausal Form: Markov Logic Networks
First-Order Logic • Advantages: • Compact representation a wide variety of knowledge • Flexible and modularly incorporate a wide range of domain • knowledge • Disadvantages: • No possibility to handle uncertainty • No handling of imperfection and contradictory knowledge Markov Logic Networks
Outline • Part 1: Why we need Markov Logic Networks (MLN’s) • Markov Networks • First-Order Logic • Conclusion and Motivation • Part 2: How the MLN’s work? • Part 3: Are they better than other methods? Markov Logic Networks
Conclusion and Motivation Markov Networks First-Order Logic Efficiently handling uncertainty Tolerant against imperfection and contradictory knowledge Compact representation a wide variety of knowledge Flexible and modularly incorporate a wide range of domain knowledge → Combination of Markov Networks and First-Order Logic to use the advantages of both Markov Logic Networks
Outline • Part 1: Why we need Markov Logic Networks (MLN’s) • Markov Networks • First-Order Logic • Conclusion and Motivation • Part 2: How the MLN’s work? • Part 3: Are they better than other methods? Markov Logic Networks
Markov Logic Network Description of the problem Translation in First-Order Logic Construction of a MLN-”Template” Compute whatever you want Derive a concrete MLN for a given Set of Constants Markov Logic Networks
Markov Logic Network Description of the problem Translation in First-Order Logic Construction of a MLN-”Template” Compute whatever you want Derive a concrete MLN for a given Set of Constants Markov Logic Networks
Markov Logic Network- Translation in First-Order Logic - If two employees are friends then the probability is high that both play all day long or both do not. In Clausal Form: If someone plays all day long then the chance for her/him is high to get fired. In Clausal Form: Markov Logic Networks
Markov Logic Network Description of the problem Translation in First-Order Logic Construction of a MLN-”Template” Compute whatever you want Derive a concrete MLN for a given Set of Constants Markov Logic Networks
Markov Logic Network • Each formula matches one clique • Each formula owns a weight that reflects the importance of this formula • If a world violates one formula then it is less probable but not impossible • Concrete: The weight of this formula will be ignored (that means the weight is 0) Markov Logic Networks
Markov Logic Network Formula to compare: Three Assumptions: • Unique Names • Domain closure • Known functions Markov Logic Networks
Markov Logic Network Description of the problem Translation in First-Order Logic Construction of a MLN-”Template” Compute whatever you want Derive a concrete MLN for a given Set of Constants Markov Logic Networks
Markov Logic Network Grounding: (with Constants c1 and c2) => => => Elimination of the existential quantifier Elimination of the universalquantifier Elimination of the functions MarkovLogic Networks
Markov Networks Friends(A,B) Friends(A,A) Plays(A) Plays(B) Friends(B,B) Friends(B,A) Constants: Alice (A) and Bob (B) Markov Logic Networks
Markov Logic Network Friends(A,B) Friends(A,A) Plays(A) Plays(A) Plays(B) Plays(B) Friends(B,B) Friends(B,A) Fired(A) Fired(B) Constants: Alice (A) and Bob (B) Markov Logic Networks
Markov Logic Network Friends(A,B) Friends(A,A) Plays(A) Plays(B) Friends(B,B) Friends(B,A) Fired(A) Fired(B) Markov Logic Networks
Markov Logic Network Description of the problem Translation in First-Order Logic Construction of a MLN-”Template” Compute whatever you want Derive a concrete MLN for a given Set of Constants Markov Logic Networks
Markov Logic Network What is the probability that Alice and Bob are friends, both play playstation all day long but both are not getting fired? Friends(A,B)=1 Friends(A,A)=1 Plays(A)=1 Plays(B)=1 Friends(B,B)=1 Friends(B,A)=1 Fired(A)=0 Fired(B)=0 Markov Logic Networks
Markov Logic Network What happens if limω→∞? If all formulas fulfilled: If not all formulas fulfilled: => Markov Logic Networks
Markov Logic Network What happens if limω→∞? If all formulas fulfilled: Markov Logic Networks
Markov Logic Network What happens if limω→∞? If not all formulas fulfilled: Markov Logic Networks
Markov Logic Network What is the probability that a formula F1 holds given that formula F2 does? Markov Logic Networks
Markov Logic Network Learning the weights: It is #P-complete to count the number of true groundings => approximation is necessary => using the pseudo-likelihood: Markov Logic Networks
Outline • Part 1: Why we need Markov Logic Networks (MLN’s) • Markov Networks • First-Order Logic • Conclusion and Motivation • Part 2: How the MLN’s work? • Part 3: Are they better than other methods? Markov Logic Networks
Experiment I Setting: • Using a database describing the Department of Computer Science and Engineering at the University of Washington • 12 predicates (e.g. Professor, Student, Area, AdvisedBy, …) • 2707 constants • 96 formulas (Knowledge base was provided by four volunteers who not know the database but were member of the department) • The whole database was divided into five subsets for each area (AI, graphics, programming languages, systems, theory) => in the end 521 true ground atoms of possible 58,457 Markov Logic Networks
Experiment II Testing: • leave-one-out over the areas • Prediction of AdvisedBy(x,y) • Either with all or only partial (except Student(x) and Professor(x)) information • Drawing the precision/recall curves • Computation of the area under the curve (AUC) Markov Logic Networks
Experiment III MLN was compared with: • Logic (only logical KB without probability) • Probability (only probability relations without special knowledge representations) • Naïve Bayes (NB) and Bayesian Network (BN) • Inductive logic programming (automatically development of the KB) • CLAUDIEN (clausal discovery engine) Markov Logic Networks
Results Markov Logic Networks
Results I all Areas AI Area Markov Logic Networks
Results II graphics area prog. Language Area MarkovLogic Networks
Results III systems area theory Area Markov Logic Networks
Sample applications • Link Prediction • Link-Based Clustering • Social Network Modeling • … Markov Logic Networks
Conclusion • MLN’s are a simple way to combine first-order logic and probability • They can be seen as a template for construction ordinary Markov Networks • Clauses can be learned by CLAUDIEN • Empirical tests with real-world data and knowledge are promising for the use of the MLN’s Markov Logic Networks
Literature Richardson, M., & Domingos, P. (2006). Markov logic networks. Machine Learning Journal, 62, 107-136. Markov Logic Networks