DeepDive Model Dongfang Xu Ph.D student, School of Information, University of Arizona Dec 13, 2015.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Bayesian Network and Influence Diagram A Guide to Construction And Analysis.
Social networks, in the form of bibliographies and citations, have long been an integral part of the scientific process. We examine how to leverage the.
UML (Sequence Diagrams, Collaboration and State Chart Diagrams) Presentation By - SANDEEP REDDY CHEEDEPUDI (Student No: ) - VISHNU CHANDRADAS (Student.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
An Approach to Evaluate Data Trustworthiness Based on Data Provenance Department of Computer Science Purdue University.
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Interactive Generation of Integrated Schemas Laura Chiticariu et al. Presented by: Meher Talat Shaikh.
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graph.
Report on Intrusion Detection and Data Fusion By Ganesh Godavari.
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
Learning with Bayesian Networks David Heckerman Presented by Colin Rickert.
1 Department of Computer Science and Engineering, University of South Carolina Issues for Discussion and Work Jan 2007  Choose meeting time.
1 LM Approaches to Filtering Richard Schwartz, BBN LM/IR ARDA 2002 September 11-12, 2002 UMASS.
Ranking by Odds Ratio A Probability Model Approach let be a Boolean random variable: document d is relevant to query q otherwise Consider document d as.
Design of an Intrusion Response System using Evolutionary Computation Rohit Parti.
Scalable Text Mining with Sparse Generative Models
Building Knowledge-Driven DSS and Mining Data
1 Learning with Bayesian Networks Author: David Heckerman Presented by Yan Zhang April
Introduction to Machine Learning Approach Lecture 5.
Correlation Nabaz N. Jabbar Near East University 25 Oct 2011.
Bayesian Decision Theory Making Decisions Under uncertainty 1.
Modeling and Finding Abnormal Nodes (chapter 2) 駱宏毅 Hung-Yi Lo Social Network Mining Lab Seminar July 18, 2007.
EVALUATION David Kauchak CS 451 – Fall Admin Assignment 3 - change constructor to take zero parameters - instead, in the train method, call getFeatureIndices()
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
Modeling (Chap. 2) Modern Information Retrieval Spring 2000.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
A Markov Random Field Model for Term Dependencies Donald Metzler W. Bruce Croft Present by Chia-Hao Lee.
Agenda Introduction Overview of White-box testing Basis path testing
Report on Intrusion Detection and Data Fusion By Ganesh Godavari.
Geo597 Geostatistics Ch9 Random Function Models.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Tuffy Scaling up Statistical Inference in Markov Logic using an RDBMS
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Lesson Overview Lesson Overview What Is Science? Lesson Overview 1.1 What Is Science?
The Expert System Shell SPIRIT Presented by Poom Samaharn I-MMIS
LAC group, 16/06/2011. So far...  Directed graphical models  Bayesian Networks Useful because both the structure and the parameters provide a natural.
Sampling distributions rule of thumb…. Some important points about sample distributions… If we obtain a sample that meets the rules of thumb, then…
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Learning to Share Meaning in a Multi-Agent System (Part I) Ganesh Padmanabhan.
DeepDive Introduction Dongfang Xu Ph.D student, School of Information, University of Arizona Sept 10, 2015.
Post-Ranking query suggestion by diversifying search Chao Wang.
Lesson Overview Lesson Overview What Is Science? Lesson Overview 1.1 What Is Science?
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
An Ontological Approach to Financial Analysis and Monitoring.
1 An infrastructure for context-awareness based on first order logic 송지수 ISI LAB.
Introduction on Graphic Models
DeepDive Case Study Dongfang Xu School of Information.
Pattern Recognition. What is Pattern Recognition? Pattern recognition is a sub-topic of machine learning. PR is the science that concerns the description.
Artificial Intelligence Knowledge Representation.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
An Introduction to Markov Logic Networks in Knowledge Bases
A Brief Introduction to Distant Supervision
Markov Logic Networks for NLP CSCI-GA.2591
Associative Query Answering via Query Feature Similarity
Factor Graph in DeepDive
Lecture 7: Knowledge Base Construction: From dark data to insights
Deep Belief Nets and Ising Model-Based Network Construction
Causal Models Lecture 12.
Learning Probabilistic Graphical Models Overview Learning Problems.
Discriminative Probabilistic Models for Relational Data
Probabilistic Databases with MarkoViews
Sanguthevar Rajasekaran University of Connecticut
Presentation transcript:

DeepDive Model Dongfang Xu Ph.D student, School of Information, University of Arizona Dec 13, 2015

Agenda Overview Factor Graph Learning &Inference Reference DeepDive: A Data Management System for Automatic Knowledge Base Construction. Ce Zhang.Ph.D. Dissertation, University of Wisconsin-Madison, 2015.

Overview What is Deep Dive? DeepDive is a new type of data management system that enables one to tackle extraction, integration, and prediction problems in a single system. DeepDive makes good use of uncertainty to improve predictions during the probabilistic inference step. For example, DeepDive may find a certain mention of "Barack" is only 60% likely to actually refer to "Barack Obama", and use this fact to discount the impact of that mention on the final result for the entity "Barack Obama")

Overview

What users/developers need do? 1. Generation and Extraction ---Users schema and Correlation schema (correlation schema captures correlations among tuples in the user schema)Users schema ---Extraction Features. (100K features) With weighted value. 2. Distant Supervision ---One way is for user to create training data. ---Take the real fact, and then label each training data true or false. 3. Inference and Learning

Overview  User Schema

Overview What system will do? 1.Generation and Extraction ---Extracted mentions (entity), candidate relation( based on features and supervise rule) ---Entity linking (Sometimes has glossaries or a database of all known entities; Sometimes need sophisticated machine learning approaches, with weighted value and boolean value) --- Label some of these pairs as true or false according to the supervision rules. (with weighted value and boolean value)

Overview What system will do? 1.Generation and Extraction 2. Distant Supervision --- Make use of an already existing database to collect examples for the relation. --- Use these examples to automatically generate our training data, including positive and negative training data (harder process).

Agenda Overview Factor Graph Learning &Inference Reference DeepDive: A Data Management System for Automatic Knowledge Base Construction. Ce Zhang.Ph.D. Dissertation, University of Wisconsin-Madison, 2015.

Learning and Inference step 1.Factor graph grounding DeepDive heavily relies on factor graphs, one type of probabilistic graphical models, for its statistical inference and learning phase. A Factor graph has two types of nodes: Variable notes and factor Notes. Factor Graph

Learning and Inference step Both the features extracted and domain knowledge (inference rule In factor from a factor graph) integrated need a weight to indicate how strong an indicator they are to the target task. ---One way to do that is for the user to manually specify the weight. ---another more easy, consistent, and effective way is for DeepDive to automatically learn the weight with machine learning techniques. (Through an iterative way)

Learning and Inference step 1.Factor graph grounding ---Variables, which can be used to quantitatively describe an event. Specifically, describe the tuple in users schema. The variables can be evidence variables when their value is known (from training data or user defined), or query variables when their value should be predicted. ---Factor (correlation relation, from correlation schema), is a function of variables, and is used to evaluate the relations among variable(s). The main task that DeepDive conducts on factor graphs is statistical inference, i.e., for a given node, what is the marginal probability that this node takes the value 1? Factor Graph

Learning and Inference step 1.Factor graph grounding The variable nodes of the factor graph are connected to factors according to inference rules specified by the user, who also defines the factor functions which describe how the variables are related. The user can specify whether the factor weights should be constant or learned by the system. Inference rules are edges in graph. Each rule consists of three components: The input query specifies the variables to create (variable notes); The factor function (factor notes); The factor weight describes the confidence in the relationship expressed by the factor. Factor Graph

Agenda Overview Factor Graph Learning &Inference

Learning and Inference step 2. How it works? A Each variable can take value 0 or 1, and let’s say there are two variables. So we have four possible worlds (a combination of varaible(s)). B Define the probability of a possible world through factor functions. We give different weight to factor functions, to express the relative influence of each factor on the probability Learning &Inference Pr(I) ∝ measure{w1f1(v1, v2) + w2f2(v2)}.

Learning and Inference step 2. How it works? B + The probability of a possible world graph is then defined to be proportional to some measure of weighted combination of factor functions. C Now, we can perform marginal inference on factor graphs of one variable taking a particular value. A marginal inference is to infer the probability of one variable taking a particular value. This is similar to marginal probability and joint probability. Learning &Inference

Learning and Inference step 2. How it works? In DeepDive, you can assign factor weights manually, or you can let DeepDive learn weights automatically. In order to learn weights automatically, you must have enough training data available. DeepDive chooses weights that agree most with the training data. Formally, the training data is just set of possible worlds, and we choose weights by maximizing the probabilities of these possible worlds. Learning &Inference

DeepDive Resource

Thank you! Q&A