A Brief Introduction to Distant Supervision

Slides:



Advertisements
Similar presentations
Distant Supervision for Relation Extraction without Labeled Data CSE 5539.
Advertisements

Overview of the TAC2013 Knowledge Base Population Evaluation: Temporal Slot Filling Mihai Surdeanu with a lot help from: Hoa Dang, Joe Ellis, Heng Ji,
Linking Entities in #Microposts ROMIL BANSAL, SANDEEP PANEM, PRIYA RADHAKRISHNAN, MANISH GUPTA, VASUDEVA VARMA INTERNATIONAL INSTITUTE OF INFORMATION TECHNOLOGY,
Problem Semi supervised sarcasm identification using SASI
Event Extraction Using Distant Supervision Kevin Reschke, Martin Jankowiak, Mihai Surdeanu, Christopher D. Manning, Daniel Jurafsky 30 May 2014 Language.
LEDIR : An Unsupervised Algorithm for Learning Directionality of Inference Rules Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: From EMNLP.
Person Name Disambiguation by Bootstrapping Presenter: Lijie Zhang Advisor: Weining Zhang.
An Overview of Text Mining Rebecca Hwa 4/25/2002 References M. Hearst, “Untangling Text Data Mining,” in the Proceedings of the 37 th Annual Meeting of.
Semantic Video Classification Based on Subtitles and Domain Terminologies Polyxeni Katsiouli, Vassileios Tsetsos, Stathes Hadjiefthymiades P ervasive C.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Part I: Classification and Bayesian Learning
Introduction to Machine Learning Approach Lecture 5.
Modeling Missing Data in Distant Supervision for Information Extraction Alan Ritter Luke Zettlemoyer Mausam Oren Etzioni 1.
Jan 4 th 2013 Event Extraction Using Distant Supervision Kevin Reschke.
(ACM KDD 09’) Prem Melville, Wojciech Gryc, Richard D. Lawrence
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
Recognition of Multi-sentence n-ary Subcellular Localization Mentions in Biomedical Abstracts G. Melli, M. Ester, A. Sarkar Dec. 6, 2007
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
A Bootstrapping Method for Building Subjectivity Lexicons for Languages with Scarce Resources Author: Carmen Banea, Rada Mihalcea, Janyce Wiebe Source:
Constructing Knowledge Graph from Unstructured Text Image Source: Kundan Kumar Siddhant Manocha.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Instance Filtering for Entity Recognition Advisor : Dr.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
Template-Based Event Extraction Kevin Reschke – Aug 15 th 2013 Martin Jankowiak, Mihai Surdeanu, Dan Jurafsky, Christopher Manning.
Ontology based Information Extraction
Multiple Instance Learning for Sparse Positive Bags Razvan C. Bunescu Machine Learning Group Department of Computer Sciences University of Texas at Austin.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
DeepDive Introduction Dongfang Xu Ph.D student, School of Information, University of Arizona Sept 10, 2015.
DeepDive Model Dongfang Xu Ph.D student, School of Information, University of Arizona Dec 13, 2015.
Shallow Parsing for South Asian Languages -Himanshu Agrawal.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
Data Mining and Decision Support
MACHINE LEARNING 3. Supervised Learning. Learning a Class from Examples Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
Concept-Based Analysis of Scientific Literature Chen-Tse Tsai, Gourab Kundu, Dan Roth UIUC.
Learning to Extract CSCI-GA.2590 Ralph Grishman NYU.
Relation Extraction (RE) via Supervised Classification See: Jurafsky & Martin SLP book, Chapter 22 Exploring Various Knowledge in Relation Extraction.
The University of Illinois System in the CoNLL-2013 Shared Task Alla RozovskayaKai-Wei ChangMark SammonsDan Roth Cognitive Computation Group University.
Learning Relational Dependency Networks for Relation Extraction
Automatically Labeled Data Generation for Large Scale Event Extraction
Queensland University of Technology
Semi-Supervised Clustering
NELL Knowledge Base of Verbs
Measuring Monolinguality
Korean version of GloVe Applying GloVe & word2vec model to Korean corpus speaker : 양희정 date :
Concept Grounding to Multiple Knowledge Bases via Indirect Supervision
Relation Extraction CSCI-GA.2591
Supervised Time Series Pattern Discovery through Local Importance
Restricted Boltzmann Machines for Classification
Lecture 24: Relation Extraction
Research in Computational Molecular Biology , Vol (2008)
By Dan Roth and Wen-tau Yih PowerPoint by: Reno Kriz CIS
Reading Report: Open QA Systems
Giuseppe Attardi Dipartimento di Informatica Università di Pisa
Distant supervision for relation extraction without labeled data
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Factor Graph in DeepDive
Distributed Representation of Words, Sentences and Paragraphs
Introduction Task: extracting relational facts from text
Automatic Detection of Causal Relations for Question Answering
Family History Technology Workshop
Statistical n-gram David ling.
Enriching Taxonomies With Functional Domain Knowledge
Rachit Saluja 03/20/2019 Relation Extraction with Matrix Factorization and Universal Schemas Sebastian Riedel, Limin Yao, Andrew.
INTRODUCTION TO Machine Learning
Dan Roth Department of Computer Science
INTRODUCTION TO Machine Learning 3rd Edition
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Presenter: Yu Chen Computer Science Department
Presentation transcript:

A Brief Introduction to Distant Supervision Jin Mao Postdoc, School of Information, University of Arizona March 8th, 2016

Background and Framework OUTLINE Background and Framework An Example—Relation Extraction Major Challenges

Background and Framework Supervised Learning requires a large amount of labeled samples. Human annotating is tedious and costly in terms of time and economy.

Background and Framework Solution A set of Seeds –-- semi-supervision Existing Resources, such as knowledge bases(e.g., Wikipedia), databases, ---ontology Distant supervision: Use a distant existing knowledge base to generate training samples(Craven & Kumlien, 1999). E.g., Snow et al. (2005) exploited WordNet to extract hypernym (is-a) relations between entities.

Background and Framework A high level Framework (Roth et al., 2013)

Background and Framework A simple distant supervision algorithm(Poon et al., 2015) Require: A set of sentences, with entity mentions identified Require: A database of relation triples (entity, relation, entity) 1: For each relation triple, find all sentences containing the entity pair 2: Annotate those sentences with the corresponding relation 3: Sample unannotated sentences with co-occuring proteins as negative examples 4: Train a classifier using the annotated dataset 5: return the resulting classifier

An Example—Relation Extraction(Mintz et al., 2009) Knowledge base: Freebase, a training set of relations and entity pairs that participate in those relations. Entity: all entities are identified in sentences using a named entity tagger that labels persons, organizations and locations. Annotate: a sentence contains two entities and those entities are an instance of one of the Freebase relations Negative Samples: randomly selecting 1% entity pairs that do not appear in any Freebase relation. (sentences holding the entity pairs.)

An Example—Relation Extraction(Mintz et al., 2009) Classifier: A multiclass logistic regression classifier Input: an entity pair and a feature vector Output: a relation name and a confidence score based on the probability of the entity pair belonging to that relation Multiclass: to learn noisy feature

An Example—Relation Extraction(Mintz et al., 2009) Features Named Entity Tag Feature Lexical features Each lexical feature consists of the conjunction of all these components Make the number of features fewer, Large Corpus

An Example—Relation Extraction(Mintz et al., 2009) Features Syntactic features conjunction

An Example—Relation Extraction(Mintz et al., 2009) Features Syntactic features conjunction

Major Challenges Missing: incomplete knowledge base Noisy: Incorrect Pairs

Major Challenges Noise Reduction Often only a small fraction of co-occurrence matches indeed express the relation of the fact tuple. For example, the arguments of the fact tuple (“Barack Obama", born- in, “Honululu") could match in true positive contexts like”Barack Obama was born in Honululu", as well as false positive contexts like “Barack Obama visited Honululu". improve the quality of the training data by reducing the amount of noise.

Major Challenges Methods (Intxaurrondo, 2013): Noise Reduction Methods (Intxaurrondo, 2013): mention frequency pointwise mutual information similarity between the centroids of all relation mentions and each individual mention (MC) More complex methods (Roth et al., 2013) at-least-one constraints topic-based models pattern correlations

Major Challenges Missing Data Data is not missing at random (NMAR). Cannot be predicted by other variables. Dependent upon the data

Major Challenges Missing Data (Ritter et al., 2013) propose a new latent-variable approach that models missing data.

References Craven, M., & Kumlien, J. (1999, August). Constructing biological knowledge bases by extracting information from text sources. In ISMB (Vol. 1999, pp. 77-86). Intxaurrondo, A., Surdeanu, M., de Lacalle, O. L., & Agirre, E. (2013). Removing noisy mentions for distant supervision. Procesamiento del lenguaje natural, 51, 41-48. Mintz, M., Bills, S., Snow, R., & Jurafsky, D. (2009, August). Distant supervision for relation extraction without labeled data. In Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP: Volume 2-Volume 2 (pp. 1003-1011). Association for Computational Linguistics. Poon, H., Toutanova, K., & Quirk, C. (2015). Distant supervision for cancer pathway extraction from text. In Pac. Symp. Biocomput. (pp. 120-131). Reschke, K., Jankowiak, M., Surdeanu, M., Manning, C. D., & Jurafsky, D. (2014). Event Extraction Using Distant Supervision. In LREC (pp. 4527-4531). Ritter, A., Zettlemoyer, L., & Etzioni, O. (2013). Modeling missing data in distant supervision for information extraction. Transactions of the Association for Computational Linguistics, 1, 367-378. Roth, B., Barth, T., Wiegand, M., & Klakow, D. (2013, October). A survey of noise reduction methods for distant supervision. In Proceedings of the 2013 workshop on Automated knowledge base construction (pp. 73-78). ACM.

Thank you! More communication or error correction, contact