Adding Common Sense into Artificial Intelligence Common Sense Computing Initiative Software Agents Group MIT Media Lab.

Slides:



Advertisements
Similar presentations
Henry Lieberman MIT Media Lab AI in End-User Software Engineering Henry Lieberman, MIT (presenting) Raphael Hoffmann, Michael Toomim, U. Washington, Seattle.
Advertisements

Reasoning From (Not Quite) Text
1 Latent Semantic Mapping: Dimensionality Reduction via Globally Optimal Continuous Parameter Modeling Jerome R. Bellegarda.
ConceptNet: A Wonderful Semantic World
Made with OpenOffice.org 1 Sentiment Classification using Word Sub-Sequences and Dependency Sub-Trees Pacific-Asia Knowledge Discovery and Data Mining.
Interactive Generation of Integrated Schemas Laura Chiticariu et al. Presented by: Meher Talat Shaikh.
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
GIS 200 Introduction to GIS Buildings. Poly Streams, Line Wells, Point Roads, Line Zoning,Poly MAP SHEETS.
TFIDF-space  An obvious way to combine TF-IDF: the coordinate of document in axis is given by  General form of consists of three parts: Local weight.
1 Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang, Assistant Professor Dept. of Computer Science & Information Engineering National Central.
1 Discovering Unexpected Information from Your Competitor’s Web Sites Bing Liu, Yiming Ma, Philip S. Yu Héctor A. Villa Martínez.
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Projects in the Intelligent User Interfaces Group Frank Shipman Associate Director, Center for the Study of Digital Libraries.
Multimedia Data Mining Arvind Balasubramanian Multimedia Lab (ECSS 4.416) The University of Texas at Dallas.
Discovering Computers 2010
Overview of Search Engines
CYC: The Common Sense Knowledge Base By: Jeremy Berman Alok Sahgal Dr. Reed CSC 550.
Common Sense Computing MIT Media Lab Interaction Challenges for Agents with Common Sense Henry Lieberman MIT Media Lab Cambridge, Mass. USA
Data Mining Chun-Hung Chou
1 Vector Space Model Rong Jin. 2 Basic Issues in A Retrieval Model How to represent text objects What similarity function should be used? How to refine.
ConceptNet 5 Jason Gaines 1. Overview What Is ConceptNet 5? History Structure Demonstration Questions Further Information 2.
This week: overview on pattern recognition (related to machine learning)
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
RuleML-2007, Orlando, Florida1 Towards Knowledge Extraction from Weblogs and Rule-based Semantic Querying Xi Bai, Jigui Sun, Haiyan Che, Jin.
Integrating Educational Technology into the Curriculum
Learning Object Metadata Mining Masoud Makrehchi Supervisor: Prof. Mohamed Kamel.
An Introduction to Artificial Intelligence and Knowledge Engineering N. Kasabov, Foundations of Neural Networks, Fuzzy Systems, and Knowledge Engineering,
Using Common Sense Reasoning to Create Intelligent Mobile Applications Software Agents Group MIT Media Lab.
Push Singh & Tim Chklovski. AI systems need data – lots of it! Natural language processing: Parsed & sense-tagged corpora, paraphrases, translations Commonsense.
Artificial Intelligence
SHARING EXPERIENCES SEOUL Commonsense Computing 상식을 갖춘 인공지능 This slide is prepared by MIT Media Lab Commonsense Computing Group and Hyemin Chung.
Join the Conversation: Active Listening on Social Media By Lauren Cleland New Media Specialist, Explore Georgia #TeamGaSocial.
Lifestyle Modeling and Visualization Catherine Havasi, Ryan MacDowell, Rob Speer, and Marko Popovic.
Latent Semantic Analysis Hongning Wang Recap: vector space model Represent both doc and query by concept vectors – Each concept defines one dimension.
Understanding The Semantics of Media Chapter 8 Camilo A. Celis.
Ranking in Information Retrieval Systems Prepared by: Mariam John CSE /23/2006.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
AnalogySpace: Reducing the Dimensionality of Common Sense Knowledge Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: From AAAI 2008 Robert.
1 SY DE 542 Navigation and Organization Prototyping Basics Feb 28, 2005 R. Chow
The Structure of Information Retrieval Systems LBSC 708A/CMSC 838L Douglas W. Oard and Philip Resnik Session 1: September 4, 2001.
Gene Clustering by Latent Semantic Indexing of MEDLINE Abstracts Ramin Homayouni, Kevin Heinrich, Lai Wei, and Michael W. Berry University of Tennessee.
Simultaneously Learning and Filtering Juan F. Mancilla-Caceres CS498EA - Fall 2011 Some slides from Connecting Learning and Logic, Eyal Amir 2006.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. Externally growing self-organizing maps and its application to database visualization and exploration.
Web Search and Text Mining Lecture 5. Outline Review of VSM More on LSI through SVD Term relatedness Probabilistic LSI.
Commonsense Reasoning in and over Natural Language Hugo Liu, Push Singh Media Laboratory of MIT The 8 th International Conference on Knowledge- Based Intelligent.
Link Distribution on Wikipedia [0407]KwangHee Park.
MIT Artificial Intelligence Laboratory — Research Directions The START Information Access System Boris Katz
11/23/00UNU/IAS/UNL Centre1 The Universal Networking Language United Nations University Institute of Advanced Studies United Networking Language ® UNU/IAS.
Knowledge Structure Vijay Meena ( ) Gaurav Meena ( )
Artificial Intelligence: Research and Collaborative Possibilities a presentation by: Dr. Ernest L. McDuffie, Assistant Professor Department of Computer.
Understanding Naturally Conveyed Explanations of Device Behavior Michael Oltmans and Randall Davis MIT Artificial Intelligence Lab.
Matrix Factorization & Singular Value Decomposition Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Artificial Intelligence, simulation and modelling.
Some Thoughts to Consider 5 Take a look at some of the sophisticated toys being offered in stores, in catalogs, or in Sunday newspaper ads. Which ones.
GoRelations: an Intuitive Query System for DBPedia Lushan Han and Tim Finin 15 November 2011
Sentiment Analysis Using Common- Sense and Context Information Basant Agarwal 1,2, Namita Mittal 2, Pooja Bansal 2, and Sonal Garg 2 1 Department of Computer.
Artificial Intelligence Knowledge Representation.
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
ITT_04101 COMPUTER APPLICATIONS Gaper M CIT
Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
Trends in NL Analysis Jim Critz University of New York in Prague EurOpen.CZ 12 December 2008.
An introduction to Amazon AI
Jonatas Wehrmann, Willian Becker, Henry E. L. Cagnini, and Rodrigo C
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Software Mr. Singh.
Software engineering USER INTERFACE DESIGN.
Word Embedding Word2Vec.
Introduction to Sentiment Analysis
Artificial Intelligence 2004 Speech & Natural Language Processing
Information Retrieval
Presentation transcript:

Adding Common Sense into Artificial Intelligence Common Sense Computing Initiative Software Agents Group MIT Media Lab

Why do computers need common sense? Conversation works because of unspoken assumptions People tend not to provide information they consider extraneous (Grice, 1975) Understanding language requires understanding connections

What can computers do with common sense? Understand the context of what the user wants Fill in missing information using background knowledge Discover trends in what people mean, not just what they say But how do we collect it?

A Brief Outline What is OMCS? What is ConceptNet? Using AnalogySpace for Inference Using Blending for Intuition OMCS Applications

Open Mind Common Sense Project Collecting common sense from internet volunteers since 2000 We have over 1,000,000 pieces of English language knowledge from 15,000 contributors Multilingual –Additional resources in Chinese, Portuguese, Korean, Japanese, and Dutch –In-progress: Spanish and Hungarian Users consider 87% of statements used in ConceptNet to be true

“A coat is used for keeping warm.” “People want to be respected.” “The sun is very hot.” “The last thing you do when you cook dinner is wash your dishes.” “People want good coffee.” What kind of knowledge?

Where does the knowledge come from? Contributors on our Web site (openmind.media.mit.edu) Games that collect knowledge

What is ConceptNet? A semantic network representation of the OMCS database (Liu and Singh, 2004) Over the years, used for: affect sensing, photo and video storytelling, text prediction, goal-oriented interfaces, speech recognition, task prediction, … ConceptNet 4.0 – Over 300,000 connections between ~80,000 concepts – Natural language processing tools to help line up your data with ConceptNet

An Example

Creation of ConceptNet A shallow parser turns natural language sentences into ConceptNet assertions 59 top-level patterns for English, such as “You would use {NP} to {VP}” {NP} and {VP} candidates identified by a chart parser

Representation Statement: expresses a fact in natural language Assertion: asserts that a relation exists between two concepts Concepts: sets of related phrases –identified by lemmatizing (or stemming) and removing stop words Relations: one of 25: –IsA, UsedFor, HasA, CapableOf, Desires, CreatedBy, AtLocation, CausesDesire, …

Example

Reliability Reliability increases when more users affirm that a statement is true –by entering equivalent statements independently –by rating existing statements on the Web Each assertion gets a weight according to how many users support it

Polarity Allows predicates that express true, negative information: “Pigs cannot fly” Negated assertions are represented by negative weights Reliability and polarity are independent

AnalogySpace Technique for learning, reasoning, and analyzing using common sense AnalogySpace can: –generalize from sparsely-collected knowledge –confirm or question existing knowledge –classify information in a knowledge base in a variety of ways Can use the same technique in other domains: businesses, people, communities, opinions

AnalogySpace Overview Finds patterns in knowledge Builds a representation in terms of those patterns Finds additional knowledge using the combination of those patterns Uses dimensionality reduction via Singular Value Decomposition

Input to the SVD Input to SVD: matrix of concepts vs. features Feature: a concept, a relation, and an open slot, e.g., (..., MadeOf, metal) Concepts × features = assertions

The Input Matrix For consistency, we scale each concept to unit Euclidean magnitude

Running the SVD

The Truncated SVD Truncating the SVD smoothes over sparse data.

Good vs. Bad

Reasoning with AnalogySpace Similarity represented by dot products of concepts (AA T ) –Approximately the cosine of their angle

Reasoning with AnalogySpace Predictions represented by dot products of concepts with features

Contributors are in the loop

Ad-hoc Categories

What can we use common sense for? A “sanity check” on natural language Text prediction Affect sensing Recommender systems “Knowledge management”

Common Sense in Context We don’t just use common sense to make more common sense Helps a system make sense of everyday life –Making connections in domain-specific information –Understanding free text –Bridging different knowledge sources

Digital Intuition Add common sense intuition Using similar techniques to make connections and inference between data sets Create a shared “Analogy”Space from two data sets using Blending

Blending Two data sets are combined in a way to maximize the interaction between the data sets They are weighted by a factor: C = (1 – f)A + fB

Blending Creates a New Representation With f = 0 or 1, equivalent to projecting one dataset into the other’s space In the middle, representation determined by both datasets.

No overlap = no interaction A’s singular valuesB’s singular values

Overlap -> Nonlinear Interaction (Veering)

Overlap -> Nonlinear Interaction

SVD over Multiple Data Sets Convert all data sets to matrices Find a rough alignment between the matrices –Some rows or features Find a blending factor –Maximize veering or interaction Run the AnalogySpace process jointly

Blends of Multiple Data Sets You can blend more than two things –Simple blending heuristic: scale all your data so that their largest singular vectors are equal

Applications Inference over domain specific data Word sense disambiguation Data visualization and analysis Finance

Tools we Distribute The OMCS database ConceptNet Divisi In development: the Luminoso visualizer

The Common Sense Computing Initiative Web: Thank you!