AI and Machine Learning

Slides:



Advertisements
Similar presentations
Learning from Observations Chapter 18 Section 1 – 3.
Advertisements

CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Huffman code and ID3 Prof. Sin-Min Lee Department of Computer Science.
IT 433 Data Warehousing and Data Mining
Decision Tree Approach in Data Mining
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Learning From Observations
Three kinds of learning
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Part I: Classification and Bayesian Learning
Machine Learning CPS4801. Research Day Keynote Speaker o Tuesday 9:30-11:00 STEM Lecture Hall (2 nd floor) o Meet-and-Greet 11:30 STEM 512 Faculty Presentation.
Data Mining Joyeeta Dutta-Moscato July 10, Wherever we have large amounts of data, we have the need for building systems capable of learning information.
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
Inductive learning Simplest form: learn a function from examples
Mohammad Ali Keyvanrad
Decision Trees & the Iterative Dichotomiser 3 (ID3) Algorithm David Ramos CS 157B, Section 1 May 4, 2006.
1 CO Games Development 2 Week 19 Probability Trees + Decision Trees (Learning Trees) Gareth Bellaby.
LEARNING DECISION TREES Yılmaz KILIÇASLAN. Definition - I Decision tree induction is one of the simplest, and yet most successful forms of learning algorithm.
Learning from Observations Chapter 18 Through
CHAPTER 18 SECTION 1 – 3 Learning from Observations.
CS 445/545 Machine Learning Spring, 2013 See syllabus at
Learning from Observations Chapter 18 Section 1 – 3, 5-8 (presentation TBC)
Learning from Observations Chapter 18 Section 1 – 3.
Learning from observations
Machine Learning, Decision Trees, Overfitting Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 14,
CS 445/545 Machine Learning Winter, 2014 See syllabus at
Decision Tree Learning Presented by Ping Zhang Nov. 26th, 2007.
Data Mining and Decision Support
1 Classification: predicts categorical class labels (discrete or nominal) classifies data (constructs a model) based on the training set and the values.
CIS 335 CIS 335 Data Mining Classification Part I.
WHAT IS DATA MINING?  The process of automatically extracting useful information from large amounts of data.  Uses traditional data analysis techniques.
Chapter 18 Section 1 – 3 Learning from Observations.
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
Machine Learning Lecture 1: Intro + Decision Trees Moshe Koppel Slides adapted from Tom Mitchell and from Dan Roth.
WHAT IS DATA MINING?  The process of automatically extracting useful information from large amounts of data.  Uses traditional data analysis techniques.
Friday’s Deliverable As a GROUP, you need to bring 2N+1 copies of your “initial submission” –This paper should be a complete version of your paper – something.
Learning From Observations Inductive Learning Decision Trees Ensembles.
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Anifuddin Azis LEARNING. Why is learning important? So far we have assumed we know how the world works Rules of queens puzzle Rules of chess Knowledge.
CSE573 Autumn /09/98 Machine Learning Administrative –Last topic: Decision Tree Learning Reading: 5.1, 5.4 Last time –finished NLP sample system’s.
Learning from Observations
Learning from Observations
Machine Learning Inductive Learning and Decision Trees
CS 445/545 Machine Learning Winter, 2017
DECISION TREES An internal node represents a test on an attribute.
Introduce to machine learning
CO Games Development 2 Week 22 Trees
Prepared by: Mahmoud Rafeek Al-Farra
CS 4700: Foundations of Artificial Intelligence
Presented By S.Yamuna AP/CSE
Chapter 6 Classification and Prediction
CS 445/545 Machine Learning Spring, 2017
Data Mining Lecture 11.
Unsupervised Learning and Autoencoders
Classification and Prediction
Machine Learning Chapter 3. Decision Tree Learning
Overview of Machine Learning
Machine Learning Chapter 3. Decision Tree Learning
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Learning from Observations
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
©Jiawei Han and Micheline Kamber
Learning from Observations
A task of induction to find patterns
Machine Learning: Decision Tree Learning
A task of induction to find patterns
Version Space Machine Learning Fall 2018.
Presentation transcript:

AI and Machine Learning

The Turing Test

Turing attack How can we show a machine is Intelligent? Let A = machine. Let C = Intelligent. Let B = someone that “we” claim is intelligent. How can we show A = C? Hmm. It’s subjective? Well most (normal) say B = C. So if can we show that A = B, then we can show that A = C!

Turing attack So subjective criteria when normal can be reasoned with transitivity! E.g., let C = Greatest basketball player ever. Most “normal” say Jordan. Let Jordan = B. Now is Lebron = the greatest ever? Let Lebron = A. Well we have to show that A = C, or A = B. Let’s assume A = B. Do we wind up in a contradiction? If so A!=B. If not A = B and A = C. Is Lebron the greatest?

Chinese Room Argument

Chinese Room Argument Searle defines: Strong AI = Understanding Weak AI = pattern recognition and mappings What do you think?

Loebner prize Mitsuka The Loebner Prize is an annual competition in artificial intelligence that awards prizes to the computer programs considered by the judges to be the most human-like. The format of the competition is that of a standard Turing test. No one has ever won the silver or gold medal!

Loebner bronze winner in 2013, 2016, 2017, 2018 Mitsuko Loebner bronze winner in 2013, 2016, 2017, 2018

Unsupervised vs. Supervised Learning Supervised learning is aided by training data and human correction. Here’s some training data. Learn the patterns. Make your best guess at what the patterns are. We’ll feed you test data to figure out if you’ve understood it. If you stray of course we’ll correct you and retrain. Examples include Decision Trees and Neural Networks. Unsupervised learning is uncorrected and runs on data. It can’t classify things “yet”. But is very good at clustering and anomaly detection.

Decision Tree Learning Task: Given: collection of examples (x, f(x)) Return: a function h (hypothesis) that approximates f h is a decision tree Input: an object or situation described by a set of attributes (or features) Output: a “decision” – the predicts output value for the input. The input attributes and the outputs can be discrete or continuous. We will focus on decision trees for Boolean classification: each example is classified as positive or negative.

Can we learn how counties vote? New York Times April 16, 2008 Decision Trees: a sequence of tests. Representation very natural for humans. Style of many “How to” manuals and trouble-shooting procedures.

Decision Tree What is a decision tree? A tree with two types of nodes: Decision nodes Leaf nodes Decision node: Specifies a choice or test of some attribute with 2 or more alternatives;  every decision node is part of a path to a leaf node Leaf node: Indicates classification of an example

Inductive Learning Example Etc. Instance Space X: Set of all possible objects described by attributes (often called features). Target Function f: Mapping from Attributes to Target Feature (often called label) (f is unknown) Hypothesis Space H: Set of all classification rules hi we allow. Training Data D: Set of instances labeled with Target Feature What is the best Variable (Feature) to use as an indicator of a BigTip?

Entropy & Information Gain Obviously  Don’t Panic. Entropy is just a way to measure disorder = uncertainty in data and uncertainty in “data mappings”.

Information Gain

Excel example

Decision Tree Example: “BigTip” Food Price Speedy no yes great mediocre yuck adequate high Our data Is the decision tree we learned consistent? Definition of components Internal nodes: tests on attributes Branches: attribute value (one branch for each value) Leaves: assign classification Understanding example tree Classify Example 1: yes Classify Example 2: yes Classify Example 3: no Classify Example 4: yes Classify: $x_2=(Food=great, Chat=no, Fast=no, Price=adequate, Bar=no)$ $\rightarrow$ yes Example represents function: (Food=great) AND ((Fast=yes) OR ((Fast=no) AND (Price=adequate)))  Simplify to: (Food=great) AND ((Fast=yes) OR (Price=adequate)) How would you represent: A AND B A OR B A XOR B Yes, it agrees with all the examples! Data: Not all 2x2x3 = 12 tuples Also, some repeats! These are literally “observations.”

Top-Down Induction of Decision Tree: Big Tip Example Node “done” when uniform label or “no further uncertainty.” 10 8 7 4 3 1 2 5 6 9 10 examples: 6+ Food 4- y 10 8 7 4 3 1 2 m g No No Speedy y n 6 5 9 4 2 Definition of components Internal nodes: tests on attributes Branches: attribute value (one branch for each value) Leaves: assign classification Understanding example tree Classify Example 1: yes Classify Example 2: yes Classify Example 3: no Classify Example 4: yes Classify: $x_2=(Food=great, Chat=no, Fast=no, Price=adequate, Bar=no)$ $\rightarrow$ yes Example represents function: (Food=great) AND ((Fast=yes) OR ((Fast=no) AND (Price=adequate)))  Simplify to: (Food=great) AND ((Fast=yes) OR (Price=adequate)) How would you represent: A AND B A OR B A XOR B Price a h Yes 10 8 7 3 1 Yes No 4 2 How many + and - examples per subclass, starting with y? Let’s consider next the attribute Speedy

An example

An example

An example