Goodfellow: Chap 1 Introduction

Slides:



Advertisements
Similar presentations
ARCHITECTURES FOR ARTIFICIAL INTELLIGENCE SYSTEMS
Advertisements

1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Artificial Intelligence
G5BAIM Artificial Intelligence Methods Graham Kendall Neural Networks.
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
Comp 5013 Deep Learning Architectures Daniel L. Silver March,
Brief History of AI Augusta Ada ( ) work on Charles Babbage’s early mechanical general-purpose computer, the Analytical Engine. Her.
Artificial Intelligence CS 363 Kawther Abas Lecture 1 Introduction 5/4/1435.
A shallow introduction to Deep Learning
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
Introduction to Artificial Intelligence and Soft Computing
Neural Networks in Computer Science n CS/PY 231 Lab Presentation # 1 n January 14, 2005 n Mount Union College.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Neural Networks Demystified by Louise Francis Francis Analytics and Actuarial Data Mining, Inc.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Chapter 8: Adaptive Networks
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 6: Applying backpropagation to shape recognition Geoffrey Hinton.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Chapter 6 Neural Network.
Object Recognizing. Deep Learning Success in 2012 DeepNet and speech processing.
Pattern Recognition. What is Pattern Recognition? Pattern recognition is a sub-topic of machine learning. PR is the science that concerns the description.
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
Artificial Neural Networks By: Steve Kidos. Outline Artificial Neural Networks: An Introduction Frank Rosenblatt’s Perceptron Multi-layer Perceptron Dot.
Chapter 13 Artificial Intelligence. Artificial Intelligence – Figure 13.1 The Turing Test.
Artificial Intelligence
Bassem Makni SML 16 Click to add text 1 Deep Learning of RDF rules Semantic Machine Learning.
Sparse Coding: A Deep Learning using Unlabeled Data for High - Level Representation Dr.G.M.Nasira R. Vidya R. P. Jaia Priyankka.
Introduction to Machine Learning, its potential usage in network area,
2/13/2018 4:38 AM © Microsoft Corporation. All rights reserved. MICROSOFT MAKES NO WARRANTIES, EXPRESS, IMPLIED OR STATUTORY, AS TO THE INFORMATION IN.
Big data classification using neural network
Who is the “Father of Deep Learning”?
Introduction to Neural Networks
Neural Network Architecture Session 2
Convolutional Neural Network
The Relationship between Deep Learning and Brain Function
Fundamentals of Information Systems, Sixth Edition
Deep Learning Amin Sobhani.
Goodfellow: Chap 1 Introduction
Deep Learning Insights and Open-ended Questions
Article Review Todd Hricik.
Matt Gormley Lecture 16 October 24, 2016
Lecture 24: Convolutional neural networks
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Neural Networks Dr. Peter Phillips.
What is an ANN ? The inventor of the first neuro computer, Dr. Robert defines a neural network as,A human brain like system consisting of a large number.
Neural Networks CS 446 Machine Learning.
Intelligent Information System Lab
Machine Learning Ali Ghodsi Department of Statistics
Artificial Intelligence in Healthcare
Goodfellow: Chap 1 Introduction
Machine Learning Today: Reading: Maria Florina Balcan
Goodfellow: Chap 6 Deep Feedforward Networks
Introduction to Artificial Intelligence and Soft Computing
of the Artificial Neural Networks.
Goodfellow: Chapter 14 Autoencoders
Perceptron as one Type of Linear Discriminants
[Figure taken from googleblog
A Proposal Defense On Deep Residual Network For Face Recognition Presented By SAGAR MISHRA MECE
Artificial Intelligence Lecture No. 28
Chapter 11 Practical Methodology
Machine Learning: Lecture 4
Machine Learning: UNIT-2 CHAPTER-1
Fundamentals of Neural Networks Dr. Satinder Bal Gupta
Artificial Intelligence
The Network Approach: Mind as a Web
David Kauchak CS158 – Spring 2019
CS855 Overview Dr. Charles Tappert.
Goodfellow: Chapter 14 Autoencoders
Patterson: Chap 1 A Review of Machine Learning
Presentation transcript:

Goodfellow: Chap 1 Introduction Dr. Charles Tappert The information here, although greatly condensed, comes almost entirely from the chapter content.

Artificial Intelligence Early Days – AI solved problems easy for computers but difficult for humans Problems described by formal, math rules The AI challenge was to solve problems easy for humans but difficult to describe formally Recognizing spoken words or faces in images Deep learning is abut solving these more intuitive problems Allowing computers to learn from experience Building a hierarchy of concepts, each defined by simpler concepts

Machine Learning Early AI successes were in formal environments IBM’s Deep Blue beats world champion Kasparov Knowledge-based AI attempted to hard-code world knowledge in formal languages Difficulties faced suggested need for ability to acquire knowledge by extracting patterns from real-world raw data This capability is known as machine learning Examples are logistic regression and naïve Bayes

Data Representation Performance of machine learning algorithms depends heavily on the data representation The features/attributes characterizing the data This dependence on representations is a general phenomenon in computer science and even in daily life Many AI tasks can be solved by designating the right set of features

Representation Learning One solution to the data representation problem is to have machine learning discover the representation Example: Autoencoder – combination of encoder to convert the input data into another representation, and decoder to convert back to the original format Goal is usually to separate the factors of variation that explain the observed data To disentangle and discard those not of interest

Deep Learning Deep learning solves the representation problem by introducing representations expressed in terms of simpler representations Deep learning involves a hierarchy of concepts that allows the computer to learn complicated concepts by building them out of simpler ones Graphically the concepts are built on top of each other with many layers The quintessential example of a deep learning model is the feedforward deep network or MLP

Illustration of Deep Learning Model

Deep Learning Deep learning is a machine learning approach to AI that allows computers to improve with experience and data Deep learning achieves great power and flexibility by learning to represent the world as a nested hierarchy of concepts, with each concept defined in relation to simpler ones The relationship among these different AI disciplines is shown in the following figure

Venn Diagram: Deep Learning

Flow Chart: Deep Learning

Goodfellow Textbook Part I Part II Part III Basic math tools and machine learning concepts Part II Established deep learning algorithms Part III More speculative ideas for future research

Goodfellow Textbook Organization

Three Waves of Development of Deep Learning 1940s-1960s: Early Neural Networks (Cybernetics?) Rosenblatt’s perceptron – developed from Hebb’s synaptic strengthening ideas and McCulloch-Pitts Neuron Key idea – variations of stochastic gradient descent Wave killed by Minsky 1969, lead to “AI Winter” 1980s-1990s: Connectionism Rumelhart, et al. Key idea – backpropagation 2006-present: Deep Learning Started with Hinton’s deep belief network Key idea – hierarchy of many layers in the neural network

First Two Waves

Deep Learning Increasing Dataset Sizes Since the 1990s machine learning systems have been used successfully in commercial applications But regarded as being more of an art than a technology Deep learning is regarded more and more as a technology – the amount of development skill reduces as the amount of training data increases The age of “Big Data” has made data collection easier As of 2016, rule of thumb is that supervised deep learning algorithms need around 5k labeled samples per category Performance is expected to match or exceed the human with a dataset of at least 10 million labeled samples

Deep Learning Increasing Dataset Sizes

Deep Learning Increasing Dataset Sizes Example inputs from MNIST dataset

Deep Learning Increasing Model Sizes A key reason deep learning networks are wildly successful today is that greater computation resources have allowed rapidly increasing model sizes Biological neurons are not especially densely connected and the number of connections per neuron in machine learning models have been within an order of magnitude of mammalian brains for decades As for the increase in the total number of model neurons, which doubles roughly every 2.4 years, we have decades to go at this rate before reaching the number on neurons in the human brain

Deep Learning Increasing Model Sizes: Connections per Neuron 9 = Commodity Off-The-Shelf High Performance Computing (COTS HPC) technology, 2013

Deep Learning Increasing Model Sizes: Number of Neurons 19 = Commodity Off-The-Shelf High Performance Computing (COTS HPC) technology, 2013

Deep Learning Increasing Accuracy, Complexity and Real-World Impact Since the 1980s machine learning systems have consistently improved recognition accuracy And applied with success to broader sets of applications In terms of complexity Early models recognized individual objects in small images Today, we process large high-resolution photos and typically recognize 1000 different categories of objects A dramatic moment in the meteoric rise of deep learning came when a convolutional network won the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) for the first time – Krizhevsky, et al., 2012

Deep Learning Increasing Accuracy, Complexity and Real-World Impact ImageNet Large Scale Visual Recognition Challenge (ILSVRC) now consistently won by deep networks