Computer Vision. DARPA Challenge Seeks Robots To Drive Into Disasters. DARPA's Robotics Challenge offers a $2 million prize if you can build a robot capable.

Slides:



Advertisements
Similar presentations
Artificial Intelligence 12. Two Layer ANNs
Advertisements

By: Mani Baghaei Fard.  During recent years number of moving vehicles in roads and highways has been considerably increased.
Chapter 4: Image Enhancement
Virtual Dart: An Augmented Reality Game on Mobile Device Supervisor: Professor Michael R. Lyu Prepared by: Lai Chung Sum Siu Ho Tung.
CPSC 425: Computer Vision (Jan-April 2007) David Lowe Prerequisites: 4 th year ability in CPSC Math 200 (Calculus III) Math 221 (Matrix Algebra: linear.
Yiming Zhang SUNY at Buffalo TRAFFIC SIGN RECOGNITION WITH COLOR IMAGE.
Raster Data. The Raster Data Model The Raster Data Model is used to model spatial phenomena that vary continuously over a surface and that do not have.
Bits are Not just for Numbers or Characters Computers store characters as bits or binary digits. Characters from the English-language keyboard can be represented.
1 5. Video Object Tracking and Processing To achieve augmented reality, computer generated graphics should be shown together with the live video In addition,
Redaction: redaction: PANAKOS ANDREAS. An Interactive Tool for Color Segmentation. An Interactive Tool for Color Segmentation. What is color segmentation?
1/22/04© University of Wisconsin, CS559 Spring 2004 Last Time Course introduction Image basics.
College Algebra Fifth Edition James Stewart Lothar Redlin Saleem Watson.
Prepared by: - Mr. T.R.Shah, Lect., ME/MC Dept., U. V. Patel College of Engineering. Ganpat Vidyanagar. Digital Image Processing & Machine Vision – An.
Intelligent Ground Vehicle Competition Navigation Michael Lebson - James McLane - Image Processing Hamad Al Salem.
Visual Representation of Information
A Brief Overview of Computer Vision Jinxiang Chai.
SCCS 4761 Introduction What is Image Processing? Fundamental of Image Processing.
Satellites in Our Pockets: An Object Positioning System using Smartphones Justin Manweiler, Puneet Jain, Romit Roy Choudhury TsungYun
1 Ethics of Computing MONT 113G, Spring 2012 Session 11 Graphics on the Web Limits of Computer Science.
Technology and digital images. Objectives Describe how the characteristics and behaviors of white light allow us to see colored objects. Describe the.
Applying Colour Theory to Visual Design.  Color Theory is the art of mixing colors to achieve desired effects.  The way colors are combined can be used.
Welcome Topic: Pixels A.M.Meshkatur Rahman Class: vii Roll: 07.
Objective Understand concepts used to create digital graphics. Course Weight : 15% Part Three : Concepts of Digital Graphics.
Vector vs. Bitmap
Digital Media Dr. Jim Rowan ITEC 2110 Wednesday, September 4.
Digital Cameras And Digital Information. How a Camera works Light passes through the lens Shutter opens for an instant Film is exposed to light Film is.
September 23, 2014Computer Vision Lecture 5: Binary Image Processing 1 Binary Images Binary images are grayscale images with only two possible levels of.
September 5, 2013Computer Vision Lecture 2: Digital Images 1 Computer Vision A simple two-stage model of computer vision: Image processing Scene analysis.
Computer Science Department Pacific University Artificial Intelligence -- Computer Vision.
DIEGO AGUIRRE COMPUTER VISION INTRODUCTION 1. QUESTION What is Computer Vision? 2.
Welcome Students Research Experience for Teachers (R.E.T.) Teachers seek out promising students from high school to connect to the engineering program.
Digital Media Dr. Jim Rowan ITEC Up Next! In the next several lectures we will be covering these topics: –Vector graphics –Bitmapped graphics –Color.
School of Engineering and Computer Science Victoria University of Wellington Copyright: Peter Andreae, VUW Image Recognition COMP # 18.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
Digital Media Dr. Jim Rowan ITEC 2110 Chapter 3. Roll call.
Lecture 7: Intro to Computer Graphics. Remember…… DIGITAL - Digital means discrete. DIGITAL - Digital means discrete. Digital representation is comprised.
DIGITAL IMAGE. Basic Image Concepts An image is a spatial representation of an object An image can be thought of as a function with resulting values of.
Kirk Bishop. What is Photosynth? Photosynth is a software application that analyzes digital photographs to build a three-dimensional models.
AN INTELLIGENT ASSISTANT FOR NAVIGATION OF VISUALLY IMPAIRED PEOPLE N.G. Bourbakis*# and D. Kavraki # #AIIS Inc., Vestal, NY, *WSU,
Digital Media Dr. Jim Rowan ITEC 2110 Images: Chapters 3, 4 & 5.
 By Bob “The Bird” Fiske & Anita “The Snail” Cost.
Content-Based Image Retrieval (CBIR) By: Victor Makarenkov Michael Marcovich Noam Shemesh.
Visual Computing Computer Vision 2 INFO410 & INFO350 S2 2015
1 Machine Vision. 2 VISION the most powerful sense.
A Tutorial on using SIFT Presented by Jimmy Huff (Slightly modified by Josiah Yoder for Winter )
Intelligent Robotics Today: Vision & Time & Space Complexity.
Digital Media Dr. Jim Rowan ITEC 2110 Chapter 3. Roll call.
Fundamentals of Digital Images & Photography. Pixels & Colors The pixel (a word invented from "picture element") is the basic unit of programmable color.
Machine Learning BY UZMA TUFAIL MCS : section (E) ROLL NO: /31/2016.
Graphics and Image Data Representations 1. Q1 How images are represented in a computer system? 2.
An Introduction to Digital Image Processing Dr.Amnach Khawne Department of Computer Engineering, KMITL.
Chapter 1 Lesson 3 Mapping Technology How are maps made? What are GPS and GIS?
Graphics Basic Concepts 1.  A graphic is an image or visual representation of an object.  A visual representation such as a photo, illustration or diagram.
Coin Recognition Using MATLAB - Emad Zaben - Bakir Hasanein - Mohammed Omar.
Another Example: Circle Detection
CS 115: Computing for The Socio-Techno Web
Introduction to Skin and Face Detection
Binary Notation and Intro to Computer Graphics
Vocabulary byte - The technical term for 8 bits of data.
Vocabulary byte - The technical term for 8 bits of data.
Quadtree representations
Opportunities of Scale, Part 2
Color Images.
CSC020 Computer Graphics Assignment 2 1.
Computer Vision Lecture 5: Binary Image Processing
Pearson Lanka (Pvt) Ltd.
Representing Images 2.6 – Data Representation.
Colors Computers build colors from Red, Green, and Blue; not Red, Blue, and Yellow. RGB = Red Green Blue Creating Colors Red + Blue = Purple No Red, No.
Magnetic Resonance Imaging
Artificial Intelligence 12. Two Layer ANNs
Presentation transcript:

Computer Vision

DARPA Challenge Seeks Robots To Drive Into Disasters. DARPA's Robotics Challenge offers a $2 million prize if you can build a robot capable of driving and using human tools. DARPA Wants to build rescue robots for the military? The new challenge; to develop robots that can navigate disaster-response scenarios.challenge Robots will be required to complete a number of discrete tasks, including traveling across rubble, removing debris from a blocked entryway, climbing a ladder, and entering and driving a car.

Bigdog Petman Robot Swag Aquatic synthetic life Current capabilities show that robot physical forms are quite advanced.

Why robots? They can go places humans can’t or shouldn’t go! – Natural disasters areas too dangerous for people – Uninhabitable environments – Nuclear, chemical or biological disasters

Why, then, if my robot is so useful is my android paranoid?

As mobile and dexterous as my robot is it needs a way to know where it is and how to get to a desired location or to know the location of objects in the surrounding environment.

This task is referred to as localization. Where am I? How did I get here? How do I get out of here, or go to the next place? Localization also refers to knowledge of the position or location of important objects in the environment.

Our robot could use GPS, except after a natural disasters like hurricanes or earthquakes nothing is where it is supposed to be. Major features of a city or locale may no longer exist.

After the Tsunami in Japan all GPS references to buildings and streets was useless

The same was true after this earthquake in China

Our robot may be in an area where GPS is unavailable – under the sea, in a building, a geographical location with no reference points.

This building could collapse at any time. Perfect for an android.

But, the android would need a method of creating navigational information for these new situations.

In general, important objects within the environment may be destroyed, moved from their original location, temporarily positioned or too insignificant to have GPS reference data.

The Goal for our robot is to autonomously develop an independent and real time understanding of its location and position vis-à-vis existing locations and objects using visual information.

The problem is to develop the machine’s capability to visually recognize objects within the immediate environment. However: – Humans see and immediately recognize specific objects; trees, faces, buildings, etc. We automatically know where we are in relation to these objects. – This is the result of many complex neurological and cognitive processes.

Computers don’t see things the way humans do. Computers see light and dark, color intensity. Computers store numbers and analyze (think) in equations.

We need a way for computers to identify objects in the environment as reference points for locating themselves and other objects in the environment. They will not see objects as we do. Though, they will be able to consistently recognize an object as having been “seen” before and associate the object with a specific location and position.

Robots see using digital cameras. Images are stored as digital information. We need our robot to recognized objects within the digital images of that object taken a different times and at different angles. Regardless of ambient conditions of light, weather, time of day, rotation of image and angle at which the picture was taken

As mentioned, images are stored as digital information. This information can be stored as an array of numbers that represent individual dots of the image called pixels. For grayscale images similar to “black and white” photographs the intensity of blackness is stored as a number from 0(black) to 255(white)

For color images the image information can be stored as an array of numbers that represent individual pixels as well. Instead of storing information on the intensity of blackness, information is stored about the levels of red, green and blue colors for each pixel (RGB). The intensity for these colors is stored as a number from 0(maximum) to 255(minimum)

Color images can be easily converted to Grayscale by weighting the values for Red, Green, and Blue and summing the resulting values. Often color will no provide useful information for our process, but increase the processing burden. We will work in grayscale.

We need our robot to recognized an object within the digital images of that object taken a different times and different angles. Matching an object’s features across different images is a common problem in computer vision. When you have different images with different scales and rotations, you need to use an program called Scale Invariant Feature Transform or SIFT.

The SIFT algorithm analyses sections of a digital image and identifies regions of significant change in grayscale referred to as SIFT points. These SIFT points are then described by the size and intensity of the region’s changes and the general direction (orientation) of these changes. The region’s location and its relative position to other SIFT points is recorded.

Through this process SIFT creates a unique “fingerprint” for an object that is independent of many ambient features of a photograph or image.

Photograph of rooftops in [Italy?]

Converted to grayscale and SIFT process applied The SIFT point are detected

This slide displays information about each SIFT point.

Two photos which share objects

We have found SIFT points that match from the two photos!!!

Using SIFT our robot can identify a geographic or architectural feature and associate the image with a position either by GPS or another positioning technology.

Even though our goal is for our robot to employ an autonomous locating ability separate from GPS technology, we can use GPS technology in the development of our system.

Current research employees geolocalization using GPS and Google street view images. I can take a photograph on any street in downtown Orlando and using SIFT and other technologies determine my location by comparing SIFT points in my photograph and SIFT points in images found in Google street view for Orlando.

The success rate in obtaining a good location is quite high. Ultimately my paranoid android may be calmed by the certain knowledge of his location in the world.

Frankly, the mathematics involved in these tasks is far beyond the scope of any of your current classes. Calculus, Statistic and probability theory as well as finite mathematics are employed. People who are able to develop Artificial intelligence and robotic capabilities are and will be some of the most sought after employees in the modern work force. All These technologies are still developing technologies and creative fundamental work still needs to be done.

If this type of work interests you, please contact: Dr. Mubarak Shah or Dr. Neils Lobo for more information. Questions