1 Unsupervised and Transfer Learning Challenge Unsupervised and Transfer Learning Challenge Isabelle Guyon Clopinet, California.

Slides:



Advertisements
Similar presentations
Active Learning Challenge Active Learning Challenge Isabelle Guyon (Clopinet, California) Gavin Cawley (University of East Anglia,
Advertisements

Semi-Supervised Learning in Gigantic Image Collections
Location Recognition Given: A query image A database of images with known locations Two types of approaches: Direct matching: directly match image features.
Advanced topics.
Rajat Raina Honglak Lee, Roger Grosse Alexis Battle, Chaitanya Ekanadham, Helen Kwong, Benjamin Packer, Narut Sereewattanawoot Andrew Y. Ng Stanford University.
TL by Kernel Meta Learning Fabio Aiolli University of Padova (Italy) F. Aiolli - Transfer Learning by Kernel Meta-Learning - ICML WS on Unsupervised and.
ImageNet Classification with Deep Convolutional Neural Networks
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
1 Part 1: Classical Image Classification Methods Kai Yu Dept. of Media Analytics NEC Laboratories America Andrew Ng Computer Science Dept. Stanford University.
Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009
Unsupervised Transfer Classification Application to Text Categorization Tianbao Yang, Rong Jin, Anil Jain, Yang Zhou, Wei Tong Michigan State University.
Deep Learning.
Discriminative and generative methods for bags of features
Causality Workbenchclopinet.com/causality Results of the Causality Challenge Isabelle Guyon, Clopinet Constantin Aliferis and Alexander Statnikov, Vanderbilt.
Self Taught Learning : Transfer learning from unlabeled data Presented by: Shankar B S DMML Lab Rajat Raina et al, CS, Stanford ICML 2007.
Recognition: A machine learning approach
Small Codes and Large Image Databases for Recognition CVPR 2008 Antonio Torralba, MIT Rob Fergus, NYU Yair Weiss, Hebrew University.
Unsupervised Learning With Neural Nets Deep Learning and Neural Nets Spring 2015.
1 Jun Wang, 2 Sanjiv Kumar, and 1 Shih-Fu Chang 1 Columbia University, New York, USA 2 Google Research, New York, USA Sequential Projection Learning for.
Semi-Supervised Learning in Gigantic Image Collections Rob Fergus (NYU) Yair Weiss (Hebrew U.) Antonio Torralba (MIT) TexPoint fonts used in EMF. Read.
K-means Based Unsupervised Feature Learning for Image Recognition Ling Zheng.
AN ANALYSIS OF SINGLE- LAYER NETWORKS IN UNSUPERVISED FEATURE LEARNING [1] Yani Chen 10/14/
Agenda Introduction Bag-of-words models Visual words with spatial location Part-based models Discriminative methods Segmentation and recognition Recognition-based.
What is the Best Multi-Stage Architecture for Object Recognition Kevin Jarrett, Koray Kavukcuoglu, Marc’ Aurelio Ranzato and Yann LeCun Presented by Lingbo.
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Comp 5013 Deep Learning Architectures Daniel L. Silver March,
Transfer Learning From Multiple Source Domains via Consensus Regularization Ping Luo, Fuzhen Zhuang, Hui Xiong, Yuhong Xiong, Qing He.
Overcoming Dataset Bias: An Unsupervised Domain Adaptation Approach Boqing Gong University of Southern California Joint work with Fei Sha and Kristen Grauman.
Computer Vision CS 776 Spring 2014 Recognition Machine Learning Prof. Alex Berg.
Boris Babenko Department of Computer Science and Engineering University of California, San Diego Semi-supervised and Unsupervised Feature Scaling.
Marcin Marszałek, Ivan Laptev, Cordelia Schmid Computer Vision and Pattern Recognition, CVPR Actions in Context.
1 Unsupervised and Transfer Learning Challenge Can Machines Transfer Knowledge from Task to Task? Isabelle Guyon Clopinet, California.
A shallow introduction to Deep Learning
Semi-supervised Learning on Partially Labeled Imbalanced Data May 16, 2010 Jianjun Xie and Tao Xiong.
Crowdsourcing for Spoken Dialogue System Evaluation Ling 575 Spoken Dialog April 30, 2015.
Presented by: Mingyuan Zhou Duke University, ECE June 17, 2011
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
Video Google: A Text Retrieval Approach to Object Matching in Videos Josef Sivic and Andrew Zisserman.
Active Learning Challenge Active Learning Workshop.
Transfer Learning for Image Classification Group No.: 15 Group member : Feng Cai Sauptik Dhar Sauptik.
Handwritten digit recognition
Competitions in machine learning: the fun, the art, and the science Isabelle Guyon Clopinet, Berkeley, California
CS 1699: Intro to Computer Vision Support Vector Machines Prof. Adriana Kovashka University of Pittsburgh October 29, 2015.
Stochastic Unsupervised Learning on Unlabeled Data July 2, 2011 Presented by Jianjun Xie – CoreLogic Collaborated with Chuanren Liu, Yong Ge and Hui Xiong.
1 End-to-End Learning for Automatic Cell Phenotyping Paolo Emilio Barbano, Koray Kavukcuoglu, Marco Scoffier, Yann LeCun April 26, 2006.
Combining Evolutionary Information Extracted From Frequency Profiles With Sequence-based Kernels For Protein Remote Homology Detection Name: ZhuFangzhi.
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
ChaLearn HEP competitions Isabelle Guyon, ChaLearn and LRI, Université Paris Sud, Orsay for.
Gist 2.3 John H. Phan MIBLab Summer Workshop June 28th, 2006.
CSC321 Lecture 24 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
Deep Belief Network Training Same greedy layer-wise approach First train lowest RBM (h 0 – h 1 ) using RBM update algorithm (note h 0 is x) Freeze weights.
CSC321 Lecture 27 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
Presented by: Isabelle Guyon Machine Learning Research.
AutoML challenge Isabelle Guyon
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Yann LeCun Other Methods and Applications of Deep Learning Yann Le Cun The Courant Institute of Mathematical Sciences New York University
Unsupervised Learning of Video Representations using LSTMs
Mini Places Challenge Adrià Recasens, Nov 21.
Restricted Boltzmann Machines for Classification
Introduction to Data Science Lecture 7 Machine Learning Overview
Vincent Granville, Ph.D. Co-Founder, DSC
Unsupervised Learning and Autoencoders
Deep Learning Workshop
State-of-the-art face recognition systems
CS 1674: Intro to Computer Vision Scene Recognition
Gesture Recognition Challenge
Representation Learning with Deep Auto-Encoder
MILESTONE RESULTS Mar. 1st, 2007
Presentation transcript:

1 Unsupervised and Transfer Learning Challenge Unsupervised and Transfer Learning Challenge Isabelle Guyon Clopinet, California IJCNN 2011 San Jose, California Jul. 31, Aug. 5, 2011

2 Unsupervised and Transfer Learning Challenge Challenge protocol and implemetation: Web platform: Server made available by Prof. Joachim Buhmann, ETH Zurich, Switzerland. Computer admin.: Peter Schueffler. Webmaster: Olivier Guyon, MisterP.net, France. Protocol review and advising: David W. Aha, Naval Research Laboratory, USA. Gideon Dror, Academic College of Tel-Aviv Yaffo, Israel. Vincent Lemaire, Orange Research Labs, France. Gavin Cawley, University of east Anglia, UK. Olivier Chapelle, Yahoo!, California, USA. Gerard Rinkus, Brandeis University, USA. Ulrike von Luxburg, MPI, Germany. David Grangier, NEC Labs, USA. Andrew Ng, Stanford Univ., Palo Alto, California, USA Graham Taylor, NYU, New-York. USA. Quoc V. Le, Stanford University, USA. Yann LeCun, NYU. New-York, USA. Danny Silver, Acadia Univ., Canada. Beta testing and baseline methods: Gideon Dror, Academic College of Tel-Aviv Yaffo, Israel. Vincent Lemaire, Orange Research Labs, France. Gregoire Montavon, TU Berlin, Germany. Data donors: Handwriting recognition (AVICENNA) -- Reza Farrahi Moghaddam, Mathias Adankon, Kostyantyn Filonenko, Robert Wisnovsky, and Mohamed Chériet (Ecole de technologie supérieure de Montréal, Quebec) contributed the dataset of Arabic manuscripts. The toy example (ULE) is the MNIST handwritten digit database made available by Yann LeCun and Corinna Costes. Object recognition (RITA) -- Antonio Torralba, Rob Fergus, and William T. Freeman, collected and made available publicly the 80 million tiny image dataset. Vinod Nair and Geoffrey Hinton collected and made available publicly the CIFAR datasets. See the techreport Learning Multiple Layers of Features from Tiny Images, by Alex Krizhevsky, 2009, for details. Human action recognition (HARRY) -- Ivan Laptev and Barbara Caputo collected and made publicly available the KTH human action recognition datasets. Marcin Marszałek, Ivan Laptev and Cordelia Schmid collected and made publicly available the Hollywood 2 dataset of human actions and scenes. Text processing (TERRY) -- David Lewis formatted and made publicly available the RCV1-v2 Text Categorization Test Collection. Ecology (SYLVESTER) -- Jock A. Blackard, Denis J. Dean, and Charles W. Anderson of the US Forest Service, USA, collected and made available the (Forest cover type) dataset. Credits

3 Unsupervised and Transfer Learning Challenge What is the problem?

4 Unsupervised and Transfer Learning Challenge Can learning about...

5 Unsupervised and Transfer Learning Challenge help us learn about…

6 Unsupervised and Transfer Learning Challenge What is Transfer Learning?

7 Unsupervised and Transfer Learning Challenge Vocabulary Target task labels Source task labels

8 Unsupervised and Transfer Learning Challenge Vocabulary Target task labels Source task labels

9 Unsupervised and Transfer Learning Challenge Vocabulary Target task labels Source task labels Domains the same? Labels available? Tasks the same?

10 Unsupervised and Transfer Learning Challenge Taxonomy of transfer learning Adapted from: A survey on transfer learning, Pan-Yang, Transfer Learning Unsupervised TL Semi-supervised TL Inductive TL No labels in both source and target domains Labels avail. ONLY in source domain Labels available in target domain No labels in source domain Labels available in source domain Transductive TL Cross-task TL Same source and target task Different source and target tasks Self-taught TL Multi-task TL

11 Unsupervised and Transfer Learning Challenge Challenge setting

12 Unsupervised and Transfer Learning Challenge Challenge setting Adapted from: A survey on transfer learning, Pan-Yang, Transfer Learning Unsupervised TL Semi-supervised TL Inductive TL No labels in both source and target domains Labels avail. ONLY in source domain Labels available in target domain No labels in source domain Labels available in source domain Transductive TL Cross-task TL Same source and target task Different source and target tasks Self-taught TL Multi-task TL

13 Unsupervised and Transfer Learning Challenge Dec 2010-April Goal: Learning data representations or kernels. Phase 1: Unsupervised learning (Dec 25, 2010-Mar 3, 2011) Phase 2: Cross-task transfer learning (Mar. 4, 2011-Apr. 15, 2011) Prizes: $ free registrations + travel awards Dissemination: ICML and IJCNN. Proceedings in JMLR W&CP. Evaluators Challenge target task labels Challenge data Validation data Development data Validation target task labels Competitors Data represen- tations

14 Unsupervised and Transfer Learning Challenge Dec 2010-April Goal: Learning data representations or kernels. Phase 1: Unsupervised learning (Dec 25, 2010-Mar 3, 2011) Phase 2: Cross-task transfer learning (Mar. 4, 2011-Apr. 15, 2011) Prizes: $ free registrations + travel awards Dissemination: ICML and IJCNN. Proceedings in JMLR W&CP. Evaluators Challenge target task labels Challenge data Validation data Development data Validation target task labels Source task labels Competitors Data represen- tations

15 Unsupervised and Transfer Learning Challenge Datasets of the challenge

16 Unsupervised and Transfer Learning Challenge Evaluation

17 Unsupervised and Transfer Learning Challenge AUC score For each set of samples queried, we assess the predictions of the learning machine with the Area under the ROC curve.

18 Unsupervised and Transfer Learning Challenge Area under the Learning Curve (ALC) Linear interpolation. Horizontal extrapolation.

19 Unsupervised and Transfer Learning Challenge Classifier used Linear discriminant: f(x) = w. x =  i w i x i Hebbian learning: X = (p, N) training data matrix Y  {–1/p –, +1/p + } p target vector w = X’ Y = (1/p + )  k  pos x k –(1/p – )  k  neg x k

20 Unsupervised and Transfer Learning Challenge Kernel version Kernel classifier: f(x) =  k  k k(x k, x) with a linear kernel k(x k, x) = x k. x and with  k = –1/p –, if x k  neg  k = +1/p +, if x k  pos Equivalent linear discriminant f(x) = (1/p + )  k  pos x k. x – (1/p – )  k  neg x k. x = w. x with w = (1/p + )  k  pos x k – (1/p – )  k  neg x k

21 Unsupervised and Transfer Learning Challenge Methods used

22 Unsupervised and Transfer Learning Challenge No learning 1) P Validation data Task labels C prediction Pre- processed data Challenge platform

23 Unsupervised and Transfer Learning Challenge No learning 1) P Validation data Task labels C prediction Pre- processed data Select the best preprocessing based on performance on the validation tasks

24 Unsupervised and Transfer Learning Challenge No learning 1) P

25 Unsupervised and Transfer Learning Challenge No learning 2) P Challenge data Task labels C prediction Pre- processed data Use the same preprocessor for the final evaluation

26 Unsupervised and Transfer Learning Challenge Unsupervised transfer learning P R Source domain 1) Simultaneously train a preprocessor P and a re-constructor R using unlabeled data

27 Unsupervised and Transfer Learning Challenge Unsupervised transfer learning P 1)

28 Unsupervised and Transfer Learning Challenge Unsupervised transfer learning P Target domain 2) Task labels C John Use the same preprocessor for the evaluation on target domains

29 Unsupervised and Transfer Learning Challenge Supervised data representation learning Source task labels P C Source domain Sea 1) Simultaneously train a preprocessor P and a classifier C with labeled source domain data

30 Unsupervised and Transfer Learning Challenge P 1) Supervised data representation learning

31 Unsupervised and Transfer Learning Challenge P Target domain 2) Task labels C John Use the same preprocessor for the evaluation on target domains Supervised data representation learning

32 Unsupervised and Transfer Learning Challenge Variants Use all or subsets of data for training (development/validation/challenge data). Learn what preprocessing steps to apply w. validation data (not the preprocessor) then apply the method to challenge data. Learn to reconstruct noisy versions of the data. Train a kernel instead of a preprocessor.

33 Unsupervised and Transfer Learning Challenge Results

34 Unsupervised and Transfer Learning Challenge Questions Can Transfer Learning beat raw data (or simple preprocessing)? Does Deep Learning work? Do labels help (does cross-task TL beat unsupervised TL)? Is model selection possible in TL? Did consistent TL methodologies emerge? Do the results make sense? Is there code available?

35 Unsupervised and Transfer Learning Challenge Can transfer learning beat raw data? Phase 1 (6933 jobs submitted, 41 complete final entries) Phase 2 (1141 jobs sumitted, 14 complete final entries)

36 Unsupervised and Transfer Learning Challenge Results (ALC)

37 Unsupervised and Transfer Learning Challenge Does “Deep Learning” work? Evolution of performance as a function of depth on SYLVESTER LISA team, 1 st in phase 2, 4 th in phase 1

38 Unsupervised and Transfer Learning Challenge Do labels help in TL?

39 Unsupervised and Transfer Learning Challenge Is model selection possible? Phase 1 Phase 2 Use of “transfer labels”: the  -  criterion (LISA team)

40 Unsupervised and Transfer Learning Challenge Did consistent methodologies emerge?

41 Unsupervised and Transfer Learning Challenge Results (ALC)

42 Unsupervised and Transfer Learning Challenge Bottom layers: Preprocessing and feature selection

43 Unsupervised and Transfer Learning Challenge Middle layers

44 Unsupervised and Transfer Learning Challenge Top layer

45 Unsupervised and Transfer Learning Challenge Implementation

46 Unsupervised and Transfer Learning Challenge A few things that worked well Learn the preprocessing steps (not the preprocessor) – Aiolli, 1 st phase 1. As 1 st steps: eliminate low info features or keep largest PC and sphere the data, normalize, and/or standardize. Learn denoising or contrastive auto-encoders or RBMs– LISA team, 1 st phase 2. Use cluster memberships of multiple K-means – 1055A team, 2 nd phase 1 and 3 rd phase 2. Transductive PCA (as last step) – LISA.

47 Unsupervised and Transfer Learning Challenge Conclusion UL: This challenge demonstrated the potential of unsupervised learning methods used as preprocessing to supervised learning tasks. UTL: Model selection of UL hyper-parameters can be carried out with “source tasks” similar to the “target tasks”. DL: Multi-step preprocessing leading to deep architectures can be trained in a greedy bottom-up step-wise manner. Favorite methods include normalizations, PCA, clustering, and auto-encoders. A kernel method won phase 1 and a Deep Learning method won phase 2.

48 Unsupervised and Transfer Learning Challenge STEP 1: Develop a “generic” gesture recognition system that can learn new signs with a few examples. STEP 2: At conference: teach the system new signs. STEP 3: Live evaluation in front of audience. June 2011-June Challenge