Thien Anh Dinh1, Tomi Silander1, Bolan Su1, Tianxia Gong

Slides:



Advertisements
Similar presentations
Background Knowledge for Ontology Construction Blaž Fortuna, Marko Grobelnik, Dunja Mladenić, Institute Jožef Stefan, Slovenia.
Advertisements

PEBL: Web Page Classification without Negative Examples Hwanjo Yu, Jiawei Han, Kevin Chen- Chuan Chang IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING,
Using Large-Scale Web Data to Facilitate Textual Query Based Retrieval of Consumer Photos.
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Zhimin CaoThe Chinese University of Hong Kong Qi YinITCS, Tsinghua University Xiaoou TangShenzhen Institutes of Advanced Technology Chinese Academy of.
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Rapid Object Detection using a Boosted Cascade of Simple Features Paul Viola, Michael Jones Conference on Computer Vision and Pattern Recognition 2001.
AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ.
Feature/Model Selection by Linear Programming SVM, Combined with State-of-Art Classifiers: What Can We Learn About the Data Erinija Pranckeviciene, Ray.
A novel supervised feature extraction and classification framework for land cover recognition of the off-land scenario Yan Cui
Large-Scale, Real-World Face Recognition in Movie Trailers Week 2-3 Alan Wright (Facial Recog. pictures taken from Enrique Gortez)
Knowing a Good HOG Filter When You See It: Efficient Selection of Filters for Detection Ejaz Ahmed 1, Gregory Shakhnarovich 2, and Subhransu Maji 3 1 University.
São Paulo Advanced School of Computing (SP-ASC’10). São Paulo, Brazil, July 12-17, 2010 Looking at People Using Partial Least Squares William Robson Schwartz.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Graz University of Technology, AUSTRIA Institute for Computer Graphics and Vision Fast Visual Object Identification and Categorization Michael Grabner,
4 th NETTAB Workshop Camerino, 5 th -7 th September 2004 Alberto Bertoni, Raffaella Folgieri, Giorgio Valentini
Robust Real-time Object Detection by Paul Viola and Michael Jones ICCV 2001 Workshop on Statistical and Computation Theories of Vision Presentation by.
Visual Expertise Is a General Skill Maki Sugimoto University of California, San Diego November 20, 2000.
Presented by Zeehasham Rasheed
1 Automated Feature Abstraction of the fMRI Signal using Neural Network Clustering Techniques Stefan Niculescu and Tom Mitchell Siemens Medical Solutions,
Generic object detection with deformable part-based models
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
Multiclass object recognition
EADS DS / SDC LTIS Page 1 7 th CNES/DLR Workshop on Information Extraction and Scene Understanding for Meter Resolution Image – 29/03/07 - Oberpfaffenhofen.
Exploiting Ontologies for Automatic Image Annotation M. Srikanth, J. Varner, M. Bowden, D. Moldovan Language Computer Corporation
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
AUTOMATIZATION OF COMPUTED TOMOGRAPHY PATHOLOGY DETECTION Semyon Medvedik Elena Kozakevich.
Multimodal Information Analysis for Emotion Recognition
Hierarchical Annotation of Medical Images Ivica Dimitrovski 1, Dragi Kocev 2, Suzana Loškovska 1, Sašo Džeroski 2 1 Department of Computer Science, Faculty.
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
A Codebook-Free and Annotation-free Approach for Fine-Grained Image Categorization Authors Bangpeng Yao et al. Presenter Hyung-seok Lee ( 이형석 ) CVPR 2012.
Transfer Learning for Image Classification Group No.: 15 Group member : Feng Cai Sauptik Dhar Sauptik.
A Two-level Pose Estimation Framework Using Majority Voting of Gabor Wavelets and Bunch Graph Analysis J. Wu, J. M. Pedersen, D. Putthividhya, D. Norgaard,
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
Distributed Representative Reading Group. Research Highlights 1Support vector machines can robustly decode semantic information from EEG and MEG 2Multivariate.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
CoCQA : Co-Training Over Questions and Answers with an Application to Predicting Question Subjectivity Orientation Baoli Li, Yandong Liu, and Eugene Agichtein.
Data Mining, ICDM '08. Eighth IEEE International Conference on Duy-Dinh Le National Institute of Informatics Hitotsubashi, Chiyoda-ku Tokyo,
Image Classification over Visual Tree Jianping Fan Dept of Computer Science UNC-Charlotte, NC
FACE DETECTION : AMIT BHAMARE. WHAT IS FACE DETECTION ? Face detection is computer based technology which detect the face in digital image. Trivial task.
Subjectivity Recognition on Word Senses via Semi-supervised Mincuts Fangzhong Su and Katja Markert School of Computing, University of Leeds Human Language.
Text Categorization by Boosting Automatically Extracted Concepts Lijuan Cai and Tommas Hofmann Department of Computer Science, Brown University SIGIR 2003.
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Developing outcome prediction models for acute intracerebral hemorrhage patients: evaluation of a Support Vector Machine based method A. Jakab 1, L. Lánczi.
Max-Confidence Boosting With Uncertainty for Visual tracking WEN GUO, LIANGLIANG CAO, TONY X. HAN, SHUICHENG YAN AND CHANGSHENG XU IEEE TRANSACTIONS ON.
Martina Uray Heinz Mayer Joanneum Research Graz Institute of Digital Image Processing Horst Bischof Graz University of Technology Institute for Computer.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
1 Bilinear Classifiers for Visual Recognition Computational Vision Lab. University of California Irvine To be presented in NIPS 2009 Hamed Pirsiavash Deva.
Automatically Labeled Data Generation for Large Scale Event Extraction
Hybrid Deep Learning for Reflectance Confocal Microscopy Skin Images
The Relationship between Deep Learning and Brain Function
Guillaume-Alexandre Bilodeau
Object detection with deformable part-based models
System for Semi-automatic ontology construction
Bag-of-Visual-Words Based Feature Extraction
Object Localization Goal: detect the location of an object within an image Fully supervised: Training data labeled with object category and ground truth.
R-CNN region By Ilia Iofedov 11/11/2018 BGU, DNN course 2016.
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Students: Meiling He Advisor: Prof. Brain Armstrong
PEBL: Web Page Classification without Negative Examples
Introduction Computer vision is the analysis of digital images
The Assistive System Progress Report 2 Shifali Kumar Bishwo Gurung
iSRD Spam Review Detection with Imbalanced Data Distributions
Leverage Consensus Partition for Domain-Specific Entity Coreference
Using Uneven Margins SVM and Perceptron for IE
Introduction Computer vision is the analysis of digital images
CIS 519 Recitation 11/15/18.
Extracting Why Text Segment from Web Based on Grammar-gram
Presentation transcript:

Unsupervised medical image classification by combining case-based classifiers Thien Anh Dinh1, Tomi Silander1, Bolan Su1, Tianxia Gong Boon Chuan Pang2, Tchoyoson Lim2, Cheng Kiang Lee2 Chew Lim Tan1,Tze-Yun Leong1 1National University of Singapore 2National Neuroscience Institute 3Bioinformatics Institute, Singapore Good afternoon everyone! My talk today is about building an unsupervised medical image classification by combining case-based classifiers This is the joint work of the National University of Singapore and the local hospital in Singapore

Automated medical image annotation Huge amount of valuable data available in medical image databases Not fully utilized for medical treatment, research and education Medical image annotation: To extract knowledge from images to facilitate text-based retrieval of relevant images To provide a second source of opinions for clinicians on abnormality detection and pathology classification There is huge amount of valuable data available in medical image databases. However, they have not been fully utilized for medical treatment, research and education. Automated medical Image annotation technology attempts to improve this situation. It provides a way to extract knowledge from images to facilitate text-based retrieval of relevant images and In addition, it provides a second source of opinions for clinicians on abnormality detection and pathology classification

Selecting discriminative features Problem Flowchart of current methods Challenges in current methods Highly sensitive and accurate segmentation Extracting domain knowledge Automatic feature selection Time-consuming manual adjustment process  reduces usages of medical image annotation systems Extracting features Selecting discriminative features Building classifiers Labeling Many image annotation systems have been proposed to address this problem. They usually have four main components. First is the feature extraction process for detecting areas of interest, and these features are subsequently selected for building a classifiers, and finally use the classifier for labeling a new image It is commonly acknowledged that the success of the annotation system depends significantly on the feature extraction and feature selection. However, there exists many challenges in these components. For instance, automated segmentation is considered as the most challenging task in medical imaging. Developing a highly sensitive and accurate segmentation requires a lot of time and effort. Secondly, the domain knowledge is not always available or easily extracted from experts. Therefore, it is hard to manually create a particular set of features that are useful for analyzing images. On the other hand, automatic feature selection is also a non-trivial task. It is hard to even establish a space of relevant feature candidates. Even more, the set of relevant features may well vary from case to case depending on context and the particular image, so that no single set of selected features is optimal for all images As a result, constructing these annotation systems is often a difficult, time-consuming and labor-intensive process. These problems have reduced usage of medical image annotation systems in practice

Objective An automated pathology classification system for volumetric brain image slices Main highlights Eliminates the need for segmentation and semantic or annotation-based feature selection Reduces the amount of manual work for constructing an annotation system Extracts automatically and efficiently knowledge from images Improves the utilization of medical image databases In this work, We propose an automated pathology classification system for volumetric brain image slices. The main highlights our work are: first, we eliminate the need for segmentation nor semantic- or annotation-based feature selection. Consequently, the knowledge in image databases can be extracted automatically and efficiently. And finally, we hope to improve the utilization of medical image databases.

System overview Case-based classifier Gabor filters Non domain specific features Localized low-level features Ensemble learning Set of classifiers Each classifier with a random subset of features Final classification: an aggregated result In this slide, we will talk about the architecture of our proposed system. The system consists of two main components: case-based classifier and ensemble learning. Case-based classification is performed directly on previously seen data without an intermediate modeling phase. We deploy domain independent Gabor-filters that extract localized low level features from the image. However, with such a low level feature space, selection of a small set of significant features appears counterintuitive. We therefore use ensemble learning and form a collection of weak classifiers. Each classifier specialize on a different random subset of features. The final classification is then a product of these classifiers.

Sparse representation-based classifier Sparse representation-based classifier (SRC) proposed by Wright et al. for face recognition task Non-parametric sparse representation classifier SRC consists of two stages Reconstructing: a test image as a linear combination of a small number of training images Classifying: evaluating how the images belonging to different classes contribute to the reconstruction of the test image First, we will look at the case-base classifier. In our work, we deploy a sparse representation-based classifier or SRC for short. SRC was proposed by Wright et al. [12] and has achieved high performance and success for face recognition task. This is a nonparametric sparse representation classifier. The SRC consists of two stages: reconstruction and classification In reconstruction: the SRC aims at reconstructing the test image as a linear combination of a small number of training images. Classification is then done by evaluating how the images be- longing to the different classes contribute to the reconstruction of the test image.

Sparse representation-based classifier (Wright et al.) This is the pseudo-code for this classification algorithm which I just described briefly before. The inputs for this algorithm is training samples and a test sample which are represented as a matrix The final output of the algorithm is the class specific reconstruction error r_k(y) or we call It residual. In the original algorithm, from these residuals, we can decide which class a test sample belongs to. However, in our framework, we stop at this residual and we use it in the ensemble learning framework.

Image databases x1, x2,…, x1000 Sparse reconstruction New data item y ≈ a7x7 + a23x23 + a172x172 + a134x134 + a903x903 y ≈ a7x7 + a23x23 + a172x172 + a134x134 + a903x903 Class residuals Here is the concrete example of the SRC algorithm. Suppose that we have a training databases consisting of 1000 images. Given a new image, it will be sparsely reconstructed from the training cases. This is done through the optimization of the trade-off between the approximation quality and the number of old cases allowed in the reconstruction. Notice that cases in reconstruction belong to different classes (red and blue) From this, we can calculate the class specific residuals as r1 and r2 r1 = || y – (a7x7 + a172x172 + a132x134)||2 r2 = || y – (a23x23 + a903x903)||2

Ensemble of weak classifiers Combine multiple weak classifiers Take class specific residuals as confidence measures The smaller the residual for the class, the better we construct the test by just using the samples from that class To classify image y, compute average class-specific residuals of all W weak classifiers Then, multiple SRC classifiers are combined to form an ensemble. As we observe that the smaller the residual for the class, the better we construct the test by just using the samples from that class. Therefore, we take the class specific residuals as confidence measurers. Finally, .... The test sample y is assigned to the class with minimum average residual.

Domain Automatically annotate CT brain images for traumatic brain injury (TBI) TBI: major cause of death and disability Several types of hemorrhages: Extradural hematoma (EDH) Subdural hematoma (SDH) Intracerebral hemorrage (ICH) Subarachnoid hemorrhage (SAH) Intraventricular hematoma (IVH) Subdural hematoma Extradural hematoma In principle, our discussed approach is applicable to any volumetric brain images. In this work, we focus on CT brain images for traumatic brain injury (or TBI). TBI is a major cause of death and disability in the world. TBI consists several types of hemorrhages. These classification scheme is based on different location, different shape and different size of hemorrhage region.

Data Volumetric stack of 18-30 images (slices) CT brain scans of 103 patients Each scan: Volumetric stack of 18-30 images (slices) Image resolution: 512 x 512 pixels Manually assigned a hematoma type extracted from its medical text report Our data set consists of images featuring three types of TBI: EDH (24 patients), ICH (21 patients) and SDH (58 patients). Each case is in the form of a volumetric stack consisting of 18-30 slices. The dataset is manually labeled with TBI subtype and verified with findings from medical text reports. In this work, we only focus on detection of abnormal slices, thus slices without abnormali- ties are removed from the dataset. We assume that each slice only exhibits a single type of hematoma.

A volumetric CT brain scan with 19 slices Here is an example of a volumetric brain scan. It is a stack of 19 slices. A volumetric CT brain scan with 19 slices

Experimental setup Compared performances of SRC vs. SVM vs. SVM + feature selection With/without ensemble learning Run stratified ten-fold cross-validation 50 times with different random foldings Measured the average precisions and recalls Separated training and testing dataset at the case level In our experiments, we compare performances of proposed method with plain SRC, baseline classifier SVM and SVM with feature selection. And we also study the performance of ensemble learning on both SRC and SVM. We run stratified ten-fold cross-validation 50 times with different random foldings to evaluate the performance of our system. Although the system works at the slice level, training and testing dataset are separated at the case level to avoid slices from the same case being both in training and test data.

Experimental results As discussed in the previous slide, we constructed five settings: a baseline classifier SVM, SVM with automatic feature selection technique, Ensemble with SVM, and plain SRC Our Ensemble+SRC method yields promising results in classifying abnormal TBI slices (Table 1). The performance in the majority class SDH is better than in EDH and ICH. It is also interesting to see that SRC appears to perform better than SVM in this domain. Even without feature selection. ---- We use the standard support vector machine (SVM) classifier as a baseline classifier. To study the aspect of feature selection, we construct the baseline classifier with automatic feature selection method. Features are ranked according to their variance in the dataset and the top 1000 features with largest variance serve as an input space for the SVM To study the role of ensemble, we have constructed an Ensemble+SVM method that uses an ensemble architecture and SVM as a weak classifier We have also included a plain SRC method that directly uses all the features.

Experimental results when varying the ensemble size There are two parameters that might influence on the overall performance of our system. They are the size of ensemble and the number of features per classifier. This table illustrates the performance of the system when changing the size of the ensemble. All cases are given 1000 randomly selected features. The number of classifiers appears to affect both the precision and the recall. Average precision and recall of classifiers when varying the ensemble size (number of features = 1000)

Experimental results when varying the number of features per classifier But when we vary the number of feature per classifier, they appear to have a smaller effect. In this table, we use 50 weak classifiers, when increasing the number of sampled features from 500 to 2000, it does not appear to improve the results. We also observed that when the number of features is increased, the ensemble size should be increased as well to avoid overfitting. Average precisions and recalls of classifiers when varying number of features (ensemble size = 50)

Conclusion Ensemble classification framework with sparse Gabor-feature based classifier Eliminates the requirement for segmentation and supervised feature selection Reduces the need for manual adjustment Achieves reasonable results compared to segmentation dependent techniques (Gong et al.) Limitation Longer classification time when dealing with large training data Manual weighting needed for imbalanced data We have introduced an ensemble classification framework with sparse Gabor-feature based classifiers The system does not require segmentation or supervised feature selection. This reduces the need for manual work in extracting useful information from medical images. Our experiments show that in the domain of classifying TBI images, we achieve reasonable results to segmentation-dependent techniques that rely on manually selected handcrafted features Although the proposed approach does not make any modality dependent assumptions, testing it on other modalities is an obvious topic for future work. The proposed method is non-parametric, thus when more training data is used, the classification will take a longer time. Finally, manual weighting is still needed when dealing imbalanced data

Thank you

Gabor features Localize low level features from an input image Resemble the primitive features extracted by human visual cortex Extract edge like features in different scales and orientations at different locations of the image Create a Gabor filter bank with 5 frequencies and 8 orientations A 128 x 128 grayscale image: 655360 features Randomly select 4000 Gabor features to form a feature subspace No domain specific feature Automatic Generic low level features