Ranking Electrical Feeders of the New York City Power Grid Phil Gross Ansaf Salleb-Abouissi Haimonti Dutta Albert Boulanger Problem Primary electricity.

Slides:



Advertisements
Similar presentations
Adjusting Active Basis Model by Regularized Logistic Regression
Advertisements

Control Case Common Always active
An Introduction to Boosting Yoav Freund Banter Inc.
Elif Kongar*, Mahesh Baral and Tarek Sobh *Departments of Technology Management and Mechanical Engineering University of Bridgeport, Bridgeport, CT, U.S.A.
Feature selection and transduction for prediction of molecular bioactivity for drug design Reporter: Yu Lun Kuo (D )
Sensor-Based Abnormal Human-Activity Detection Authors: Jie Yin, Qiang Yang, and Jeffrey Junfeng Pan Presenter: Raghu Rangan.
Proprietary information – Columbia University. All rights reserved, 2009 – Using Historical and Real-Time Data to Optimize Reliability for Power.
Adaption Adjusting Model’s parameters for a new speaker. Adjusting all parameters need a huge amount of data (impractical). The solution is to cluster.
Application of Stacked Generalization to a Protein Localization Prediction Task Melissa K. Carroll, M.S. and Sung-Hyuk Cha, Ph.D. Pace University, School.
An Overview of Machine Learning
Predicting Electricity Distribution Feeder Failures using Machine Learning Marta Arias 1, Hila Becker 1,2 1 Center for Computational Learning Systems 2.
Simulation Metamodeling using Dynamic Bayesian Networks in Continuous Time Jirka Poropudas (M.Sc.) Aalto University School of Science and Technology Systems.
Review of : Yoav Freund, and Robert E
Introduction to Boosting Slides Adapted from Che Wanxiang( 车 万翔 ) at HIT, and Robin Dhamankar of Many thanks!
Standard electrode arrays for recording EEG are placed on the surface of the brain. Detection of High Frequency Oscillations Using Support Vector Machines:
Text Classification With Support Vector Machines
Properties of Machine Learning Applications for Use in Metamorphic Testing Chris Murphy, Gail Kaiser, Lifeng Hu, Leon Wu Columbia University.
Robust Real-time Object Detection by Paul Viola and Michael Jones ICCV 2001 Workshop on Statistical and Computation Theories of Vision Presentation by.
Cancer classification using Machine Learning Techniques on Microarray Data Yongjin Park 1 and Ming-Chi Tsai 2 1 Department of Biology, Computational Biology.
Data Mining By Archana Ketkar.
The Implicit Mapping into Feature Space. In order to learn non-linear relations with a linear machine, we need to select a set of non- linear features.
Parameterizing Random Test Data According to Equivalence Classes Chris Murphy, Gail Kaiser, Marta Arias Columbia University.
Active Learning Strategies for Compound Screening Megon Walker 1 and Simon Kasif 1,2 1 Bioinformatics Program, Boston University 2 Department of Biomedical.
The Social Web: A laboratory for studying s ocial networks, tagging and beyond Kristina Lerman USC Information Sciences Institute.
Randomized Variable Elimination David J. Stracuzzi Paul E. Utgoff.
AbstractAbstract Adjusting Active Basis Model by Regularized Logistic Regression Ruixun Zhang Peking University, Department of Statistics and Probability.
AdaBoost Robert E. Schapire (Princeton University) Yoav Freund (University of California at San Diego) Presented by Zhi-Hua Zhou (Nanjing University)
Storage Allocation in Prefetching Techniques of Web Caches D. Zeng, F. Wang, S. Ram Appeared in proceedings of ACM conference in Electronic commerce (EC’03)
Data Mining Joyeeta Dutta-Moscato July 10, Wherever we have large amounts of data, we have the need for building systems capable of learning information.
Anomaly detection with Bayesian networks Website: John Sandiford.
EADS DS / SDC LTIS Page 1 7 th CNES/DLR Workshop on Information Extraction and Scene Understanding for Meter Resolution Image – 29/03/07 - Oberpfaffenhofen.
General Information Course Id: COSC6342 Machine Learning Time: TU/TH 10a-11:30a Instructor: Christoph F. Eick Classroom:AH123
1 Information Geometry on Classification Logistic, AdaBoost, Area under ROC curve Shinto Eguchi – – ISM seminor on 17/1/2001 This talk is based on one.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
The Disputed Federalist Papers: Resolution via Support Vector Machine Feature Selection Olvi Mangasarian UW Madison & UCSD La Jolla Glenn Fung Amazon Inc.,
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
Data Mining with Oracle using Classification and Clustering Algorithms Proposed and Presented by Nhamo Mdzingwa Supervisor: John Ebden.
Wang Y 1,2, Damaraju S 1,3,4, Cass CE 1,3,4, Murray D 3,4, Fallone G 3,4, Parliament M 3,4 and Greiner R 1,2 PolyomX Program 1, Department.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
1 Ensembles An ensemble is a set of classifiers whose combined results give the final decision. test feature vector classifier 1classifier 2classifier.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
Xiangnan Kong,Philip S. Yu Multi-Label Feature Selection for Graph Classification Department of Computer Science University of Illinois at Chicago.
1 Chapter 6. Classification and Prediction Overview Classification algorithms and methods Decision tree induction Bayesian classification Lazy learning.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
CSSE463: Image Recognition Day 33 This week This week Today: Classification by “boosting” Today: Classification by “boosting” Yoav Freund and Robert Schapire.
Discriminative Frequent Pattern Analysis for Effective Classification By Hong Cheng, Xifeng Yan, Jiawei Han, Chih- Wei Hsu Presented by Mary Biddle.
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Carnegie Mellon School of Computer Science Language Technologies Institute CMU Team-1 in TDT 2004 Workshop 1 CMU TEAM-A in TDT 2004 Topic Tracking Yiming.
NTU & MSRA Ming-Feng Tsai
KAIST TS & IS Lab. CS710 Know your Neighbors: Web Spam Detection using the Web Topology SIGIR 2007, Carlos Castillo et al., Yahoo! 이 승 민.
Application of machine learning to RCF decision procedures Zongyan Huang.
Combining Evolutionary Information Extracted From Frequency Profiles With Sequence-based Kernels For Protein Remote Homology Detection Name: ZhuFangzhi.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
Machine Learning in CSC 196K
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
Knowledge Discovery and Data Mining 19 th Meeting Course Name: Business Intelligence Year: 2009.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science John Cavazos J Eliot B Moss Architecture and Language Implementation Lab University.
Gist 2.3 John H. Phan MIBLab Summer Workshop June 28th, 2006.
General Information Course Id: COSC6342 Machine Learning Time: TU/TH 1-2:30p Instructor: Christoph F. Eick Classroom:AH301
1 Ensembles An ensemble is a set of classifiers whose combined results give the final decision. test feature vector classifier 1classifier 2classifier.
Intelligent and Adaptive Systems Research Group A Novel Method of Estimating the Number of Clusters in a Dataset Reza Zafarani and Ali A. Ghorbani Faculty.
Introduction to Machine Learning, its potential usage in network area,
In Search of the Optimal Set of Indicators when Classifying Histopathological Images Catalin Stoean University of Craiova, Romania
Source: Procedia Computer Science(2015)70:
Finding Clusters within a Class to Improve Classification Accuracy
Classification of class-imbalanced data
IEEE World Conference on Computational Intelligence – WCCI 2010
3.1.1 Introduction to Machine Learning
Presentation transcript:

Ranking Electrical Feeders of the New York City Power Grid Phil Gross Ansaf Salleb-Abouissi Haimonti Dutta Albert Boulanger Problem Primary electricity distribution cables (feeders,  in diagram), connecting substations to local transformers, fail regularly Successful prediction of which feeder is likely to fail next would have major benefits for both operations and maintenance of the electrical grid [1] Outage-Derived Data Sets (ODDS) are a mixture of feeder attribute snapshots from just before failures, labeled as +1, and snapshots of all feeders in their current state, labeled as -1 Ranking Methods Supervised Ranking Problem: Given a set of labeled examples, we would like a scoring hypothesis h that will rank all positive examples above all negative examples. We use Area under the ROC [2] to evaluate our ranking results. SVM: Svmlight, linear kernel; Output scores not thresholded, in order to generate ranking; Vapnik[3] Rankboost: Freund, Iyer, Schapire, Singer [4] MartiRank: Long and Servedio[5], Gross, et al. [6] Center for Computational Learning Systems, Columbia University Experiment Compare the effectiveness of several machine learning techniques for generating feeder rankings Focus is on summer performance, when electricity consumption is much higher due to air conditioner usage, and feeder failures far more frequent Each example has 237 attributes Models were trained over a fixed 45-day window previous to date under consideration Testing periods were 7 and 15 following date under consideration We want to test if feeders that actually failed were ranked high on the susceptibility list Results Rankboost and SVM Score Ranker have comparable results, while MartiRank has slightly worse performance Prediction over 7 days appears to be better than prediction over 15 days The system appears to exhibit concept drift, where the amount of historical data for optimal prediction occasionally plummets All algorithms agree on the time periods when such concept drifts occur Future Work : Use of multivariate feature selection mechanisms Better ways of aggregating time series data Integrating analysis of feeder subcomponents (cable sections, joints, transformers) Decision Tree-based ranking methods (Probability Estimate Trees (PETs) Tracking and Prediction of concept drift Estimation of feeder Time To Failure Recognition of imminent failure from dynamic data Comparative AUC Results Concept Drift: Optimal Window Sizes Examined 35 datasets of 7 days each Attempted to predict each set of outages using up to 6 previous datasets References [1] P. Gross, A. Salleb-Aouissi, H. Dutta, A. Boulanger. Susceptibility Ranking of Electrical Feeders: A Case Study. CCLS Tech Report CCLS-08-04, [2] A. P. Bradley. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognition, 30(7):1145–1159, July [3] V. N. Vapnik. The nature of statistical learning theory. Springer-Verlag New York, Inc., New York, NY, USA, [4] Y. Freund, R. D. Iyer, R. E. Schapire, and Y. Singer. An efficient boosting algorithm for combining preferences. In ICML ’98: Proceedings of the Fifteenth International Conference on Machine Learning, pages 170–178, San Francisco, CA, USA, Morgan Kaufmann Publishers Inc. [5] P. M. Long and R. A. Servedio. Martingale boosting. In 18th Annual Conference on Learning Theory, Bertinoro, Italy, June 27-30, 2005, pages 79–94. Springer, [6] P. Gross et al. Predicting electricity distribution feeder failures using machine learning susceptibility analysis. In The Eighteenth Conference on Innovative Applications of Artificial Intelligence IAAI-06, Boston, Massachusetts, Acknowledgements Thanks to Maggie Chow, Arthur Kressner, Serena Lee, and many others at Consolidated Edison Company of New York, Charles Lawson at Digimedia, Marta Arias at Universitat Politècnica de Catalunya and David Waltz, Roger Anderson, Cynthia Rudin, Bert Huang, Hatim Diab, Sam Lee and Leon Wu at the Center for Computational Learning Systems, Columbia University, New York. Funding for this work is provided by a contract between Columbia University and the Consolidated Edison Company of New York. Summer Temperature Effect Data a mixture of static attributes describing cable structure and connections, and dynamic attributes describing the feeder’s environ- ment over time