Metadata-driven Threat Classification of Network Endpoints Appearing in Malware Andrew G. West and Aziz Mohaisen (Verisign Labs) July 11, 2014 – DIMVA.

Slides:



Advertisements
Similar presentations
Wenke Lee and Nick Feamster Georgia Tech Botnet and Spam Detection in High-Speed Networks.
Advertisements

3.02H Publishing a Website 3.02 Develop webpages..
A Survey of Botnet Size Measurement PRESENTED: KAI-HSIANG YANG ( 楊凱翔 ) DATE: 2013/11/04 1/24.
Detecting Malicious Flux Service Networks through Passive Analysis of Recursive DNS Traces Roberto Perdisci, Igino Corona, David Dagon, Wenke Lee ACSAC.
An Introduction of Botnet Detection – Part 2 Guofei Gu, Wenke Lee (Georiga Tech)
RB-Seeker: Auto-detection of Redirection Botnet Presenter: Yi-Ren Yeh Authors: Xin Hu, Matthew Knysz, Kang G. Shin NDSS 2009 The slides is modified from.
1 CANTINA : A Content-Based Approach to Detecting Phishing Web Sites WWW Yue Zhang, Jason Hong, and Lorrie Cranor.
Design and Evaluation of a Real-Time URL Spam Filtering Service
Kindred Domains: Detecting and Clustering Botnet Domains Using DNS Traffic Matt Thomas Data Architect, Verisign Labs.
Understanding the Network-Level Behavior of Spammers Mike Delahunty Bryan Lutz Kimberly Peng Kevin Kazmierski John Thykattil By Anirudh Ramachandran and.
Mining Behavior Models Wenke Lee College of Computing Georgia Institute of Technology.
Botnets Abhishek Debchoudhury Jason Holmes. What is a botnet? A network of computers running software that runs autonomously. In a security context we.
Threat infrastructure: proxies, botnets, fast-flux
Bayesian Bot Detection Based on DNS Traffic Similarity Ricardo Villamarín-Salomón, José Carlos Brustoloni Department of Computer Science University of.
BOTNETS & TARGETED MALWARE Fernando Uribe. INTRODUCTION  Fernando Uribe   IT trainer and Consultant for over 15 years specializing.
Norman SecureSurf Protect your users when surfing the Internet.
Automated malware classification based on network behavior
Presentation by Kathleen Stoeckle All Your iFRAMEs Point to Us 17th USENIX Security Symposium (Security'08), San Jose, CA, 2008 Google Technical Report.
11 The Ghost In The Browser Analysis of Web-based Malware Reporter: 林佳宜 Advisor: Chun-Ying Huang /3/29.
URLDoc: Learning to Detect Malicious URLs using Online Logistic Regression Presented by : Mohammed Nazim Feroz 11/26/2013.
PhishNet: Predictive Blacklisting to Detect Phishing Attacks Pawan Prakash Manish Kumar Ramana Rao Kompella Minaxi Gupta Purdue University, Indiana University.
B OTNETS T HREATS A ND B OTNETS DETECTION Mona Aldakheel
WARNINGBIRD: A Near Real-time Detection System for Suspicious URLs in Twitter Stream.
Jay Stokes, Microsoft Research John Platt, Microsoft Research Joseph Kravis, Microsoft Network Security Michael Shilman, ChatterPop, Inc. ALADIN: Active.
Niels Provos and Panayiotis Mavrommatis Google Google Inc. Moheeb Abu Rajab and Fabian Monrose Johns Hopkins University 17 th USENIX Security Symposium.
Discovery of Emergent Malicious Campaigns in Cellular Networks Nathaniel Boggs, Wei Wang, Suhas Mathur, Baris Coskun, Carol Pincock © 2013 AT&T Intellectual.
Beyond Blacklists: Learning to Detect Malicious Web Sites from Suspicious URLs Justin Ma, Lawrence Saul, Stefan Savage, Geoff Voelker Computer Science.
1 All Your iFRAMEs Point to Us Mike Burry. 2 Drive-by downloads Malicious code (typically Javascript) Downloaded without user interaction (automatic),
BotMiner: Clustering Analysis of Network Traffic for Protocol- and Structure-Independent Botnet Detection Guofei Gu, Roberto Perdisci, Junjie Zhang, and.
Speaker:Chiang Hong-Ren Botnet Detection by Monitoring Group Activities in DNS Traffic.
Reporter: Li, Fong Ruei National Taiwan University of Science and Technology 9/19/2015Slide 1 (of 32)
Symantec Targeted Attack Protection 1 Stopping Tomorrow’s Targeted Attacks Today iPuzzlebiz
Automated Classification and Analysis of Internet Malware M. Bailey J. Oberheide J. Andersen Z. M. Mao F. Jahanian J. Nazario RAID 2007 Presented by Mike.
Automatically Generating Models for Botnet Detection Presenter: 葉倚任 Authors: Peter Wurzinger, Leyla Bilge, Thorsten Holz, Jan Goebel, Christopher Kruegel,
What’s New in WatchGuard XCS v9.1 Update 1. WatchGuard XCS v9.1 Update 1  Enhancements that improve ease of use New Dashboard items  Mail Summary >
Not So Fast Flux Networks for Concealing Scam Servers Theodore O. Cochran; James Cannady, Ph.D. Risks and Security of Internet and Systems (CRiSIS), 2010.
Week 10-11c Attacks and Malware III. Remote Control Facility distinguishes a bot from a worm distinguishes a bot from a worm worm propagates itself and.
Wide-scale Botnet Detection and Characterization Anestis Karasaridis, Brian Rexroad, David Hoeflin In First Workshop on Hot Topics in Understanding Botnets,
1 Behind Phishing: An Examination of Phisher Modi Operandi Speaker: Jun-Yi Zheng 2010/05/10.
Understanding the Human Network Martin Kruger LCDR Jodie Gooby November 2008.
Leveraging Asset Reputation Systems to Detect and Prevent Fraud and Abuse at LinkedIn Jenelle Bray Staff Data Scientist Strata + Hadoop World New York,
Cross-Analysis of Botnet Victims: New Insights and Implication Seungwon Shin, Raymond Lin, Guofei Gu Presented by Bert Huang.
Understanding the Network-Level Behavior of Spammers Author: Anirudh Ramachandran, Nick Feamster SIGCOMM ’ 06, September 11-16, 2006, Pisa, Italy Presenter:
Detecting Phishing in s Srikanth Palla Ram Dantu University of North Texas, Denton.
© 2009 WatchGuard Technologies WatchGuard ReputationAuthority Rejecting Unwanted & Web Traffic at the Perimeter.
Studying Spamming Botnets Using Botlab
Security Analytics Thrust Anthony D. Joseph (UCB) Rachel Greenstadt (Drexel), Ling Huang (Intel), Dawn Song (UCB), Doug Tygar (UCB)
Search Worms, ACM Workshop on Recurring Malcode (WORM) 2006 N Provos, J McClain, K Wang Dhruv Sharma
The Koobface Botnet and the Rise of Social Malware Kurt Thomas David M. Nicol
Trends and Lessons from Three Years Fighting Malicious Extensions Nav Jagpal, Eric Dingle, Jean-Philippe, Gravel Panayiotis, Mavrommatis Niels, Provos.
BotMiner: Clustering Analysis of Network Traffic for Protocol- and Structure-Independent Botnet Detection Presented by D Callahan.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
1 Modeling and Measuring Botnets David Dagon, Wenke Lee Georgia Institute of Technology Cliff C. Zou Univ. of Central Florida Funded by NSF CyberTrust.
A Framework for Detection and Measurement of Phishing Attacks Reporter: Li, Fong Ruei National Taiwan University of Science and Technology 2/25/2016 Slide.
Speaker: Hom-Jay Hom Date:2009/10/20 Botnet Research Survey Zhaosheng Zhu. et al July 28-August
Spamming Botnets: Signatures and Characteristics Yinglian Xie, Fang Yu, Kannan Achan, Rina Panigrahy, Microsoft Research, Silicon Valley Geoff Hulten,
2009/6/221 BotMiner: Clustering Analysis of Network Traffic for Protocol- and Structure- Independent Botnet Detection Reporter : Fong-Ruei, Li Machine.
Microsoft NDA Material Adwait Joshi Sr. Technical Product Manager Microsoft Corporation.
Brett Stone-Gross, Marco Cova, Lorenzo Cavallaro, Bob Gilbert, Martin Szydlowski, Richard Kemmerer, Christopher Kruegel, and Giovanni Vigna Proceedings.
Unveiling Zeus Automated Classification of Malware Samples Abedelaziz Mohaisen Omar Alrawi Verisign Inc, VA, USA Verisign Labs, VA, USA
Internet Vulnerabilities & Criminal Activity Internet Forensics 12.1 April 26, 2010 Internet Forensics 12.1 April 26, 2010.
Under the Shadow of sunshine
Domain Reputation Hussien Othman.
3.02H Publishing a Website 3.02 Develop webpages..
A lustrum of malware network communication: Evolution & insights
BotCatch: A Behavior and Signature Correlated Bot Detection Approach
4.02 Develop web pages using various layouts and technologies.
4.02 Develop web pages using various layouts and technologies.
The Domain Abuse Activity Reporting System (DAAR)
TRANCO: A Research-Oriented Top Sites Ranking Hardened Against Manipulation By Prudhvi raju G id:
Presentation transcript:

Metadata-driven Threat Classification of Network Endpoints Appearing in Malware Andrew G. West and Aziz Mohaisen (Verisign Labs) July 11, 2014 – DIMVA – Egham, United Kingdom

Verisign Public PRELIMINARY OBSERVATIONS 2 C&C instruction Drop sites Program code 1. HTTP traffic is ubiquitous in malware 2. Network identifiers are relatively persistent x Migration has economic consequences 4. Not all contacted endpoints are malicious

Verisign Public OUR APPROACH AND USE-CASES Our approach Obtained 28,000 *expert-labeled* endpoints, including both threats (of varying severity) and non-threats Learn *static* endpoint properties that indicate class: Lexical: URL structure WHOIS: Registrars and duration n-gram: Token patterns Reputation: Historical learning USE-CASES: Analyst prioritization; automatic classification TOWARDS: Efficient and centrally administrated blacklisting End-game takeaways Prominent role of DDOS and “shared” use services 99.4% accuracy at binary task; 93% at severity prediction 3

Verisign Public Classifying Endpoints with Network Metadata 1. Obtaining and labeling malware samples 2. Basic/qualified corpus properties 3. Feature derivation [lexical, WHOIS, n-gram, reputation] 4. Model-building and performance 4

Verisign Public SANDBOXING & LABELING WORKFLOW 5 Malware samples AUTONOMOUS AutoMal (sandbox) processes registry filesystem network PCAP file HTTP endpoint parser Potential Threat-indicators ANALYST-DRIVEN Non-threat High-threat Med-threat Low-threat Is there a benign use-case? How severe is the threat? domain.com Can we generalize this claim?

Verisign Public EXPERT-DRIVEN LABELING Where do malware samples come from? Organically, customers, industry partners 93k samples → 203k endpoints → 28k labeled 6 LEVELEXAMPLES LOW-THREAT“Nuisance” malware; ad-ware MED-THREATUntargeted data theft; spyware; banking trojans HIGH-THREATTargeted data theft; corporate and state espionage Fig: Num. of malware MD5s per corpus endpoint os.solvefile.com 1901 binaries Verisign analysts select potential endpoints Expert labeling is not very common [1] Qualitative insight via reverse engineering Selection biases

Verisign Public Classifying Endpoints with Network Metadata 1. Obtaining and labeling malware samples 2. Basic/qualified corpus properties 3. Feature derivation [lexical, WHOIS, n-gram, reputation] 4. Model-building and performance 7

Verisign Public BASIC CORPUS STATISTICS 4 of 5 endpoint labels can be generalized to domain granularity Some 4067 unique second level domains (SLDs) in data; statistical weighting 8 TOTAL28,077 DOMAINS21, % high-threat med-threat low-threat non-threat 5, ,139 4, % 0.5% 52.8% 19.4% URLS7, % high-threat med-threat low-threat non-threat 318 1,299 2,005 3, % 18.6% 28.6% 48.3% Tab: Corpus composition by type and severity 73% of endpoints are threats (63% estimated in full set)

Verisign Public QUALITATIVE ENDPOINT PROPERTIES System doesn’t care about content; for audience benefit… What lives at malicious endpoints? Binaries: Complete program code; pay-per-install Botnet C&C: Instruction sets of varying creativity Drop sites: HTTP POST to return stolen data What lives at “benign” endpoints? Reliable websites (connectivity tests or obfuscation) Services reporting on infected hosts (IP, geo-location, etc.) Advertisement services (click-fraud malware) Images (hot-linked for phishing/scare-ware purposes) 9

Verisign Public COMMON SECOND-LEVEL DOMAINS Non-dedicated and shared-use settings are problematic 10 THREATSNON-THREATS SLD# # 3322.ORG 2,172 YTIMG.COM 1,532 NO-IP.BIZ 1,688 PSMPT.COM 1,277 NO-IP.ORG 1,060 BAIDU.COM 920 ZAPTO.ORG 719 GOOGLE.COM 646 NO-IP.INFO 612 AKAMAI.NET 350 PENTEST[…].TK 430 YOUTUBE.COM 285 SURAS-IP.COM ORG 243 Tab: Second-level domains (SLDs) parent to the most number of endpoints, by class. 6 of 7 top threat SLDs are DDNS services; cheap and agile Sybil accounts as a labor sink; cheaply serving content along distinct paths Motivation for reputation

Verisign Public Classifying Endpoints with Network Metadata 1. Obtaining and labeling malware samples 2. Basic/qualified corpus properties 3. Feature derivation [lexical, WHOIS, n-gram, reputation] 4. Model-building and performance 11

Verisign Public LEXICAL FEATURES: DOMAIN GRANULARITY DOMAIN TLD Why is ORG bad? Cost-effective TLDs Non-threats in COM/NET DOMAIN LENGTH & DOMAIN ALPHA RATIO Address memorability Lack of DGAs? (SUB)DOMAIN DEPTH Having one subdomain ( sub.domain.com ) often indicates shared-use settings; indicative of threats 12 Fig: Class patterns by TLD after normalization; Data labels indicate raw quantities

Verisign Public LEXICAL FEATURES: URL GRANULARITY Some features require URL paths to calculate: URL EXTENSION Extensions not checked Executable file types Standard textual web content & images 63 “other” file types; some appear fictional URL LENGTH & URL DEPTH Similar to domain case; not very indicative 13 Fig: Behavioral distribution over file extensions (URLs only); Data labels indicate raw quantity

Verisign Public WHOIS DERIVED FEATURES 55% zone coverage; zones nearly static DOMAIN REGISTRAR* Related work on spammers [2] MarkMonitor’s customer base and value-added services Laggard’s often exhibit low cost, weak enforcement, or bulk registration support [2] 14 Fig: Behavioral distribution over popular registrars (COM/NET/CC/TV) * DISCLAIMER: Recall that SLDs of a single customer may dramatically influence a registrar’s behavioral distribution. In no way should this be interpreted as an indicator of registrar quality, security, etc.

Verisign Public WHOIS DERIVED FEATURES DOMAIN AGE 40% of threats <1 year 2.5 years for threats vs non-threat Recall shared-use settings Economic factors, which in turn relates to… DOMAIN REG PERIOD Rarely more than 5 years for threat domains DOMAIN AUTORENEW 15 Fig: CDF for domain age (reg. to malware label)

Verisign Public N-GRAM ANALYSIS 16 mailnewsapisfreeeasy koreadateyahoosoftmicro onlinewinsupdateportwinsoft Tab: Dictionary tokens most indicative of threat domains

Verisign Public DOMAIN BEHAVIORAL REPUTATION DOMAIN REPUTATION Calculate “opinion” objects based on Beta probability distribution over a binary feedback model [3] Reputation bounded on [0,1], initialized at 0.5 Novel non-threat SLDs are exceedingly rare Area-between-curve indicates SLD behavior is quite consistent CAVEAT: Dependent on accurate prior classifications 17 Fig: CDF for domain reputation. All reputations held at any point in time are plotted.

Verisign Public Classifying Endpoints with Network Metadata 1. Obtaining and labeling malware samples 2. Basic/qualified corpus properties 3. Feature derivation [lexical, WHOIS, n-gram, reputation] 4. Model-building and performance 18

Verisign Public FEATURE LIST & MODEL BUILDING Random-forest model Decision tree ensemble Missing features Human-readable output WHOIS features (external data) are covered by others in problem space. Results presented w/10×10 cross-fold validation 19 FEATURETYPEIG DOM_REPUTATION real0.749 DOM_REGISTRAR enum0.211 DOM_TLD enum0.198 DOM_AGE real0.193 DOM_LENGTH int0.192 DOM_DEPTH int0.186 URL_EXTENSION enum0.184 DOM_TTL_RENEW int0.178 DOM_ALPHA real0.133 URL_LENGTH int0.028 [snip 3 ineffective features]..… Tab: Feature list as sorted by information-gain metric

Verisign Public PERFORMANCE EVALUATION (BINARY) BINARY TASK 99.47% accurate 148 errors in 28k cases No mistakes until 80% recall ROC-AUC 20 Fig: (inset) Entire precision-recall curve; (outset) focusing on the interesting portion of that PR curve

Verisign Public PERFORMANCE EVALUATION (SEVERITY) Severity task under- emphasized for ease of presentation % accurate Role of DOM_REGISTRAR ROC-AUC Prioritization is viable Classed→ Labeled ↓ NonLowMedHigh Non Low Med High Tab: Confusion matrix for severity task

Verisign Public DISCUSSION Remarkable in its simplicity; metadata works! [4] Being applied in production Scoring potential indicators discovered during sandboxing Preliminary results comparable to offline ones Gamesmanship Account/Sybil creation inside services with good reputation Use collaborative functionality to embed payloads on benign content (wikis, comments on news articles); what to do? Future work DNS traffic statistics to risk-assess false positives More DDNS emphasis: Monitor “A” records and TTL values Malware family identification (also expert labeled) 22

Verisign Public CONCLUSIONS “Threat indicators” and associated blacklists… … an established and proven approach (VRSN and others) … require non-trivial analyst labor to avoid false positives Leveraged 28k expert labeled domains/URLs contacted by malware during sandboxed execution Observed DDNS and shared-use services are common (cheap and agile for attackers), consequently an analyst labor sink Utilized cheap static metadata features over network endpoints Outcomes and applications Exceedingly accurate (99%) at detecting threats; reasonable at predicting severity (93%) Prioritize, aid, and/or reduce analyst labor 23

Verisign Public REFERENCES & ADDITIONAL READING [01] Mohaisen et al. “A Methodical Evaluation of Antivirus Scans and Labels”, WISA ‘13. [02] Hao et al. “Understanding the Domain Registration Behavior of Spammers”, IMC ‘13. [03] Josang et al. “The Beta Reputation System”, Bled eCommerce ‘02. [04] Hao et al. "Detecting Spammers with SNARE: Spatio-temporal Network-level Automatic Reputation Engine", USENIX Security ‘09. [05] Felegyhazi et al. “On the Potential of Proactive Domain Blacklisting”, LEET ‘10. [06] McGrath et al. “Behind Phishing: An Examination of Phisher Modi Operandi”, LEET ‘08. [07] Ntoulas et al. “Detecting Spam Webpages Through Content Analysis”, WWW ‘06. [08] Chang et al. “Analyzing and Defending Against Web-based Malware”, ACM Computing Surveys ‘13. [09] Provos et al. “All Your iFRAMEs Point to Us”, USENIX Security ‘09. [10] Antonakakis et al. “Building a Dynamic Reputation System for DNS”, USENIX Security ‘10. [11] Bilge et al. “EXPOSURE: Finding Malicious Domains Using Passive DNS …”, NDSS ‘11. [12] Gu et al. “BotSniffer: Detecting Botnet Command and Control Channels … ”, NDSS ‘08. 24

© 2014 VeriSign, Inc. All rights reserved. VERISIGN and other trademarks, service marks, and designs are registered or unregistered trademarks of VeriSign, Inc. and its subsidiaries in the United States and in foreign countries. All other trademarks are property of their respective owners.

Verisign Public RELATED WORK Endpoint analysis in security contexts Shallow URL properties leveraged in spam [5], phishing [6] Our results find the URL sets and feature polarity to be unique Mining content at endpoints; looking for commercial intent [7] Do re-use domain registration behavior of spammers [2] Sandboxed execution is an established approach [8] Machine assisted tagging alarmingly inconsistent [1] Network signatures of malware Google Safe Browsing [9]; drive-by-downloads, no passive endpoints Lots of work at DNS server level; a specialized perspective [10,11] Network flows as basis for C&C traffic [12] 26