DITL & APNIC George Michaelson APNIC 26, Christchurch.

Slides:



Advertisements
Similar presentations
Hadoop at ContextWeb February ContextWeb: Traffic Traffic – up to 6 thousand Ad requests per second. Comscore Trend Data:
Advertisements

LNL CMS M.Biasotto, Padova, 24 aprile Farm monitoring Massimo Biasotto - LNL.
1 Introduction of IPv6 in Vienna University's LAN, ACOnet, VIX, ccTLD name service Wilfried Wöber, UniVie - ACOnet for FLIP-6, San Jose, CR Octoberber.
Measuring IPv6 Deployment Geoff Huston George Michaelson
1 Technical Area highlights and priorities APNIC Member Meeting 29 February 2008 APNIC 25, Taipei George Michaelson.
APNIC Update AfriNIC 10 May2009. Overview APNIC 27 policy outcomes APNIC Members and Stakeholder Survey IPv6 Program Research and development activities.
MoFo and Disaster Recovery August 24, 2004 Bruce Friedman Infrastructure Manager
NGS computation services: API's,
DITL George Michaelson APNIC
Whats New In Dream Report 4.5 Renee Sikes Applications Engineer Dream Report Brand Manager.
Extreme Performance with Oracle Data Warehousing
Computing Infrastructure
SHARKFEST '09 | Stanford University | June 15–18, 2009 The Reality of 10G Analysis Presented by: Network Critical Wednesday, June 17 th, :30 pm –
Chapter 1: Introduction to Scaling Networks
FIBRE-BR Meeting GENI I&M Marcelo Pinheiro. Agenda GENI Overview GENI User groups GENI I&M Use Cases GENI I&M Services.
HetnetIP Ethernet BackHaul Configuration Automation Demo.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
Measuring IPv6 Geoff Huston APNIC Labs, February 2014.
Merit’s CALEA Compliance Architecture and Platform, “OpenCALEA” Mary Eileen McLaughlin, Merit - Director Technical Operations Manish Karir, Merit - Research.
Introduction to Network Analysis and Sniffer Pro
Shared File Service VM Forum January, SFS Topics Targeted Usage Security Accessing CIFS Shares Availability & Protection Monitoring Pricing.
The new APNIC DNS generation system. Previous System Direct access to backend whois.db files – Constructed radix tree in memory from domain objects –
Internet Traffic Patterns Learning outcomes –Be aware of how information is transmitted on the Internet –Understand the concept of Internet traffic –Identify.
I/O Channels I/O devices getting more sophisticated e.g. 3D graphics cards CPU instructs I/O controller to do transfer I/O controller does entire transfer.
High Performance Computing Course Notes High Performance Storage.
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
Peer-to-peer archival data trading Brian Cooper and Hector Garcia-Molina Stanford University.
Passive traffic measurement Capturing actual Internet packets in order to measure: –Packet sizes –Traffic volumes –Application utilisation –Resource utilisation.
Developing PANDORA Mark Corbould Director, IT Business Systems.
Empowering Business in Real Time. © Copyright 2009, OSIsoft Inc. All rights Reserved. Data Center & IT Monitoring Use Cases Regional Seminar Series Carolyn.
D-Link International Private Limited Training and Staff Development Department Module : Network Attached Storage Module : Network Attached Storage.
CSE 321b Computer Organization (2) تنظيم الحاسب (2) 3 rd year, Computer Engineering Winter 2015 Lecture #4 Dr. Hazem Ibrahim Shehata Dept. of Computer.
Computer Security: Principles and Practice First Edition by William Stallings and Lawrie Brown Lecture slides by Lawrie Brown Chapter 8 – Denial of Service.
DNS root server deployments George Michaelson DNS operations SIG APNIC17/APRICOT 2004 Feb KL, Malaysia.
COEN 252 Computer Forensics
Step-by-Step Intrusion Detection using TCPdump SHADOW.
Hosted by Case Study - Storage Consolidation Steve Curry Yahoo Inc.
Cybersecurity Coordination and Cooperation Colloquium (f41lf3st 2015) 17 June 2015 Tallinna Tehnickaülikool, Tallinn, Estonia IPv6 Golden Networks Jeroen.
Inventory:OCSNG + GLPI Monitoring: Zenoss 3
COEN 252 Computer Forensics Collecting Network-based Evidence.
Planning and Designing Server Virtualisation.
DNF Security | Cabot Blvd, Hayward, CA | | Presenter: Joe Cuellar Sales Manager - DNF Security Company/Product.
You’ll get a lot more from the course if you ask questions. If I don’t know the answer, I will find out before the end of the course.
| nectar.org.au NECTAR TRAINING Module 9 Backing up & Packing up.
Measuring IPv6 Deployment Geoff Huston George Michaelson
Block1 Wrapping Your Nugget Around Distributed Processing.
Tony McGregor RIPE NCC Visiting Researcher The University of Waikato DAR Active measurement in the large.
Securing and Monitoring 10GbE WAN Links Steven Carter Center for Computational Sciences Oak Ridge National Laboratory.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
Clustering In A SAN For High Availability Steve Dalton, President and CEO Gadzoox Networks September 2002.
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
Operating Systems Proj.. Background A firewall is an information technology (IT) security device which is configured to permit, deny or proxy data connections.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
The Client-Server Model And the Socket API. Client-Server (1) The datagram service does not require cooperation between the peer applications but such.
 The End to the Means › (According to IBM ) › 03.ibm.com/innovation/us/thesmartercity/in dex_flash.html?cmp=blank&cm=v&csr=chap ter_edu&cr=youtube&ct=usbrv111&cn=agus.
“OpenCALEA” Pragmatic Cost Effective CALEA Compliance Manish Karir, Merit - Research and Development.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
© 2015 Pittsburgh Supercomputing Center Opening the Black Box Using Web10G to Uncover the Hidden Side of TCP CC PI Meeting Austin, TX September 29, 2015.
Sven Ubik, Aleš Friedl CESNET TNC 2009, Malaga, Spain, 11 June 2009 Experience with passive monitoring deployment in GEANT2 network.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
CPS110: Networks Landon Cox March 25, Network hardware reality  Lots of different network interface cards (NICs)  3Com/Intel, Ethernet/802.11x.
Video Security Design Workshop:
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
“OpenCALEA” Pragmatic Cost Effective CALEA Compliance
Diskpool and cloud storage benchmarks used in IT-DSS
Module 2: DriveScale architecture and components
Distributed Network Traffic Feature Extraction for a Real-time IDS
OASIS Deployment: VideoCollective
PerformanceBridge Application Suite and Practice 2.0 IT Specifications
CS 295: Modern Systems Organizing Storage Devices
Presentation transcript:

DITL & APNIC George Michaelson APNIC 26, Christchurch

Overview What is DITL? How APNIC deployed data-capture Outcomes Where to from here?

What is DITL Day In The Life of the Internet –24hours of data capture –A snapshot of what really goes on –Collect, not sample or just measure Opportunity for long-life research –Compare to past years, future years –what if questions on data after the event –Data archiving, future unplanned uses Rights of use has to be managed

What is DITL (cont) Organized by CAIDA –Cooperative Association for Internet Data Analysis –Safe harbour for data Can observe different T&C to access, publish Long-term data storage –Data storage, collation facilitated by OARC thumper RAID TB filestore S/W support, operations/management

24 hours of data capture!!! …actually, it was 48 hours –Worldwide coordination required some timezone arithmetic …which some of us got a bit wrong! -(ie me…) –Previous experience showed getting complete data for a given 24h window was difficult without more data to select from Yes. Its a lot of data –APNICs contribution was ~ 300Gb alone

Participation 156 points of measurement 24 agencies –Roots, ccTLD, RIR, R&D 2Tb of data collected over 48hr+ window APNICs contribution –329Gb –All NS, primary (in-region) and secondary (cctLD, other RIR) –Passed via SSH to OARC over 96h period

Lessons learned Practice makes perfect.. –Timezone arithmetic –Smaller blocks, sent more often, parallelize Cant scale disk to infinity –Capture designed for 2+ weeks retention –DNS load growth exceeded expectations Sample or Measure? –Jury still out: both have merits, depending –Why not do both? Get samples from capture (used for measurement)

How to capture? Originally used on-server tcpdump –Higher CPU load, can impact service –More chance of packetloss (design of packet filter in kernel drops for the collector, not the real socket endpoint) –Servers not designed for this scale of data capture Clearly needs re-design for DITL –More disk –More CPU –More flexibility

Passive TAP approach Rejected switch-level portspan –Imposes switchload, uncomfortable with remote deployment and switch reconfig Selected a 1Gig capable TAP, copper –Futureproof for re-scaling (current service delivery is 100mbit) –Fibre service delivery not yet common in locations APNIC provides service to –Vendor-backed 1-packet cost to switchover Dual power supply, failure mode is passive connector Alternate second tap on unit, permits onsite checks –No kernel mods, no OS dependencies Libpcap, tcpdump well understood

APNIC DNS measurements 5+ year investment in DNS measurements –sample-based, not measurement of all DNS –15min duty cycle, 1min packet collector Using tcpdump already –Counting by economy, V4/V6, type Also OARC DSC –Measurement of all DNS –Distributed collector, central repository model Ssh, rsync, push model to feed repository –Libpcap based collector (lib-level tcpdump) –XML data retention, mapped into simple CSV file –Graphing tools on web (perl) Both Ideal to move to TAP based model –No loss of historical data series

The APNIC DNS measurement node Redhat EL -series OS Mirrored 750Gb disks –Typically 3-day data retention, up to 10 days for smaller traffic node 3-ethernet model (PCI card) –Management I/f distinct from capture port –Node can measure up to 2 DNS servers onsite Packet capture feeds sample, DSC collections –Pushed to central collector on management I/f

DITL outcomes Published at CAIDA website eg: DITL 2008 Analysis Sebastian Castro Cooperative Association for Internet Data Analysis - CAIDA San Diego Supercomputer Center, University of California, San Diego A Few highlights..