Grid Datafarm and File System Services

Slides:



Advertisements
Similar presentations
National Institute of Advanced Industrial Science and Technology Belle/Gfarm Grid Experiment at SC04 Osamu Tatebe Grid Technology Research Center, AIST.
Advertisements

A Proposal of Capacity and Performance Assured Storage in The PRAGMA Grid Testbed Yusuke Tanimura 1) Hidetaka Koie 1,2) Tomohiro Kudoh 1) Isao Kojima 1)
Gfarm v2 and CSF4 Osamu Tatebe University of Tsukuba Xiaohui Wei Jilin University SC08 PRAGMA Presentation at NCHC booth Nov 19,
National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center Data Grids for Collection Federation Reagan W. Moore University.
Grid Datafarm Architecture for Petascale Data Intensive Computing Osamu Tatebe Grid Technology Research Center, AIST On behalf of the Gfarm project
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
The Google File System. Why? Google has lots of data –Cannot fit in traditional file system –Spans hundreds (thousands) of servers connected to (tens.
GGF Toronto Spitfire A Relational DB Service for the Grid Peter Z. Kunszt European DataGrid Data Management CERN Database Group.
Lee Center Workshop, May 19, 2006 Distributed Objects System with Support for Sequential Consistency.
Homework 2 In the docs folder of your Berkeley DB, have a careful look at documentation on how to configure BDB in main memory. In the docs folder of your.
GRID COMPUTING: REPLICATION CONCEPTS Presented By: Payal Patel.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
Hadoop Distributed File System by Swathi Vangala.
1 A Basic R&D for an Analysis Framework Distributed on Wide Area Network Hiroshi Sakamoto International Center for Elementary Particle Physics (ICEPP),
Presenter: Dipesh Gautam.  Introduction  Why Data Grid?  High Level View  Design Considerations  Data Grid Services  Topology  Grids and Cloud.
Hadoop & Condor Dhruba Borthakur Project Lead, Hadoop Distributed File System Presented at the The Israeli Association of Grid Technologies.
Data GRID Activity in Japan Yoshiyuki WATASE KEK (High energy Accelerator Research Organization) Tsukuba, Japan
High Performance File System Service for Cloud Computing Kenji Kobayashi, Osamu Tatebe University of Tsukuba, JAPAN.
Building Hierarchical Grid Storage Using the GFarm Global File System and the JuxMem Grid Data-Sharing Service Gabriel Antoniu, Lo ï c Cudennec, Majd Ghareeb.
The JuxMem-Gfarm Collaboration Enhancing the JuxMem Grid Data Sharing Service with Persistent Storage Using the Gfarm Global File System Gabriel Antoniu,
A Hierarchical MapReduce Framework Yuan Luo and Beth Plale School of Informatics and Computing, Indiana University Data To Insight Center, Indiana University.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
Introduction to HDFS Prasanth Kothuri, CERN 2 What’s HDFS HDFS is a distributed file system that is fault tolerant, scalable and extremely easy to expand.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
Large Scale Parallel File System and Cluster Management ICT, CAS.
Worldwide Fast File Replication on Grid Datafarm Osamu Tatebe 1, Youhei Morita 2, Satoshi Matsuoka 3, Noriyuki Soda 4, Satoshi Sekiguchi 1 1 Grid Technology.
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
GGF9 GFS WG BOF10/07/2003, Chicago Grid File System Group Proposal BOF Osamu Tatebe (AIST) Jane Xu (IBM) Arun Jagatheesan (SDSC)
Worldwide File Replication on Grid Datafarm Osamu Tatebe and Satoshi Sekiguchi Grid Technology Research Center, National Institute of Advanced Industrial.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
Introduction to HDFS Prasanth Kothuri, CERN 2 What’s HDFS HDFS is a distributed file system that is fault tolerant, scalable and extremely easy to expand.
Challenges of deploying Wide-Area-Network Distributed Storage System under network and reliability constraints – A case study
 Introduction  Architecture NameNode, DataNodes, HDFS Client, CheckpointNode, BackupNode, Snapshots  File I/O Operations and Replica Management File.
Google File System Robert Nishihara. What is GFS? Distributed filesystem for large-scale distributed applications.
National Institute of Advanced Industrial Science and Technology Gfarm Grid File System for Distributed and Parallel Data Computing Osamu Tatebe
Database CNAF Barbara Martelli Rome, April 4 st 2006.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition, Lecture 24: GFS.
ECE 456 Computer Architecture Lecture #9 – Input/Output Instructor: Dr. Honggang Wang Fall 2013.
PSM, Database requirements for POOL (File catalog performance requirements) Maria Girone, IT-DB Strongly based on input from experiments: subject.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Distributed File System. Outline Basic Concepts Current project Hadoop Distributed File System Future work Reference.
Design of File System Directory Services Osamu Tatebe Grid Technology Research Center, AIST GFS-WG, GGF10 March 2004, Berlin GGF10 GFS-WG March 2004, Berlin.
File System Directory Services Osamu Tatebe Grid Technology Research Center, AIST GFS-WG, F2F meeting April 2004, San Jose GFS-WG April 2004, San Jose.
1 Student Date Time Wei Li Nov 30, 2015 Monday 9:00-9:25am Shubbhi Taneja Nov 30, 2015 Monday9:25-9:50am Rodrigo Sanandan Dec 2, 2015 Wednesday9:00-9:25am.
An Introduction to GPFS
S. Pardi Computing R&D Workshop Ferrara 2011 – 4 – 7 July SuperB R&D on going on storage and data access R&D Storage Silvio Pardi
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
National Institute of Advanced Industrial Science and Technology Gfarm v2: A Grid file system that supports high-performance distributed and parallel data.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
Performance measurement of transferring files on the federated SRB
Grid File System Working Group
Grid File System WG GGF11, Honolulu June 8-9, 2004.
U.S. ATLAS Grid Production Experience
MapReduce: Simplified Data Processing on Large Clusters
LCG Deployment in Japan
PROOF – Parallel ROOT Facility
Realization of a stable network flow with high performance communication in high bandwidth-delay product network Y. Kodama, T. Kudoh, O. Tatebe, S. Sekiguchi.
Distributed Data Access and Resource Management in the D0 SAM System
Appro Xtreme-X Supercomputers
Auburn University COMP7330/7336 Advanced Parallel and Distributed Computing MapReduce - Introduction Dr. Xiao Qin Auburn.
Osamu Tatebe Grid Technology Research Center, AIST
Grid Canada Testbed using HEP applications
Hadoop Distributed Filesystem
Support for ”interactive batch”
Data services in gLite “s” gLite and LCG.
PROOF - Parallel ROOT Facility
by Mikael Bjerga & Arne Lange
Presentation transcript:

Grid Datafarm and File System Services Osamu Tatebe Grid Technology Research Center, National Institute of Advanced Industrial Science and Technology (AIST)

ATLAS/Grid Datafarm project: CERN LHC Experiment ~2000 physicists from 35 countries Detector for ALICE experiment Detector for LHCb experiment ATLAS Detector 40mx20m 7000 Tons LHC Perimeter 26.7km Truck Collaboration between KEK, AIST, Titech, and ICEPP, U Tokyo

Petascale Data-intensive Computing Requirements Peta/Exabyte scale files Scalable parallel I/O throughput > 100GB/s, hopefully > 1TB/s within a system and between systems Scalable computational power > 1TFLOPS, hopefully > 10TFLOPS Efficiently global sharing with group-oriented authentication and access control Resource Management and Scheduling System monitoring and administration Fault Tolerance / Dynamic re-configuration Global Computing Environment

Grid Datafarm (1): Global virtual file system [CCGrid 2002] World-wide virtual file system Transparent access to dispersed file data in a Grid Map from virtual directory tree to physical file Fault tolerance and access-concentration avoidance by file replication /grid ggf jp aist gtrc file1 file3 file2 file4 Virtual Directory Tree mapping File replica creation Grid File System

Grid Datafarm (2): High-performance data processing [CCGrid 2002] World-wide parallel and distributed processing Aggregate of files = superfile Data processing of superfiles = parallel and distributed data processing of member files Local file view File-affinity scheduling World-wide Parallel & distributed processing Virtual CPU Grid File System Newspapers in a year (superfile) 365 newspapers

Extreme I/O bandwidth support example: gfgrep - parallel grep gfmd gfarm:input Host1.ch Host2.ch Host3.ch Host4.jp Host5.jp % gfrun –G gfarm:input gfgrep –o gfarm:output regexp gfarm:input File affinity scheduling Host2.ch open(“gfarm:input”, &f1) create(“gfarm:output”, &f2) set_view_local(f1) set_view_local(f2) Host4.jp gfgrep input.2 input.4 output.4 output.2 output.5 output.3 output.1 Host1.ch Host5.jp Host3.ch grep regexp input.1 input.5 input.3 close(f1); close(f2) KEK.JP CERN.CH

Design of AIST Gfarm Cluster I Cluster node (High density and High performance) 1U, Dual 2.8GHz Xeon, GbE 800GB RAID with 4 3.5” 200GB HDDs + 3ware RAID 97 MB/s on writes, 130 MB/s on reads 80-node experimental cluster (operational from Feb 2003) Force10 E600 181st position in TOP500 (520.7 GFlops, peak 1000.8 GFlops) 70TB Gfarm file system with 384 IDE disks 7.7 GB/s on writes, 9.8 GB/s on reads for a 1.7TB file 1.6 GB/s (= 13.8 Gbps) on file replication of a 640GB file with 32 streams

World-wide Grid Datafarm Testbed Titech Tsukuba U AIST KEK Kasetsert U, Thiland SDSC Indiana U Total disk capacity: 80 TB, disk I/O bandwidth: 12 GB/s

Gfarm filesystem metadata File status File ID Owner, file type, access permission, access times Num. of fragments, a command history File fragment status File ID, fragment index Fragment file size, checksum type, checksum Directories List of file IDs and logical filenames Replica catalog File ID, fragment index, filesystem node Filesystem node status hostname, architecture, #CPUs, . . . File status Virtual File system Metadata Services File fragment Directories Replica Location Services Replica catalog Filesystem node Gfarm filesystem metadata

Filesystem metadata operation No direct manipulation Metadata is consistently managed via file operations only open() refers to the metadata close() updates or checks the metadata rename(), unlink(), chown(), chmod(), utime(), . . . New replication API Creation and deletion Inquiry and management