National Institute of Advanced Industrial Science and Technology Gfarm Grid File System for Distributed and Parallel Data Computing Osamu Tatebe

Slides:



Advertisements
Similar presentations
National Institute of Advanced Industrial Science and Technology Belle/Gfarm Grid Experiment at SC04 Osamu Tatebe Grid Technology Research Center, AIST.
Advertisements

Reports from Resource Breakout PRAGMA 16 KISTI, Korea.
A Proposal of Capacity and Performance Assured Storage in The PRAGMA Grid Testbed Yusuke Tanimura 1) Hidetaka Koie 1,2) Tomohiro Kudoh 1) Isao Kojima 1)
Gfarm v2 and CSF4 Osamu Tatebe University of Tsukuba Xiaohui Wei Jilin University SC08 PRAGMA Presentation at NCHC booth Nov 19,
SALSA HPC Group School of Informatics and Computing Indiana University.
High Performance Computing Course Notes Grid Computing.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
Job submission architectures in GRID environment Masamichi Ando M1 Student Taura Lab. Department of Information Science and Technology.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
Bondyakov A.S. Institute of Physics of ANAS, Azerbaijan JINR, Dubna.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks GINGIN Grid Interoperation on Data Movement.
Pooja Shetty Usha B Gowda.  Network File Systems (NFS)  Drawbacks of NFS  Parallel Virtual File Systems (PVFS)  PVFS components  PVFS application.
Presenter: Dipesh Gautam.  Introduction  Why Data Grid?  High Level View  Design Considerations  Data Grid Services  Topology  Grids and Cloud.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
1 School of Computer, National University of Defense Technology A Profile on the Grid Data Engine (GridDaEn) Xiao Nong
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Miguel Branco CERN/University of Southampton Enabling provenance on large-scale e-Science applications.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
Data GRID Activity in Japan Yoshiyuki WATASE KEK (High energy Accelerator Research Organization) Tsukuba, Japan
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
File and Object Replication in Data Grids Chin-Yi Tsai.
07:44:46Service Oriented Cyberinfrastructure Lab, Introduction to BOINC By: Andrew J Younge
High Performance File System Service for Cloud Computing Kenji Kobayashi, Osamu Tatebe University of Tsukuba, JAPAN.
Storage Tank in Data Grid Shin, SangYong(syshin, #6468) IBM Grid Computing August 23, 2003.
The Grid System Design Liu Xiangrui Beijing Institute of Technology.
InstantGrid: A Framework for On- Demand Grid Point Construction R.S.C. Ho, K.K. Yin, D.C.M. Lee, D.H.F. Hung, C.L. Wang, and F.C.M. Lau Dept. of Computer.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Building Hierarchical Grid Storage Using the GFarm Global File System and the JuxMem Grid Data-Sharing Service Gabriel Antoniu, Lo ï c Cudennec, Majd Ghareeb.
The JuxMem-Gfarm Collaboration Enhancing the JuxMem Grid Data Sharing Service with Persistent Storage Using the Gfarm Global File System Gabriel Antoniu,
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
Towards Exascale File I/O Yutaka Ishikawa University of Tokyo, Japan 2009/05/21.
Supporting Molecular Simulation-based Bio/Nano Research on Computational GRIDs Karpjoo Jeong Konkuk Suntae.
Worldwide File Replication on Grid Datafarm Osamu Tatebe and Satoshi Sekiguchi Grid Technology Research Center, National Institute of Advanced Industrial.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
EVGM081 Multi-Site Virtual Cluster: A User-Oriented, Distributed Deployment and Management Mechanism for Grid Computing Environments Takahiro Hirofuchi,
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
KISTI & Belle experiment Eunil Won Korea University On behalf of the Belle Collaboration.
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
© 2008 Open Grid Forum File Catalog Development in Japan e-Science Project GFS-WG, OGF24 Singapore Hideo Matsuda Osaka University.
30/07/2005Symmetries and Spin - Praha 051 MonteCarlo simulations in a GRID environment for the COMPASS experiment Antonio Amoroso for the COMPASS Coll.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
1 Grid Activity Summary » Grid Testbed » CFD Application » Virtualization » Information Grid » Grid CA.
7. Grid Computing Systems and Resource Management
Operating System. Chapter 1: Introduction What is an Operating System? Mainframe Systems Desktop Systems Multiprocessor Systems Distributed Systems Clustered.
Rights Management for Shared Collections Storage Resource Broker Reagan W. Moore
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Considerations on Using CernVM-FS for Datasets Sharing Within Various Research Communities Catalin Condurache STFC RAL UK ISGC, Taipei, 18 March 2016.
A System for Monitoring and Management of Computational Grids Warren Smith Computer Sciences Corporation NASA Ames Research Center.
An Introduction to GPFS
29/04/2008ALICE-FAIR Computing Meeting1 Resulting Figures of Performance Tests on I/O Intensive ALICE Analysis Jobs.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
National Institute of Advanced Industrial Science and Technology Gfarm v2: A Grid file system that supports high-performance distributed and parallel data.
Performance measurement of transferring files on the federated SRB
Introduction to Distributed Platforms
U.S. ATLAS Grid Production Experience
PROOF – Parallel ROOT Facility
iGAP: Integrative Grid-enabled Genome Annotation Pipeline
Overview of the Encyclopedia of Life (EOL) Project
Grid Computing.
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Study course: “Computing clusters, grids and clouds” Andrey Y. Shevel
Grid Datafarm and File System Services
Grid Canada Testbed using HEP applications
Presentation transcript:

National Institute of Advanced Industrial Science and Technology Gfarm Grid File System for Distributed and Parallel Data Computing Osamu Tatebe Grid Technology Research Center, AIST APAN Workshop on Exploring eScience Aug 26, 2005 Taipei, Taiwan

National Institute of Advanced Industrial Science and Technology [Background] Petascale Data Intensive Computing Detector for ALICE experiment Detector for LHCb experiment High Energy Physics CERN LHC, KEK-B Belle ~MB/collision, 100 collisions/sec ~PB/year 2000 physicists, 35 countries Astronomical Data Analysis data analysis of the whole data TB~PB/year/telescope Subaru telescope 10 GB/night, 3 TB/year

National Institute of Advanced Industrial Science and Technology Petascale Data-intensive Computing Requirements Peta/Exabyte scale files, millions of millions of files Scalable computational power > 1TFLOPS, hopefully > 10TFLOPS Scalable parallel I/O throughput > 100GB/s, hopefully > 1TB/s within a system and between systems Efficiently global sharing with group-oriented authentication and access control Fault Tolerance / Dynamic re-configuration Resource Management and Scheduling System monitoring and administration Global Computing Environment

National Institute of Advanced Industrial Science and Technology Goal and feature of Grid Datafarm Goal Dependable data sharing among multiple organizations High-speed data access, High-performance data computing Grid Datafarm Gfarm Grid File System – Global dependable virtual file system Federates scratch disks in PCs Parallel & distributed data computing Associates Computational Grid with Data GridFeatures Secured based on Grid Security Infrastructure Scalable depending on data size and usage scenarios Data location transparent data access Automatic and transparent replica selection for fault tolerance High-performance data access and computing by accessing multiple dispersed storages in parallel (file affinity scheduling)

National Institute of Advanced Industrial Science and Technology Gfarm file system (1) Virtual file system that federates local disks of cluster nodes or Grid nodes Enables transparent access using Global namespace to dispersed file data in a Grid Supports fault tolerance and avoid access concentration by automatic and transparent replica selection It can be shared among all cluster nodes and clients Gfarm File System /gfarm ggfjp aistgtrc file1file3 file2 file4 file1file2 File replica creation Global namespace mapping

National Institute of Advanced Industrial Science and Technology Gfarm file system (2) A file can be shared among all nodes and clients Physically, it may be replicated and stored on any file system node Applications can access it regardless of its location In cluster environment, shared secret key is used for authentication GridFTP, samba, NFS server Gfarm metadata server Compute node Client PC Note PC /gfarm metadata Gfarm file system … File A File B File C File A File B File C File B

National Institute of Advanced Industrial Science and Technology Grid-wide configuration Grid-wide file system by integrating local disks in several areas GSI authentication It can be shared among all cluster nodes and clients GridFTP and samba servers in each site GridFTP server, samba server, (NFS server) Metaserver node Compute & fs node … GridFTP server, samba server, (NFS server) Compute & fs node … GridFTP server, samba server, (NFS server) Compute & fs node … US JapanSingapore /gfarm Gfarm Grid file system

National Institute of Advanced Industrial Science and Technology Feature of Gfarm file system A file can be stored on any file system (compute) node (Distributed file system) A file can be replicated and stored on different nodes (Fault tolerant, access concentration tolerant) When there is a file replica on a compute node, it can be accessed without overhead (High performance, scalable I/O)

National Institute of Advanced Industrial Science and Technology More Scalable I/O Performance CPU Gfarm file system Cluster, Grid File A network Job A File A User’s viewPhysical execution view in Gfarm (file-affinity scheduling) File B Job A Job B File B File system nodes = compute nodes Shared network file system Do not separate storage and CPU (SAN not necessary) Move and execute program instead of moving large-scale data Scalable file I/O by exploiting local I/O User A submits that accessesis executed on a node that has User B submits that accessesis executed on a node that has

National Institute of Advanced Industrial Science and Technology Gfarm TM Data Grid middleware Open source development Gfarm TM version released on May 17 th, 2005 ( ) Read-write mode support, more support for existing binary applications, metadata cache server A shared file system in a cluster or a grid Accessibility from legacy applications without any modification Standard protocol support by scp, GridFTP server, samba server,... application Gfarm client library Metadata server CPU... gfsd gfmdslapd Compute and file system nodes Existing applications can access Gfarm file system without any modification using LD_PRELOAD of syscall hooking library or GfarmFS-FUSE

National Institute of Advanced Industrial Science and Technology Gfarm TM Data Grid middleware (2) libgfarm – Gfarm client library Gfarm API gfmd, slapd – Metadata server Namespace, replica catalog, host information, process information gfsd – I/O server Remote file access application Gfarm client library Metadata server CPU... gfsd gfmdslapd Compute and file system nodes File, host information Remote file access

National Institute of Advanced Industrial Science and Technology Access from legacy applications libgfs_hook.so – system call hooking library It emulates to mount Gfarm file system at /gfarm hooking open(2), read(2), write(2), … When it accesses under /gfarm, call appropriate Gfarm API Otherwise, call ordinal system call Re-link not necessary by specifying LD_PRELOAD Linux, FreeBSD, NetBSD, … Higher portability than developing kernel module Mounting Gfarm file system GfarmFS-FUSE enables to mount Gfarm file system using FUSE mechanism in Linux released on Jul 12, 2005 Need to develop a kernel module for other OSs Need volunteers

National Institute of Advanced Industrial Science and Technology Gfarm – Application and performance result

National Institute of Advanced Industrial Science and Technology Scientific Application (1) ATLAS Data Production Distribution kit (binary) Atlfast – fast simulation Input data stored in Gfarm file system not NFS G4sim – full simulation (Collaboration with ICEPP, KEK) Belle Monte-Carlo/Data Production Online data processing Distributed data processing Realtime histgram display 10 M events generated in a few days using a 50-node PC cluster (Collaboration with KEK, U-Tokyo)

National Institute of Advanced Industrial Science and Technology Scientific Application (2) Astronomical Object Survey Data analysis on the whole archive 652 GBytes data observed by SUBARU telescope Large configuration data from Lattice QCD Three sets of hundreds of gluon field configurations on a 24^3*48 4-D space-time lattice (3 sets x MB x 800 = GB) Generated by the CP-PACS parallel computer at Center for Computational Physics, Univ. of Tsukuba (300Gflops x years of CPU time)

National Institute of Advanced Industrial Science and Technology Performance result of parallel grep 25 GBytes text file Xeon 2.8GHz/512KB, 2GB memory NFS340 sec (sequential grep) Gfarm15 sec (16 fs nodes, 16 parallel processes) 22.6 times superlinear speed up Gfarm file system Compute node Compute node Compute node... NFS Compute node *Gfarm file system consists of local disks of compute nodes

National Institute of Advanced Industrial Science and Technology GridFTP data transfer performance Two GridFTP servers can provide almost peak performance (1 Gbps) Local disk vs Gfarm (1~2 nodes) ftpd Client

National Institute of Advanced Industrial Science and Technology Gaussian 03 in Gfarm Ab initio quantum chemistry Package Install once and run everywhere No modification required to access Gfarm Test415 (IO intensive test input) 1h 54min 33sec (NFS) 1h 0min 51sec (Gfarm) Parallel analysis of all 666 test inputs using 47 nodes Write error! (NFS) Due to heavy IO load 17h 31m 02s (Gfarm) Quite good scalability of IO performance Elapsed time can be reduced by re-ordering test inputs NFS vs Gfarm Compute node Compute node Compute node... NFS vs Gfarm Compute node *Gfarm consists of local disks of compute nodes

National Institute of Advanced Industrial Science and Technology Bioinformatics in Gfarm iGAP (Integrative Genome Annotation Pipeline) - A suite of bioinformatics software for protein structural and functional annotation - More than 140 complete or partial proteomes analyzed iGAP on Gfarm - Install once and run everywhere using Gfarm’s high performance file replication and transfer - no modifications required to use distributed compute and storage resource Burkholderia mallei (Bacteria) Gfarm makes it possible to use iGAP to analyze the complete proteome (available 9/28/04) of the bacteria Burkholderia mallei, a known biothreat agent, on distributed resources. This is a collaboration under PRAGMA and the data is available through Participating sites: SDSC/UCSD (US), BII (Singapore), Osaka Univ, AIST (Japan), Konkuk Univ, Kookmin Univ, KISTI (Korea)

National Institute of Advanced Industrial Science and Technology Protein sequences Prediction of : signal peptides (SignalP, PSORT) transmembrane (TMHMM, PSORT) coiled coils (COILS) low complexity regions (SEG) Structural assignment of domains by PSI-BLAST profiles on FOLDLIB Structural assignment of domains by 123D on FOLDLIB Structural assignment of domains by WU-BLAST Data Warehouse Functional assignment by PFAM, NR assignments FOLDLIB Building FOLDLIB: PDB chains SCOP domains PDP domains CE matches PDB vs. SCOP 90% sequence non-identical minimum size 25 aa coverage (90%, gaps <30, ends<30) Domain location prediction by sequence structure info sequence info Step 1 Step 2 Step 3 Step 4 Step 5 Step 6 NR, PFAM SCOP, PDB

National Institute of Advanced Industrial Science and Technology Cluster configuration of Worldwide iGAP/Gfarm data analysis NPTypeProc Spee d (GHz) Memor y (GB) Hard Disk (GB) OS Kookmi n 81Athlon RH 7.3 Konkuk81 Intel P RH 7.3 KISTI81 Intel P RH 7.3 UCSD52 Intel Xeon Rocks 3.2 BII41 Intel P RH 8.0 Osaka101 Intel P RH 7.2 Total

National Institute of Advanced Industrial Science and Technology Preliminary performance result Multiple cluster data analysis 4-node cluster A + 4-node cluster B min 4-node cluster A min NFS Gfarm

National Institute of Advanced Industrial Science and Technology

Development Status and Future Plan Gfarm – Grid file system Global virtual file system A dependable network shared file system in a cluster or a grid High performance data computing support Associates Computational Grid with Data Grid Gfarm Grid software Version released on May 17, 2005 ( Version 1.2 available real soon now Existing programs can access Gfarm file system using syscall hooking library or GfarmFS-FUSE Distribute analysis shows scalable I/O performance iGAP/Gfarm – bioinformatics package Gaussian 03 – Ab initio quantum chemistry package Standardization effort with GGF Grid File System WG (GFS-WG)