Www.eminerals.org Experience of the SRB in support of collaborative grid computing Martin Dove University of Cambridge.

Slides:



Advertisements
Similar presentations
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Advertisements

Using eScience to calibrate our tools: parameterisation of quantum mechanical calculations with grid technologies Kat Austen Dept. of Earth Sciences, University.
3rd Campus Grid SIG Meeting. Agenda Welcome OMII Requirements document Grid Data Group HTC Workshop Research Computing SIG? AOB Next meeting (AG)
Peter Berrisford RAL – Data Management Group SRB Services.
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
High Performance Computing Course Notes Grid Computing.
GridRPC Sources / Credits: IRISA/IFSIC IRISA/INRIA Thierry Priol et. al papers.
Daresbury Laboratory Enabling Science with Grid Technology Jamie Rintelman, Kerstin Kleese-Van Dam, Rik Tyer STFC-Daresbury Laboratory; Daresbury, Cheshire,
RCAC Research Computing Presents: DiaGird Overview Tuesday, September 24, 2013.
6th Biennial Ptolemy Miniconference Berkeley, CA May 12, 2005 Distributed Computing in Kepler Ilkay Altintas Lead, Scientific Workflow Automation Technologies.
OxGrid, A Campus Grid for the University of Oxford Dr. David Wallom.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
BLOGGING MEETS COMPUTATIONAL CHEMISTRY Dr Kieron Taylor University of Southampton*
Cyberinfrastructure for Rapid Prototyping Capability Tomasz Haupt, Anand Kalyanasundaram, Igor Zhuk, Vamsi Goli Mississippi State University GeoResouces.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
Connecting OurGrid & GridSAM A Short Overview. Content Goals OurGrid: architecture overview OurGrid: short overview GridSAM: short overview GridSAM: example.
Flexibility and user-friendliness of grid portals: the PROGRESS approach Michal Kosiedowski
A summary of the outputs of the ARCHER Project David Groenewegen, Nick Nicholas and Anthony Beitz ARCHER Project.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Crystal-25 April The Rising Power of the Web Browser: Douglas du Boulay, Clinton Chee, Romain Quilici, Peter Turner, Mathew Wyatt. Part of a.
1 Overview of the Application Hosting Environment Stefan Zasada University College London.
Grid tool integration within the eMinerals project Mark Calleja.
1 All-Hands Meeting 2-4 th Sept 2003 e-Science Centre The Data Portal Glen Drinkwater.
Ian C. Smith ULGrid – Experiments in providing a campus grid.
Crystal25 Hunter Valley, Australia, 11 April 2007 Crystal25 Hunter Valley, Australia, 11 April 2007 JAINIS (JCU and Indiana Instrument Services): A Grid.
© 2006 Open Grid Forum Enabling Pervasive Grids The OGF GIN Effort Erwin Laure GIN-CG co-chair, EGEE Technical Director
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Code Applications Tamas Kiss Centre for Parallel.
Grid Architecture William E. Johnston Lawrence Berkeley National Lab and NASA Ames Research Center (These slides are available at grid.lbl.gov/~wej/Grids)
6/12/99 Java GrandeT. Haupt1 The Gateway System This project is a collaborative effort between Northeast Parallel Architectures Center (NPAC) Ohio Supercomputer.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Building the e-Minerals Minigrid Rik Tyer, Lisa Blanshard, Kerstin Kleese (Data Management Group) Rob Allan, Andrew Richards (Grid Technology Group)
The NGS Grid Portal David Meredith NGS + Grid Technology Group, e-Science Centre, Daresbury Laboratory, UK
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
ISERVOGrid Architecture Working Group Brisbane Australia June Geoffrey Fox Community Grids Lab Indiana University
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.
Michael Doherty RAL UK e-Science AHM 2-4 September 2003 SRB in Action.
Leveraging the InCommon Federation to access the NSF TeraGrid Jim Basney Senior Research Scientist National Center for Supercomputing Applications University.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Introduction to The Storage Resource.
E-Curator: A Web-based Curatorial Tool Ian Brown, Mona Hess Sally MacDonald, Francesca Millar Yean-Hoon Ong, Stuart Robson Graeme Were UCL Museums & Collections.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
GridChem Architecture Overview Rion Dooley. Presentation Outline Computational Chemistry Grid (CCG) Current Architectural Overview CCG Future Architectural.
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
Partnerships in Innovation: Serving a Networked Nation Grid Technologies: Foundations for Preservation Environments Portals for managing user interactions.
Grid Remote Execution of Large Climate Models (NERC Cluster Grid) Dan Bretherton, Jon Blower and Keith Haines Reading e-Science Centre
The National Grid Service Mike Mineter.
The NGS Grid Portal David Meredith NGS + Grid Technology Group, e-Science Centre, Daresbury Laboratory, UK
The Storage Resource Broker and.
Holding slide prior to starting show. Lessons Learned from the GECEM Portal David Walker Cardiff University
Library Efforts and IT Infrastructure Services Paul Conway & Mike Pickett Duke University
© Geodise Project, University of Southampton, Workflow Support for Advanced Grid-Enabled Computing Fenglian Xu *, M.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
Collection-Based Persistent Archives Arcot Rajasekar, Richard Marciano, Reagan Moore San Diego Supercomputer Center Presented by: Preetham A Gowda.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 22 February 2006.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Grid Services for Digital Archive Tao-Sheng Chen Academia Sinica Computing Centre
PLG-Data and rimrock Services as Building
Simulation Production System
Problem: Ecological data needed to address critical questions are dispersed, heterogeneous, and complex Solution: An internet-based mechanism to discover,
Grid Portal Services IeSE (the Integrated e-Science Environment)
Gordon Erlebacher Florida State University
Introduction to the SHIWA Simulation Platform EGI User Forum,
Presentation transcript:

Experience of the SRB in support of collaborative grid computing Martin Dove University of Cambridge

A voyage of discovery ‣ We aimed to focus on grid computing to support molecular-scale simulations... ‣... but discovered the important role of data and information delivery ‣ We thought that the SRB would provide a means to archive data... ‣... but discovered that it could be much more useful than that The SRB has radically changed our view of how we should carry out the scientific process

My view of eScience Computing grids Data grids Collaborative grids

Science beyond the lab-book ‣ Management of too many tasks ‣ Management of the resultant data deluge ‣ Sharing the information content with collaborators ‣ Maintaining accuracy and verification

Crystal structure of CaCO 3 ‣ Note layers of Ca cations and CO 3 carbonate molecular ions ‣ Carbonate ions start to spin and then tumble at high temperature

Expansion of calcite Neutron diffraction experiments small decrease in a 5% increase in c

Rock-salt structure of BaCO 3 Note disordered positions of oxygen atoms

BaCO 3 : lattice parameters Molecular dynamics simulations on the NGS

Challenge for the researcher ‣ Short-term collation of the data ‣ Longer-term management of the data ‣ Sharing the data with collaborators

SRB and grid computing ‣ It was important to build the data grid – in our case the SRB – into the heart of the computing grid environment ‣ Then we needed tools to make the integration of the data and compute grids seamless, and which are easy to use – non- intrusive

Profile of our users ‣ They want maximum control over their work processes – they don’t want to access them through portals or GUI’s ‣ They also don’t want their applications pre-wrapped as services: they want to have complete control over their applications, e.g. to add capability ‣ They know what they are doing... ‣... and they don’t want to be told how to do things!

Internet Researcher Compute clusters Cluster JobMgr Globus Application server Access to external facilities and grids Desktop pools Condor JobMgr Globus Campus grids Condor JobMgr Globus Parallel (HPC) clusters Cluster JobMgr Globus Data vault Globus is used a) to provide user authentication via digital certificates b) job submission middleware Our data grid is based on the San Diego Storage Resource Broker The application server provides databases and server capabilities for the SRB, metadata tools, and job submission tool

Job submission process ‣ We have developed RMCS to run the job submission process ‣ It integrates with the use of the data grid, specifically with the SRB ‣ RMCS can be run from the user’s desktop via a shell-command client tool

4. Job runs on grid compute resources Application server 1. Upload data files and application to data vault 5. Metadata is sent to the application server 6. Output files are transferred to the data vault 7. Researcher interacts with the metadata database to extract core output values 2. Submit job to grid via RMCS 3. Data files and application are transferred to the grid resource Researcher Data vault

Parameter sweeps ‣ implement bulk file upload to the SRB or other data grid ‣ generate set of RMCS input files ‣ submit all the RMCS jobs We have perl programs that Bulk job creation and submission is a one-command procedure

Scientific collaboration Classical molecular dynamics methods Quantum mechanical methods

Data and information ? XML data representation instead

Data vault Upload XML data files to data vault for sharing with collaborator Researcher A Researcher B View information content of data files using ccViz Access Grid with JMAST Instant messaging SciSpace.net

SRB: some early positives ‣ When we started, it was the only show in town to facilitate easy data sharing ‣ It was affordable in terms of capital and person ££££ ‣ It is easily extended through addition of new vaults ‣ It proved easy to use

Anecdote: Lucy’s project ‣ Lucy learned to use the SRB-based data grid very easily ‣ Using our data tools, she was able to provide me with remote access to the information content of her data very easily Lucy was a third-year project student, and we let her perform her project using all our grid infrastructure with no compromises

Some caveats ‣ We didn’t actually need to federate or distribute different data sources... ‣... and by distributing our data we discovered that such an approach gives an unnecessary weak link and issues of ownership ‣ We didn’t need the access-control tools, nor the data replication tools, in which case some of the infrastructure was heavier than needed

So what is different now? ‣ We now expect to be able to share their data with collaborators... ‣... and we expect this to be easy (ie not via a multi-stage process) ‣ We now routinely produce complete archives of all files associated with a study easily and automatically, rather than have stuff dumped to our desktops ‣ And we now expect a single place to deposit data, and for this process to be easy and automatic

Summary ‣ The SRB was critical to the successes of the eMinerals project ‣ The SRB was easy to use, and affordable ‣ We have developed some tools on top of the SRB to make access, display of data, and access control easier (eg webdav access, web interface) ‣ The SRB has radically changed the way we think about managing data – but I don’t think that this was an easy change to acquire

Credits Cambridge: Kat Austen, Richard Bruin, Mark Calleja, Gen-Tao Chiang, Ian Frame, Peter Murray-Rust, Toby White, Andrew Walker STFC: Kerstin Kleese van Dam, Phil Couch, Tom Mortimer-Jones, Rik Tyer Funded by NERC