NCeSS International Conference Manchester Progress with e-Science? Malcolm Atkinson e-Science Institute & e-Science Envoy www.nesc.ac.uk 29 th June 2006.

Slides:



Advertisements
Similar presentations
Delivering User Needs: A middleware perspective Steven Newhouse Director.
Advertisements

ASYCUDA Overview … a summary of the objectives of ASYCUDA implementation projects and features of the software for the Customs computer system.
1 From Grids to Service-Oriented Knowledge Utilities research challenges Thierry Priol.
© Fraunhofer Institute SCAI and other members of the SIMDAT consortium Data Grids for Process and Product Development using Numerical Simulation and Knowledge.
Advanced Grid Research Workshops through European and Asian Co-operation The Project
Jorge Gasós Grid Technologies Unit European Commission The EU e Infrastructures Programme Workshop, Beijing, June 2005.
Research Infrastructures WP 2012 Call 10 e-Infrastructures part Topics: Construction of new infrastructures (or major upgrades) – implementation.
DG INFSO- Grid Research & Infrastructures: W. Boch, M. Campolargo 1 Delivery of Industrial-strength Grid Middleware: establishing an effective European.
An open source approach for grids Bob Jones CERN EU DataGrid Project Deputy Project Leader EU EGEE Designated Technical Director
Delivery of Industrial Strength Middleware Federated Strengths Agility & Coordination Prof. Malcolm Atkinson Director 21 st January 2004.
Fighting Malaria With The Grid. Computing on The Grid The Internet allows users to share information across vast geographical distances. Using similar.
Professor Dave Delpy Chief Executive of Engineering and Physical Sciences Research Council Research Councils UK Impact Champion Competition vs. Collaboration:
The National Grid Service Mike Mineter.
Chinese Delegation visit Malcolm Atkinson Director 18 th November 2004.
AHM 2006 Talk 18 th September 2006 NanoCMOSgriD Meeting the Design Challenges of Nano-CMOS Electronics Meeting the Design Challenges of Nano-CMOS Electronics:
Research Councils ICT Conference Welcome Malcolm Atkinson Director 17 th May 2004.
National e-Science Centre Glasgow e-Science Hub Opening: Remarks NeSCs Role Prof. Malcolm Atkinson Director 17 th September 2003.
National e-Science Centre & e-Science Institute Malcolm Atkinson Director 2 nd March 2005.
DISC e-Science and Future Technology Conference Sheffield Malcolm Atkinson Director & e-Science Envoy National e-Science Centre & e-Science Institute
SWITCH Visit to NeSC Malcolm Atkinson Director 5 th October 2004.
OMII-UK Steven Newhouse, Director. © 2 OMII-UK aims to provide software and support to enable a sustained future for the UK e-Science community and its.
E-(Social) Science: Where next? 4th International e-Social Science Conference Manchester Malcolm Atkinson Director e-Science Institute UK e-Science Envoy.
Directors Forum NWeS Centre Malcolm Atkinson Director e-Science Institute & e-Science Envoy 12 th April 2007.
Colonial Life IT Modernization Overview Alan Whelchel Assistant Vice President Global Business Technology UNUM / Colonial Life.
Martin Schuurmans Chair EIT The EIT Sustainable Growth and Competitiveness through Innovation.
1 Ideas About the Future of HPC in Europe “The views expressed in this presentation are those of the author and do not necessarily reflect the views of.
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
Towards a Virtual European Supercomputing Infrastructure Vision & issues Sanzio Bassini
10 th October 2013 The delivery of 21 st century services – the implications for the evolution of the Healthcare Science workforce Joan Fletcher.
GENI: Global Environment for Networking Innovations Larry Landweber Senior Advisor NSF:CISE Joint Techs Madison, WI July 17, 2006.
An overview of the EGEE project Bob Jones EGEE Technical Director DTI International Technology Service-GlobalWatch Mission CERN – June 2004.
UK e-Science OGC Technical Committee Edinburgh Malcolm Atkinson Director & e-Science Envoy e-Science Institute 28 th June 2006.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
UK e-Science People & Data WWW Conference Edinburgh Malcolm Atkinson Director & e-Science Envoy e-Science Institute 25 th May 2006.
Ischia, Italy July Session 1 Second Part Day 1 Monday 10 th July Malcolm Atkinson.
The EPSRC Sustainable Urban Environment Programme Philippa Hemmings 27 October 2010.
Computing in Atmospheric Sciences Workshop: 2003 Challenges of Cyberinfrastructure Alan Blatecky Executive Director San Diego Supercomputer Center.
SICSA student induction day, 2009Slide 1 Social Simulation Tutorial Session 6: Introduction to grids and cloud computing International Symposium on Grid.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
Towards a European network for digital preservation Ideas for a proposal Mariella Guercio, University of Urbino.
IST programme Cultural Heritage in Tomorrow ’s Knowledge Society FUTURE PLANS FP6 Cultural Heritage in Tomorrow ’s Knowledge Society FUTURE PLANS FP6 RRRESE.
Strategic Priorities of the NWE INTERREG IVB Programme Harry Knottley, UK representative in the International Working Party Lille, 5th March 2007.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE – paving the way for a sustainable infrastructure.
Bob Jones Technical Director CERN - August 2003 EGEE is proposed as a project to be funded by the European Union under contract IST
INFSO-SSA International Collaboration to Extend and Advance Grid Education e-IRG: Education & Training Malcolm Atkinson National e-Science Centre.
David Carr The Wellcome Trust Data management and sharing: the Wellcome Trust’s approach Economic & Social Data Service conference.
SEEK Welcome Malcolm Atkinson Director 12 th May 2004.
Chapter 5 McGraw-Hill/Irwin Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
August 3, March, The AC3 GRID An investment in the future of Atlantic Canadian R&D Infrastructure Dr. Virendra C. Bhavsar UNB, Fredericton.
Marv Adams Chief Information Officer November 29, 2001.
7. Grid Computing Systems and Resource Management
INFSO-SSA International Collaboration to Extend and Advance Grid Education Education and Training at IST Conference, Helsinki Malcolm Atkinson Director.
Toward a common data and command representation for quantum chemistry Malcolm Atkinson Director 5 th April 2004.
1 Framework Programme 7 Overview. 2 The Programmes within FP7 IDEAS European Research Counsel ERC PEOPLE Marie Curie Measures Initial Training Life-long.
Technology-enhanced Learning: EU research and its role in current and future ICT based learning environments Pat Manson Head of Unit Technology Enhanced.
The National Grid Service Mike Mineter.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI strategy and Grand Vision Ludek Matyska EGI Council Chair EGI InSPIRE.
"The views expressed in this presentation are those of the author and do not necessarily reflect the views of the European Commission" Global reach of.
NERC e-Science Meeting Malcolm Atkinson Director & e-Science Envoy UK National e-Science Centre & e-Science Institute 26 th April 2006.
All Hands Meeting 2005 BIRN-CC: Building, Maintaining and Maturing a National Information Infrastructure to Enable and Advance Biomedical Research.
Clinical and Translational Science Awards (CTSAs) Future Directions and Opportunities for ACNP investigators.
Clouds , Grids and Clusters
National e-Infrastructure Vision
EGEE support for HEP and other applications
UK e-Science People & Data WWW Conference Edinburgh Malcolm Atkinson Director & e-Science Envoy e-Science Institute 25th May 2006.
2.6.5 – International Co-operation
EGI Webinar - Introduction -
Brian Matthews STFC EOSCpilot Brian Matthews STFC
Presentation transcript:

NCeSS International Conference Manchester Progress with e-Science? Malcolm Atkinson e-Science Institute & e-Science Envoy 29 th June 2006

Overview A Context E-Science, Service-oriented Architectures & Services International & UK Scene Projects, Activities & Technologies A Wealth of Opportunities & Challenges Causes of Data Growth Interpretational challenges Software challenges Crucial Issues Usability & Abstraction Interoperation & Federations Cultural challenges of e-Science

What is e-Science? Goal: to enable better research in all disciplines Method: Develop collaboration supported by advanced distributed computation to generate, curate and analyse rich data resources From experiments, observations and simulations Quality management, preservation and reliable evidence to develop and explore models and simulations Computation and data at all scales Trustworthy, economic, timely and relevant results to enable dynamic distributed collaboration Facilitating collaboration with information and resource sharing Security, trust, reliability, accountability, manageability and agility Our challenge is to develop an integrated approach to all three

Are we making any progress? General purpose e-Infrastructure Supports most research requirements Convenient & reliable Amortises costs over wide communities General purpose e-Science methods Conduct of multi-disciplinary collaboration Distributed collaboration Use of diverse information sources Use of advanced computational methods

Web Services Generic & Industry Supported Mechanism for presenting any resource so it can be used remotely Data resources Computation resources Instruments Research processes & procedures Wide range of choices as to how it may be done Easy for provider to make localised decisions Quick deployment Onus on users to compose services Little opportunity to amortise costs across services An obvious first step & components of e-Infrastructure

A Grid Computing Timeline I-Way: SuperComputing US Grid Forum forms at SC 98 Grid Forums merge, form GGF OGSA-WG formed OGSA v1.0 Anatomy paper Physiology paper European & AP Grid Forums UK e-Science programme starts DARPA funds Globus Toolkit & Legion EU funds UNICORE project US DoE pioneers grids for scientific research NSF funds National Technology Grid NASA starts Information Power Grid DARPA funds Globus Toolkit & Legion EU funds UNICORE project US DoE pioneers grids for scientific research NSF funds National Technology Grid NASA starts Information Power Grid Today: Grid solutions are common for HPC Grid-based business solutions are becoming common Required technologies & standards are evolving Today: Grid solutions are common for HPC Grid-based business solutions are becoming common Required technologies & standards are evolving Japan government funds: Business Grid project NAREGI project Japan government funds: Business Grid project NAREGI project Source: Hiro Kishimoto GGF17 Keynote May 2006 GGF & EGA form OGF

What is a Grid? License Printer A grid is a system consisting of Distributed but connected resources and Software and/or hardware that provides and manages logically seamless access to those resources to meet desired objectives A grid is a system consisting of Distributed but connected resources and Software and/or hardware that provides and manages logically seamless access to those resources to meet desired objectives R2AD Database Web server Data CenterCluster HandheldSupercomputer Workstation Server Source: Hiro Kishimoto GGF17 Keynote May 2006

Grid & Related Paradigms Utility Computing Computing services No knowledge of provider Enabled by grid technology Utility Computing Computing services No knowledge of provider Enabled by grid technology Distributed Computing Loosely coupled Heterogeneous Single Administration Distributed Computing Loosely coupled Heterogeneous Single Administration Cluster Tightly coupled Homogeneous Cooperative working Cluster Tightly coupled Homogeneous Cooperative working Grid Computing Large scale Cross-organizational Geographical distribution Distributed Management Grid Computing Large scale Cross-organizational Geographical distribution Distributed Management

How Are Grids Used? High-performance computing Collaborative data-sharing Collaborative design Drug discovery Financial modeling Data center automation High-energy physics Life sciences E-Business E-Science Source: Hiro Kishimoto GGF17 Keynote May 2006

Grids In Use: E-Science Examples Data sharing and integration Life sciences, sharing standard data-sets, combining collaborative data-sets Medical informatics, integrating hospital information systems for better care and better science Sciences, high-energy physics Data sharing and integration Life sciences, sharing standard data-sets, combining collaborative data-sets Medical informatics, integrating hospital information systems for better care and better science Sciences, high-energy physics Capability computing Life sciences, molecular modeling, tomography Engineering, materials science Sciences, astronomy, physics Capability computing Life sciences, molecular modeling, tomography Engineering, materials science Sciences, astronomy, physics High-throughput, capacity computing for Life sciences: BLAST, CHARMM, drug screening Engineering: aircraft design, materials, biomedical Sciences: high-energy physics, economic modeling High-throughput, capacity computing for Life sciences: BLAST, CHARMM, drug screening Engineering: aircraft design, materials, biomedical Sciences: high-energy physics, economic modeling Simulation-based science and engineering Earthquake simulation Simulation-based science and engineering Earthquake simulation Source: Hiro Kishimoto GGF17 Keynote May 2006

Grids (Potentially) Generic & Industry Supported Mechanism for presenting many heterogeneous resources so they can be used remotely Data resources Computation resources Instruments Research processes & procedures Restricting choices as to how it may be done Harder for provider to make localised decisions Deployment can be challenging Providing more homogeneity through virtualisation Should be easier to compose services More opportunity to amortise costs A component of e-Infrastructure

How much do Grids help a researcher? The grid per se doesnt provide Supported e-Science methods Supported data & information resources Computations Convenient access Grids help providers of these International & national secure e-Infrastructure Standards for interoperation Standard APIs to promote re-use But Research Support must be built What is needed? Who should do it?

Commitment to e-Infrastructure A shared resource That enables science, research, engineering, medicine, industry, … It will improve UK / European / … productivity Lisbon Accord 2000 e-Science Vision SR2000 – John Taylor Commitment by UK government Sections Always there c.f. telephones, transport, power

Globus Apache Project & CDIG CeSC (Cambridge) e-Science Institute The e-Science On The Map Today EGEE-II NGS Support Centre National Centre for e-Social Science National Institute for Environmental e-Science NERC e-Science Centre OMII-UK Funded centres National Grid Service Digital Curation Centre

And much more besides Supported data & information services E.g. EBI, MIMAS, EDINA, Zoo-DB, … Community efforts E.g. GridPP, AstroGrid, NERC DataGrid Access to new scientific facilities E.g. Satellite observations, Met-Office data E.g. Diamond E.g. CEASAR, HPCx & HECTOR Communities developing code & tools Pilot projects in most research areas

A Complex multi-provider System Strengths Resilience – no over-dependence on one set of decisions Dynamic & rapid response to need Resources gathered through many channels Specific user requirements understood Weaknesses Complexity & possibly unnecessary duplication Reinforces old silos Composition and amortisation harder Is more coordination Desirable? Feasible?

© OMII-UK & National Grid Service: Life Sciences Gateway SOAPLab Services Core NGS Middleware Bioinformatics & Life Science Users Hardware Resources OMII: e-Infrastructure Services OGSA-DAI GridSAM GRIMOIRES Taverna Upperware Underware Middleware Talk to us about other gateways Computational Chemistry Engineering Image Analysis …

The National Grid Service Core sites White Rose (Leeds) Manchester Oxford CCLRC Partner sites Bristol Cardiff Lancaster Access to HPC facilities HPCx CSAR Capacity CPUs 30+ Tera Bytes Specialist facilities Cardiff 4x16 proc SGI Bristol: Intel Lancaster SUN Cluster Services: Job submission (GT2) Storage service (SRB) Oracle Service OGSA-DAI

Enabling Grids for E-sciencE INFSO-RI Ian Bird, SA1, EGEE Final Review th May 2006 A global, federated e-Infrastructure EGEE infrastructure ~ 200 sites in 39 countries ~ CPUs > 5 PB storage > concurrent jobs per day > 60 Virtual Organisations EUIndiaGrid EUMedGrid SEE-GRID EELA BalticGrid EUChinaGrid OSG NAREGI Related projects & collaborations are where the future expansion of resources will come from ProjectAnticipated resources (initial estimates) Related Infrastructure projects SEE-grid6 countries, 17 sites, 150 cpu EELA5 countries, 8 sites, 300 cpu EUMedGrid6 countries BalticGrid3 countries, fewx100 cpu EUChinaGridTBC Collaborations OSG30 sites, cpu ARC15 sites, 5000 cpu DEISASupercomputing resources

Enabling Grids for E-sciencE INFSO-RI Ian Bird, SA1, EGEE Final Review th May 2006 Use of the infrastructure Sustained & regular workloads of >30K jobs/day spread across full infrastructure doubling/tripling in last 6 months – no effect on operations Massive data transfers > 1.5 GB/s

About Grid Computing Now! A Knowledge Transfer Network project funded by the DTI Technology Programme aimed at transferring knowledge about Grid Computing Technologies to Public and Private Sectors in the UK. Partnership between Intellect, the UK Hi-Tech Trade Association; National e-Science Centre, a world leader in Grid Computing research; and CNR Ltd, a consultancy focused on SME organisations and business intermediaries. Substantial number of industrial, business and academic partners Website Background Information Industry News/Events User Case Studies Events programme Technical Overviews Multiple vendor perspectives User Case Studies Sector Agenda Healthcare; Government; Telecoms; Services; etc.. User Community Network with peers Find useful contacts Contribute experience

INFSO-SSA Invest in People Training – Targeted – Immediate goals – Specific skills – Building a workforce Education – Pervasive – Long term and sustained – Generic conceptual models – Developing a culture Both are needed Society Graduates EducationInnovation Invests PreparesCreate Enriches Organisation Skilled Workers TrainingServices & Applications Invests PreparesDevelop Strengthens

Biomedical Research Informatics Delivered by Grid Enabled Services Synteny Grid Service blast + Portal

Overview of e-Infrastructure eDiaMoND: Screening for Breast Cancer 1 Trust Many Trusts Collaborative Working Audit capability Epidemiology Other Modalities -MRI -PET -Ultrasound Better access to Case information And digital tools Supplement Mentoring With access to digital Training cases and sharing Of information across clinics Letters Radiology reporting systems eDiaMoND Grid 2ndary Capture Or FFD Case Information X-Rays and Case Information Digital Reading SMF Case and Reading Information CADTemporal Comparison Screening Electronic Patient Records Assessment/ Symptomatic Biopsy Case and Reading Information Symptomatic/Assessment Information Training Manage Training Cases Perform Training SMF CAD 3D Images Patients Provided by eDiamond project: Prof. Sir Mike Brady et al.

climateprediction.net and GENIE Largest climate model ensemble >45,000 users, >1,000,000 model years 10K 2K Response of Atlantic circulation to freshwater forcing

Courtesy of David Gavaghan & IB Team Integrative Biology Tackling two Grand Challenge research questions: What causes heart disease? How does a cancer form and grow? Together these diseases cause 61% of all UK deaths Will build a powerful, fault-tolerant Grid infrastructure for biomedical science Enabling biomedical researchers to use distributed resources such as high-performance computers, databases and visualisation tools to develop complex models of how these killer diseases develop.

Courtesy of David Gavaghan & IB Team IB Partners

Foundations of Collaboration Strong commitment by individuals To work together To take on communication challenges Mutual respect & mutual trust Distributed technology To support information interchange To support resource sharing To support data integration To support trust building Sufficient time Common goals Complementary knowledge, skills & data Can we predict when it will work? Can we find remedies when it doesnt?

CARMEN - Scales of Integration resolving the neural code from the timing of action potential activity determining ion channel contribution to the timing of action potentials examining integration within networks of differing dimensions Understanding the brain may be the greatest informatics challenge of the 21 st century

CARMEN Consortium Leadership & Infrastructure Colin Ingram Paul Watson Leslie SmithJim Austin

CARMEN Consortium International Partners Daniel Gardner (Cornell) Lead for the US NIH, Neuroscience Information Framework and Brain ML Shiro Usui (RIKEN Brain Science Institute) Lead for the Japan Node of the International Neuroinformatics Coordinating Facility Sten Grillner (Karolinska Institute) Chairman of the OECD, International Neuroinformatics Coordinating Facility Ad Aertsen (Freiburg) Neural network modelling and large-scale simulations George Gerstein (Pennsylvania) Analysis of spike pattern trains

CARMEN Consortium Commercial Partners - applications in the pharmaceutical sector - interfacing of data acquisition software - application of infrastructure - commercialisation of tools

16 th March 2006 NanoCMOSgriD Meeting the Design Challenges of Nano-CMOS Electronics The Challenge Toshiba 04 Device diversification 90nm: HP, LOP, LSTP 45nm: UTB SOI 32nm: Double gate

16 th March 2006 NanoCMOSgriD Meeting the Design Challenges of Nano-CMOS Electronics Advanced Processor Technologies Group (APTGUM) Device Modelling Group (DMGUG) Electronic Systems Design Group (ESDGUS) Intelligent Systems Group (ISGUY) National e-Science Centre (NeSC) Microsystems Technology Group (MSTGUG) Mixed-Mode Design Group in IMNS (MMDGUE) e-Science NorthWest Centre (eSNW) University Partners

16 th March 2006 NanoCMOSgriD Meeting the Design Challenges of Nano-CMOS Electronics Industrial Partners Global EDS vendor and world TCAD leader 600 licences of grid implementation, model implementation UK fabless design company and world microprocessor leader Core IP, simulation tools, staff time UK fabless design company and world mixed mode leader Additional PhD studentship for mixed mode design Global semiconductor player with strong UK presence Access to technology, device data, processing Global semiconductor player with strong UK presence Access to technology, device data, processing Global semiconductor player with UK presence CASE studentship, interconnects Trade association of the microelectronics industry in the UK Recruiting new industrial partners and dissemination

Compound Causes of (Geo)Data Growth Faster devices Cheaper devices Higher-resolution all ~ Moores law Increased processor throughput more derived data Cheaper & higher-volume storage Remote data more accessible Public policy to make research data available Bandwidth increases Latency doesnt get less though

Interpretational Challenges Finding & Accessing data Variety of mechanisms & policies Interpreting data Variety of forms, value systems & ontologies Independent provision & ownership Autonomous changes in availability, form, policy, … Processing data Understanding how it may be related Devising models that expose the relationships Presenting results Humans need either Derived small volumes of statistics Visualisations

Interpretational Challenges Finding & Accessing data Variety of mechanisms & policies Interpreting data Variety of forms, value systems & ontologies Independent provision & ownership Autonomous changes in availability, form, policy, … Processing data Understanding how it may be related Devising models that expose the relationships Presenting results Humans need either Derived small volumes of statistics Visualisations

Interpretational Challenges Finding & Accessing data Variety of mechanisms & policies Interpreting data Variety of forms, value systems & ontologies Independent provision & ownership Autonomous changes in availability, form, policy, … Processing data Understanding how it may be related Devising models that expose the relationships Presenting results Humans need either Derived small volumes of statistics Visualisations

Collaboration Essential to assemble experts Multi-discipline, Multi-organisation, Multi-national, Multi-scale Hard to achieve Instinctive competition Trust slow to build Communication is difficult Requirements Leadership Investment New culture Technology Cross commercial – academic boundaries Address these issues Focus here

Towards Accessible e-Science High-level tools Abstraction Metadata-driven interpretation & guidance Well-defined semantics Automated data management is key Convenient user-controlled composition Lego-brick convenience + Precision control Understood by scientists, engineers, modellers, diagnosticians, … Responsibility & Credit Provenance tracking automated & understood Culture developed for consortia and collaboration

Built on Dependable Infrastructure Global Federations Dependable and persistent facilities Always there & always on Consistent for mobile users Consistent for mobile code & mobile information Affordable e-Infrastructure Based on well-established standards Based on well-honed operating procedures Investment preserved through stability Utility improved through agile development Amortised across disciplines and nations Consistent framework for multiple provisions Trustworthy management of information

Collaboration Collaboration is a Key Issue Multi-disciplinary Multi-national Academia & industry Trustworthy data sharing key for collaboration Plenty of opportunities for research and innovation Establish common frameworks where possible Islands of stability – reference points for data integration Establish international standards and cooperative behaviour Extend incrementally Trustworthy code & service sharing also key

Federation Federation is a Key Issue Multi-organisation Multi-purpose Multi-national Academia & industry Build shared standards and ontologies Requires immense effort Requires critical mass of adoption Requires sufficient time Trustworthy code & e-Infrastructure sharing Economic & social necessity Balance independence & self-motivation With social & technical constraints

Major Intellectual Challenges Require Many approaches to be integrated Many minds engaged Many years of effort Using the Systems Requires well-tuned models Well-tuned relationships between systems & people Flexibility, adaptability & agility Enabling this Is itself a major intellectual challenge

Matched Systems Communities of Researchers E-Infrastructure Facilities Each will evolve Can you help make this co-evolution successful?