Presentation is loading. Please wait.

Presentation is loading. Please wait.

A new collaborative scientific initiative at Harvard.

Similar presentations


Presentation on theme: "A new collaborative scientific initiative at Harvard."— Presentation transcript:

1 A new collaborative scientific initiative at Harvard.

2 One-Slide IIC Proposal-driven, from within Harvard “Projects” focus on areas where computers are key to new science; widely applicable results Technical focus “Branches”  Instrumentation  Databases & Provenance  Analysis & Simulations  Visualization  Distributed Computing (e.g. GRID, Semantic Web) Matrix organization: “Projects” by “Branches” Education: Train Future Consumers & Producers of Computational Science Proposal-driven, from within Harvard “Projects” focus on areas where computers are key to new science; widely applicable results Technical focus “Branches”  Instrumentation  Databases & Provenance  Analysis & Simulations  Visualization  Distributed Computing (e.g. GRID, Semantic Web) Matrix organization: “Projects” by “Branches” Education: Train Future Consumers & Producers of Computational Science Goal: Fill the void in, highly value, and learn from, the emerging field of “computational science.”

3 “Astronomical Medicine” A joint venture of FAS-Astronomy & HMS/BWH-Surgical Planning Lab; Work shown here is from the 2005 Junior Thesis of Michelle Borkin, Harvard College.

4 Filling the “Gap” between Science and Computer Science Increasingly, core problems in science require computational solution Typically hire/“home grow” computationalists, but often lack the expertise or funding to go beyond the immediate pressing need Focused on finding elegant solutions to basic computer science challenges Often see specific, “applied” problems as outside their interests Scientific disciplines Computer Science departments

5 “Workflow” & “Continuum”

6 Workflow ExamplesAstronomyPublic Health “ Collect ” TelescopeMicroscope, Stethoscope, Survey COLLECT “ National Virtual Observatory ” / COMPLETE CDC Wonder “ Analyze ” Study the density structure of a star- forming glob of gas Find a link between one factory ’ s chlorine runoff & disease ANALYZE Study the density structure of all star- forming gas in … Study the toxic effects of chlorine runoff in the U.S. “ Collaborate ” Work with your student COLLABORATE Work with 20 people in 5 countries, in real-time “ Respond ” Write a paper for a Journal. RESPOND Write a paper, the quantitative results of which are shared globally, digitally.

7 IIC branches address shared “workflow” challenges Challenges common to data-intensive science Data acquisition Data processing, storage, and access Deriving meaningful insight from large datasets Maximizing understanding through visual representation Sharing knowledge and computing resources across geographically dispersed researchers Instrumentation Analysis & Simulations Databases/ Provenance Distributed Computing Visualization IIC branches

8 Continuum “Pure” Discipline Science (e.g. Einstein) “Pure” Computer Science (e.g. Turing) “Computational Science” Missing at Most Universities

9 IIC Organization: Research and Education Assoc Dir, Instrumentation Assoc Dir, Visualization Assoc Dir, Analysis & Simulation Provost IIC Director Assoc Provost Dir of Admin & Operations Project 1 (Proj Mgr 1) Project 2 (Proj Mgr 2) Project 3 (Proj Mgr 3) Dir of Education & Outreach    Etc. CIO (systems) Knowledge mgmt Education & Outreach staff Dean, Physical Sciences Dir of Research Assoc Dir, Databases/Data Provenance Assoc Dir, Distributed Computing

10 IIC Organization: Admin and Operations Provost IIC Director Dir of Research Assoc Provost Dir of Admin & Operations Dir of Education & Outreach Dean, Physical Sciences Admin Finance Development Facilities HR Note: admin roles expected to be played by 1-2 staff members at outset; staff will grow with overall IIC growth

11

12 Barnard’s Perseus COMPLETE/IRAS N dust

13 IRAS N dust H-  emission,WHAM/SHASSA Surveys (see Finkbeiner 2003) HH 2MASS/NICER Extinction

14 Numerical Simulation of Star Formation Bate, Bonnell & Bromm 2002 (UKAFF) MHD turbulence gives “t=0” conditions; Jeans mass=1 M sun 50 M sun, 0.38 pc, n avg =3 x 10 5 ptcls/cc forms ~50 objects T=10 K SPH, no B or  movie=1.4 free-fall times

15 Goal: Statistical Comparison of “Real” and “Synthesized” Star Formation Figure based on work of Padoan, Nordlund, Juvela, et al. Excerpt from realization used in Padoan & Goodman 2002.

16 Measuring Motions: Molecular Line Maps

17 Alves, Lada & Lada 1999 Radio Spectral-Line Survey Radio Spectral-line Observations of Interstellar Clouds

18 Velocity from Spectroscopy 1.5 1.0 0.5 0.0 -0.5 Intensity 400350300250200150100 "Velocity" Observed Spectrum All thanks to Doppler Telescope  Spectrometer

19 1.5 1.0 0.5 0.0 -0.5 Intensity 400350300250200150100 "Velocity" Observed Spectrum Telescope  Spectrometer All thanks to Doppler Velocity from Spectroscopy

20 Barnard’s Perseus COMPLETE/FCRAO W( 13 CO)

21 “Astronomical Medicine” Excerpts from Junior Thesis of Michelle Borkin (Harvard College); IIC Contacts: AG (FAS) & Michael Halle (HMS/BWH/SPL)

22 IC 348

23 “Astronomical Medicine”

24

25 After “Medical Treatment” Before “Medical Treatment”

26 3D Slicer Demo IIC contacts: Michael Halle & Ron Kikinis

27 IIC Research Branches Improved data acquisition. Novel hardware approaches (e.g. GPUs, sensors). Development of efficient algorithms. Cross-disciplinary comparative tools (e.g. statistical). Management, and rapid retrieval, of data. “Research reproducibility” …where did the data come from? How? e-Science aspects of large collaborations. Sharing of data and computational resources and tools in real-time. Physically meaningful combination of diverse data types. InstrumentationAnalysis & Simulations Databases/ Provenance Distributed Computing Visualization IIC projects will bring together IIC experts from relevant branches with discipline scientists to address a pressing computing challenge facing the discipline, that has broad application

28 3D Slicer

29 Distributed Computing & Large Databases: Large Synoptic Survey Telescope Optimized for time domain scan mode deep mode 7 square degree field 6.5m effective aperture 24th mag in 20 sec > 5 Tbyte/night Real-time analysis Simultaneous multiple science goals Simultaneous multiple science goals IIC contact: Christopher Stubbs (FAS)

30 Relative optical survey power based on A  = 270 LSST design

31 AstronomyHigh Energy Physics LSSTSDSS2MASSMACHODLSBaBarAtlasRHIC First year of operation 20111998200119921999199820071999 Run-time data rate to storage (MB/sec) 5000 Peak 500 Avg 8.3 1 1 2.7 60 (zero- suppressd) 6* 540* 120* ( ’ 03) 250* ( ’ 04) Daily average data rate (TB/day) 200.020.0160.0080.0120.660.03 ( ’ 03) 10 ( ’ 04) Annual data store (TB) 20003.6610.253007000200 ( ’ 03) 500 ( ’ 04) Total data store capacity (TB) 20,000 (10 yrs) 20024.58210,000100,000 (10 yrs) 10,000 (10 yrs) Peak computational load (GFLOPS) 140,000100 111.000.6002,000100,0003,000 Average computational load (GFLOPS) 140,0001020.7000.0302,000100,0003,000 Data release delay acceptable 1 day moving 3 months static 2 months 6 months1 year6 hrs (trans) 1 yr (static) 1 day (max) <1 hr (typ) Few days100 days Real-time alert of event30 secnone <1 hour1 hrnone Type/number of processors TBD1GHz Xeon 18 450MHz Sparc 28 60-70MHz Sparc 10 500MH z Pentium 5 Mixed/ 5000 20GHz/ 10,000 Pentium/ 2500

32 Challenges at the LHC For each experiment (4 total): 10’s of Petabytes/year of data logged 2000 + Collaborators 40 Countries 160 Institutions (Universities, National Laboratories) CPU intensive Global distribution of data Test with « Data Challenges »

33 Earth Simulator Atmospheric Chemistry Group LHC Exp. Astronomy Grav. Wave Nuclear Exp. Current accelerator Exp. CPU vs. Collaboration Size

34 interactive physics analysis batch physics analysis batch physics analysis detector event summary data raw data event reprocessing event reprocessing event simulation event simulation analysis objects (extracted by physics topic) Data Handling and Computation for Physics Analysis event filter (selection & reconstruction) event filter (selection & reconstruction) processed data les.robertson@cern.ch CERN

35 Workflow a.k.a. The Scientific Method (in the Age of the Age of High-Speed Networks, Fast Processors, Mass Storage, and Miniature Devices) IIC contact: Matt Welsh, FAS


Download ppt "A new collaborative scientific initiative at Harvard."

Similar presentations


Ads by Google