4 Daniel E. Atkins, Chair, University of Michigan Kelvin K. Droegemeier, University of Oklahoma Stuart I. Feldman, IBM Hector Garcia-Molina, Stanford University Michael L. Klein, University of Pennsylvania David G. Messerschmitt, University of California at Berkeley Paul Messina, California Institute of Technology Jeremiah P. Ostriker, Princeton University Margaret H. Wright,New York University Setting the Stage
5 Cyberinfrastructure Promise Ubiquitous, digital knowledge environments that are both interactive and functionally complete………… (Atkins report) revolutionize the processes of discovery, learning and innovation across the science and engineering frontier.
6 Cyberinfrastructure Characteristics Community-Focused –virtual organizations –distributed, –collaborative Scale and Scope –Multidisciplinary –International –Supporting data- and compute-intensive applications –High-end to desktop –Heterogenuous Common Technology & Policy Platform(s) –Interoperability –Supports characteristics above
7 Cyberinfrastructure Early Adopters Network for Earthquake Engineering Simulation (NEES) National Ecological Observatory Network (NEON) Biomedical Informatics Research Network (BIRN) Extensible Terascale Facility (ETF)
8 CalTechArgonne Extensible Terascale Facility A Heterogenuous Grid Environment SDSC Los Angeles Chicago NCSA PSC Existing ETF Partners Hubs New Partners
9 Hardware Integrated CI System meeting the needs of a community of communities Grid Services & Middleware Development Tools & Libraries Applications Environmental Science High Energy Physics Proteomics/Genomics … Domain- specific Cybertools (software) Shared Cybertools (software) Distributed Resources (computation, communication storage, etc.) Education and Training Discovery & Innovation
10 Shared Cybertools (Middleware Tools and Services) Basic Services Security, Scheduling, Data Services, Database Services, User Services, Application Management Services, Autonomy and Monitoring Services, Information Services, Composition Service, Messaging Service Application Level Services People Collaboration, Resource Collaboration, Decision-Making Services, Knowledge Discovery Services, Workflow Services, Universal Access
11 Challenging Context Institutional & Infrastructural Ecology –Technological change more rapid than institutional change Broadening Participation Community-Building Seamless Integration of New and Old –Balancing upgrades of existing and creation of new resources –Legacy data/models Providing sustainable support
12 Plan of Action Focused, cross-cutting attention on cyberinfrastructure – not business as usual Internal NSF planning now underway - active discussion on specific cyberinfrastructure issues Community building - broad consultations with scientific/engineering communities is intensifying 2003 workshops and town hall meetings – management models NSF FY05 budget planning for cyberinfrastructure underway
13 CISE/ANIR CISE - Computer and Information Science and Engineering Directorate 3 Goals –to enable the U.S. to uphold a position of world leadership in computing, communications, and information science and engineering –To promote understanding of the principles and uses of advanced computing, communications and information systems in service to society –To contribute to universal, transparent and affordable participation in an information-based society.
14 CISE/ANIR 5 Divisions –Advanced Computational Infrastructure & Research (ACIR) –Advanced Networking Infrastructure & Research (ANIR) –Computer-Communications Research (CCR) –Experimental & Integrative Activities (EIA) –Information & Intelligent Systems (IIS)
15 CISE/ANIR ANIR - The programs focus on the fundamental science and technology needed to facilitate the efficient, high-speed transfer of information through networks and distributed systems. Projects funded span the entire spectrum, from network design and performance evaluation to middleware and software frameworks in support of applications running on top of networks and distributed systems.
16 ANIR Programs in 2003 Network Research (NR) Program Special Projects (SP) Program High Performance Network Connections for Science and Engineering Research (HPNC) Network Research Testbeds (NRT) Experimental Infrastructure Networks (EIN) NSF Middleware Initiative (NMI) Strategic Technologies for the Internet (STI) Other Cross Directorate –ITR (Information Technology Research) –REU (Research Experiences for Undergraduates) –Career
17 Network Research Testbeds Characteristics –Disruptive technologies and approaches –Hybrid and experimental designs –End-device research –Core technology development –New protocol research –Alternative network architectures –Testbed implementations $9M spent in 2003 funds 10 awards given (3 are collaborative)
18 NRT Awards Collaborative Research: A Unified Experimental Environment for Diverse Networks –PI: Paul Barford (UWisc) and Jay Lapreau (Utah) for $4.6M Collaborative: Testing and Benchmarking Methodologies for Future Network Security Mechanisms –Pis: George Kesidis, Karl Levitt, Vern Paxson (ICSI), Catherine Rosenburg (Purdue) for $5.3M Orbit: Open-Access Research Testbed for Next-Generation Wireless Networks –PI: Dipankar Raychaudt (Rutgers) for $5.8M Agile and Efficient Ultra-Wideband Wireless Network Testbed for Challenged Environments –PI: Moe Win (MIT) for $3.1M
19 NRT Awards in 2003 Heterogeneous Wireless Access Network Test Bed –PI: Paul Kolodzy (Stevens) for $800k Scalable Testbed for Next-Generation Mobile Wireless Networking Technologies –PI: Ravive Bagrodia (UCLA) for $5.4M National Radio Networking Research Testbed (NRNRT) –PI: Gary Minden (Kansa) for $1.8M
20 Experimental Infrastructure Networks (EIN) Complimentary program to NRT Characteristics –Control/management of the infrastructure end-to-end –End-to-end performance/support w/ dedicated provisioning –Pre-market technologies, experimental h/w, alpha software –Significant collaboration vertically and across sites –Persistence, w/ repeatable network experiments and/or reconfigurability –Experimental protocols, configurations, and approaches for high throughput, low latency, large bursts $10M spent in 2003 funds
21 EIN Awards in 2003 Collaborative Research: Dynamic Resource Allocation for GMPLS: Optical Networks (DRAGON) –PIs: Jerry Sobieski UMd), Herbert Schorr (Tom Lehman - ISI East), and Bijan Jabbari (GMU) for $6.5M Collaborative Research: End-to-End Provisioned Optical Network Tested for Large-Scale eScience Applications (CHEETAH) –PIs: Malathi Veeraraghavan (UVA), John Blondin (NC St.), and Nageswara Rao (ORNL) for $3.5M Collaborative Research: Cyber Defense Technology Experimental Research Network (DETER) –Pis: Shankar Sastry, and Clifford Neuman for $6M Collaborative Research: PlanetLab: An Overlay Testbed for Disruptive Network Services –Larry Peterson, David Culler and Thomas Anderson for $3.2M WAN-in-Lab: PI - Steven Low, Caltech,(co-funded w/ EIA) for $600k
22 NSF Middleware Initiative Purpose - To design, develop, deploy and support a set of reusable, expandable set of middleware functions and services that benefit applications in a networked environment Program created by Alan Blatecky 2003 marks the 3rd set of awards
NSF Middleware Initiative Program Awards 20 awards totaling $9M –10 “System Integrator” awards Focus – to further develop the integration and support infrastructure of middleware for the longer term Themes - extending and deepening current activities, and expanding into new areas –10 smaller awards focused on near-term capabilities and tool development Focus – to encourage the development of additional new middleware components and capabilities for the NMI program
24 NMI “SI” Awards in 2003 Disseminating and Supporting Middleware Infrastructure: Engaging and Expanding Scientific Grid Communities –PI: Randy Butler (NCSA) for $1.8M Designing and Building a National Middleware Infrastructure (NMI-2) –PI: Carl Kesselman (USC/ISI) for $1.8M An Integrative Testing Framework for Grid Middleware and Grid Environments –PI: Miron Livny (U Wisc) for $1.8M Extending Integrated Middleware to Collaborative Environments in Research and Education –PI: Klingenstein (UCAID) for $1.8M Instruments and Sensors as Network Services: Instruments as First Class Members of the Grid –PI: Rick McMullen (Indiana) for $1.6M Collaborative Proposal: Middleware for Grid Portal Development –Pis: Pierce (Indiana - Lead), Alameda, Severance, Thomas, and von Laszewski for $3M
25 NMI - Looking Ahead There will be an NMI solicitation in 2004 Exact funding level not set NMI program is expected to be a primary focus area under CISE’s new division - Shared Cyberinfrastructure October 23 Review at NSF for existing activities among the Grids Center and EDIT teams
26 Strategic Technologies for the Internet (STI) Theme areas in 2003 –complex network monitoring, problem detection and resolution mechanisms –development of automated and advanced network tools –networked applications tools or network-based middleware –creation of usable and widely deployable networking applications that promote collaborative research and information sharing. –innovative access network technologies $5.2M in FY 2003 funds 14 awards
27 STI Awards in 2003 A Security Architecture for IP Telephony PI: Dipak Ghosal (UC Davis) for $160k Network Startup Resource Center (NSRC) PI: Joanne Hugi (Oregon) for $385k Plethora: A Wide-Area Read-Write Object Repository for the Internet PI: Suresh Jagganath (Purdue) for $550k Marist Grid Collaboration in Support of Advanced Internet and Research Applications PI: Kamran Khan (Marist) for $500k Implementation of a Handle/DNS Server PI: Robert Khan (CNRI) for $500k
28 STI Awards in 2003 Effective Diagnostic Strategies for Wide Area Networks PI: Matt Mathis (CMU) for $900k Development of an Infrastructure for Real-Time Super Media over the Internet PI: Mathew Mutka (Michigan St.) for $750k Viable Network Defense for Scientific Research Institutions PI: Vern Paxson (ICSI) for $900k The Strategic Technology Astronomy Research Team (START) Collaboratory: Broadening Participation in Authentic Astronomy Research PI: Morteza Rahimi (NWU) for $600k Network Measurement, Monitoring and Analysis in Cluster Computing PI: Tony Rimovsky (UIUC) for $800k
29 STI Awards in 2003 Towards more Secure Inter-Domain Routing PI: Aviel Rubin (JHU) for $615k XCP Development Aaron Falk (ISI) for $640k Media Aware Congestion Control Herbert Schorr (USC) for $710k Self-Organizing Spectrum Allocation David Staelin (MIT) for $520k
30 $10 M TeraGrid Extensions Program FY 2003 Open to all academic institutions and FFRDCs Provide connection to one of the TeraGrid Hubs –Hardware for connection including router hardware for hubs, router hardware for institution and fiber connection to Los Angeles or Chicago 13 Proposals received on June 9 Proposals panel review on June 26, 27 3 Awards announced on October 29 –Additional hub placed in Atlanta
31 FY 2003 Terascale Extension Awards $3.9 M to Oak Ridge National Laboratory (ORNL) to establish a 10-Gbps network connection through a new ETF network hub to be located in Atlanta. –connecting ORNL's High Flux Isotope Reactor and Spallation Neutron Source instruments, as well as Center for Computational Sciences that houses leading edge high-end computing and storage resources $3 M to Indiana University and Purdue University to build a 20- gigabit-per-second connection from those institutions through Indianapolis to the existing ETF hub in Chicago. –6.26 teraflops of computing capability; 400 terabytes of data storage capacity; visualization resources; specialized instrumentation including the Purdue Terrestrial Observatory and a number of life science data sets. $3.2 M to The University of Texas, the Texas Advanced Computing Center (TACC) to establish a 10-gigabit-per-second through the new ETF hub in Atlanta. –access to high-end computers capable of 6.2 teraflops, its new terascale visualization system, the center's 2.8-petabyte mass storage system and geoscience data collections.
32 Extensible Terascale Facility 2003
33 Finally, a words on the CISE Reorganization Drivers - CISE’s growing importance Enormous impact of IT on all areas of NSF National leader in computer, information science and engineering Critical to success of cybertrust, cyberscience, national and homeland defense, and continued economic growth Conceptual Drivers Atkins report: “CISE must be deeply involved as a technology user and as a technology leader for the overall [CI] program”. CISE Advisory Committee: “CISE organization structure should promote intellectual integrity and cohesiveness, should be flexible and fundamental enough to last 10 years”.
34 Reorg Objectives Capitalize on emerging scientific opportunities Build on CISE’s and NSF’s organizational successes Ensure integration of education with research in all CISE activities Sharpen programmatic focus and enhance budget flexibility Enhance intellectual coherence in CISE cross-cutting, thematic areas Increase research and education opportunities that will broaden participation in CISE activities
35 Additional Objectives Decrease the number of CISE programs thereby combating the tendency to reductionism in programmatic activity while promoting more integration Encourage the growth of cohesive scientific communities in important or emerging sub-disciplines Provide for more flexibility in defining and redefining cross-cutting priorities and emerging priorities of national and/or societal interest Empower CISE staff with the organizational, budget, and management support necessary to remain focused on key programmatic activities and goals Improve effectiveness and efficiency of CISE business practices by reducing program overlap and enhancing program synergy Position the Directorate for continued, strong budgetary growth and programmatic evolution.
36 Proposed Divisions and Areas of Responsibility Division of Computing & Communication Foundations (CCF) –Formal and mathematical foundations –Foundations of computing processes and artifacts –Emerging models for technology and computation Division of Computer and Network Systems (CNS) –Computer systems –Networks –Computing research infrastructure Division of Information and Intelligent Systems (IIS) –Systems in context –Understanding, inference, & data –Data-driven science Division of Shared Cyberinfrastructure (SCI) –Infrastructure planning, constructing, commissioning, & operations Timeline - sometime between now and end-of-year
37 For more info on NSF awards Fastlane.nsf.gov