Title: CyberInfrastructure and Office of CyberInfrastructure OCI to: SURA Information Technology Committee
1CyberInfrastructureandOffice of
CyberInfrastructure(OCI)to SURA Information
Technology Committee Meeting
- José Muñoz , Ph.D.
- Office of CyberInfrastructure
- Deputy Office Director
- Senior Science Advisor
2Outline
- NSF organizational changes
- CyberInfrastructure (CI) at NSF
- Strategic Planning
- New HPC Acquisition
- Related OCI activities
- Summary
3 CyberInfrastructure (CI) Governance
- Dr. Dan Atkins (University of Michigan) selected
as new OCI Director - served as Chair of NSF's Blue-Ribbon Advisory
Panel on Cyberinfrastructure Revolutionizing
Science and Engineering Through
Cyberinfrastructure (2003) - tenure begins June 2006
- CyberInfrastructure Council (CIC) created
- NSF ADs and ODs, chaired by Dr. Bement (NSF
Dir.) - CIC responsible for shared stewardship and
ownership of NSFs cyberinfrastructure portfolio - Advisory Committee for NSFs CI activities and
portfolio created - candidate members have been identified.
- first meeting June 2006?
4CI Governance
- NSF High-End Computing Coordinating Group
- representatives from each NSF Directorate
- Chaired by OCI
- Meets every other week
- SCI to OCI Realignment
- SCI was moved from CISE to Office of the
Director and became Office of CyberInfrastructure
(OCI) - Budget transferred
- Ongoing projects and personnel transferred
- OCI is focused on provision of production-quality
CI for research and education - CISE remains focused on basic CS research and
education mission - CI software and hardware
- Collaborate with other NSF Directorates
- CI Strategic Planning Process Underway
- NSF CyberInfrastructure Vision document
5Office of CyberInfrastructure
Debra Crawford Office Director (Interim) José
Muñoz Dep. Office Dir.
Judy Hayden Priscilla Bezdek Mary Daley Irene
Lombardo Allison Smith
ANL RP IU RP PU RP ORNL RP TACC RP MRI REU Sites
ETF GIG EIN IRNC Condor NMI Integ. Optiputer
CI-TEAM EPSCOR GriPhyN Disun CCG NMI
HPC Acq. NCSA Core NCSA RP PSC RP
STI NMI Dev. CyberSecurity
SDSC Core SDSC RP
Kevin Thompson Program Director
Guy Almes Program Director
Fillia Makedon Program Director
Doug Gatchell Program Director
Miriam Heller Program Director
Steve Meacham Program Director
SBE CyberTools SBE POC
Frank Scioli Program Director
(Vacancy) Program Director (Software)
Vittal Rao Program Director
6CI Vision Document4 Interrelated Plans
Data, Data Analysis Visualization
High Performance Computing
7Strategic Plan (FY 2006 2010)
Completed in Summer 2006
8CyberInfrastructure Budgets
OCI Budget 127M (FY06)
NSF 2006 CI Budget
NMI
TESTBEDS
HPC ACQ.
25
IRNC
Research
FY07 182.42 (Request)
directorates
OCI
75
CI-TEAM
CORE
ETF
HPC hardware acquisitions, OM, and user support
as a fraction of NSFs overall CI budget
9Recent CI Activities
- HPC Solicitation Released September 27, 2005
- Performance Benchmarks Identified November 9
- Proposals were due 10Feb06
- Continuing Work on CI Vision Document
- Four, NSF-wide SWOT Teams developing Strategic
and Implementation Plans covering various aspects
of CI - http//www.nsf.gov/dir/index.jsp?orgOCI
- Ongoing Interagency Discussions
- Committee on Science
- Office of Science and Technology Policy
- Agency-to-agency (DARPA, HPCMOD, DOE/OS, NNSA,
NIH) -
10Acquisition Strategy
Science and engineering capability (logrithmic
scale)
Track 1 system(s)
Track 2 systems
Typical university HPC systems
FY06
FY09
FY08
FY07
FY10
11HPC Acquisition Activities
- HPC acquisition will be driven by the needs of
the SE community - RFI held for interested Resource Providers and
HPC vendors on 9 Sep 2005 - First in a series of HPC SE requirements
workshops held 20-21 Sep 2005 - Generated Application Benchmark Questionnaire
- Attended by 77 scientists and engineers
12TeraGrid What is It?
- Integration of services provided by grid
technologies - Distributed, open architecture.
- GIG responsible for integration
- Software integration (including the common
software stack, CTSS) - Base infrastructure (security, networking, and
operations) - User support
- Community engagement (including the Science
Gateways activities) - 8 Resource Providers (with separate awards)
- PSC, TACC, NCSA, SDSC, ORNL, Indiana, Purdue,
Chicago/ANL - Several other institutions participate in
TeraGrid as a sub-awardees of the GIG - New sites may join as Resource Partners
- TeraGrid
- Provides a unified user environment to support
high-capability, production-quality
cyberinfrastructure services for science and
engineering research. - Provides new SE opportunities by making new
ways of using distributed resources and services
- Examples of services include
- HPC
- Data collections
- Visualization servers
- Portals
13NSF Middleware Initiative (NMI)
- Develop, deploy and sustain a set of reusable and
expandable - middleware functions that benefit many science
and engineering - applications in a networked environment.
- "open standards
- international collaboration
- sustainable
- scalable and securable
- Program Solicitations between 2001-2004 funded
over 40 development awards and a series of
integration awards - Integration award highlights include NMI Grids
Center (e.g. Build and Test), Campus Middleware
Services (e.g. Shibolleth) and Nanohub - OCI made a major award in middleware in November
2005 to Foster/Kesselman - "Community Driven Improvement of Globus
Software", 13.3M award over 5 years
142005 IRNC Awards
International Research Network Connections
- Awards
- TransPAC2 (U.S. Japan and beyond)
- GLORIAD (U.S. China Russia Korea)
- Translight/PacificWave (U.S. Australia)
- TransLight/StarLight (U.S. Europe)
- WHREN (U.S. Latin America)
15Learning and Our 21st Century CI
WorkforceCI-TEAM Demonstration Projects
- Input 70 projects / 101 proposals / 17 (24)
projects were collaborative - Outcomes
- 15.7 success rate 11 projects (14 proposals)
awarded up to 250,000 over 1-2 years related to
BIO, CISE, EHR, ENG, GEO, MPS - Broadening Participation for CI Workforce
Development - Alvarez (FIU) CyberBridges
- Crasta (VaTech) Project-Centric Bioinformatics
- Fortson (Adler) CI-Enabled 21st c. Astronomy
Training for HS Science Teachers - Fox (IU) - Bringing Minority Serving Institution
Faculty into CI e-Science Communities - Gordon (OSU) Leveraging CI to Scale-Up a
Computational Science U/G Curriculum - Panoff (Shodor) Pathways to CyberInfrastructure
CI through Computational Science - Takai (SUNY Stonybrook) High School Distributed
Search for Cosmic Rays (MARIACHI) - Developing and Implementing CI Resources for CI
Workforce Development - DiGiano (SRI) Cybercollaboration Between
Scientists and Software Developers - Figueiredo (UFl) In-VIGO/Condor-G MW for
Coastal Estuarine Science CI Training - Regli (Drexel) CI for Creation and Use of
Multi-Disciplinary Engineering Models - Simpson (PSU) CI-Based Engineering Repositories
for Undergraduates (CIBER-U)
New CI-TEAM SOLICITATION COMING MARCH 2006
16How it all fits together
HPC
ETF CORE NMI
ETF CORE
COVO
LWD
CI-TEAM
CI-TEAM ETF IRNC
17NSF HECURA 2004
- FY 2004 NSF/DARPA/DOE activity focused on
research in - Languages
- Compilers
- Libraries
- 100 proposals submitted in July 2005
- 82 projects submitted by 57 US academic
institutions and non-profit organizations - Includes no-cost national lab and industrial lab
collaborators - Nine projects were awarded
- Tools and libraries for high-end computing
- Resource management
- Reliability of high-end systems
18NSF HECURA 2005/2006 FOCUS
- I/O, file and storage systems design for
efficient, high throughput data storage,
retrieval and management in the HEC environment.
- hardware and software tools for design,
simulation, benchmarking, performance measurement
and tuning of file and storage systems.
19HECURA 2005/2006 SCOPE(CISE)
- File Systems Research
- Future File Systems related protocols
- I/O middleware
- Quality of Service
- Security
- Management, reliability, and availability at
scale - Archives/Backups as extensions to file systems
- Novel storage devices for the IO stack
- I/O Architectures
- Hardware and software tools for design,
simulation of I/O, file and storage systems. - Efficient benchmarking, tracing, performance
measurement and tuning tools of I/O, file and
storage systems
20Benchmarking
- Broad inter-agency interest
- Use of benchmarking for performance prediction
- valuable when target systems are not readily
available either because - Inaccessible (e.g. secure)
- Does not exist at sufficient scale
- In various stages of design
- Useful for what-if analysis
- Suppose I double the memory on my Redstorm?
- Nirvana (e.g. Snavely/SDSC)
- Abstract away the application application
signatures - Platform independent
- Abstract away the hardware platform signature
- Convolve the signatures to provide an assessment
21HPC Benchmarking
- HPC Challenge Benchmarks (http//icl.cs.utk.edu/h
pcc/) - HPL - the Linpack TPP benchmark which measures
the floating point rate of execution for solving
a linear system of equations. - DGEMM - measures the floating point rate of
execution of double precision real matrix-matrix
multiplication. - STREAM - a simple synthetic benchmark program
that measures sustainable memory bandwidth (in
GB/s) and the corresponding computation rate for
simple vector kernel. - PTRANS (parallel matrix transpose) - exercises
the communications where pairs of processors
communicate with each other simultaneously. It is
a useful test of the total communications
capacity of the network. - RandomAccess - measures the rate of integer
random updates of memory (GUPS). - FFTE - measures the floating point rate of
execution of double precision complex
one-dimensional Discrete Fourier Transform (DFT).
- Communication bandwidth and latency - a set of
tests to measure latency and bandwidth of a
number of simultaneous communication patterns
based on b_eff (effective bandwidth benchmark).
22DARPA HPCS
- Partitioned Global Address Space (PGAS)
programming paradigm - Intended to support scaling to 1000s of
processors - Co-Array Fortran
- Unified Parallel C
- Crays Chapel
- IBMs X10
- Suns Fortress
- DARPA HPCS productivity activities
- HPC specific programming environments?
?
23OCI INVESTMENT HIGHLIGHTS
- Leadership Class High-Performance Computing
System Acquistion (50M) - Data- and Collaboration-Intensive Software
Services (25.7M) - Conduct applied research and development
- Perform scalability/reliability tests to explore
tool viability - Develop, harden and maintain software tools and
services - Provide software interoperability
24HPC Acquisition - Track 1
- Increased funding will support first phase of a
petascale system acquisition - Over four years NSF anticipates investing 200M
- Acquisition is crititical to NSFs multi-year
plan to deploy and support world-class HPC
environment - Collaborating with sister agencies with a stake
in HPC - DARPA, HPCMOD, DOE/OS, DOE/NNSA, NIH
25CI SummaryThe Tide that Raises All Boats
- CI impacts and enables the broad spectrum of
science and engineering activities - NSF CI deployment/acquisition activities must be
driven by the needs of the science, engineering
and education communities - CI is more than just big iron
- Many opportunities to work with other federal
agencies that develop, acquire and/or use various
CI resources - Work required in all aspects of CI software
- application and middleware for petascale systems
- systems software
- CI has been, and will continue to be, an
effective mechanism for broadening participation
26In the end
Its all about the SCIENCE
27(No Transcript)