CyberInfrastructure and Office of CyberInfrastructure OCI to: SURA Information Technology Committee - PowerPoint PPT Presentation

1 / 27
About This Presentation
Title:

CyberInfrastructure and Office of CyberInfrastructure OCI to: SURA Information Technology Committee

Description:

Science and engineering capability (logrithmic scale) Typical university ... CI impacts and enables the broad spectrum of science and engineering activities ... – PowerPoint PPT presentation

Number of Views:83
Avg rating:3.0/5.0
Slides: 28
Provided by: josl2
Learn more at: http://sura.org
Category:

less

Transcript and Presenter's Notes

Title: CyberInfrastructure and Office of CyberInfrastructure OCI to: SURA Information Technology Committee


1
CyberInfrastructureandOffice of
CyberInfrastructure(OCI)to SURA Information
Technology Committee Meeting
  • José Muñoz , Ph.D.
  • Office of CyberInfrastructure
  • Deputy Office Director
  • Senior Science Advisor

2
Outline
  • NSF organizational changes
  • CyberInfrastructure (CI) at NSF
  • Strategic Planning
  • New HPC Acquisition
  • Related OCI activities
  • Summary

3
CyberInfrastructure (CI) Governance
  • Dr. Dan Atkins (University of Michigan) selected
    as new OCI Director
  • served as Chair of NSF's Blue-Ribbon Advisory
    Panel on Cyberinfrastructure Revolutionizing
    Science and Engineering Through
    Cyberinfrastructure (2003)
  • tenure begins June 2006
  • CyberInfrastructure Council (CIC) created
  • NSF ADs and ODs, chaired by Dr. Bement (NSF
    Dir.)
  • CIC responsible for shared stewardship and
    ownership of NSFs cyberinfrastructure portfolio
  • Advisory Committee for NSFs CI activities and
    portfolio created
  • candidate members have been identified.
  • first meeting June 2006?

4
CI Governance
  • NSF High-End Computing Coordinating Group
  • representatives from each NSF Directorate
  • Chaired by OCI
  • Meets every other week
  • SCI to OCI Realignment
  • SCI was moved from CISE to Office of the
    Director and became Office of CyberInfrastructure
    (OCI)
  • Budget transferred
  • Ongoing projects and personnel transferred
  • OCI is focused on provision of production-quality
    CI for research and education
  • CISE remains focused on basic CS research and
    education mission
  • CI software and hardware
  • Collaborate with other NSF Directorates
  • CI Strategic Planning Process Underway
  • NSF CyberInfrastructure Vision document

5
Office of CyberInfrastructure
Debra Crawford Office Director (Interim) José
Muñoz Dep. Office Dir.
Judy Hayden Priscilla Bezdek Mary Daley Irene
Lombardo Allison Smith
ANL RP IU RP PU RP ORNL RP TACC RP MRI REU Sites
ETF GIG EIN IRNC Condor NMI Integ. Optiputer
CI-TEAM EPSCOR GriPhyN Disun CCG NMI
HPC Acq. NCSA Core NCSA RP PSC RP
STI NMI Dev. CyberSecurity
SDSC Core SDSC RP
Kevin Thompson Program Director
Guy Almes Program Director
Fillia Makedon Program Director
Doug Gatchell Program Director
Miriam Heller Program Director
Steve Meacham Program Director
SBE CyberTools SBE POC
Frank Scioli Program Director
(Vacancy) Program Director (Software)
Vittal Rao Program Director
6
CI Vision Document4 Interrelated Plans
Data, Data Analysis Visualization
High Performance Computing
7
Strategic Plan (FY 2006 2010)
Completed in Summer 2006
8
CyberInfrastructure Budgets
OCI Budget 127M (FY06)
NSF 2006 CI Budget
NMI
TESTBEDS
HPC ACQ.
25
IRNC
Research
FY07 182.42 (Request)
directorates
OCI
75
CI-TEAM
CORE
ETF
HPC hardware acquisitions, OM, and user support
as a fraction of NSFs overall CI budget
9
Recent CI Activities
  • HPC Solicitation Released September 27, 2005
  • Performance Benchmarks Identified November 9
  • Proposals were due 10Feb06
  • Continuing Work on CI Vision Document
  • Four, NSF-wide SWOT Teams developing Strategic
    and Implementation Plans covering various aspects
    of CI
  • http//www.nsf.gov/dir/index.jsp?orgOCI
  • Ongoing Interagency Discussions
  • Committee on Science
  • Office of Science and Technology Policy
  • Agency-to-agency (DARPA, HPCMOD, DOE/OS, NNSA,
    NIH)


10
Acquisition Strategy
Science and engineering capability (logrithmic
scale)
Track 1 system(s)
Track 2 systems
Typical university HPC systems
FY06
FY09
FY08
FY07
FY10
11
HPC Acquisition Activities
  • HPC acquisition will be driven by the needs of
    the SE community
  • RFI held for interested Resource Providers and
    HPC vendors on 9 Sep 2005
  • First in a series of HPC SE requirements
    workshops held 20-21 Sep 2005
  • Generated Application Benchmark Questionnaire
  • Attended by 77 scientists and engineers

12
TeraGrid What is It?
  • Integration of services provided by grid
    technologies
  • Distributed, open architecture.
  • GIG responsible for integration
  • Software integration (including the common
    software stack, CTSS)
  • Base infrastructure (security, networking, and
    operations)
  • User support
  • Community engagement (including the Science
    Gateways activities)
  • 8 Resource Providers (with separate awards)
  • PSC, TACC, NCSA, SDSC, ORNL, Indiana, Purdue,
    Chicago/ANL
  • Several other institutions participate in
    TeraGrid as a sub-awardees of the GIG
  • New sites may join as Resource Partners
  • TeraGrid
  • Provides a unified user environment to support
    high-capability, production-quality
    cyberinfrastructure services for science and
    engineering research.
  • Provides new SE opportunities by making new
    ways of using distributed resources and services
  • Examples of services include
  • HPC
  • Data collections
  • Visualization servers
  • Portals

13
NSF Middleware Initiative (NMI)
  • Develop, deploy and sustain a set of reusable and
    expandable
  • middleware functions that benefit many science
    and engineering
  • applications in a networked environment.
  • "open standards
  • international collaboration
  • sustainable
  • scalable and securable
  • Program Solicitations between 2001-2004 funded
    over 40 development awards and a series of
    integration awards
  • Integration award highlights include NMI Grids
    Center (e.g. Build and Test), Campus Middleware
    Services (e.g. Shibolleth) and Nanohub
  • OCI made a major award in middleware in November
    2005 to Foster/Kesselman
  • "Community Driven Improvement of Globus
    Software", 13.3M award over 5 years

14
2005 IRNC Awards
International Research Network Connections
  • Awards
  • TransPAC2 (U.S. Japan and beyond)
  • GLORIAD (U.S. China Russia Korea)
  • Translight/PacificWave (U.S. Australia)
  • TransLight/StarLight (U.S. Europe)
  • WHREN (U.S. Latin America)

15
Learning and Our 21st Century CI
WorkforceCI-TEAM Demonstration Projects
  • Input 70 projects / 101 proposals / 17 (24)
    projects were collaborative
  • Outcomes
  • 15.7 success rate 11 projects (14 proposals)
    awarded up to 250,000 over 1-2 years related to
    BIO, CISE, EHR, ENG, GEO, MPS
  • Broadening Participation for CI Workforce
    Development
  • Alvarez (FIU) CyberBridges
  • Crasta (VaTech) Project-Centric Bioinformatics
  • Fortson (Adler) CI-Enabled 21st c. Astronomy
    Training for HS Science Teachers
  • Fox (IU) - Bringing Minority Serving Institution
    Faculty into CI e-Science Communities
  • Gordon (OSU) Leveraging CI to Scale-Up a
    Computational Science U/G Curriculum
  • Panoff (Shodor) Pathways to CyberInfrastructure
    CI through Computational Science
  • Takai (SUNY Stonybrook) High School Distributed
    Search for Cosmic Rays (MARIACHI)
  • Developing and Implementing CI Resources for CI
    Workforce Development
  • DiGiano (SRI) Cybercollaboration Between
    Scientists and Software Developers
  • Figueiredo (UFl) In-VIGO/Condor-G MW for
    Coastal Estuarine Science CI Training
  • Regli (Drexel) CI for Creation and Use of
    Multi-Disciplinary Engineering Models
  • Simpson (PSU) CI-Based Engineering Repositories
    for Undergraduates (CIBER-U)

New CI-TEAM SOLICITATION COMING MARCH 2006
16
How it all fits together
HPC
ETF CORE NMI
ETF CORE
COVO
LWD
CI-TEAM
CI-TEAM ETF IRNC
17
NSF HECURA 2004
  • FY 2004 NSF/DARPA/DOE activity focused on
    research in
  • Languages
  • Compilers
  • Libraries
  • 100 proposals submitted in July 2005
  • 82 projects submitted by 57 US academic
    institutions and non-profit organizations
  • Includes no-cost national lab and industrial lab
    collaborators
  • Nine projects were awarded
  • Tools and libraries for high-end computing
  • Resource management
  • Reliability of high-end systems

18
NSF HECURA 2005/2006 FOCUS
  • I/O, file and storage systems design for
    efficient, high throughput data storage,
    retrieval and management in the HEC environment.
  • hardware and software tools for design,
    simulation, benchmarking, performance measurement
    and tuning of file and storage systems.

19
HECURA 2005/2006 SCOPE(CISE)
  • File Systems Research
  • Future File Systems related protocols
  • I/O middleware
  • Quality of Service
  • Security
  • Management, reliability, and availability at
    scale
  • Archives/Backups as extensions to file systems
  • Novel storage devices for the IO stack
  • I/O Architectures
  • Hardware and software tools for design,
    simulation of I/O, file and storage systems.
  • Efficient benchmarking, tracing, performance
    measurement and tuning tools of I/O, file and
    storage systems

20
Benchmarking
  • Broad inter-agency interest
  • Use of benchmarking for performance prediction
  • valuable when target systems are not readily
    available either because
  • Inaccessible (e.g. secure)
  • Does not exist at sufficient scale
  • In various stages of design
  • Useful for what-if analysis
  • Suppose I double the memory on my Redstorm?
  • Nirvana (e.g. Snavely/SDSC)
  • Abstract away the application application
    signatures
  • Platform independent
  • Abstract away the hardware platform signature
  • Convolve the signatures to provide an assessment

21
HPC Benchmarking
  • HPC Challenge Benchmarks (http//icl.cs.utk.edu/h
    pcc/)
  • HPL - the Linpack TPP benchmark which measures
    the floating point rate of execution for solving
    a linear system of equations.
  • DGEMM - measures the floating point rate of
    execution of double precision real matrix-matrix
    multiplication.
  • STREAM - a simple synthetic benchmark program
    that measures sustainable memory bandwidth (in
    GB/s) and the corresponding computation rate for
    simple vector kernel.
  • PTRANS (parallel matrix transpose) - exercises
    the communications where pairs of processors
    communicate with each other simultaneously. It is
    a useful test of the total communications
    capacity of the network.
  • RandomAccess - measures the rate of integer
    random updates of memory (GUPS).
  • FFTE - measures the floating point rate of
    execution of double precision complex
    one-dimensional Discrete Fourier Transform (DFT).
  • Communication bandwidth and latency - a set of
    tests to measure latency and bandwidth of a
    number of simultaneous communication patterns
    based on b_eff (effective bandwidth benchmark).

22
DARPA HPCS
  • Partitioned Global Address Space (PGAS)
    programming paradigm
  • Intended to support scaling to 1000s of
    processors
  • Co-Array Fortran
  • Unified Parallel C
  • Crays Chapel
  • IBMs X10
  • Suns Fortress
  • DARPA HPCS productivity activities
  • HPC specific programming environments?

?
23
OCI INVESTMENT HIGHLIGHTS
  • Leadership Class High-Performance Computing
    System Acquistion (50M)
  • Data- and Collaboration-Intensive Software
    Services (25.7M)
  • Conduct applied research and development
  • Perform scalability/reliability tests to explore
    tool viability
  • Develop, harden and maintain software tools and
    services
  • Provide software interoperability

24
HPC Acquisition - Track 1
  • Increased funding will support first phase of a
    petascale system acquisition
  • Over four years NSF anticipates investing 200M
  • Acquisition is crititical to NSFs multi-year
    plan to deploy and support world-class HPC
    environment
  • Collaborating with sister agencies with a stake
    in HPC
  • DARPA, HPCMOD, DOE/OS, DOE/NNSA, NIH

25
CI SummaryThe Tide that Raises All Boats
  • CI impacts and enables the broad spectrum of
    science and engineering activities
  • NSF CI deployment/acquisition activities must be
    driven by the needs of the science, engineering
    and education communities
  • CI is more than just big iron
  • Many opportunities to work with other federal
    agencies that develop, acquire and/or use various
    CI resources
  • Work required in all aspects of CI software
  • application and middleware for petascale systems
  • systems software
  • CI has been, and will continue to be, an
    effective mechanism for broadening participation

26
In the end
Its all about the SCIENCE
27
(No Transcript)
Write a Comment
User Comments (0)
About PowerShow.com