ACCI Recommendations on Long Term Cyberinfrastructure Issues: Building Future Development

Similar documents
Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21)

Cyberinfrastructure!

OneUConn IT Service Delivery Vision

National Science and Technology Council. Interagency Working Group on Digital Data

ESFRI Strategic Roadmap & RI Long-term sustainability an EC overview

Business Model for Global Platform for Big Data for Official Statistics in support of the 2030 Agenda for Sustainable Development

EPA Near-port Community Capacity Building: Tools and Technical Assistance for Collaborative Solutions

irods at TACC: Secure Infrastructure for Open Science Chris Jordan

Journey Towards Science DMZ. Suhaimi Napis Technical Advisory Committee (Research Computing) MYREN-X Universiti Putra Malaysia

Evaluating and Improving Cybersecurity Capabilities of the Electricity Critical Infrastructure

The EuroHPC strategic initiative

Introduction to FREE National Resources for Scientific Computing. Dana Brunson. Jeff Pummill

Metadata Framework for Resource Discovery

SPARC 2 Consultations January-February 2016

Preamble. A Strategic Plan for the Internet2 Community Spring 2008

CANARIE Mandate Renewal Proposal

New Zealand Government IBM Infrastructure as a Service

Scientific Computing at SLAC. Amber Boehnlein

SLATE. Services Layer at the Edge. First Meeting of the National Research Platform Montana State University August 7-8, 2017

The Science and Technology Roadmap to Support the Implementation of the Sendai Framework for Disaster Risk Reduction

Indiana University Research Technology and the Research Data Alliance

Regional Centers Framework Update. Regional Project Evaluation Committee March 24, 2017

3.0 NETWORX ARCHITECTURE FOR IP-BASED SERVICES (L ) (M.2.1) (a), M.2.1.1(a))

High Performance Computing from an EU perspective

The U.S. National Spatial Data Infrastructure

Accelerate Your Enterprise Private Cloud Initiative

Higher Education in Texas: Serving Texas Through Transformational Education, Research, Discovery & Impact

National Research Data Cloud

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Department of Homeland Security Updates

I D C T E C H N O L O G Y S P O T L I G H T

einfrastructures Concertation Event

Research Infrastructures and Horizon 2020

Data publication and discovery with Globus

Georgia State University Cyberinfrastructure Plan

RSA Solution Brief. Managing Risk Within Advanced Security Operations. RSA Solution Brief

Overview of NIPP 2013: Partnering for Critical Infrastructure Security and Resilience October 2013

STRATEGIC PLAN

DT-ICT : Big data solutions for energy

Jetstream: Adding Cloud-based Computing to the National Cyberinfrastructure

Creating a Digital Preservation Network with Shared Stewardship and Cost

Proposition to participate in the International non-for-profit Industry Association: Energy Efficient Buildings

Eastern Regional Network (ERN) Barr von Oehsen Internet2 Tech Exchange 10/16/2018

The Office of Infrastructure Protection

Potential for Technology Innovation within the Internet2 Community: A Five-Year View

First Session of the Asia Pacific Information Superhighway Steering Committee, 1 2 November 2017, Dhaka, Bangladesh.

XSEDE s Campus Bridging Project Jim Ferguson National Institute for Computational Sciences

IT Town Hall Meeting

Science-as-a-Service

CONCLUSIONS AND RECOMMENDATIONS

December 10, Statement of the Securities Industry and Financial Markets Association. Senate Committee on Banking, Housing, and Urban Development

Implementing Executive Order and Presidential Policy Directive 21

The National Center for Genome Analysis Support as a Model Virtual Resource for Biologists

Effective: 12/31/17 Last Revised: 8/28/17. Responsible University Administrator: Vice Chancellor for Information Services & CIO

Deanship of Academic Development. Comprehensive eportfolio Strategy for KFU Dr. Kathryn Chang Barker Director, Department of Professional Development

Grid Computing a new tool for science

RESOLUTION 47 (Rev. Buenos Aires, 2017)

Mapping to the National Broadband Plan

Cybersecurity-Related Information Sharing Guidelines Draft Document Request For Comment

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Improving Network Infrastructure to Enable Large Scale Scientific Data Flows and Collaboration (Award # ) Klara Jelinkova Joseph Ghobrial

Defense Security Service. Strategic Plan Addendum, April Our Agency, Our Mission, Our Responsibility

UNC Campus Security Initiatives Update. Business Affairs Committee May 9, 2017

Marine Transportation System Resilience: A Federal Agency Perspective

XSEDE Campus Bridging Tools Rich Knepper Jim Ferguson

ACF Interoperability Human Services 2.0 Overview. August 2011 David Jenkins Administration for Children and Families

Connectivity Services, Autobahn and New Services

Advancing Library Cyberinfrastructure for Big Data Sharing and Reuse. Zhiwu Xie

UNCLASSIFIED. UNCLASSIFIED R-1 Line Item #49 Page 1 of 10

21ST century enterprise. HCL Technologies Presents. Roadmap for Data Center Transformation

Call for Participation in AIP-6

Annual Report for the Utility Savings Initiative

Symantec Data Center Transformation

NC Education Cloud Feasibility Report

GN3plus External Advisory Committee. White Paper on the Structure of GÉANT Research & Development

DDN Annual High Performance Computing Trends Survey Reveals Rising Deployment of Flash Tiers & Private/Hybrid Clouds vs.

National Earthquake Hazards Reduction Program

CHAMELEON: A LARGE-SCALE, RECONFIGURABLE EXPERIMENTAL ENVIRONMENT FOR CLOUD RESEARCH

Edward Seidel Director, National Center for Supercomputing Applications Founder Prof. of Physics, U of Illinois

Woolwich CHC adopts the CIW as a strategic planning tool. A tool to shift the conversation. Using the CIW to advance community health and wellbeing

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

White Paper. View cyber and mission-critical data in one dashboard

Developing Applications with Networking Capabilities via End-to-End Software Defined Networking (DANCES)

Preparatory process of the second High-level United Nations Conference on South-South Cooperation

Overview of the Cybersecurity Framework

National Open Source Strategy

WSIS Forum Open Consultation Process

SOLUTION BRIEF RSA ARCHER IT & SECURITY RISK MANAGEMENT

Accessibility Design Guidelines

The Future of the Internet

Principles for a National Space Industry Policy

The US National Near-Earth Object Preparedness Strategy and Action Plan

The library s role in promoting the sharing of scientific research data

SDN/DANCES Project Update Developing Applications with Networking Capabilities via End-to-end SDN (DANCES)

e-infrastructures in FP7 INFO DAY - Paris

PREPARE FOR TAKE OFF. Accelerate your organisation s journey to the Cloud.

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

Open Grid Forum. OGF s Role in the Community

Modernizing Healthcare IT for the Data-driven Cognitive Era Storage and Software-Defined Infrastructure

Response to Industry Canada Consultation Developing a Digital Research Infrastructure Strategy

Transcription:

ACCI Recommendations on Long Term Cyberinfrastructure Issues: Building Future Development Jeremy Fischer Indiana University 9 September 2014 Citation: Fischer, J.L. 2014. ACCI Recommendations on Long Term Cyberinfrastructure Issues: Building Future Development. Indiana University, Bloomington, IN. September, 2014. Available at: http://hdl.handle.net/2022/18713

In 2009, the National Science Foundation s (NSF) Advisory Committee for Cyberinfrastructure (ACCI) created six task forces to make recommendations to the NSF in the strategic areas of cyberinfrastructure: Campus Bridging, Data, Grand Challenges and Virtual Organizations, High Performance Computing, Software and Tools, and Work Force Development. These task forces were led by ACCI members and consisted of members of the industrial and academic communities. Over a two-year period, through meetings, workshops, position and white papers, and other outreach, these task forces gathered ideas and information in their respective areas and composed a final report outlining recommendations for furthering cyberinfrastructure to support NSF research. In looking at the development of the next-generation cyberinfrastructure for NSF research, it is helpful to have the recommendations of the previous ACCI task forces. These recommendations may give a strong foundation for future designs. In this paper, the relevant findings of the various ACCI task forces will be covered and included. With this guidance as the bedrock, the next generation cyberinfrastructure will benefit from the experience of a wide swath of NSF researchers and contributors. Grand Challenges [1] NSF should work with the Department of Energy and other agencies in creation of an Interagency Working Group con CS&E or generally on Computational Science and Engineering, including Data-Intensive Computing, in the spirit of other NSF-wide working groups. This broad-based Working Group could provide input leading to important interagency collaborations on new programs, particularly in HPC, and could lead to more focused and efficient use of resources to address the Grand Challenges facing our nation. It is recommended that NSF, through OCI, continue to give priority to funding a sustained and diverse set of HPC and innovative equipment resources to support the wide range of needs within the research community. These needs include support for the development of technologies to meet the foremost challenges in HPC, such as the power-aware and application-sensitive architectures, new numerical algorithms to efficiently use petascale and exascale architectures and data flow and data analysis at the extreme scale. Recommendation from the NSF Advisory Committee for Cyberinfrastructure Task Force on Data and Visualization [2] Recognize data infrastructure and services as essential research assets fundamental to today s science and as long-term investments in national prosperity. Make specific budget provisions for the establishment and maintenance of data sets and services and the associated software and visualization tools infrastructure. Software for Science and Engineering [3] NSF should take leadership in promoting verification, validation, sustainability, and reproducibility through software developed with federal support. NSF support for software should entail collaborations among all of its divisions, related federal agencies, and private industry NSF should develop a consistent policy on open source software that promotes scientific discovery and encourages innovation. High Performance Computing [4] Develop a sustainable model to provide the academic research community with access, by 2015 2016, to a rich mix of HPC systems that: 1

o Deliver sustained performance of 20 100 petaflops on a broad range of science and engineering applications; o Are integrated into a comprehensive national CI environment; and o Are supported at national, regional, and/or campus levels. Invest now to prepare for exascale systems that will be available by 2018 2020. NSF should consider the use of co-design partnerships to provide the HPC systems and data CI needed to enable data-driven science. Establish a continuing process for soliciting community input on plans for HPC investments Campus Bridging [5] As part of a strategy of coherence between the NSF and campus cyberinfrastructure and reducing reimplementation of multiple authentication systems, the NSF should encourage the use of the InCommon Federation global federated system by using it in the services it deploys and supports, unless there are specific technical or risk management barriers. The NSF should create a new program funding high-speed (currently 10 Gbps) connections from campuses to the nearest landing point for a national network backbone. The design of these connections must include support for dynamic network provisioning services and must be engineered to support rapid movement of large scientific data sets. The NSF should establish a national cyberinfrastructure software roadmap. Through the Software Infrastructure for Sustained Innovation (SI 2 ) or other programs, the NSF should seek to systematically fund the creation and ongoing development and support of a suite of critical cyberinfrastructure software that identifies and establishes this roadmap, including cyberinfrastructure software for authentication and access control; computing cluster management; data movement; data sharing; data, metadata, and provenance management; distributed computation / cycle scavenging; parallel computing libraries; network performance analysis / debugging; VO collaboration; and scientific visualization. Funding for personnel should be a strong portion of such a strategy. The NSF should fund activities that support the evolution and maturation of cyberinfrastructure through careful analyses of needs (in advance of creating new cyberinfrastructure facilities) and outcomes (during and after the use of cyberinfrastructure facilities). The NSF should establish and fund processes for collecting disciplinary community requirements and planning long-term cyberinfrastructure software roadmaps to support disciplinary community research objectives. The NSF should likewise fund studies of cyberinfrastructure experiences to identify attributes leading to impact, and recommend a set of metrics for the development, deployment, and operation of cyberinfrastructure, including a set of guidelines for how the community should judge cyberinfrastructure technologies in terms of their technology readiness. All NSF-funded cyberinfrastructure implementations should include analysis of effectiveness including formal user surveys. All studies of cyberinfrastructure needs and outcomes, including ongoing studies of existing cyberinfrastructure In revisiting these recommendations from 2009, examining the progress in the years since, and finding the gaps between 2009 s recommendations and 2014 s practices, we can establish an updated framework for building the next generation cyberinfrastructure. Part of Indiana University s ongoing commitment to the XSEDE project and notably the Campus Bridging program is to help extend the scientific cyberinfrastructure seamlessly into more organizations as well as into organizations that are currently not well served or represented. Designing the next generation of cyberinfrastructure with that in mind might allow for both broader and deeper research that goes beyond the traditional organizations represented. 2

Indiana University has also significantly contributed in other ways to the recommendations the ACCI task forces provided. IU has partnered with the Texas Advanced Computing Center (TACC) to further the goals of creating collaborative space for data analysis and management. [6] Wrangler will provide fast, reliable, and replicated storage for researchers, a necessary part of cyberinfrastructure for managing and sharing data. IU has also been a major contributor to the Apache Airavata project. In contributing to Airavata, IU is helping to create sustainable, easily implemented, free software for executing and managing computational jobs and workflows on distributed computing resources including local clusters, supercomputers, national grids, academic and commercial clouds. [7] In implementing Airavata, IU and its collaborators are helping to lower the barriers to research computing and to simplify specific workflows. As part of the goals of supporting grand challenges, IU houses and supports one of two computation and data storage clusters in support of the ATLAS Midwest Tier 2 Center (MWT2). These systems provide a high-throughput, data-intensive analysis and simulation facility for proton collisions recorded by the ATLAS detector at the Large Hadron Collider (LHC), based at the CERN Laboratory in Geneva, Switzerland. [8] In combination with the Open Science Grid, whose Grid Operations Center is based at IU, these resources are interconnected with other Tier 1 and Tier 2 centers, creating a single service for researchers. Creating an infrastructure that is more easily accessible to a broader range of researchers at all types of universities should be a primary goal of any design ideas. In that process, designers must take into account the jumps in data size and the need to transport it quickly and easily for collaboration. Current state of the art focuses on reliability and robustness, and has made significant gains in providing solid services in a fashion familiar to a very large number of users. In order for the national cyberinfrastructure ecology to evolve, it must reach new research communities, including the social sciences and the humanities and support novel types of analyses. The design goals for this process must keep an emphasis on reducing barriers to collaboration, dealing with the increasing scale of data sets and analyses, and supporting new and infrastructures for analyses in a broad range of scientific domains. The end result should be a design for the next generation that concentrates on being inclusive, reliable, fast, and seamless. 3

References [1] NSF Advisory Committee for Cyberinfrastructure Task Force on Grand Challenges. Final Report. March 2011. http://www.nsf.gov/cise/aci/taskforces/taskforcereport_grandchallenges.pdf [2] NSF Advisory Committee for Cyberinfrastructure Task Force on Data and Visualization. Final Report. March 2011. http://www.nsf.gov/cise/aci/taskforces/taskforcereport_data.pdf [3] NSF Advisory Committee for Cyberinfrastructure Task Force on Software for Science and Engineering. Final Report. March 2011. http://www.nsf.gov/cise/aci/taskforces/taskforcereport_software.pdf [4] NSF Advisory Committee for Cyberinfrastructure Task Force on High Performance Computing. Final Report. March 2011. http://www.nsf.gov/cise/aci/taskforces/taskforcereport_hpc.pdf [5] NSF Advisory Committee for Cyberinfrastructure Task Force on Campus Bridging. Final Report. March 2011. http://www.nsf.gov/cise/aci/taskforces/taskforcereport_campusbridging.pdf [6] IU plays key role in two new supercomputing systems. IT News. January 14, 2014. http://itnews.iu.edu/articles/2014/iu-plays-key-role-in-two-new-national-supercomputing-systems.php [7] Apache Airavata. Retrieved September 9, 2014. http://airavata.apache.org/ [8] MWT2 : ATLAS Midwest Tier 2 Center. Laboratory for Advanced Numerical Simulations. Retrieved September 9, 2014. http://www.mcs.anl.gov/research/lans/research/index.php?p=proj_detail&id=86 Acknowledgements This document was developed with support from National Science Foundation (NSF) grant OCI-1053575 for XSEDE: extreme Science and Engineering Discovery Environment. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the NSF. More information is available at http://xsede.org This document was also made possible with support by the Indiana University Pervasive Technology Institute. More information is available at http://pti.iu.edu 4