3,104 results on '"Cyberinfrastructure"'
Search Results
2. Enhancing Vulnerability Prioritization in Cloud Computing Using Multi-View Representation Learning.
- Author
-
Ullman, Steven, Samtani, Sagar, Zhu, Hongyi, Lazarine, Ben, Chen, Hsinchun, and Nunamaker Jr., Jay F.
- Subjects
VIRTUAL machine systems ,WEB services ,CLOUD computing ,BEHAVIORAL research ,SECURITY personnel ,DEEP learning - Abstract
Cybersecurity is a present and growing concern that needs to be addressed with both behavioral and design-oriented research. Public cloud providers such as Amazon Web Services and federal funding agencies such as the National Science Foundation have invested billions of dollars into developing high-performance computing resources accessible to users through configurable virtual machine (VM) images. This approach offers users the flexibility of changing and updating their environment for their computational needs. Despite the substantial benefits, users often introduce thousands of vulnerabilities by installing open-source software packages and misconfiguring file systems. Given the scale of vulnerabilities, security personnel struggle to identify and prioritize vulnerable assets for remediation. In this research, we designed a novel unsupervised deep learning-based Multi-View Combinatorial-Attentive Autoencoder (MV-CAAE) to capture multi-dimensional vulnerability data and automatically identify groups of similar vulnerable compute instances to help facilitate the development of targeted remediation strategies. We rigorously evaluated the proposed MV-CAAE against state-of-the-art methods in three technical clustering experiments. Experiment results indicate that the MV-CAAE achieves V-measure scores (metric of cluster quality) 8 percent-48 percent higher than benchmark methods. We demonstrated the practical value through a comprehensive case study by clustering vulnerable VMs and gathering qualitative feedback from experienced security professionals through semi-structured interviews. The results indicated that clustering vulnerable assets can help prioritize vulnerable instances for remediation and enhance decision-making tasks. The present design-research work also contributes to our theoretical knowledge of cyber-defense. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Moving CyberGIS education forward: Knowing what matters and how it is decided.
- Author
-
Zhang, Zhe, King, Jerad, Wang, Shaowen, Sinton, Diana, Wilson, John, and Shook, Eric
- Subjects
- *
EDUCATIONAL resources , *CYBERINFRASTRUCTURE , *INFORMATION science , *CURRICULUM , *STUDENTS - Abstract
Maintaining educational resources and training materials as timely, current, and aligned with the needs of students, practitioners, and other users of geospatial technologies is a persistent challenge. This is particularly problematic within CyberGIS, a subfield of Geographic Information Science and Technology (GIS&T) that involves high‐performance computing and advanced cyberinfrastructure to address computation‐ and data‐intensive problems. In this study, we analyzed and compared content from two open educational resources: (1) a popular online web resource that regularly covers CyberGIS‐related topics (GIS Stack Exchange) and (2) existing and proposed content in the GIS&T Body of Knowledge. While current curricula may build a student's conceptual understanding of CyberGIS, there is a noticeable lack of resources for practical implementation of CyberGIS tools. The results highlight discrepancies between the attention and frequency of CyberGIS topics according to a popular online help resource and the CyberGIS academic community. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Development of a cost-efficient automated wildlife camera network in a European Natura 2000 site.
- Author
-
Kissling, W. Daniel, Evans, Julian C., Zilber, Rotem, Breeze, Tom D., Shinneman, Stacy, Schneider, Lindy C., Chalmers, Carl, Fergus, Paul, Wich, Serge, and Geelen, Luc H.W.T.
- Subjects
CONVOLUTIONAL neural networks ,WILDLIFE monitoring ,BIODIVERSITY monitoring ,OPTICAL sensors ,REMOTE sensing ,DEEP learning ,ANIMAL populations - Abstract
• Automating wildlife monitoring with wireless 4G cameras and end-to-end data streams. • Remote monitoring of sensor performance, API handling and automated task management. • Deep learning for automated identification of focal species and human detection. • Total cost saving of >40 % through automation, AI and less regular site visits. • Enabling technologies allow scaling-up of a cost-efficient biodiversity monitoring. Modern approaches with advanced technology can automate and expand the extent and resolution of biodiversity monitoring. We present the development of an innovative system for automated wildlife monitoring in a coastal Natura 2000 nature reserve of the Netherlands with 65 wireless 4G wildlife cameras which are deployed autonomously in the field with 12 V/2A solar panels, i.e. without the need to replace batteries or manually retrieve SD cards. The cameras transmit images automatically (through a mobile network) to a sensor portal, which contains a PostgreSQL database and functionalities for automated task scheduling and data management, allowing scientists and site managers via a web interface to view images and remotely monitor sensor performance (e.g. number of uploaded files, battery status and SD card storage of cameras). The camera trap sampling design combines a grid-based sampling stratified by major habitats with the camera placement along a traditional monitoring route, and with an experimental set-up inside and outside large herbivore exclosures. This provides opportunities for studying the distribution, habitat use, activity, phenology, population structure and community composition of wildlife species and allows comparison of traditional with novel monitoring approaches. Images are transferred via application programming interfaces to external services for automated species identification and long-term data storage. A deep learning model for species identification was tested and showed promising results for identifying focal species. Furthermore, a detailed cost analysis revealed that establishment costs of the automated system are higher but the annual operating costs much lower than those for traditional camera trapping, resulting in the automated system being >40 % more cost-efficient. The developed end-to-end data pipeline demonstrates that continuous monitoring with automated wildlife camera networks is feasible and cost-efficient, with multiple benefits for extending the current monitoring methods. The system can be applied in open habitats of other nature reserves with mobile network coverage. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. An Examination and Survey of Data Confidentiality Issues and Solutions in Academic Research Computing
- Author
-
Peisert, Sean
- Subjects
cyberinfrastructure ,confidentiality ,research computing ,sensitive data ,HIPAA ,security ,CUI ,FISMA ,privacy - Abstract
In research academic computing, may be natural to emphasize data integrity over confidentiality. However, there are numerous categories of academic research that do have data confidentiality requirements, from research that is simply embargoed until a future publication date to research that contains industry-owned proprietary information or is subject to government regulation. The contents of this report are based on numerous community conversations with leaders in academic research IT, data librarians, computer science researchers, computer security professionals, and others with roles involving using or enabling the use of sensitive data in academic research. The report discusses challenges to conducting research on data that is in some way sensitive, and solutions that are being used or could be used to address those challenges and enable the research to take place. Those solutions include technical solutions as well as administrative and procedural issues. The report concludes with recommendations to campuses on issues to consider in order to enable research on sensitive data while ensuring security and privacy as well as usability and usefulness of the environment hosting that data.
- Published
- 2023
6. Development of a cost-efficient automated wildlife camera network in a European Natura 2000 site
- Author
-
W. Daniel Kissling, Julian C. Evans, Rotem Zilber, Tom D. Breeze, Stacy Shinneman, Lindy C. Schneider, Carl Chalmers, Paul Fergus, Serge Wich, and Luc H.W.T. Geelen
- Subjects
API ,Artificial intelligence ,Biodiversity survey ,Convolutional neural network ,Cyberinfrastructure ,Dunes ,Ecology ,QH540-549.5 - Abstract
Modern approaches with advanced technology can automate and expand the extent and resolution of biodiversity monitoring. We present the development of an innovative system for automated wildlife monitoring in a coastal Natura 2000 nature reserve of the Netherlands with 65 wireless 4G wildlife cameras which are deployed autonomously in the field with 12 V/2A solar panels, i.e. without the need to replace batteries or manually retrieve SD cards. The cameras transmit images automatically (through a mobile network) to a sensor portal, which contains a PostgreSQL database and functionalities for automated task scheduling and data management, allowing scientists and site managers via a web interface to view images and remotely monitor sensor performance (e.g. number of uploaded files, battery status and SD card storage of cameras). The camera trap sampling design combines a grid-based sampling stratified by major habitats with the camera placement along a traditional monitoring route, and with an experimental set-up inside and outside large herbivore exclosures. This provides opportunities for studying the distribution, habitat use, activity, phenology, population structure and community composition of wildlife species and allows comparison of traditional with novel monitoring approaches. Images are transferred via application programming interfaces to external services for automated species identification and long-term data storage. A deep learning model for species identification was tested and showed promising results for identifying focal species. Furthermore, a detailed cost analysis revealed that establishment costs of the automated system are higher but the annual operating costs much lower than those for traditional camera trapping, resulting in the automated system being >40 % more cost-efficient. The developed end-to-end data pipeline demonstrates that continuous monitoring with automated wildlife camera networks is feasible and cost-efficient, with multiple benefits for extending the current monitoring methods. The system can be applied in open habitats of other nature reserves with mobile network coverage.
- Published
- 2024
- Full Text
- View/download PDF
7. Calibrating Spatial Stratified Heterogeneity for Heavy-Tailed Distributed Data.
- Author
-
Hu, Bisong, Wu, Tingting, Yin, Qian, Wang, Jinfeng, Jiang, Bin, and Luo, Jin
- Subjects
- *
DISTRIBUTED databases , *DATABASES , *CYBERINFRASTRUCTURE , *CONSISTENCY models (Computers) , *DISTRIBUTED computing - Abstract
The phenomena with within-strata characteristics that are more similar than between-strata characteristics are ubiquitous (e.g., land-use types and image classifications). It can be summarized as spatial stratified heterogeneity (SSH), which is measured and attributed using the geographical detector (Geodetector) q-statistic. SSH is typically calibrated by stratification and hundreds of algorithms have been developed. Little is discussed about the conditions of the methods. In this work, a novel stratification method based on head/tail breaks is introduced for the purpose of better capturing the SSH of geographical variables with a heavy-tailed distribution. Compared to conventional sample-based stratifications, the presented approach is a population-based optimized stratification that indicates an underlying scaling property in geographical spaces. It requires no prior knowledge or auxiliary variables and supports a naturally determined number of strata instead of being subjectively preset. In addition, our approach reveals the inherent hierarchical structure of geographical variables, characterizes its dominant components across all scales, and provides the potential to make the stratification meaningful and interpretable. The advantages were illustrated by several case studies in natural and social sciences. The proposed approach is versatile and flexible so that it can be applied for the stratification of both geographical and nongeographical variables and is conducive to advancing SSH-related studies as well. This study provides a new way of thinking for advocating spatial heterogeneity or scaling law and advances our understanding of geographical phenomena. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Software Training Outreach In HEP.
- Author
-
Malik, Sudhir, Cordero, Danelix, Elmer, Peter, LaMee, Adam, and Cecire, Ken
- Subjects
- *
PARTICLE physics , *HIGH school teachers , *CYBERINFRASTRUCTURE , *LARGE Hadron Collider , *PHYSICS education - Abstract
The NSF-funded IRIS-HEP "Training, Education & Outreach" program and QuarkNet are partnering to enable and expand software training for the high school teachers with a goal to tap, grow and diversify the talent pipeline from K-12 students for future cyberinfrastructure. The Institute for Research and Innovation in Software for High Energy Physics (IRIS-HEP) is a software institute that aims to develop the state-of-the-art software cyberinfrastructure for the High Luminosity Large Hadron Collider (HL-LHC) at CERN and other planned HEP experiments of the 2020's. QuarkNet provides professional development to K-12 physics teachers in particle physics content and teaching methods. The two projects have recently built a collaborative relationship where a well-established community of QuarkNet K-12 teachers has access to a wide training on software tools via its Data and Coding Camps supported by IRISHEP. The paper highlights the synergistic efforts and future plans. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Coffea-Casa: Building composable analysis facilities for the HL-LHC.
- Author
-
Albin, Sam, Attebury, Garhan, Bloom, Kenneth, Bockelman, Brian, Lundstedt, Carl, Shadura, Oksana, and Thiltges, John
- Subjects
- *
LARGE Hadron Collider , *DATA analysis , *CYBERINFRASTRUCTURE , *MACHINE learning , *GRAPHICS processing units - Abstract
The large data volumes expected from the High Luminosity LHC (HL-LHC) present challenges to existing paradigms and facilities for end-user data analysis. Modern cyberinfrastructure tools provide a diverse set of services that can be composed into a system that provides physicists with powerful tools that give them straightforward access to large computing resources, with low barriers to entry. The Coffea-Casa analysis facility (AF) provides an environment for end users enabling the execution of increasingly complex analyses such as those demonstrated by the Analysis Grand Challenge (AGC) and capturing the features that physicists will need for the HL-LHC. We describe the development progress of the Coffea-Casa facility featuring its modularity while demonstrating the ability to port and customize the facility software stack to other locations. The facility also facilitates the support of batch systems while staying Kubernetes-native. We present the evolved architecture of the facility, such as the integration of advanced data delivery services (e.g. ServiceX) and making data caching services (e.g. XCache) available to end users of the facility. We also highlight the composability of modern cyberinfrastructure tools. To enable machine learning pipelines at coffee-casa analysis facilities, a set of industry ML solutions adopted for HEP columnar analysis were integrated on top of existing facility services. These services also feature transparent access for user workflows to GPUs available at a facility via inference servers while using Kubernetes as enabling technology. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Efficient interface to the GridKa tape storage system.
- Author
-
Musheghyan, Haykuhi, Petzold, Andreas, Ressmann, Doris, Konstantinov, Preslav, Lobontu, Dorin-Daniel, Gottmann, Artur, Ambroj Pérez, Samuel, and Mol, Xavier
- Subjects
- *
COMPUTER software , *COMPUTER files , *HIGH performance computing , *CYBERINFRASTRUCTURE , *COMPUTER storage devices - Abstract
Providing high performance and reliable tape storage system is GridKa's top priority. The GridKa tape storage system was recently migrated from IBM SP to High Performance Storage System (HPSS) for LHC and non-LHC HEP experiments. These are two different tape backends and each has its own design and specifics that need to be studied thoroughly. Taking into account the features and characteristics of HPSS, a new approach has been developed for flushing and staging files to and from tape storage system. This new approach allows better optimized and efficient flush and stage operations and leads to a substantial improvement in the overall performance of the GridKa tape storage system. The efficient interface that was developed to use IBM SP is now adapted to the HPSS use-case to connect the access point from experiments to the tape storage system. This contribution provides details on these changes and the results of the Tape Challenge 2022 within the new HPSS tape storage configuration. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. Democratization is a Process, not a Destination: Operationalizing Ethics and Democratization in a Cyberinfrastructure for AI Project
- Author
-
Khan, Sadia, Morales, Alfonso, Plale, Beth, Akan, Ozgur, Editorial Board Member, Bellavista, Paolo, Editorial Board Member, Cao, Jiannong, Editorial Board Member, Coulson, Geoffrey, Editorial Board Member, Dressler, Falko, Editorial Board Member, Ferrari, Domenico, Editorial Board Member, Gerla, Mario, Editorial Board Member, Kobayashi, Hisashi, Editorial Board Member, Palazzo, Sergio, Editorial Board Member, Sahni, Sartaj, Editorial Board Member, Shen, Xuemin, Editorial Board Member, Stan, Mircea, Editorial Board Member, Jia, Xiaohua, Editorial Board Member, Zomaya, Albert Y., Editorial Board Member, Ziosi, Marta, editor, Sartor, Giovanni, editor, Cunha, João Miguel, editor, Trotta, Angelo, editor, and Wicke, Philipp, editor
- Published
- 2024
- Full Text
- View/download PDF
12. Leveraging Graph Neural Networks for Botnet Detection
- Author
-
Saad, Ahmed Mohamed Saad Emam, Filipe, Joaquim, Editorial Board Member, Ghosh, Ashish, Editorial Board Member, Prates, Raquel Oliveira, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Ortis, Alessandro, editor, Hameed, Alaa Ali, editor, and Jamil, Akhtar, editor
- Published
- 2024
- Full Text
- View/download PDF
13. A framework for developing a real-time lake phytoplankton forecasting system to support water quality management in the face of global change
- Author
-
Carey, Cayelan C., Calder, Ryan S. D., Figueiredo, Renato J., Gramacy, Robert B., Lofton, Mary E., Schreiber, Madeline E., and Thomas, R. Quinn
- Published
- 2024
- Full Text
- View/download PDF
14. Improving big data governance in healthcare institutions: user experience research for honest broker based application to access healthcare big data.
- Author
-
Singh, Kanupriya, Li, Shangman, Jahnke, Isa, Alarcon, Mauro Lemus, Mosa, Abu, and Calyam, Prasad
- Subjects
- *
DATA security , *DATABASE management , *MEDICAL quality control , *GRAPHIC arts , *TASK performance , *QUALITATIVE research , *RESEARCH funding , *PRIVACY , *CLINICAL governance , *HEALTH , *INTERVIEWING , *DATA analytics , *INFORMATION resources , *QUANTITATIVE research , *SURVEYS , *RESEARCH methodology , *HEALTH care industry , *USER-centered system design , *ACCESS to information , *MEDICAL ethics , *TIME - Abstract
Data users (researchers, scientists) in healthcare institutions need access to integrated healthcare data to conduct timely analysis of diseases to serve the right population at the right time. However, preserving patient privacy and timely access to quality healthcare data is a critical challenge. Current healthcare data governance systems are largely manual. Besides, processing process data requests is extremely slow, often taking months. To address this gap, we designed an honest-broker-based healthcare application to support data users in accessing healthcare data securely and to design a comprehendible process of data governance for data users. This study applied two iterations of a user experience (UX) evaluation of an honest broker prototype. Results show that participants found the new system promising for their research prospects. Implications suggest that technological knowledge should not be a requirement for using healthcare applications to promote broader adoption in the community. This study highlights the necessity of a process to balance the control of access to sensitive data between data providers and users as well as to educate data users on data privacy. Iterative UX studies can be a fruitful approach in gradually uncovering problems and improving the design of complex systems. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. SPASTC: a Spatial Partitioning Algorithm for Scalable Travel-time Computation.
- Author
-
Michels, A. C., Park, J., Kang, J.-Y., and Wang, S.
- Subjects
- *
PARALLEL algorithms , *COMPUTER storage devices , *HOSPITAL beds , *CYBERINFRASTRUCTURE , *PARALLEL programming - Abstract
Travel-time computation with large transportation networks is often computationally intensive for two main reasons: 1) large computer memory is required to handle large networks; and 2) calculating shortest-distance paths over large networks is computing intensive. Therefore, previous research tends to limit their spatial extent to reduce computational intensity or resolve computational intensity with advanced cyberinfrastructure. In this context, this article describes a new Spatial Partitioning Algorithm for Scalable Travel-time Computation (SPASTC) that is designed based on spatial domain decomposition with computer memory limit explicitly considered. SPASTC preserves spatial relationships required for travel-time computation and respects a user-specified memory limit, which allows efficient and large-scale travel-time computation within the given memory limit. We demonstrate SPASTC by computing spatial accessibility to hospital beds across the conterminous United States. Our case study shows that SPASTC achieves significant efficiency and scalability making the travel-time computation tens of times faster. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. Towards a National Data Architecture for Cultural Collections: Designing the Australian Cultural Data Engine.
- Author
-
Fensham, Rachel, Sumner, Tyne Daile, Cutter, Nat, Buchanan, George, Liu, Rui, Munoz, Justin, Smithies, James, Zheng, Ivy, Carlin, David, Champion, Erik, Craig, Hugh, East, Scott, Hay, Chris, Given, Lisa M., Macarthur, John, McMeekin, David, Mendelssohn, Joanna, and van der Plaat, Deborah
- Subjects
INFORMATION architecture ,ARCHITECTURAL design ,DATA management ,RESEARCH questions ,CYBERINFRASTRUCTURE ,VIRTUAL work teams ,INTERNETWORKING - Abstract
This article summarises the aims, methods, information architecture, outputs, and innovations of the Australian Cultural Data Engine (ACD-Engine), a project that harnesses leading cultural databases to build bridges to research, industry, and government. The project investigates digital heritage collections, data ontologies, and interoperability, building an information architecture to enhance the open sharing of Australian cultural data. Working with a cross-disciplinary team, the ACD-Engine establishes conceptual and technical frameworks for better understanding the platforms and uses of cultural data across a range of national and international contexts. This new cyber-infrastructure advances cultural data aggregation and interoperability whilst prioritising data quality and domain distinctiveness to answer new research questions across disciplines. As such, the ACD-Engine provides a novel approach to data management and data modelling in the arts and humanities that has significant implications for digital collections, digital humanities, and data analytics. Cultural databases are complicated beasts: rich in their contents and yet often idiosyncratic, siloed, and precarious. This article outlines an innovative workflow and information architecture designed to harness the interoperability of digital resources/records for cultural analytics research without obliterating distinctive domain knowledges. [ABSTRACT FROM AUTHOR]
- Published
- 2024
17. Creating intelligent cyberinfrastructure for democratizing AI.
- Author
-
Panda, Dhabaleswar K., Chaudhary, Vipin, Fosler‐Lussier, Eric, Machiraju, Raghu, Majumdar, Amit, Plale, Beth, Ramnath, Rajiv, Sadayappan, Ponnuswamy, Savardekar, Neelima, and Tomko, Karen
- Subjects
CYBERINFRASTRUCTURE ,ARTIFICIAL intelligence ,ANIMAL ecology ,CLASSROOM environment - Abstract
Artificial intelligence (AI) has the potential for vast societal and economic gain; yet applications are developed in a largely ad hoc manner, lacking coherent, standardized, modular, and reusable infrastructures. The NSF‐funded Intelligent CyberInfrastructure with Computational Learning in the Environment AI Institute ("ICICLE") aims to fundamentally advance edge‐to‐center, AI‐as‐a‐Service, achieved through intelligent cyberinfrastructure (CI) that spans the edge‐cloud‐HPC computing continuum, plug‐and‐play next‐generation AI and intelligent CI services, and a commitment to design for broad accessibility and widespread benefit. This design is foundational to the institute's commitment to democratizing AI. The institute's CI activities are informed by three high‐impact domains: animal ecology, digital agriculture, and smart foodsheds. The institute's workforce development and broadening participation in computing efforts reinforce the institute's commitment to democratizing AI. ICICLE seeks to serve as the national nexus for AI and intelligent CI, and welcomes engagement across its wide set of programs. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. OpenAltimetry - rapid analysis and visualization of Spaceborne altimeter data
- Author
-
Khalsa, Siri Jodha S, Borsa, Adrian, Nandigam, Viswanath, Phan, Minh, Lin, Kai, Crosby, Christopher, Fricker, Helen, Baru, Chaitan, and Lopez, Luis
- Subjects
Earth Sciences ,Cyberinfrastructure ,Data visualization ,Data discovery ,Altimetry ,ICESat-2 ,Earth sciences - Abstract
NASA's Ice, Cloud, and land Elevation Satellite-2 (ICESat-2) carries a laser altimeter that fires 10,000 pulses per second towards Earth and records the travel time of individual photons to measure the elevation of the surface below. The volume of data produced by ICESat-2, nearly a TB per day, presents significant challenges for users wishing to efficiently explore the dataset. NASA's National Snow and Ice Data Center (NSIDC) Distributed Active Archive Center (DAAC), which is responsible for archiving and distributing ICESat-2 data, provides search and subsetting services on mission data products, but providing interactive data discovery and visualization tools needed to assess data coverage and quality in a given area of interest is outside of NSIDC's mandate. The OpenAltimetry project, a NASA-funded collaboration between NSIDC, UNAVCO and the University of California San Diego, has developed a web-based cyberinfrastructure platform that allows users to locate, visualize, and download ICESat-2 surface elevation data and photon clouds for any location on Earth, on demand. OpenAltimetry also provides access to elevations and waveforms for ICESat (the predecessor mission to ICESat-2). In addition, OpenAltimetry enables data access via APIs, opening opportunities for rapid access, experimentation, and computation via third party applications like Jupyter notebooks. OpenAltimetry emphasizes ease-of-use for new users and rapid access to entire altimetry datasets for experts and has been successful in meeting the needs of different user groups. In this paper we describe the principles that guided the design and development of the OpenAltimetry platform and provide a high-level overview of the cyberinfrastructure components of the system.
- Published
- 2022
19. Knowledge machines : digital transformations of the sciences and humanities.
- Author
-
Meyer, Eric T. and Schroeder, Ralph
- Subjects
Communication in learning and scholarship -- Technological innovations ,Cyberinfrastructure ,Interdisciplinary research ,Internet research ,Open access publishing ,Research -- Data processing ,Research -- Technological innovations - Published
- 2015
20. EasyScienceGateway: A new framework for providing reproducible user environments on science gateways.
- Author
-
Michels, Alexander, Padmanabhan, Anand, Li, Zhiyu, and Wang, Shaowen
- Subjects
CYBERINFRASTRUCTURE ,COMMUNITY life - Abstract
Summary: Science gateways have become a core part of the cyberinfrastructure ecosystem by increasing access to computational resources and providing community platforms for sharing and publishing education and research materials. While science gateways represent a promising solution for computational reproducibility, common methods for providing users with their user environments on gateways present challenges which are difficult to overcome. This article presents EasyScienceGateway: a new framework for providing user environments on science gateways to resolve these challenges, provides the technical details on implementing the framework on a science gateway based on Jupyter Notebook, and discusses our experience applying the framework to the CyberGIS‐Jupyter and CyberGIS‐Jupyter for Water gateways. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
21. Entangled collaborations: tensions in cross-disciplinary user experience studies in cyberinfrastructure projects.
- Author
-
Singh, Kanu Priya, Jahnke, Isa, and Calyam, Prasad
- Abstract
Cyberinfrastructure research develops and deploys solutions that benefit cyberinfrastructure and the broader scientific community. Designing novel cyberinfrastructure solutions is inherently complex and requires collaborative relations between heterogeneous scientific stakeholders, governing bodies, and organisations. This study investigated the problems and pitfalls experienced in the cross-disciplinary collaboration of three groups: computer scientists (Group 1), User Experience (UX) researchers (Group 2), and domain scientists (Group 3 – bioinformatics, and health informatics researchers) who worked together to build cyberinfrastructure applications. Using participatory action research (PAR), we studied the dynamics of conducting UX research. The main results indicate four tensions that impacted the collaborative practices of cross-disciplinary groups using UX studies in cyberinfrastructure projects: (1) contradictory views on the quality of prototype development, (2) mental models of the work processes (know how) varied among the different groups (3) clarity of feedback was lacking, and (4) the usability problem was perceived to be with the users. Our results highlight the significance of aligning UX and computer science research goals and actively engaging involved cyberinfrastructure research members to meet user expectations. Findings reveal the nuanced ways in which computer science and UX work processes become entangled during the research process and shape cyberinfrastructure development. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. CyVerse: Cyberinfrastructure for open science.
- Author
-
Swetnam, Tyson L., Antin, Parker B., Bartelme, Ryan, Bucksch, Alexander, Camhy, David, Chism, Greg, Choi, Illyoung, Cooksey, Amanda M., Cosi, Michele, Cowen, Cindy, Culshaw-Maurer, Michael, Davey, Robert, Davey, Sean, Devisetty, Upendra, Edgin, Tony, Edmonds, Andy, Fedorov, Dmitry, Frady, Jeremy, Fonner, John, and Gillan, Jeffrey K.
- Subjects
- *
OPEN scholarship , *CYBERINFRASTRUCTURE , *ARTIFICIAL intelligence , *SOFTWARE as a service , *CLOUD computing , *LIFE sciences - Abstract
CyVerse, the largest publicly-funded open-source research cyberinfrastructure for life sciences, has played a crucial role in advancing data-driven research since the 2010s. As the technology landscape evolved with the emergence of cloud computing platforms, machine learning and artificial intelligence (AI) applications, CyVerse has enabled access by providing interfaces, Software as a Service (SaaS), and cloud-native Infrastructure as Code (IaC) to leverage new technologies. CyVerse services enable researchers to integrate institutional and private computational resources, custom software, perform analyses, and publish data in accordance with open science principles. Over the past 13 years, CyVerse has registered more than 124,000 verified accounts from 160 countries and was used for over 1,600 peer-reviewed publications. Since 2011, 45,000 students and researchers have been trained to use CyVerse. The platform has been replicated and deployed in three countries outside the US, with additional private deployments on commercial clouds for US government agencies and multinational corporations. In this manuscript, we present a strategic blueprint for creating and managing SaaS cyberinfrastructure and IaC as free and open-source software. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
23. Cyber-agricultural systems for crop breeding and sustainable production.
- Author
-
Sarkar, Soumik, Ganapathysubramanian, Baskar, Singh, Arti, Fotouhi, Fateme, Kar, Soumyashree, Nagasubramanian, Koushik, Chowdhary, Girish, Das, Sajal K., Kantor, George, Krishnamurthy, Adarsh, Merchant, Nirav, and Singh, Asheesh K.
- Subjects
- *
PLANT breeding , *SUSTAINABILITY , *DIGITAL twins , *CYBER physical systems , *AGRICULTURE , *COMPUTATIONAL neuroscience , *ANIMAL breeding , *PRECISION farming - Abstract
The cyber-agricultural system (CAS) integrates cybersystems with the physical world of agriculture via sensing, modeling, and actuation, and leverages the three pillars of functional cyber-physical systems (CPSs): computation, control, and communication: Advances in computation (i.e., ubiquitous, multimodal sensing, modeling/reasoning enabled by complex computation capabilities, and off-the-shelf deep learning models) have opened up numerous opportunities in CAS. Progress in control/actuation is characterized by advanced agricultural machinery and the rise of agricultural robotics (e.g., dexterous manipulation and harvesting, interactive sensing, precision spraying, mechanical operations, and weed culling). Advanced communication is enabling sensors, actuators, and platforms to coordinate and collaborate using internet of things (IoT) principles/tools. The cyber-agricultural system (CAS) represents an overarching framework of agriculture that leverages recent advances in ubiquitous sensing, artificial intelligence, smart actuators, and scalable cyberinfrastructure (CI) in both breeding and production agriculture. We discuss the recent progress and perspective of the three fundamental components of CAS – sensing, modeling, and actuation – and the emerging concept of agricultural digital twins (DTs). We also discuss how scalable CI is becoming a key enabler of smart agriculture. In this review we shed light on the significance of CAS in revolutionizing crop breeding and production by enhancing efficiency, productivity, sustainability, and resilience to changing climate. Finally, we identify underexplored and promising future directions for CAS research and development. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
24. A Geospatial Decision Support System for Supporting the Assessment of Land Degradation in Europe.
- Author
-
Manna, Piero, Agrillo, Antonietta, Bancheri, Marialaura, Di Leginio, Marco, Ferraro, Giuliano, Langella, Giuliano, Mileti, Florindo Antonio, Riitano, Nicola, and Munafò, Michele
- Subjects
LAND degradation ,DECISION support systems ,DATA modeling ,CITIES & towns ,LAND use planning ,CROWDSOURCING ,CYBERINFRASTRUCTURE - Abstract
Nowadays, Land Degradation Neutrality (LDN) is on the political agenda as one of the main objectives in order to respond to the increasing degradation processes affecting soils and territories. Nevertheless, proper implementation of environmental policies is very difficult due to a lack of the operational, reliable and easily usable tools necessary to support political decisions when identifying problems, defining the causes of degradation and helping to find possible solutions. It is within this framework that this paper attempts to demonstrate a new valuable web-based operational LDN tool as a component of an already running Spatial Decision Support System (S-DSS) developed on a Geospatial Cyberinfrastructure (GCI). The tool could be offered to EU administrative units (e.g., municipalities) so that they may better evaluate the state and the impact of land degradation in their territories. The S-DSS supports the acquisition, management and processing of both static and dynamic data, together with data visualization and on-the-fly computing, in order to perform modelling, all of which is potentially accessible via the Web. The land degradation data utilized to develop the LDN tool refer to the SDG 15.3.1 indicator and were obtained from a platform named Trends.Earth, designed to monitor land change by using earth observations, and post-processed to correct some of the major artefacts relating to urban areas. The tool is designed to support land planning and management by producing data, statistics, reports and maps for any EU area of interest. The tool will be demonstrated through a short selection of practical case studies, where data, tables and stats are provided to challenge land degradation at different spatial extents. Currently, there are WEBGIS systems to visualize land degradation maps but—to our knowledge—this is the first S-DSS tool enabling customized LDN reporting at any NUTS (nomenclature of territorial units for statistics) level for the entire EU territory. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
25. LearnSphere: A Learning Data and Analytics Cyberinfrastructure.
- Author
-
Stamper, John, Pavlik Jr., Philip I., Moore, Steven, Koedinger, Kenneth, and Rosé, Carolyn P.
- Subjects
CYBERINFRASTRUCTURE ,GRAPHICAL user interfaces ,EDUCATIONAL technology ,RESEARCH personnel ,REPRODUCIBLE research ,DATA mining - Abstract
LearnSphere is a web-based data infrastructure designed to transform scientific discovery and innovation in education. It supports learning researchers in addressing a broad range of issues including cognitive, social, and motivational factors in learning, educational content analysis, and educational technology innovation. LearnSphere integrates previously separate educational data and analytic resources developed by participating institutions. The web-based workflow authoring tool, Tigris, allows technical users to contribute sophisticated analytic methods, and learning researchers can adapt and apply those methods using graphical user interfaces, importantly, without additional programming. As part of our use-driven design of LearnSphere, we built a community through workshops and summer schools on educational data mining. Researchers interested in particular student levels or content domains can find student data from elementary through higher-education and across a wide variety of course content such as math, science, computing, and language learning. LearnSphere has facilitated many discoveries about learning, including the importance of active over passive learning activities and the positive association of quality discussion board posts with learning outcomes. LearnSphere also supports research reproducibility, replicability, traceability, and transparency as researchers can share their data and analytic methods along with links to research papers. We demonstrate the capabilities of LearnSphere through a series of case studies that illustrate how analytic components can be combined into research workflow combinations that can be developed and shared. We also show how open web-accessible analytics drive the creation of common formats to streamline repeated analytics and facilitate wider and more flexible dissemination of analytic tool kits. [ABSTRACT FROM AUTHOR]
- Published
- 2024
26. Practical security solutions for serial data bus networks
- Author
-
Rogers, Matthew and Rasmussen, Kasper
- Subjects
Cyberinfrastructure ,Computer security - Abstract
For decades, manufacturers incorporated computers into vehicles, weapon systems, and numerous pieces of modern infrastructure, networking them together via serial data bus networks. These networks took direct control away from the operator in the name of efficiency and safety, but we now know this bargain came at the cost of security. By having no authentication in the design, anyone communicating on the network can create harmful effects. The natural solution to this problem is to redesign these networks to support authentication, but abandoning decades of vital infrastructure would be unfathomably expensive. The only option is to continue using our existing equipment, but to develop a way to add security in an efficient and cost effective manner. However, there are an innumerable number of ways to design and implement a serial data bus network. This thesis lays out a methodology for taking a serial data bus network, understanding it, and designing a security patch for it. To illustrate the utility of our methodology, we implement two proof-of-concept intrusion detection systems. The first improves upon existing CAN detection techniques by using a standardized data field to abstract the state of the system. The second takes advantage of the architecture of MIL-STD-1553 to more efficiently detect new messages on the network. Our protocol-specific approach is emphasized by our development of new attacks on the CAN protocol. Finally, this thesis produces an evaluation framework for the prevention and detection systems increasingly being deployed to these vehicle and weapon system networks. Our framework generates a score for defenders, such that they know exactly what attacks and attack vectors their system can detect and prevent. By looking at multiple protocols, from ground vehicles to fighter aircraft, we highlight how a protocol targeted security approach creates more accurate, more efficient, and more practical technology, to detect even the most advanced attackers.
- Published
- 2022
27. A review of Earth Artificial Intelligence
- Author
-
Sun, Ziheng, Sandoval, Laura, Crystal-Ornelas, Robert, Mousavi, S Mostafa, Wang, Jinbo, Lin, Cindy, Cristea, Nicoleta, Tong, Daniel, Carande, Wendy Hawley, Ma, Xiaogang, Rao, Yuhan, Bednar, James A, Tan, Amanda, Wang, Jianwu, Purushotham, Sanjay, Gill, Thomas E, Chastang, Julien, Howard, Daniel, Holt, Benjamin, Gangodagamage, Chandana, Zhao, Peisheng, Rivas, Pablo, Chester, Zachary, Orduz, Javier, and John, Aji
- Subjects
Information and Computing Sciences ,Artificial Intelligence ,Machine Learning and Artificial Intelligence ,Data Science ,Networking and Information Technology R&D (NITRD) ,Bioengineering ,Geosphere ,Hydrology ,Atmosphere ,Artificial intelligence/machine learning ,Big data ,Cyberinfrastructure ,Earth Sciences ,Engineering ,Geochemistry & Geophysics ,Earth sciences ,Information and computing sciences - Abstract
In recent years, Earth system sciences are urgently calling for innovation on improving accuracy, enhancing model intelligence level, scaling up operation, and reducing costs in many subdomains amid the exponentially accumulated datasets and the promising artificial intelligence (AI) revolution in computer science. This paper presents work led by the NASA Earth Science Data Systems Working Groups and ESIP machine learning cluster to give a comprehensive overview of AI in Earth sciences. It holistically introduces the current status, technology, use cases, challenges, and opportunities, and provides all the levels of AI practitioners in geosciences with an overall big picture and to “blow away the fog to get a clearer vision” about the future development of Earth AI. The paper covers all the majorspheres in the Earth system and investigates representative AI research in each domain. Widely used AI algorithms and computing cyberinfrastructure are briefly introduced. The mandatory steps in a typical workflow of specializing AI to solve Earth scientific problems are decomposed and analyzed. Eventually, it concludes with the grand challenges and reveals the opportunities to give some guidance and pre-warnings on allocating resources wisely to achieve the ambitious Earth AI goals in the future.
- Published
- 2022
28. How Reproducibility Will Accelerate Discovery Through Collaboration in Physio-Logging.
- Author
-
Czapanskiy, Max and Beltran, Roxanne
- Subjects
bio-logging ,cyberinfrastructure ,ecoinformatics ,ecophysiology ,technical debt - Abstract
What new questions could ecophysiologists answer if physio-logging research was fully reproducible? We argue that technical debt (computational hurdles resulting from prioritizing short-term goals over long-term sustainability) stemming from insufficient cyberinfrastructure (field-wide tools, standards, and norms for analyzing and sharing data) trapped physio-logging in a scientific silo. This debt stifles comparative biological analyses and impedes interdisciplinary research. Although physio-loggers (e.g., heart rate monitors and accelerometers) opened new avenues of research, the explosion of complex datasets exceeded ecophysiologys informatics capacity. Like many other scientific fields facing a deluge of complex data, ecophysiologists now struggle to share their data and tools. Adapting to this new era requires a change in mindset, from data as a noun (e.g., traits, counts) to data as a sentence, where measurements (nouns) are associate with transformations (verbs), parameters (adverbs), and metadata (adjectives). Computational reproducibility provides a framework for capturing the entire sentence. Though usually framed in terms of scientific integrity, reproducibility offers immediate benefits by promoting collaboration between individuals, groups, and entire fields. Rather than a tax on our productivity that benefits some nebulous greater good, reproducibility can accelerate the pace of discovery by removing obstacles and inviting a greater diversity of perspectives to advance science and society. In this article, we 1) describe the computational challenges facing physio-logging scientists and connect them to the concepts of technical debt and cyberinfrastructure, 2) demonstrate how other scientific fields overcame similar challenges by embracing computational reproducibility, and 3) present a framework to promote computational reproducibility in physio-logging, and bio-logging more generally.
- Published
- 2022
29. Smart One Water: An Integrated Approach for the Next Generation of Sustainable and Resilient Water Systems
- Author
-
Sinha, Sunil K., Babbar-Sebens, Meghna, Dzombak, David, Gardoni, Paolo, Watford, Bevlee, Scales, Glenda, Grigg, Neil, Westerhof, Edgar, Thompson, Kenneth, and Meeker, Melissa
- Published
- 2023
- Full Text
- View/download PDF
30. Using Dynamic Data-Driven Cyberinfrastructure for Next-Generation Wildland Fire Intelligence
- Author
-
Altintas, Ilkay, Block, Jessica, Crawl, Daniel L., Callafon, Raymond A. de, Darema, Frederica, editor, Blasch, Erik P., editor, Ravela, Sai, editor, and Aved, Alex J., editor
- Published
- 2023
- Full Text
- View/download PDF
31. Implementation Examples of Big Data Management Systems for Remote Sensing
- Author
-
Di, Liping, Yu, Eugene, Di, Liping, and Yu, Eugene
- Published
- 2023
- Full Text
- View/download PDF
32. Remote Sensing Big Data Collection Challenges and Cyberinfrastructure and Sensor Web Solutions
- Author
-
Di, Liping, Yu, Eugene, Di, Liping, and Yu, Eugene
- Published
- 2023
- Full Text
- View/download PDF
33. Development of Large-Scale Scientific Cyberinfrastructure and the Growing Opportunity to Democratize Access to Platforms and Data
- Author
-
Luettgau, Jakob, Scorzelli, Giorgio, Pascucci, Valerio, Taufer, Michela, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Streitz, Norbert A., editor, and Konomi, Shin'ichi, editor
- Published
- 2023
- Full Text
- View/download PDF
34. CyEd: A Cyberinfrastructure for Computer Education
- Author
-
Abdelhamid, Sherif, Mallari, Tanner, Stower, Tristen, Kacprzyk, Janusz, Series Editor, Gomide, Fernando, Advisory Editor, Kaynak, Okyay, Advisory Editor, Liu, Derong, Advisory Editor, Pedrycz, Witold, Advisory Editor, Polycarpou, Marios M., Advisory Editor, Rudas, Imre J., Advisory Editor, Wang, Jun, Advisory Editor, Guralnick, David, editor, Auer, Michael E., editor, and Poce, Antonella, editor
- Published
- 2023
- Full Text
- View/download PDF
35. A systematic literature review of the application of user experience studies in cyberinfrastructure for scientific research
- Author
-
Singh, Kanu Priya
- Published
- 2024
- Full Text
- View/download PDF
36. Collaborative Use of Sensor Networks and Cyberinfrastructure to Understand Complex Ecosystem Interactions in a Tropical Rainforest: Challenges and Lessons Learned.
- Author
-
Rundel, Philip W., Harmon, Thomas C., Fernandez-Bou, Angel S., and Allen, Michael F.
- Subjects
- *
TROPICAL ecosystems , *RAIN forests , *SENSOR networks , *LEAF-cutting ants , *CYBERINFRASTRUCTURE , *ECOSYSTEMS - Abstract
Collaborations between ecosystem ecologists and engineers have led to impressive progress in developing complex models of biogeochemical fluxes in response to global climate change. Ecology and engineering iteratively inform and transform each other in these efforts. Nested data streams from local sources, adjacent networks, and remote sensing sources together magnify the capacity of ecosystem ecologists to observe systems in near real-time and address questions at temporal and spatial scales that were previously unobtainable. We describe our research experiences working in a Costa Rican rainforest ecosystem with the challenges presented by constant high humidity, 4300 mm of annual rainfall, flooding, small invertebrates entering the tiniest openings, stinging insects, and venomous snakes. Over the past two decades, we faced multiple challenges and learned from our mistakes to develop a broad program of ecosystem research at multiple levels of integration. This program involved integrated networks of diverse sensors on a series of canopy towers linked to multiple belowground soil sensor arrays that could transport sensor data streams from the forest directly to an off-site location via a fiber optic cable. In our commentary, we highlight three components of our work: (1) the eddy flux measurements using canopy towers; (2) the soil sensor arrays for measuring the spatial and temporal patterns of CO2 and O2 fluxes at the soil–atmosphere interface; and (3) focused investigations of the ecosystem impact of leaf-cutter ants as "ecosystem engineers" on carbon fluxes. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
37. Overcoming barriers to enable convergence research by integrating ecological and climate sciences: the NCAR–NEON system Version 1.
- Author
-
Lombardozzi, Danica L., Wieder, William R., Sobhani, Negin, Bonan, Gordon B., Durden, David, Lenz, Dawn, SanClements, Michael, Weintraub-Leff, Samantha, Ayres, Edward, Florian, Christopher R., Dahlin, Kyla, Kumar, Sanjiv, Swann, Abigail L. S., Zarakas, Claire M., Vardeman, Charles, and Pascucci, Valerio
- Subjects
- *
CLIMATOLOGY , *SYSTEMS theory , *USER interfaces , *CYBERINFRASTRUCTURE , *ECOSYSTEMS , *SCIENTIFIC community , *BIOSPHERE - Abstract
Global change research demands a convergence among academic disciplines to understand complex changes in Earth system function. Limitations related to data usability and computing infrastructure, however, present barriers to effective use of the research tools needed for this cross-disciplinary collaboration. To address these barriers, we created a computational platform that pairs meteorological data and site-level ecosystem characterizations from the National Ecological Observatory Network (NEON) with the Community Terrestrial System Model (CTSM) that is developed with university partners at the National Center for Atmospheric Research (NCAR). This NCAR–NEON system features a simplified user interface that facilitates access to and use of NEON observations and NCAR models. We present preliminary results that compare observed NEON fluxes with CTSM simulations and describe how the collaboration between NCAR and NEON that can be used by the global change research community improves both the data and model. Beyond datasets and computing, the NCAR–NEON system includes tutorials and visualization tools that facilitate interaction with observational and model datasets and further enable opportunities for teaching and research. By expanding access to data, models, and computing, cyberinfrastructure tools like the NCAR–NEON system will accelerate integration across ecology and climate science disciplines to advance understanding in Earth system science and global change. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
38. A Novel False Measurement Data Detection Mechanism for Smart Grids.
- Author
-
Shahid, Muhammad Awais, Ahmad, Fiaz, Nawaz, Rehan, Khan, Saad Ullah, Wadood, Abdul, and Albalawi, Hani
- Subjects
- *
ELECTRICAL load , *DUMMY variables , *CYBERINFRASTRUCTURE , *TEST systems , *RADIAL distribution function , *RISK communication , *NONLINEAR oscillators , *DUMBBELLS - Abstract
With the growing cyber-infrastructure of smart grids, the threat of cyber-attacks has intensified, posing an increased risk of compromised communication links. Of particular concern is the false data injection (FDI) attack, which has emerged as a highly dangerous cyber-attack targeting smart grids. This paper addresses the limitations of the variable dummy value model proposed in the authors previous work and presents a novel defense methodology called the nonlinear function-based variable dummy value model for the AC power flow network. The proposed model is evaluated using the IEEE 14-bus test system, demonstrating its effectiveness in detecting FDI attacks. It has been shown that previous detection techniques are unable to detect FDI attacks, whereas the proposed method is shown to be successful in the detection of such attacks, guaranteeing the security of the smart grid's measurement infrastructure. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
39. GIS-Based Scientific Workflows for Automated Spatially Driven Sea Level Rise Modeling.
- Author
-
Tang, Wenwu, Hearne, Heidi S., Slocum, Zachery, and Chen, Tianyang
- Abstract
Sea level rise (SLR) poses a significant threat to shorelines and the environment in terms of flooding densely populated areas and associated coastal ecosystems. Scenario analysis is often used to investigate potential SLR consequences, which can help stakeholders make informed decisions on climate change mitigation policies or guidelines. However, SLR scenario analysis requires considerable geospatial data analytics and repetitive execution of SLR models for alternative scenarios. Having to run SLR models many times for scenario analysis studies leads to heavy computational needs as well as a large investment of time and effort. This study explores the benefits of incorporating cyberinfrastructure technologies, represented by scientific workflows and high-performance computing, into spatially explicit SLR modeling. We propose a scientific workflow-driven approach to modeling the potential loss of marshland in response to different SLR scenarios. Our study area is the central South Carolina coastal region, USA. The scientific workflow approach allows for automating the geospatial data processing for SLR modeling, while repetitive modeling and data analytics are accelerated by leveraging high-performance and parallel computing. With support from automation and acceleration, this scientific workflow-driven approach allows us to conduct computationally intensive scenario analysis experiments to evaluate the impact of SLR on alternative land cover types including marshes and tidal flats as well as their spatial characteristics. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
40. An Examination and Survey of Data Confidentiality Issues and Solutions in Academic Research Computing
- Author
-
Peisert, Sean
- Subjects
cyberinfrastructure ,confidentiality ,research computing ,sensitive data ,HIPAA ,security ,CUI ,FISMA ,privacy - Abstract
In research academic computing, may be natural to emphasize data integrity over confidentiality. However, there are numerous categories of academic research that do have data confidentiality requirements, from research that is simply embargoed until a future publication date to research that contains industry-owned proprietary information or is subject to government regulation. The contents of this report are based on numerous community conversations with leaders in academic research IT, data librarians, computer science researchers, computer security professionals, and others with roles involving using or enabling the use of sensitive data in academic research. The report discusses challenges to conducting research on data that is in some way sensitive, and solutions that are being used or could be used to address those challenges and enable the research to take place. Those solutions include technical solutions as well as administrative and procedural issues. The report concludes with recommendations to campuses on issues to consider in order to enable research on sensitive data while ensuring security and privacy as well as usability and usefulness of the environment hosting that data.
- Published
- 2021
41. Towards Developing an Open-Infrastructure for Assessing the Progress, Success, and Impacts of CyberInfrastructure Projects
- Author
-
Arora, Ritu and Sondhi, Sukrit
- Published
- 2024
- Full Text
- View/download PDF
42. Evaluating Return on Investment for Cyberinfrastructure Using the International Integrated Reporting <IR> Framework
- Author
-
Snapp-Childs, Winona G., Hart, David L., Costa, Claudia M., Wernert, Julie A., Jankowski, Harmony E., Towns, John W., and Stewart, Craig A.
- Published
- 2024
- Full Text
- View/download PDF
43. From Vision to Evaluation: A Metrics Framework for the ACCESS Allocations Service
- Author
-
Hart, David L., Deems, Stephen L., and Herriott, Laura T.
- Published
- 2024
- Full Text
- View/download PDF
44. The Data Analytics Framework for XDMoD
- Author
-
Weeden, Aaron, White, Joseph P., DeLeon, Robert L., Rathsam, Ryan, Simakov, Nikolay A., Saeli, Conner, and Furlani, Thomas R.
- Published
- 2024
- Full Text
- View/download PDF
45. The EarthLife Consortium API: an extensible, open-source service for accessing fossil data and taxonomies from multiple community paleodata resources
- Author
-
Uhen, Mark D., Buckland, Philip I., Goring, Simon J., Jenkins, Julian P., and Williams, John W.
- Subjects
community-curated data resource (CCDR) ,cyberinfrastructure ,database ,informatics ,paleobiodiversity ,paleobiogeography ,paleoecology ,taxonomy - Abstract
Paleobiologists and paleoecologists interested in studying biodiversity dynamics over broad spatial and temporal scales have built multiple community-curated data resources, each emphasizing a particular spatial domain, timescale, or taxonomic group(s). This multiplicity of data resources is understandable, given the enormous diversity of life across Earth's history, but creates a barrier to achieving a truly global understanding of the diversity and distribution of life across time. Here we present the Earth Life Consortium Application Programming Interface (ELC API), a lightweight data service designed to search and retrieve fossil occurrence and taxonomic information from across multiple paleobiological resources. Key endpoints include Occurrences (returns spatiotemporal locations of fossils for selected taxa), Locales (returns information about sites with fossil data), References (returns bibliographic information), and Taxonomy (returns names of subtaxa associated with selected taxa). Data objects are returned as JSON or CSV format. The ELC API supports tectonic-driven shifts in geographic position back to 580 Ma using services from Macrostrat and GPlates. The ELC API has been implemented first for the Paleobiology Database and Neotoma Paleoecology Database, with a test extension to the Strategic Environmental Archaeology Database. The ELC API is designed to be readily extensible to other paleobiological data resources, with all endpoints fully documented and following open-source standards (e.g., Swagger, OGC). The broader goal is to help build an interlinked and federated ecosystem of paleobiological and paleoenvironmental data resources, which together provide paleobiologists, macroecologists, biogeographers, and other interested scientists with full coverage of the diversity and distribution of life across time.
- Published
- 2021
46. Insights on Sustainability of Earth Science Data Infrastructure Projects
- Author
-
Arika Virapongse, James Gallagher, and Basil Tikoff
- Subjects
earth science data ,sustainability ,cyberinfrastructure ,databases ,funding ,community ,Science (General) ,Q1-390 - Abstract
We studied 11 long-term data infrastructure projects, most of which focused on the Earth Sciences, to understand characteristics that contributed to their project sustainability. Among our sample group, we noted the existence of three different types of project groupings: Database, Framework, and Middleware. Most efforts started as federally funded research projects, and our results show that nearly all became organizations in order to become sustainable. Projects were often funded for short time scales but had the long-term burden of sustaining and supporting open science, interoperability, and community building–activities that are difficult to fund directly. This transition from ‘project’ to ‘organization’ was challenging for most efforts, especially in regard to leadership change and funding issues. Some common approaches to sustainability were identified within each project grouping. Framework and Database projects both relied heavily on the commitment to, and contribution from, a disciplinary community. Framework projects often used bottom-up governance approaches to maintain the active participation and interest of their community. Database projects succeeded when they were able to position themselves as part of the core workflow for disciplinary-specific scientific research. Middleware projects borrowed heavily from sustainability models used by software companies, while maintaining strong scientific partnerships. Cyberinfrastructure for science requires considerable resources to develop and sustain itself, and much of these resources are provided through in-kind support from academics, researchers, and their institutes. It is imperative that more work is done to find appropriate models that help sustain key data infrastructure for Earth Science over the long-term.
- Published
- 2024
- Full Text
- View/download PDF
47. Transforming Science through Cyberinfrastructure: NSF's vision for the U.S. cyberinfrastructure ecosystem for science and engineering in the 21st century.
- Author
-
Parashar, Manish, Friedlander, Amy, Gianchandani, Erwin, and Martonosi, Margaret
- Subjects
- *
CYBERINFRASTRUCTURE , *HIGH performance computing , *ENGINEERING , *PROGRESS , *SCIENTIFIC computing - Abstract
The article explores how science can be transformed through the use of advanced cyberinfrastructure (CI), describing the National Science Foundation's (NSF) vision for science and engineering in the 21st century. The authors describe how CI enabled research that dramatically accelerated efforts to respond to the novel coronavirus disease in 2019. Various disruptive applications for the technology are also described including high-fidelity modeling, online data processing and novel data-centric applications.
- Published
- 2022
- Full Text
- View/download PDF
48. The situational window for boundary-spanning infrastructure professions: Making sense of cyberinfrastructure emergence.
- Author
-
Hayes, Cassandra, Kulkarni, Chaitra, and Kee, Kerk F
- Subjects
CYBERINFRASTRUCTURE ,PROFESSIONS ,PATTERNMAKING ,TWENTY-first century ,PROFESSIONALIZATION - Abstract
In the twenty-first century, professions are complex and difficult to define due to their fluid and interdisciplinary natures. In this study, we examined the personal career stories of professionals in the field of cyberinfrastructure (CI) to identify the narrative patterns used to make sense of CI as a boundary-spanning profession. Overall, we found that professionalization of CI is a sensemaking process of communal, retrospective storytelling. The meaning-making of CI as a profession occurred through three levels of narrative patterns: individual traits of CI professionals, situational introductions to CI, and inspirational convictions about CI. The situational level, which connected innate qualities and internal motivations with external forces to join CI as a career, was especially important to the professionalization of CI. Our findings have implications for re-examining professionalization as an ongoing sensemaking process, as well as offering guidance for recruitment and retention in critical boundary-spanning professions. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
49. Use of accounting concepts to study research: return on investment in XSEDE, a US cyberinfrastructure service.
- Author
-
Stewart, Craig A., Costa, Claudia M., Wernert, Julie A., Snapp-Childs, Winona, Bland, Marques, Blood, Philip, Campbell, Terry, Couvares, Peter, Fischer, Jeremy, Hancock, David Y., Hart, David L., Jankowski, Harmony, Knepper, Richard, McMullen, Donald F., Mehringer, Susan, Pierce, Marlon, Rogers, Gary, Sinkovits, Robert S., and Towns, John
- Abstract
This paper uses accounting concepts—particularly the concept of Return on Investment (ROI)—to reveal the quantitative value of scientific research pertaining to a major US cyberinfrastructure project (XSEDE—the eXtreme Science and Engineering Discovery Environment). XSEDE provides operational and support services for advanced information technology systems, cloud systems, and supercomputers supporting non-classified US research, with an average budget for XSEDE of US$20M+ per year over the period studied (2014–2021). To assess the financial effectiveness of these services, we calculated a proxy for ROI, and converted quantitative measures of XSEDE service delivery into financial values using costs for service from the US marketplace. We calculated two estimates of ROI: a Conservative Estimate, functioning as a lower bound and using publicly available data for a lower valuation of XSEDE services; and a Best Available Estimate, functioning as a more accurate estimate, but using some unpublished valuation data. Using the largest dataset assembled for analysis of ROI for a cyberinfrastructure project, we found a Conservative Estimate of ROI of 1.87, and a Best Available Estimate of ROI of 3.24. Through accounting methods, we show that XSEDE services offer excellent value to the US government, that the services offered uniquely by XSEDE (that is, not otherwise available for purchase) were the most valuable to the facilitation of US research activities, and that accounting-based concepts hold great value for understanding the mechanisms of scientific research generally. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
50. OpenAltimetry-rapid analysis and visualization of Spaceborne altimeter data
- Author
-
Khalsa, Siri Jodha S, Borsa, Adrian, Nandigam, Viswanath, Phan, Minh, Lin, Kai, Crosby, Christopher, Fricker, Helen, Baru, Chaitan, and Lopez, Luis
- Subjects
Cyberinfrastructure ,Data visualization ,Data discovery ,Altimetry ,ICESat-2 - Abstract
AbstractNASA’s Ice, Cloud, and land Elevation Satellite-2 (ICESat-2) carries a laser altimeter that fires 10,000 pulses per second towards Earth and records the travel time of individual photons to measure the elevation of the surface below. The volume of data produced by ICESat-2, nearly a TB per day, presents significant challenges for users wishing to efficiently explore the dataset. NASA’s National Snow and Ice Data Center (NSIDC) Distributed Active Archive Center (DAAC), which is responsible for archiving and distributing ICESat-2 data, provides search and subsetting services on mission data products, but providing interactive data discovery and visualization tools needed to assess data coverage and quality in a given area of interest is outside of NSIDC’s mandate. The OpenAltimetry project, a NASA-funded collaboration between NSIDC, UNAVCO and the University of California San Diego, has developed a web-based cyberinfrastructure platform that allows users to locate, visualize, and download ICESat-2 surface elevation data and photon clouds for any location on Earth, on demand. OpenAltimetry also provides access to elevations and waveforms for ICESat (the predecessor mission to ICESat-2). In addition, OpenAltimetry enables data access via APIs, opening opportunities for rapid access, experimentation, and computation via third party applications like Jupyter notebooks. OpenAltimetry emphasizes ease-of-use for new users and rapid access to entire altimetry datasets for experts and has been successful in meeting the needs of different user groups. In this paper we describe the principles that guided the design and development of the OpenAltimetry platform and provide a high-level overview of the cyberinfrastructure components of the system.
- Published
- 2020
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.