1. The perfect neuroimaging-genetics-computation storm: collision of petabytes of data, millions of hardware devices and thousands of software tools
- Author
-
Dinov, Ivo D, Petrosyan, Petros, Liu, Zhizhong, Eggert, Paul, Zamanyan, Alen, Torri, Federica, Macciardi, Fabio, Hobel, Sam, Moon, Seok Woo, Sung, Young Hee, Jiang, Zhiguo, Labus, Jennifer, Kurth, Florian, Ashe-McNalley, Cody, Mayer, Emeran, Vespa, Paul M, Van Horn, John D, Toga, Arthur W, and for the Alzheimer’s Disease Neuroimaging Initiative
- Subjects
Biomedical and Clinical Sciences ,Health Sciences ,Psychology ,Networking and Information Technology R&D (NITRD) ,Bioengineering ,Data Science ,Biomedical Imaging ,Genetics ,2.6 Resources and infrastructure (aetiology) ,Generic health relevance ,Algorithms ,Alzheimer Disease ,Brain ,Brain Injuries ,Chronic Pain ,Computational Biology ,Female ,Genome-Wide Association Study ,Genomics ,Humans ,Information Dissemination ,Internet ,Irritable Bowel Syndrome ,Male ,Middle Aged ,Neuroimaging ,Software ,Workflow ,Aging ,Pipeline ,Computation solutions ,Workflows ,IBS ,Pain ,Parkinson's disease ,Alzheimer's disease ,Shape ,Volume ,Analysis ,Big data ,Visualization ,Alzheimer’s Disease Neuroimaging Initiative ,Medical and Health Sciences ,Psychology and Cognitive Sciences ,Experimental Psychology ,Biomedical and clinical sciences ,Health sciences - Abstract
The volume, diversity and velocity of biomedical data are exponentially increasing providing petabytes of new neuroimaging and genetics data every year. At the same time, tens-of-thousands of computational algorithms are developed and reported in the literature along with thousands of software tools and services. Users demand intuitive, quick and platform-agnostic access to data, software tools, and infrastructure from millions of hardware devices. This explosion of information, scientific techniques, computational models, and technological advances leads to enormous challenges in data analysis, evidence-based biomedical inference and reproducibility of findings. The Pipeline workflow environment provides a crowd-based distributed solution for consistent management of these heterogeneous resources. The Pipeline allows multiple (local) clients and (remote) servers to connect, exchange protocols, control the execution, monitor the states of different tools or hardware, and share complete protocols as portable XML workflows. In this paper, we demonstrate several advanced computational neuroimaging and genetics case-studies, and end-to-end pipeline solutions. These are implemented as graphical workflow protocols in the context of analyzing imaging (sMRI, fMRI, DTI), phenotypic (demographic, clinical), and genetic (SNP) data.
- Published
- 2014