30 results on '"Taylor, Ian"'
Search Results
2. User-Oriented Partial Result Evaluation in Workflow-Based Science Gateways
- Author
-
Jaghoori, Mohammad Mahdi, Ramezani, Sara, Olabarriaga, Silvia D., Montagnat, Johan, Taylor, Ian, Other Research, Epidemiology and Data Science, and Amsterdam Public Health
- Subjects
Workflow ,Windows Workflow Foundation ,Computer science ,Distributed computing ,Default gateway ,Workflow engine ,Workflow Management Coalition ,Workflow management system ,XPDL ,Workflow technology - Abstract
Scientific workflow management systems provide a useful layer for defining and executing applications supported by science gateways. In various optimization or simulation applications that need to run for a long time, the users may be satisfied with an incomplete execution. The system should, therefore, allow users to evaluate partial results of the workflow execution. This entails performing a consolidation step, that would normally run only at the end of the workflow. In this paper, we present two new workflow patterns that formally define how the consolidation step should be executed (on partial inputs) whenever the user pro-actively requests evaluation of the partial results. This changes the traditional workflow behavior, in which every step runs once, when all its data dependencies are satisfied. We evaluate implementing these patterns in various workflow management systems and finally present a DIRAC-based implementation of this feature for the use case of a molecular docking gateway.
- Published
- 2014
3. Exploring workflow interoperability tools for neuroimaging data analysis
- Author
-
Korkhov, Vladimir, Krefting, Dagmar, Kukla, Tamas, Terstyanszky, Gabor Z., Caan, Matthan, Olabarriaga, Silvia D., Taylor, Ian, Montagnat, Johan, Amsterdam Neuroscience, Radiology and Nuclear Medicine, Amsterdam Public Health, and Epidemiology and Data Science
- Subjects
World Wide Web ,Workflow ,Windows Workflow Foundation ,Computer science ,Interoperability ,Grid ,Data science ,Workflow engine ,Workflow management system ,Field (computer science) ,Workflow technology - Abstract
Neuroimaging is a field that benefits from distributed computing infrastructures (DCIs) to perform data processing and analysis, which is often achieved using grid workflow systems. Collaborative research in neuroimaging requires ways to facilitate exchange between different groups, in particular to enable sharing, re-use and interoperability of applications implemented as workflows. The SHIWA project provides solutions to facilitate sharing and exchange of workflows between workflow systems and DCI resources. In this paper we present and analyse how the SHIWA platform was used to implement various usage scenarios in which workflow exchange supports collaboration in neuroscience. The SHIWA platform and the implemented solutions are described and analysed from the "user" perspective, in this case the workflow developers and the neuroscientists. We conclude that the platform in its current form is valuable for the foreseen usage scenarios, and we identify remaining challenges concerning management of multiple credentials and data transfers across DCIs.
- Published
- 2011
4. The Triana Workflow Environment: Architecture and Applications.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Taylor, Ian, Shields, Matthew, Wang, Ian, and Harrison, Andrew
- Abstract
In this chapter, the Triana workflow environment is described. Triana focuses on supporting services within multiple environments, such as peer-to-peer (P2P) and the Grid, by integrating with various types of middleware toolkits. This approach differs from that of the last chapter, which gave an overview of Taverna, a system designed to support scientists using Grid technology to conduct in silico experiments in biology. Taverna focuses workflow at the Web services level and addresses concerns of how such services should be presented to its users. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
5. Introduction.
- Author
-
Taylor, Ian J., Gannon, Dennis B., Gannon, Dennis, Deelman, Ewa, Shields, Matthew, and Taylor, Ian
- Abstract
Workflows for e-Science is divided into four parts, which represent four broad but distinct areas of scientific workflows. In the first part, Background, we introduce the concept of scientific workflows and set the scene by describing how they differ from their business workflow counterpart. In Part II, Application and User Perspective, we provide a number of scientific examples that currently use workflows for their e-Science experiments. In Workflow Representation and Common Structure (Part III), we describe core workflow themes, such as control flow or dataflow and the use of components or services. In this part, we also provide overviews for a number of common workflow languages, such as Petri Nets, the Business Process Execution Language (BPEL), and the Virtual Data Language (VDL), along with service interfaces. In Part IV, Frameworks and Tools, we take a look at many of the popular environments that are currently being used for e-Science applications by paying particular attention to their workflow capabilities. The following four sections describe the chapters in each part and therefore provide a comprehensive summary of the book as a whole. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
6. ASKALON: A Development and Grid Computing Environment for Scientific Workflows.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Fahringer, Thomas, Prodan, Radu, Duan, Rubing, Hofer, Jüurgen, Nadeem, Farrukh, Nerieri, Francesco, Podlipnig, Stefan, Jun Qin, Siddiqui, Mumtaz, Truong, Hong-Linh, Villazon, Alex, and Wieczorek, Marek
- Abstract
Most existing Grid application development environments provide the application developer with a nontransparent Grid. Commonly, application developers are explicitly involved in tedious tasks such as selecting software components deployed on specific sites, mapping applications onto the Grid, or selecting appropriate computers for their applications. Moreover, many programming interfaces are either implementation-technology-specific (e.g., based on Web services [24]) or force the application developer to program at a low-level middleware abstraction (e.g., start task, transfer data [22, 153]). While a variety of graphical workflow composition tools are currently being proposed, none of them is based on standard modeling techniques such as Unified Modeling Language (UML). [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
7. Looking into the Future of Workflows: The Challenges Ahead.
- Author
-
Taylor, Ian J., Gannon, Dennis B., Shields, Matthew, and Deelman, Ewa
- Abstract
In this chapter, we take a step back from the individual applications and software systems and attempt to categorize the types of issues that we are facing today and the challenges we see ahead. This is by no means a complete picture of the challenges but rather a set of observations about the various aspects of workflow management. In a broad sense, we are organizing our thoughts in terms of the different workflow systems discussed in this book, from the user interface down to the execution environment. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
8. Sedna: A BPEL-Based Environment for Visual Scientific Workflow Modeling.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Wassermann, Bruno, Emmerich, Wolfgang, Butchart, Ben, Cameron, Nick, Liang Chen, and Patel, Jignesh
- Abstract
Scientific Grid computing environments are increasingly adopting the Open Grid Services Architecture (OGSA), which is a service-oriented architecture for Grids. With the proliferation of OGSA, Grids effectively consist of a collection of Grid services, Web services with certain extensions providing additional support for state and life cycle management. Hence, the need arises for some means of composing these basic services into larger workflows in order to, for example, express a scientific experiment. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
9. Expressing Workflow in the Cactus Framework.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, and Goodale, Tom
- Abstract
The Cactus Framework [15, 73, 167] is an open-source, modular, portable, programming environment for collaborative HPC computing. It was designed and written specifically to enable scientists and engineers to perform the large-scale simulations needed for their science. From the outset, Cactus has followed two fundamental tenets: respecting user needs and embracing new technologies. The framework and its associated components must be driven from the beginning by user requirements. This has been achieved by developing, supporting, and listening to a large user base. Among these needs are ease of use, portability, the ability to support large and geographically diverse collaborations and to handle enormous computing resources, visualization, file IO, and data management. It must also support the inclusion of legacy code, as well as a range of programming languages. It is essential that any living framework be able to incorporate new and developing cutting edge computation technologies and infrastructure, with minimal or no disruption to its user base. Cactus is now associated with many computational science research projects, particularly in visualization, data management, and Grid computing [14]. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
10. Pegasus: Mapping Large-Scale Workflows to Distributed Resources.
- Author
-
Taylor, Ian J., Gannon, Dennis B., Shields, Matthew, Deelman, Ewa, Mehta, Gaurang, Singh, Gurmeet, Mei-Hui Su, and Vahi, Karan
- Abstract
Many scientific advances today are derived from analyzing large amounts of data. The computations themselves can be very complex and consume significant resources. Scientific efforts are also not conducted by individual scientists; rather, they rely on collaborations that encompass many researchers from various organizations. The analysis is often composed of several individual application components designed by different scientists. To describe the desired analysis, the components are assembled in a workflow where the dependencies between them are defined and the data needed for the analysis are identified. To support the scale of the applications, many resources are needed in order to provide adequate performance. These resources are often drawn from a heterogeneous pool of geographically distributed compute and data resources. Running large-scale, collaborative applications in such environments has many challenges. Among them are systematic management of the applications, their components, and the data, as well as successful and efficient execution on the distributed resources. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
11. Workflow Management in Condor.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Couvares, Peter, Kosar, Tevfik, Roy, Alain, Weber, Jeff, and Wenger, Kent
- Abstract
The Condor project began in 1988 and has evolved into a feature-rich batch system that targets high-throughput computing; that is, Condor ([262], [414]) focuses on providing reliable access to computing over long periods of time instead of highly tuned, high-performance computing for short periods of time or a small number of applications. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
12. Java CoG Kit Workflow.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, von Laszewski, Gregor, Hategan, Mihael, and Kodeboyina, Deepti
- Abstract
In order to satisfy the need for sophisticated experiment and simulation management solutions for the scientific user community, various frameworks must be provided. Such frameworks include APIs, services, templates, patterns, GUIs, command-line tools, and workflow systems that are specifically addressed towards the goal of assisting in the complex process of experiment and simulation management. Workflow by itself is just one of the ingredients for a successful experiment and simulation management tool. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
13. Taverna/myGrid: Aligning a Workflow System with the Life Sciences Community.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Oinn, Tom, Li, Peter, Kell, Douglas B., Goble, Carole, Goderis, Antoon, Greenwood, Mark, Hull, Duncan, Stevens, Robert, Turi, Daniele, and Jun Zhao
- Abstract
Bioinformatics is a discipline that uses computational and mathematical techniques to store, manage, and analyze biological data in order to answer biological questions. Bioinformatics has over 850 databases [154] and numerous tools that work over those databases and local data to produce even more data themselves. In order to perform an analysis, a bioinformatician uses one or more of these resources to gather, filter, and transform data to answer a question. Thus, bioinformatics is an in silico science. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
14. Workflow-Level Parametric Study Support by MOTEUR and the P-GRADE Portal.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Glatard, Tristan, Sipos, Gergely, Montagnat, Johan, Farkas, Zoltan, and Kacsuk, Peter
- Abstract
Many large-scale scientific applications require the processing of complete data sets made of individual data segments that can be manipulated independently following a single analysis procedure. Workflow managers have been designed for describing and controlling such complex application control flows. However, when considering very data-intensive applications, there is a large potential parallelism that should be properly exploited to ensure efficient processing. Distributed systems such as Grid infrastructures are promising for handling the load resulting from parallel data analysis and manipulation. Workflow managers can help in exploiting the infrastructure parallelism, given that they are able to handle the data flow resulting from the application's execution. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
15. Virtual Data Language: A Typed Workflow Notation for Diversely Structured Scientific Data.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Yong Zhao, Wilde, Michael, and Foster, Ian
- Abstract
When constructing workflows that operate on large and complex data sets, the ability to describe the types of both data sets and workflow procedures can be invaluable, enabling discovery of data sets and procedures, type checking and composition of procedure calls, and iteration over composite data sets. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
16. Workflow Composition: Semantic Representations for Flexible Automation.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, and Gil, Yolanda
- Abstract
Many different kinds of users may need to compose scientific workflows for different purposes. This chapter focuses on the requirements and challenges of scientific workflow composition. They are motivated by our work with two particular application domains: physics-based seismic hazard analysis (Chapter 10) and data-intensive natural language processing [238]. Our research on workflow creation spans fully automated workflow generation (Chapter 23) using artificial intelligence planning techniques for assisted workflow composition [237,276] by combining semantic representations of workflow components with formal properties of correct workflows. Other projects have used similar techniques in different domains to support workflow composition through planning and automated reasoning [286,289,415] and semantic representations (Chapter 19). As workflow representations become more declarative and expressive, they enable significant improvements in automation and assistance for workflow composition and in general for managing and automating complex scientific processes. The chapter starts off motivating and describing important requirements to support the creation of workflows. Based on these requirements, we outline the approaches that we have found effective, including separating levels of abstraction in workflow descriptions, using semantic representations of workflows and their components, and supporting flexible automation through reuse and automatic completion of user specifications of partial workflows. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
17. Protocol-Based Integration Using SSDL and π-Calculus.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Woodman, Simon, Parastatidis, Savas, and Webber, Jim
- Abstract
A "service" has become the contemporary abstraction around which modern distributed applications are designed and built. A service represents a piece of functionality that is exposed on the network. The "message" abstraction is used to create interaction patterns or protocols to represent the messaging behavior of a service. In the Web services domain, SOAP is the preferred model for encoding, transferring, and processing such messages. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
18. Adapting BPEL to Scientific Workflows.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, and Slominski, Aleksander
- Abstract
In this chapter, we examine the degree to which a de facto standard business Web services workflow language, Business Process Execution Language for Web Services (BPEL4WS), can be used to compose Grid and scientific workflows. As the Grid application models, such as Open Grid Services Architecture (OGSA) [146], move toward Web services and service-oriented architecture (SOA) [135], supporting Web services is becoming a requirement for a Grid workflow language. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
19. Petri Nets.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Hoheisel, Andreas, and Alt, Martin
- Abstract
In 1962, C.A. Petri introduced in his Ph.D. thesis [351] a formalism for describing distributed processes by extending state machines with a notion of concurrency. Due to the simple and intuitive, but at the same time formal and expressive, nature of his formalism, Petri Nets became an established tool for modelling and analyzing distributed processes in business as well as the IT sector. This chapter gives a brief introduction to the theory of Petri Nets and shows how Petri Nets can be applied for effective workflow management with regard to the choreography, orchestration, and enactment of e-Science applications. While choreography deals with the abstract modelling of applications, orchestration deals with the mapping onto concrete software components and the infrastructure. During the enactment of e-Science applications, runtime issues, such as synchronization, persistence, transaction safety, and fault management, are examined within the workflow formalism. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
20. Component Architectures and Services: From Application Construction to Scientific Workflows.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, and Gannon, Dennis
- Abstract
The idea of building computer applications by composing them out of reusable software components is a concept that emerged in the 1970s and 1980s as developers began to realize that the complexity of software was evolving so rapidly that a different approach was needed if actual software development was going to keep pace with the demands placed upon it.1 This fact had already been realized by hardware designers. By the mid 1970s, it was standard practice to build digital systems by composing them from standard, well-tested integrated circuits that encapsulated sophisticated, powerful subsystems that we easily reused in thousands of applications. By the 1990s, even the designers of integrated circuits such as microprocessors were building them by composing them from standard cell libraries that provided components such as registers and floating-point units that could be arranged on the chip and easily integrated to form a full processor. Now, multiple processor cores can be assembled on a single chip as components of larger systems. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
21. Control- Versus Data-Driven Workflows.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., and Shields, Matthew
- Abstract
Workflow is typically defined as a sequence of operations or tasks needed to manage a business process or computational activity (Chapter 1). The representation of the sequence of operations or tasks is handled in many different ways by different people and varies from simple scripting languages, through graphs represented in textual or graphical form, to mathematical representations such as Petri Nets (Chapter 13) or π-calculus (Chapter 15). [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
22. SCEC CyberShake Workflows—Automating Probabilistic Seismic Hazard Analysis Calculations.
- Author
-
Taylor, Ian J., Gannon, Dennis B., Shields, Matthew, Maechling, Philip, Deelman, Ewa, Li Zhao, Graves, Robert, Mehta, Gaurang, Gupta, Nitin, Mehringer, John, Kesselman, Carl, Callaghan, Scott, Okaya, David, Francoeur, Hunter, Gupta, Vipin, Yifeng Cui, Vahi, Karan, Jordan, Thomas, and Field, Edward
- Abstract
The Southern California Earthquake Center (SCEC) is a community of more than 400 scientists from over 54 research organizations that conducts geophysical research in order to develop a physics-based understanding of earthquake processes and to reduce the hazard from earthquakes in the Southern California region [377]. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
23. Dynamic, Adaptive Workflows for Mesoscale Meteorology.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Gannon, Dennis, Plale, Beth, Marru, Suresh, Kandaswamy, Gopi, Simmhan, Yogesh, and Shirasuna, Satoshi
- Abstract
The Linked Environments for Atmospheric Discovery (LEAD) [122] is a National Science Foundation funded1 project to change the paradigm for mesoscale weather prediction from one of static, fixed-schedule computational forecasts to one that is adaptive and driven by weather events. It is a collaboration of eight institutions,2 led by Kelvin Droegemeier of the University of Oklahoma, with the goal of enabling far more accurate and timely predictions of tornadoes and hurricanes than previously considered possible. The traditional approach to weather prediction is a four-phase activity. In the first phase, data from sensors are collected. The sensors include ground instruments such as humidity and temperature detectors, and lightning strike detectors and atmospheric measurements taken from balloons, commercial aircraft, radars, and satellites. The second phase is data assimilation, in which the gathered data are merged together into a set of consistent initial and boundary conditions for a large simulation. The third phase is the weather prediction, which applies numerical equations to measured conditions in order to project future weather conditions. The final phase is the generation of visual images of the processed data products that are analyzed to make predictions. Each phase of activity is performed by one or more application components. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
24. Case Studies on the Use of Workflow Technologies for Scientific Analysis: The Biomedical Informatics Research Network and the Telescience Project.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Lin, Abel W., Peltier, Steven T., Grethe, Jeffrey S., and Ellisman, Mark H.
- Abstract
The advent of "Grids," or Grid computing, has led to a fundamental shift in the development of applications for managing and performing computational or data-intensive analyses. A current challenge faced by the Grid community entails modeling the work patterns of domain or bench scientists and providing robust solutions utilizing distributed infrastructures. These challenges spawned efforts to develop "workflows" to manage programs and data on behalf of the end user. The technologies come from multiple scientific fields, often with disparate definitions, and have unique advantages and disadvantages, depending on the nature of the scientific process in which they are used. In this chapter, we argue that to maximize the impact of these efforts, there is value in promoting the use of workflows within a tiered, hierarchical structure where each of these emerging workflow pieces are interoperable. We present workflow models of the Telescience™ Project1 and BIRN2 architectures as frameworks that manage multiple tiers of workflows to provide tailored solutions for end-to-end scientific processes. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
25. Ecological Niche Modeling Using the Kepler Workflow System.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Pennington, Deana D., Higgins, Dan, Peterson, A. Townsend, Jones, Matthew B., Ludäscher, Bertram, and Bowers, Shawn
- Abstract
Changes in biodiversity have been linked to variations in climate and human activities [295]. These changes have implications for a wide range of socially relevant processes, including the spread of infectious disease, invasive species dynamics, and vegetation productivity [27, 70, 203, 291, 294, 376, 426]. Our understanding of biodiversity patterns and processes through space and time, scaling from genes to continents, is limited by our ability to analyze and synthesize multidimensional data effectively from sources as wide-ranging as field and laboratory experiments, satellite imagery, and simulation models. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
26. Workflows in Pulsar Astronomy.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Brooke, John, Pickles, Stephen, Carr, Paul, and Michael, Kramer
- Abstract
In this chapter, we describe the development of methods that operate on the output of the signal of a radio telescope to detect the characteristic signals of pulsars. These signals are much weaker than the noise in the signal at any given wavelength, and therefore algorithms for combining the signals in different wavelength bands must be applied. This is heavily expensive in terms of CPU power. Early versions of distributed algorithms ran on a distributed network of supercomputers connected by Internet-aware Message Passing Interface (MPI) during the period 1999-2001. Today such techniques are being integrated into workflows that automate the search process and enable sophisticated astronomical knowledge to be captured via the construction of the workflow. In particular, we address issues of parallelism within components of the workflow. Parallelism is necessary due to two constraints on workflow performance. One is the application of the workflow in real time as the signal is being processed to enable very precise measurements to be carried out on known pulsars. The other is the use of the workflow to explore large regions of parameter space in search of previously undetected pulsars. There are very severe restraints on the degree of abstraction that can currently be applied in this work since details of the architecture of the computing resource (parallel cluster or computational Grid) on which the workflows are to be run cannot be ignored in the construction of the workflow. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
27. Workflow and Biodiversity e-Science.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, and Jones, Andrew C.
- Abstract
Biodiversity e-Science is characterized by the use of a wide range of different kinds of data and by performing complex analyses on these data. In this chapter, we discuss the use of workflow systems to assist biodiversity researchers and consider how such systems can provide repeatability of experiments and other benefits.We argue that nevertheless there are also limitations to this kind of approach, and we discuss how more flexibility in a more exploratory environment could be achieved. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
28. A Case Study on the Use of Workflow Technologies for Scientific Analysis: Gravitational Wave Data Analysis.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Brown, Duncan A., Brady, Patrick R., Dietz, Alexander, Cao, Junwei, Johnson, Ben, and McNabb, John
- Abstract
Modern scientific experiments acquire large amounts of data that must be analyzed in subtle and complicated ways to extract the best results. The Laser Interferometer Gravitational Wave Observatory (LIGO) is an ambitious effort to detect gravitational waves produced by violent events in the universe, such as the collision of two black holes or the explosion of supernovae [37,258]. The experiment records approximately 1 TB of data per day, which is analyzed by scientists in a collaboration that spans four continents. LIGO and distributed computing have grown up side by side over the past decade, and the analysis strategies adopted by LIGO scientists have been strongly influenced by the increasing power of tools to manage distributed computing resources and the workflows to run on them. In this chapter, we use LIGO as an application case study in workflow design and implementation. The software architecture outlined here has been used with great efficacy to analyze LIGO data [2-5] using dedicated computing facilities operated by the LIGO Scientific Collaboration, the LIGO Data Grid. It is just the first step, however. Workflow design and implementation lies at the interface between computing and traditional scientific activities. In the conclusion, we outline a few directions for future development and provide some long-term vision for applications related to gravitational wave data analysis. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
29. Generating Complex Astronomy Workflows.
- Author
-
Taylor, Ian J., Gannon, Dennis B., Shields, Matthew, Berriman, G. Bruce, Deelman, Ewa, Good, John, Jacob, Joseph C., Katz, Daniel S., Laity, Anastasia C., Prince, Thomas A., Singh, Gurmeet, and Mei-Hui Su
- Abstract
Astronomy has a rich heritage of discovery using image data sets that cover the full range of the electromagnetic spectrum. Image data sets in one frequency range have often been studied in isolation from those in other frequency ranges. This is mostly a consequence of the diverse properties of the data collections themselves. Images are delivered in different coordinate systems, map projections, spatial samplings, and image sizes, and the pixels themselves are rarely co-registered on the sky. Moreover, the spatial extent of many astronomically important structures, such as clusters of galaxies and star formation regions, is often substantially greater than that of individual images. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
30. Scientific versus Business Workflows.
- Author
-
Taylor, Ian J., Deelman, Ewa, Gannon, Dennis B., Shields, Matthew, Barga, Roger, and Gannon, Dennis
- Abstract
The formal concept of a workflow has existed in the business world for a long time. An entire industry of tools and technology devoted to workflow management has been developed and marketed to meet the needs of commercial enterprises. The Workflow Management Coalition (WfMC) has existed for over ten years and has developed a large set of reference models, documents, and standards. Why has the scientific community not adopted these existing standards? While it is not uncommon for the scientific community to reinvent technology rather than purchase existing solutions, there are issues involved in the technical applications that are unique to science, and we will attempt to characterize some of these here. There are, however, many core concepts that have been developed in the business workflow community that directly relate to science, and we will outline them below. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.