33 results on '"Raj Kettimuthu"'
Search Results
2. Globus service enhancements for exascale applications and facilities.
- Author
-
Weijian Zheng, Jack Kordas, Tyler J. Skluzacek, Raj Kettimuthu, and Ian T. Foster
- Published
- 2024
- Full Text
- View/download PDF
3. Toward a Quantum Information System Cybersecurity Taxonomy and Testbed: Exploiting a Unique Opportunity for Early Impact.
- Author
-
Benjamin Blakely, Joaquin Chung 0001, Alec Poczatek, Ryan Syed, and Raj Kettimuthu
- Published
- 2024
- Full Text
- View/download PDF
4. What does Inter-Cluster Job Submission and Execution Behavior Reveal to Us?
- Author
-
Tirthak Patel, Devesh Tiwari, Raj Kettimuthu, William E. Allcock, Paul Rich, and Zhengchun Liu
- Published
- 2022
- Full Text
- View/download PDF
5. Leveraging In-Network Computing and Programmable Switches for Streaming Analysis of Scientific Data.
- Author
-
Ganesh C. Sankaran, Joaquin Chung 0001, and Raj Kettimuthu
- Published
- 2021
- Full Text
- View/download PDF
6. Operating Liquid-Cooled Large-Scale Systems: Long-Term Monitoring, Reliability Analysis, and Efficiency Measures.
- Author
-
Rohan Basu Roy, Tirthak Patel, Raj Kettimuthu, William E. Allcock, Paul Rich, Adam Scovel, and Devesh Tiwari
- Published
- 2021
- Full Text
- View/download PDF
7. La Résistance: Harnessing Heterogeneous Resources for Adaptive Resiliency in 6G Networks.
- Author
-
Ganesh C. Sankaran, Joaquin Chung 0001, and Raj Kettimuthu
- Published
- 2022
- Full Text
- View/download PDF
8. Petascale XCT: 3D Image Reconstruction with Hierarchical Communications on Multi-GPU Nodes.
- Author
-
Mert Hidayetoglu, Tekin Bicer, Simon Garcia De Gonzalo, Bin Ren, Vincent De Andrade, Doga Gürsoy, Raj Kettimuthu, Ian T. Foster, and Wen-mei W. Hwu
- Published
- 2020
9. Job characteristics on large-scale systems: long-term analysis, quantification, and implications.
- Author
-
Tirthak Patel, Zhengchun Liu, Raj Kettimuthu, Paul Rich, William E. Allcock, and Devesh Tiwari
- Published
- 2020
- Full Text
- View/download PDF
10. Towards 6G Networks for Rural Environments: Vision for Improving Digital Inclusion through Open Source Hardware and Software
- Author
-
Ivan Armuelles Voinov, Joaquin Chung, Raj Kettimuthu, Borja Bordel, Ramon Alcarria, and Tomas Robles
- Published
- 2022
- Full Text
- View/download PDF
11. A Review of the Solutions Ecosystem for 5G Systems on Rural and Remote Environments
- Author
-
Ivan Armuelles Voinov, Joaquin Chung, Raj Kettimuthu, Borja Bordel, Ramon Alcarria, and Tomas Robles
- Published
- 2022
- Full Text
- View/download PDF
12. Leveraging In-Network Computing and Programmable Switches for Streaming Analysis of Scientific Data
- Author
-
Raj Kettimuthu, Joaquin Chung, and Ganesh C. Sankaran
- Subjects
Stream processing ,Focus (computing) ,Load management ,SIMPLE (military communications protocol) ,Computer science ,Distributed computing ,Normalization (image processing) ,Approximation algorithm ,Load balancing (computing) ,Networking hardware - Abstract
With the emergence of programmable network devices that match the performance of fixed function devices, several recent projects have explored in-network computing, where the processing that is traditionally done outside the network is offloaded to the network devices. In-network computing has typically been applied to network functions (e.g., load balancing, NAT, and DNS), caching, data reduction/aggregation, and coordination/consensus functions. In some cases it has been used to accelerate stream-processing tasks that involve small payloads and simple operations. In this work we focus on leveraging in-network computing for stream processing of scientific datasets with large payloads that require complex operations such as floating-point computations and logarithmic functions. We demonstrate in-network computing for a real-world scientific application performing streaming normalization of a 2-D image from a light source experiment. We discuss the challenges we encountered and potential approaches to address them.
- Published
- 2021
- Full Text
- View/download PDF
13. Job Characteristics on Large-Scale Systems: Long-Term Analysis, Quantification, and Implications
- Author
-
Devesh Tiwari, Raj Kettimuthu, Paul Rich, Tirthak Patel, Zhengchun Liu, and William Allcock
- Subjects
020203 distributed computing ,Resource (project management) ,Procurement ,Computer science ,0202 electrical engineering, electronic engineering, information engineering ,Resource allocation ,020201 artificial intelligence & image processing ,Resource management ,Workload ,02 engineering and technology ,Instrumentation (computer programming) ,Data science - Abstract
HPC workload analysis and resource consumption characteristics are the key to driving better operation practices, system procurement decisions, and designing effective resource management techniques. Unfortunately, the HPC community does not have easy accessibility to long-term introspective work-load analysis and characterization for production-scale HPC systems. This study bridges this gap by providing detailed long-term quantification, characterization, and analysis of job characteristics on two supercomputers: Intrepid and Mira. This study is one of the largest of its kind – covering trends and characteristics for over three billion compute hours, 750 thousand jobs, and spanning a decade. We confirm several long-held conventional wisdom, and identify many previously undiscovered trends and its implications. We also introduce a learning based technique to predict the resource requirement of future jobs with high accuracy, using features available prior to the job submission and without requiring any application-specific tracing or application-intrusive instrumentation.
- Published
- 2020
- Full Text
- View/download PDF
14. Predicting Execution Times for Disk-based and In-Situ Parallel Data Analytics (Final Technical Report)
- Author
-
Gagan Agrawal, Gangyi Zhu, Raj Kettimuthu, Prasanna Balaprakash, Vitali Morozov, P. Sadayappan, and Tekin Bicer
- Subjects
Database ,Computer science ,Technical report ,Data analysis ,computer.software_genre ,computer - Published
- 2020
- Full Text
- View/download PDF
15. Measurements and analytics of wide-area file transfers over dedicated connections
- Author
-
Raj Kettimuthu, Qiang Liu, Satyabrata Sen, Ian Foster, Zhengchun Liu, and Nageswara S. V. Rao
- Subjects
Ethernet ,Wide area ,Computer science ,Analytics ,business.industry ,Synchronous optical networking ,Computation ,File transfer ,Lustre (file system) ,GridFTP ,business ,Computer network - Abstract
Distributed scientific and big-data computations are becoming increasingly dependent on access to remote files. Wide-area file transfers are supported by two basic schemes: (i) application-level tools, such as GridFTP, that provide transport services between file systems housed at geographically separated sites, and (ii) file systems mounted over wide-area networks, using mechanisms such as LNet routers that make them transparently available. In both cases, the file transfer performance critically depends on the configuration consisting of host, file, IO, and disk subsystems, which are complex by themselves, as well as on their complex compositions implemented using buffers and IO-network data transitions. We present extensive file transfer rate measurements collected over dedicated 10 Gbps connections with 0-366 ms round-trip times, using GridFTP and XDD file transfer tools, and Lustre file system extended over wide-area networks using LNet routers. Our test configurations are composed of: three types of host systems; XFS, Lustre, and ext3 file systems; and Ethernet and SONET wide-area connections. We present analytics based on the convexity-concavity of throughput profiles which provide insights into throughput and its superior or inferior trend compared to linear interpolations. We propose the utilization-concavity coefficient, a scalar metric that characterizes the overall performance of any file transfer method consisting of specific configuration and scheme. Our results enable performance optimizations by highlighting the significant roles of (i) buffer sizes and parallelism in GridFTP and XDD, and (ii) buffer utilization and credit mechanism in LNet routers.
- Published
- 2019
- Full Text
- View/download PDF
16. A Model Driven Intelligent Orchestration Approach to Service Automation in Large Distributed Infrastructures
- Author
-
Linda Winkler, Eun-Sung Jung, Xi Yang, Raj Kettimuthu, and Tom Lehman
- Subjects
Service (systems architecture) ,Workflow ,Software ,business.industry ,Computer science ,Distributed computing ,Orchestration ,Cloud computing ,business ,Software-defined networking ,Automation ,Abstraction (linguistics) - Abstract
Today's scientific computing applications and workflows operate on heterogeneous and vastly distributed infrastructures. Traditional human-in-the-loop service engineering approach met its insurmountable challenge in dealing with these very complex and diverse networked systems, including conventional and software defined networks, compute, storage, clouds and instruments. Orchestration is the key to integrate and coordinate the networked multi-services and automate end-to-end workflows. In this work, we present a model driven intelligent orchestration approach to this end-to-end automation, which is built upon a semantic modeling solution that supports the full stack of service integration, orchestration, abstraction, and intent and policy representation. We also present the design of a real-world orchestrator called StackV that is able to accommodate highly complex application scenarios such as Software Defined ScienceDMZ (SD-SDMZ) and Hybrid Cloud Inter-Networking (HCIN) by implementing this approach.
- Published
- 2018
- Full Text
- View/download PDF
17. Quantum Sensing for High Energy Physics
- Author
-
Cynthia Jenks, Vesna Mitrovic, Matthew Dietrich, Paul D. Lett, Petra Merkel, Kevin J. O'Brien, Volodymyr Yefremenko, Roger Rusack, Mattia Checchin, Ian Shipsey, Sae Woo Nam, Nicholas A. Peters, Alexander Romanenko, Malcolm Boshier, Michael V. Fazio, Tenzin Rabga, David Underwood, Larry Lurio, Karen Byrum, John Zasadzinski, Gensheng Wang, Benjamin J. Lawrie, Jonathan King, Hogan Nguyen, Eve Kovacs, Howard Nicholson, Jeffrey R. Guest, Robert Wagner, Xuedan Ma, Amr S. Helmy, Andrew Sonnenschein, U. Patel, Jason W. Henning, Xufeng Zhang, Valerie Taylor, Yuekun Heng, Geoffrey T. Bodwin, C. M. Posada, Andrei Nomerotski, Jessica Metcalfe, Hal Finkel, Patrick J. Fox, Yuri Alexeev, Keith Schwab, Derek F. Jackson Kimball, Nathan Woollett, Karl van Bibber, Joseph Heremans, Akito Kusaka, Harry Weerts, David Hume, Zeeshan Ahmed, Jonathan Lewis, Pavel Lougovski, Marcel Demarteau, Roger O'Brient, John F. Mitchell, Ranjan Dharmapalan, Vishnu Zutshi, Gustavo Cancelo, Przemyslaw Bienias, D. Braga, Richard Kriske, Junqi Xie, Ron Harnik, Giorgio Apollinari, Kent D. Irwin, Vladan Vuletic, Gianpaolo Carosi, R. Tschirhart, Erik Shirokoff, Zelimir Djurcic, James E. Fast, M. Crisler, Sergei Chekanov, Junjia Ding, Karl K. Berggren, Jason M. Hogan, Asimina Arvanitaki, Aaron S. Chou, Donna Kubik, Holger Mueller, Johannes Hubmayr, Andrei Gaponenko, Michael L. Norman, Raphael C. Pooser, Salman Habib, Konrad Lehnert, Nick Karonis, Aashish A. Clerk, Peter Fierlinger, Raj Kettimuthu, Monika Schleier-Smith, J. Segal, David D. Awschalom, D. Bowring, Ian C. Cloët, S. Rescia, Edward May, Misun Min, Tijana Rajh, Sandeep Miryala, Bjoern Penning, Phay J. Ho, Andrew Geraci, Gerald Gabrielse, Christopher George Tully, Supratik Guha, Maurice Garcia-Sciveres, Jie Zhang, Thomas Cecil, John M. Doyle, Sergey Perverzev, C. L. Chang, Jimmy Proudfoot, and Antonino Miceli
- Subjects
Physics ,Particle physics ,Quantum sensor ,Experimental methods ,Domain (software engineering) - Abstract
Report of the first workshop to identify approaches and techniques in the domain of quantum sensing that can be utilized by future High Energy Physics applications to further the scientific goals of High Energy Physics.
- Published
- 2018
- Full Text
- View/download PDF
18. Software-Defined Network Solutions for Science Scenarios
- Author
-
Dantong Yu, Raj Kettimuthu, Satyabrata Sen, Josh Boley, Dimitrios Katramatos, Nageswara S. V. Rao, Bradley W. Settlemyer, Hsing-bung Chen, and Qiang Liu
- Subjects
Scientific instrument ,Computer science ,Distributed computing ,Controller (computing) ,Testbed ,Floodlight ,020206 networking & telecommunications ,020207 software engineering ,02 engineering and technology ,computer.software_genre ,Workflow ,Virtual machine ,0202 electrical engineering, electronic engineering, information engineering ,Software-defined networking ,computer ,Throughput (business) - Abstract
High-performance scientific workflows utilize supercomputers, scientific instruments, and large storage systems. Their executions require fast setup of a small number of dedicated network connections across the geographically distributed facility sites. We present Software-Defined Network (SDN) solutions consisting of site daemons that use dpctl, Floodlight, ONOS, or OpenDaylight controllers to set up these connections. The development of these SDN solutions could be quite disruptive to the infrastructure, while requiring a close coordination among multiple sites; in addition, the large number of possible controller and device combinations to investigate could make the infrastructure unavailable to regular users for extended periods of time. In response, we develop a Virtual Science Network Environment (VSNE) using virtual machines, Mininet, and custom scripts that support the development, testing, and evaluation of SDN solutions, without the constraints and expenses of multi-site physical infrastructures; furthermore, the chosen solutions can be directly transferred to production deployments. By complementing VSNE with a physical testbed, we conduct targeted performance tests of various SDN solutions to help choose the best candidates. In addition, we propose a switching response method to assess the setup times and throughput performances of different SDN solutions, and present experimental results that show their advantages and limitations.
- Published
- 2018
- Full Text
- View/download PDF
19. TCP Throughput Profiles Using Measurements over Dedicated Connections
- Author
-
Nageswara S. V. Rao, Don Towlsey, Qiang Liu, Raj Kettimuthu, Ian Foster, Gayane Vardoyan, and Satyabrata Sen
- Subjects
CUBIC TCP ,Computer science ,Distributed computing ,TCP tuning ,020206 networking & telecommunications ,Provisioning ,Monotonic function ,02 engineering and technology ,Lyapunov exponent ,Computer Science::Performance ,Network congestion ,symbols.namesake ,Computer Science::Networking and Internet Architecture ,0202 electrical engineering, electronic engineering, information engineering ,symbols ,020201 artificial intelligence & image processing ,Throughput (business) ,Poincaré map - Abstract
ide-area data transfers in high-performance computing infrastructures are increasingly being carried over dynamically provisioned dedicated network connections that provide high capacities with no competing traffic. We present extensive TCP throughput measurements and time traces over a suite of physical and emulated 10 Gbps connections with 0-366 ms round-trip times (RTTs). Contrary to the general expectation, they show significant statistical and temporal variations, in addition to the overall dependencies on the congestion control mechanism, buffer size, and the number of parallel streams. We analyze several throughput profiles that have highly desirable concave regions wherein the throughput decreases slowly with RTTs, in stark contrast to the convex profiles predicted by various TCP analytical models. We present a generic throughput model that abstracts the ramp-up and sustainment phases of TCP flows, which provides insights into qualitative trends observed in measurements across TCP variants: (i) slow-start followed by well-sustained throughput leads to concave regions; (ii) large buffers and multiple parallel streams expand the concave regions in addition to improving the throughput; and (iii) stable throughput dynamics, indicated by a smoother Poincare map and smaller Lyapunov exponents, lead to wider concave regions. These measurements and analytical results together enable us to select a TCP variant and its parameters for a given connection to achieve high throughput with statistical guarantees.
- Published
- 2017
- Full Text
- View/download PDF
20. Sustained Wide-Area TCP Memory Transfers over Dedicated Connections
- Author
-
Nageswara S. V. Rao, Gayane Vardoyan, Bradley W. Settlemyer, Raj Kettimuthu, Ian Foster, and Don Towsley
- Subjects
TCP acceleration ,business.industry ,Computer science ,ComputerSystemsOrganization_COMPUTER-COMMUNICATIONNETWORKS ,InfiniBand ,TCP tuning ,Throughput ,H-TCP ,TCP congestion-avoidance algorithm ,TCP global synchronization ,TCP Friendly Rate Control ,Zeta-TCP ,business ,Throughput (business) ,Computer network - Abstract
Wide-area memory transfers between on-going computations and remote steering, analysis and visualization sites can be utilized in several High-Performance Computing (HPC) scenarios. Dedicated network connections with high capacity, low loss rates and low competing traffic, are typically provisioned over current HPC infrastructures to support such transfers. To gain insights into such transfers, we collected throughput measurements for different versions of TCP between dedicated multi-core servers over emulated 10 Gbps connections with round trip times (rtt) in the range 0-366 ms. Existing TCP models and measurements over shared links are well-known to exhibit monotonically decreasing, convex throughput profiles as rtt is increased. In sharp contrast, our these measurements show two distinct regimes: a concave profile at lower rtts and a convex profile at higher rtts. We present analytical results that explain these regimes: (a) at lower rtt, rapid throughput increase due to slow-start leads to the concave profile, and (b) at higher rtt, TCP congestion avoidance phase with slower dynamics dominates. In both cases, however, we analytically show that throughput decreases with rtt, albeit at different rates, as confirmed by the measurements. These results provide practical TCP solutions to these transfers without requiring additional hardware and software, unlike Infiniband and UDP solutions, respectively.
- Published
- 2015
- Full Text
- View/download PDF
21. Workflows at experimental facilities: Use cases from the Advanced Photon Source
- Author
-
Ian Foster, Tekin Bicer, Raj Kettimuthu, Mike Wilde, Justin Wozniak, Kyle Chard, Ben Blaiszik, Ray Osborn, Francesco De Carlo, Ian Foster, Tekin Bicer, Raj Kettimuthu, Mike Wilde, Justin Wozniak, Kyle Chard, Ben Blaiszik, Ray Osborn, and Francesco De Carlo
- Published
- 2015
- Full Text
- View/download PDF
22. Exploiting Network Parallelism for Improving Data Transfer Performance
- Author
-
Jason Zurawski, Dan Gunter, Martin Swany, Jun Yi, Ezra Kissel, and Raj Kettimuthu
- Subjects
OpenFlow ,business.industry ,Computer science ,Distributed computing ,Bandwidth (computing) ,Provisioning ,Session (computer science) ,business ,Software-defined networking ,Computer network ,Network simulation ,Data transmission - Abstract
Many scientific applications, including bulk data transfer, can achieve significantly higher performance from vir- tually loss-free dedicated resources provisioned on shared links, than from opportunistic network use. Research and Education (RaE) backbones, including the Energy Sciences Network and Internet2, provide general-purpose services to allocate dedi- cated bandwidth. However, in order to fully take advantage of this technology, applications need to move from coarse-grained "reservation" strategies, to more sophisticated control based on software defined networking (SDN) with technologies such as OpenFlow. We propose here, as one practical step in this direction, using multiple paths for the same application transfer session. This can add bandwidth from "best effort" and dedicated networks, and can also facilitate performance with applications using multiple 10G NICs over 100G capable paths.
- Published
- 2012
- Full Text
- View/download PDF
23. Accelerating Data Movement Leveraging End-System and Network Parallelism
- Author
-
Raj Kettimuthu, Jun Yi, and Venkatram Vishwanath
- Subjects
File system ,business.industry ,Computer science ,Data parallelism ,Distributed computing ,Throughput ,GridFTP ,computer.software_genre ,Shared resource ,Remote computing ,Storage area network ,Grid computing ,Wide area network ,End system ,The Internet ,business ,computer ,Computer network - Abstract
Data volumes produced by simulation, experimental and observational science is rapidly increasing. This data needs to be moved from its source to another resource for analysis, visualization and archival purposes. The destination resource could be either local or remote. The data intensive science is critically dependent upon the high-performance parallel file and storage end systems to read/write and high-speed networks to move their enormous data between local and remote computing and storage facilities. 100 Gigabit per second networks such as DOE's Advanced Network Initiative (ANI), Internet2's 100G network represent a major step forward in wide area network performance. Effective utilization of these networks requires substantial and pervasive parallelism, at the file system, end system, and network levels. Additional obstacles such as heterogeneity and time-varying conditions of network and end system arise that, if not adequately addressed, will render high performance storage and network systems extremely under-performed. In this paper, we propose a data movement system that dynamically and adaptively adjusts end systems and networks parallelisms in response to changing conditions of end systems and networks to sustain high-throughput for data transfers. We evaluate our system in multiple settings and show that (1) in a homogeneous configuration, the design can achieve better throughput for light and medium workload than GridFTP and achieve comparable throughput for heavy workload, (2) and in a heterogeneous configuration, the design can achieve several factors higher throughput for all workloads than GridFTP.
- Published
- 2012
- Full Text
- View/download PDF
24. GridFTP based real-time data movement architecture for x-ray photon correlation spectroscopy at the Advanced Photon Source
- Author
-
Alec Sandy, T. Madden, Raj Kettimuthu, Suresh Narayanan, and Michael Link
- Subjects
Speckle pattern ,Data acquisition ,Grid computing ,Computer science ,Detector ,Advanced Photon Source ,GridFTP ,Real-time data ,computer.software_genre ,Supercomputer ,computer ,Computational science - Abstract
X-ray photon correlation spectroscopy (XPCS) is a unique tool to study the dynamical properties in a wide range of materials over a wide spatial and temporal range. XPCS measures the correlated changes in the speckle pattern, produced when a coherent x-ray beam is scattered from a disordered sample, over a time series of area detector images. The technique rides on “Big Data” and relies heavily on high performance computing (HPC) techniques. In this paper, we propose a highspeed data movement architecture for moving data within the Advanced Photon Source (APS) as well as between APS and the users' institutions. We describe the challenges involved in the internal data movement and a GridFTP-based solution that enables more efficient usage of the APS beam time. The implementation of GridFTP plugin as part of the data acquisition system at the Advanced Photon Source for real time data transfer to the HPC system for data analysis is discussed.
- Published
- 2012
- Full Text
- View/download PDF
25. Monitoring and troubleshooting multi-domain networks using measurement federations: Part 2 [Guest Editorial]
- Author
-
Prasad Calyam, Constantine Dovrolis, Loki Jörgenson, Raj Kettimuthu, Brian Tierney, and Jason Zurawski
- Subjects
Computer Networks and Communications ,Electrical and Electronic Engineering ,Computer Science Applications - Published
- 2014
- Full Text
- View/download PDF
26. Lessons learned from moving earth system grid data sets over a 20 Gbps wide-area network
- Author
-
Lisa Childers, Valerio Pascucci, Alex Sim, Andrew Cherry, Kevin Harms, Dean N. Williams, Ken Raffenetti, Dan Gunter, Linda Winkler, Eli Dart, Michael Link, Bill Allcock, Keith Miller, Jeff Long, Peer-Timo Bremer, Loren Jan Wilson, Raj Kettimuthu, Jason Hick, Vijaya Natarajan, David Ressman, John Bresnahan, Jason Lee, Ian Foster, Hariri, Salim, and Keahey, Kate
- Subjects
Climate Action ,Petascale computing ,Coupled model intercomparison project ,Operations research ,Climateprediction.net ,Computer science ,IPCC Fourth Assessment Report ,Wide area network ,Systems engineering ,Petabyte ,Earth System Grid ,Supercomputer - Abstract
In preparation for the Intergovernmental Panel on Climate Change (IPCC) Fifth Assessment Report, the climate community will run the Coupled Model Intercomparison Project phase 5 (CMIP-5) experiments, which are designed to answer crucial questions about future regional climate change and the results of carbon feedback for different mitigation scenarios. The CMIP-5 experiments will generate petabytes of data that must be replicated seamlessly, reliably, and quickly to hundreds of research teams around the globe. As an end-to-end test of the technologies that will be used to perform this task, a multidisciplinary team of researchers moved a small portion (10 TB) of the multimodel Coupled Model Intercomparison Project, Phase 3 data set used in the IPCC Fourth Assessment Report from three sources-the Argonne Leadership Computing Facility (ALCF) Lawrence Livermore National Laboratory (LLNL) and National Energy Research Scientific Computing Center (NERSC)-to the 2009 Supercomputing conference (SC09) show floor in Portland, Oregon, over circuits provided by DOE's ESnet. The team achieved a sustained data rate of 15 Gb/s on a 20 Gb/s network. More important, this effort provided critical feedback on how to deploy, tune, and monitor the middleware that will be used to replicate the upcoming petascale climate datasets. We report on obstacles overcome and the key lessons learned from this successful bandwidth challenge effort. Copyright 2010 ACM.
- Published
- 2010
- Full Text
- View/download PDF
27. High Performance Data Transfer in Grid Environment Using GridFTP over InfiniBand
- Author
-
Hari Subramoni, Dhabaleswar K. Panda, Ping Lai, and Raj Kettimuthu
- Subjects
Remote direct memory access ,Grid computing ,Computer science ,ComputerSystemsOrganization_COMPUTER-COMMUNICATIONNETWORKS ,Operating system ,InfiniBand ,Local area network ,Zero-copy ,GridFTP ,Performance improvement ,computer.software_genre ,Communications protocol ,computer - Abstract
GridFTP, designed using the Globus XIO framework, is one of the most popular methods in use to perform data transfers in the grid environment. But the performance of GridFTP in WAN is limited by the relatively low communication bandwidth offered by the existing network protocols. On the other hand, modern interconnects such as InfiniBand, with many advanced communication features like zero-copy protocol and RDMA operations, can greatly improve communication efficiency. In this paper, we take on the challenge of combining the ease of use of the Globus XIO framework and the high performance achieved through InfiniBand communication, thereby natively sup-porting GridFTP over InfiniBand based networks. The Advanced Data Transfer Service (ADTS), designed in our previous work, provides the low level InfiniBand support to the Globus XIO layer. We introduce the concepts of I/Ostaging in the Globus XIO ADTS driver to achieve efficient disk based data transfers. We evaluate our designs in both LAN and WAN environments using micro benchmarks as well as communication traces from several real world applications. We also provide insights into the communication performance with some in-depth analysis. Our experimental evaluation shows a performance improvement of up to100% for ADTS based data transfers as opposed to TCP or UDP based ones in LAN and high delay WAN scenarios.
- Published
- 2010
- Full Text
- View/download PDF
28. Harnessing Multicore Processors for High-Speed Secure Transfer
- Author
-
Mike Link, Raj Kettimuthu, John Bresnahan, and Ian Foster
- Subjects
Multi-core processor ,business.industry ,Network packet ,Computer science ,Distributed computing ,Transfer (computing) ,Throughput ,Network interface ,GridFTP ,Encryption ,business ,Communications protocol - Abstract
A growing need for ultra-high-speed data transfers has motivated continued improvements in the transmission speeds of the physical network layer. As researchers develop protocols and software to operate over such networks, they often fail to account for security. The processing power required to encrypt or sign packets of data can significantly decrease transfer rates, and thus security is often sacrificed for throughput. Emerging multicore processors provide a higher ratio of CPUs to network interfaces and can, in principle, be used to accelerate encrypted transfers by applying multiple processing and network resources to a single transfer. We discuss the attributes that network protocols and software must have to exploit such systems. In particular, we study how these attributes may be applied in the GridFTP code distributed with the globus toolkit. GridFTP is a well-accepted and robust protocol for high-speed data transfer. It has been shown to scale to near-network speeds. While GridFTP can provide encrypted and protected data transfers, it historically suffers transfer performance penalties when these features are enabled. We present configurations to the Globus GridFTP server that can achieve fully encrypted high-speed data transfers.
- Published
- 2007
- Full Text
- View/download PDF
29. GridCopy: Moving Data Fast on the Grid
- Author
-
John-Paul Navarro, William Allcock, Raj Kettimuthu, Ian Foster, and Lee Liming
- Subjects
Data grid ,business.industry ,Computer science ,Interface (computing) ,Distributed computing ,Grid file ,GridFTP ,computer.software_genre ,Grid ,Semantic grid ,Grid computing ,business ,computer ,Computer network - Abstract
An important type of communication in grid and distributed computing environments is bulk data transfer. GridFTP has emerged as a de facto standard for secure, reliable, high-performance data transfer across resources on the grid. GridCopy provides a simple GridFTP client interface to users and extensible configuration that can be changed dynamically by administrators to make efficient data movement in the Grid easier for users.
- Published
- 2007
- Full Text
- View/download PDF
30. Maximum a posteriori estimation of crystallographic phases in X-ray diffraction tomography
- Author
-
Tekin Bicer, Doĝa Gürsoy, Francesco De Carlo, Raj Kettimuthu, Stuart R. Stock, and Jonathan Almer
- Subjects
Materials science ,General Mathematics ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,General Physics and Astronomy ,Advanced Photon Source ,Iterative reconstruction ,Sensitivity and Specificity ,Phase Transition ,Diffraction tomography ,X-Ray Diffraction ,Aliasing ,Maximum a posteriori estimation ,Computer Simulation ,Poisson Distribution ,Crystallography ,Models, Statistical ,General Engineering ,Reproducibility of Results ,Articles ,Undersampling ,Radiographic Image Interpretation, Computer-Assisted ,A priori and a posteriori ,Tomography ,Artifacts ,Algorithms - Abstract
A maximum a posteriori approach is proposed for X-ray diffraction tomography for reconstructing three-dimensional spatial distribution of crystallographic phases and orientations of polycrystalline materials. The approach maximizes the a posteriori density which includes a Poisson log-likelihood and an a priori term that reinforces expected solution properties such as smoothness or local continuity. The reconstruction method is validated with experimental data acquired from a section of the spinous process of a porcine vertebra collected at the 1-ID-C beamline of the Advanced Photon Source, at Argonne National Laboratory. The reconstruction results show significant improvement in the reduction of aliasing and streaking artefacts, and improved robustness to noise and undersampling compared to conventional analytical inversion approaches. The approach has the potential to reduce data acquisition times, and significantly improve beamtime efficiency.
- Published
- 2015
- Full Text
- View/download PDF
31. Monitoring and troubleshooting multi-domain networks using measurement federations [Guest Editorial]
- Author
-
Loki Jörgenson, Brian Tierney, Constantine Dovrolis, Prasad Calyam, Raj Kettimuthu, and Jason Zurawski
- Subjects
Network architecture ,Computer Networks and Communications ,Network security ,business.industry ,Computer science ,Cloud computing ,Troubleshooting ,Computer security ,computer.software_genre ,Network operations center ,Network traffic control ,Computer Science Applications ,Network management application ,Intelligent computer network ,Dynamic circuit network ,Packet loss ,Resource allocation ,The Internet ,Network performance ,Electrical and Electronic Engineering ,business ,computer ,Network management station ,Computer network - Abstract
In both the scientific and corporate worlds, users, resources, and data are often physically distributed, making networks increasingly important for all operations. Enormous progress has been made in increasing the capacity and accessibility of networking infrastructures, which in turn has fostered wider adoption of cloud and grid environments. Unfortunately, these advances have not directly translated into improved performance for all applications and users; instead, network performance problems become even more subtle and detrimental as the capacity of the network increases, and troubleshooting them on multidomain network paths is highly challenging. These problems may be as benign as congestion from other network users, or as serious as packet loss caused by one or more intermediate domain infrastructure(s) and architectural flaws.
- Published
- 2013
- Full Text
- View/download PDF
32. Enabling petascale science: data management, troubleshooting, and scalable science services
- Author
-
John Bresnahan, Dan Gunter, Carl Kesselman, Keith Beattie, A. Baranovski, Ann L. Chervenak, Brian Tierney, T. Freeman, Nick Leroy, Ian Foster, K. Keahey, Robert Schuler, Ravi Madduri, Laura Pearlman, G. Oleynik, Joshua Boverhof, Mike Link, Miron Livny, Raj Kettimuthu, and Shishir Bharathi
- Subjects
History ,Data processing ,Focus (computing) ,Process (engineering) ,Computer science ,business.industry ,Data management ,Troubleshooting ,Data science ,Computer Science Applications ,Education ,Petascale computing ,Scalability ,Resource management ,business - Abstract
The Center for Enabling Distributed Petascale Science is developing serviced to enable researchers to manage large, distributed datasets. The center projects focus on three areas: tools for reliable placement of data, issues involving failure detection and failure diagnosis in distributed systems, and scalable services that process requests to access data
- Published
- 2008
- Full Text
- View/download PDF
33. Petascale XCT: 3D Image Reconstruction with Hierarchical Communications on Multi-GPU Nodes
- Author
-
Vincent De Andrade, Bin Ren, Tekin Bicer, Mert Hidayetoglu, Simon Garcia de Gonzalo, Raj Kettimuthu, Ian Foster, Doga Gursoy, and Wen-mei W. Hwu
- Subjects
FOS: Computer and information sciences ,Iterative method ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Computed tomography ,02 engineering and technology ,Iterative reconstruction ,computer.software_genre ,01 natural sciences ,Computational science ,010309 optics ,Voxel ,0103 physical sciences ,medicine ,Projection (set theory) ,Scaling ,ComputingMethodologies_COMPUTERGRAPHICS ,Tomographic reconstruction ,medicine.diagnostic_test ,021001 nanoscience & nanotechnology ,Petascale computing ,Computer Science - Distributed, Parallel, and Cluster Computing ,Rate of convergence ,Distributed, Parallel, and Cluster Computing (cs.DC) ,0210 nano-technology ,computer - Abstract
X-ray computed tomography is a commonly used technique for noninvasive imaging at synchrotron facilities. Iterative tomographic reconstruction algorithms are often preferred for recovering high quality 3D volumetric images from 2D X-ray images, however, their use has been limited to small/medium datasets due to their computational requirements. In this paper, we propose a high-performance iterative reconstruction system for terabyte(s)-scale 3D volumes. Our design involves three novel optimizations: (1) optimization of (back)projection operators by extending the 2D memory-centric approach to 3D; (2) performing hierarchical communications by exploiting "fat-node" architecture with many GPUs; (3) utilization of mixed-precision types while preserving convergence rate and quality. We extensively evaluate the proposed optimizations and scaling on the Summit supercomputer. Our largest reconstruction is a mouse brain volume with 9Kx11Kx11K voxels, where the total reconstruction time is under three minutes using 24,576 GPUs, reaching 65 PFLOPS: 34% of Summit's peak performance.
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.