6 results on '"Montse Farreras"'
Search Results
2. Reducing Compiler-Inserted Instrumentation in Unified-Parallel-C Code Generation
- Author
-
Xavier Martorell, Ettore Tiotto, Montse Farreras, José Nelson Amaral, Michail Alvanosl, Universitat Politècnica de Catalunya. Departament d'Arquitectura de Computadors, and Universitat Politècnica de Catalunya. CAP - Grup de Computació d'Altes Prestacions
- Subjects
Computer science ,Data localization ,Program compilers ,Llenguatges de programació ,Programming languages (Electronic computers) ,Parallel computing ,computer.software_genre ,Runtime system ,Linear transformations ,Unified Parallel C ,Synchronization (computer science) ,Code generation ,Computer architecture ,Instrumentation (computer programming) ,Partitioned global address space ,Informàtica::Arquitectura de computadors::Arquitectures paral·leles [Àrees temàtiques de la UPC] ,Partitioned Global Address Space ,computer.programming_language ,Metadata ,Performance degradation ,Read/write operations ,Parallel processing (Electronic computers) ,Address space ,Processament en paral·lel (Ordinadors) ,Supercomputers Communication mechanisms ,Informàtica::Llenguatges de programació [Àrees temàtiques de la UPC] ,Prototype implementations ,Transformation based ,Operating system ,Mathematical transformations ,Compiler ,computer ,Synchronization primitive - Abstract
Programs written in Partitioned Global Address Space (PGAS) languages can access any location of the entire address space via standard read/write operations. However, the compiler have to create the communication mechanisms and the runtime system to use synchronization primitives to ensure the correct execution of the programs. However, PGAS programs may have fine-grained shared accesses that lead to performance degradation. One solution is to use the inspector-executor technique to determine which accesses are indeed remote and which accesses may be coalesced in larger remote access operations. A straightforward implementation of the inspector-executor in a PGAS system may result in excessive instrumentation that hinders performance. This paper introduces a shared-data localization transformation based on linear memory descriptors (LMADs) that reduces the amount of instrumentation introduced by the compiler into programs written in the UPC language and describes a prototype implementation of the proposed transformation. A performance evaluation, using up to 2048 cores of a POWER 775 supercomputer, allows for a prediction that applications with regular accesses can achieve up to 180% of the performance of handoptimized versions while applications with irregular accesses yield performance gain from 1.12X up to 6.3X speedup.
- Published
- 2014
- Full Text
- View/download PDF
3. ClusterSs
- Author
-
Enric Tejedor, Montse Farreras, Rosa M. Badia, David Grove, Gheorghe Almasi, and Jesús Labarta
- Subjects
Concurrent object-oriented programming ,Multi-core processor ,Computer architecture ,Procedural programming ,Computer science ,Concurrency ,Parallel programming model ,Programming paradigm ,Operating system ,Reactive programming ,computer.software_genre ,computer ,Inductive programming - Abstract
Programming for large-scale, multicore-based architectures requires adequate tools that offer ease of programming while not hindering application performance. StarSs is a family of parallel programming models based on automatic function level parallelism that targets productivity. StarSs deploys a data-flow model: it analyses dependencies between tasks and manages their execution, exploiting their concurrency as much as possible. We introduce Cluster Superscalar (ClusterSs), a new StarSs member designed to execute on clusters of SMPs. ClusterSs tasks are asynchronously created and assigned to the available resources with the support of the IBM APGAS runtime, which provides an efficient and portable communication layer based on one-sided communication.This short paper gives an overview of the ClusterSs design on top of APGAS, as well as the conclusions of a productivity study; in this study, ClusterSs was compared to the IBM X10 language, both in terms of programmability and performance. A technical report is available with the details.
- Published
- 2011
- Full Text
- View/download PDF
4. Scalable RDMA performance in PGAS languages
- Author
-
Toni Cortes, Calin Cascaval, George Almási, and Montse Farreras
- Subjects
Hardware_MEMORYSTRUCTURES ,Memory management ,Remote direct memory access ,Shared memory ,Computer architecture ,Computer science ,Scalability ,Programming paradigm ,Multiprocessing ,Distributed memory ,Parallel computing ,Partitioned global address space ,Software_PROGRAMMINGTECHNIQUES - Abstract
Partitioned Global Address Space (PGAS) languages provide a unique programming model that can span shared-memory multiprocessor (SMP) architectures, distributed memory machines, or cluster of SMPs. Users can program large scale machines with easy-to-use, shared memory paradigms.
- Published
- 2009
- Full Text
- View/download PDF
5. Combining Static and Dynamic Data Coalescing in Unified Parallel C.
- Author
-
Alvanos, Michail, Farreras, Montse, Tiotto, Ettore, Amaral, Jose Nelson, and Martorell, Xavier
- Subjects
DATA analysis ,COMPUTER programming ,COMPUTER architecture ,CRYSTALLIZED intelligence ,TELECOMMUNICATION systems - Abstract
Significant progress has been made in the development of programming languages and tools that are suitable for hybrid computer architectures that group several shared-memory multicores interconnected through a network. This paper addresses important limitations in the code generation for partitioned global address space (PGAS) languages. These languages allow fine-grained communication and lead to programs that perform many fine-grained accesses to data. When the data is distributed to remote computing nodes, code transformations are required to prevent performance degradation. Until now code transformations to PGAS programs have been restricted to the cases where both the physical mapping of the data or the number of processing nodes are known at compilation time. In this paper, a novel application of the inspector-executor model overcomes these limitations and allows profitable code transformations, which result in fewer and larger messages sent through the network, when neither the data mapping nor the number of processing nodes are known at compilation time. A performance evaluation reports both scaling and absolute performance numbers on up to 32,768 cores of a Power 775 supercomputer. This evaluation indicates that the compiler transformation results in speedups between 1.15$\times$
over a baseline and that these automated transformations achieve up to 63 percent the performance of the MPI versions. [ABSTRACT FROM AUTHOR]- Published
- 2016
- Full Text
- View/download PDF
6. All-optical packet/circuit switching-based data center network for enhanced scalability, latency, and throughput.
- Author
-
Perell?, Jordi, Spadaro, Salvatore, Ricciardi, Sergio, Careglio, Davide, Peng, Shuping, Nejabati, Reza, Zervas, George, Simeonidou, Dimitra, Predieri, Alessandro, Biancani, Matteo, S. Dorren, Harm, Lucente, Stefano, Luo, Jun, Calabretta, Nicola, Bernini, Giacomo, Ciulli, Nicola, Sancho, Jose, Iordache, Steluta, Farreras, Montse, and Becerra, Yolanda
- Subjects
OPTICAL communications ,PACKET switching ,DATA library software ,INTEGRATED circuit interconnections ,COMPUTER architecture - Abstract
Applications running inside data centers are enabled through the cooperation of thousands of servers arranged in racks and interconnected together through the data center network. Current DCN architectures based on electronic devices are neither scalable to face the massive growth of DCs, nor flexible enough to efficiently and cost-effectively support highly dynamic application traffic profiles. The FP7 European Project LIGHTNESS foresees extending the capabilities of today?s electrical DCNs through the introduction of optical packet switching and optical circuit switching paradigms, realizing together an advanced and highly scalable DCN architecture for ultra-high-bandwidth and low-latency server-to-server interconnection. This article reviews the current DC and high-performance computing (HPC) outlooks, followed by an analysis of the main requirements for future DCs and HPC platforms. As the key contribution of the article, the LIGHTNESS DCN solution is presented, deeply elaborating on the envisioned DCN data plane technologies, as well as on the unified SDN-enabled control plane architectural solution that will empower OPS and OCS transmission technologies with superior flexibility, manageability, and customizability. [ABSTRACT FROM PUBLISHER]
- Published
- 2013
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.