40 results on '"Péré, P."'
Search Results
2. Towards a graph-based foundation model for network traffic analysis
- Author
-
Van Langendonck, Louis, Castell-Uroz, Ismael, and Barlet-Ros, Pere
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Cryptography and Security ,Computer Science - Networking and Internet Architecture - Abstract
Foundation models have shown great promise in various fields of study. A potential application of such models is in computer network traffic analysis, where these models can grasp the complexities of network traffic dynamics and adapt to any specific task or network environment with minimal fine-tuning. Previous approaches have used tokenized hex-level packet data and the model architecture of large language transformer models. We propose a new, efficient graph-based alternative at the flow-level. Our approach represents network traffic as a dynamic spatio-temporal graph, employing a self-supervised link prediction pretraining task to capture the spatial and temporal dynamics in this network graph framework. To evaluate the effectiveness of our approach, we conduct a few-shot learning experiment for three distinct downstream network tasks: intrusion detection, traffic classification, and botnet classification. Models finetuned from our pretrained base achieve an average performance increase of 6.87\% over training from scratch, demonstrating their ability to effectively learn general network traffic dynamics during pretraining. This success suggests the potential for a large-scale version to serve as an operational foundational model., Comment: Pre-print of Accepted Workshop paper to 3rd GNNet, co-located with CoNEXT'24
- Published
- 2024
3. PPT-GNN: A Practical Pre-Trained Spatio-Temporal Graph Neural Network for Network Security
- Author
-
Van Langendonck, Louis, Castell-Uroz, Ismael, and Barlet-Ros, Pere
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Cryptography and Security - Abstract
Recent works have demonstrated the potential of Graph Neural Networks (GNN) for network intrusion detection. Despite their advantages, a significant gap persists between real-world scenarios, where detection speed is critical, and existing proposals, which operate on large graphs representing several hours of traffic. This gap results in unrealistic operational conditions and impractical detection delays. Moreover, existing models do not generalize well across different networks, hampering their deployment in production environments. To address these issues, we introduce PPTGNN, a practical spatio-temporal GNN for intrusion detection. PPTGNN enables near real-time predictions, while better capturing the spatio-temporal dynamics of network attacks. PPTGNN employs self-supervised pre-training for improved performance and reduced dependency on labeled data. We evaluate PPTGNN on three public datasets and show that it significantly outperforms state-of-the-art models, such as E-ResGAT and E-GraphSAGE, with an average accuracy improvement of 10.38%. Finally, we show that a pre-trained PPTGNN can easily be fine-tuned to unseen networks with minimal labeled examples. This highlights the potential of PPTGNN as a general, large-scale pre-trained model that can effectively operate in diverse network environments., Comment: Paper currently under review. Code will be made public upon acceptance. 8 pages long, 4 figures
- Published
- 2024
4. Enhanced Detection of Transdermal Alcohol Levels Using Hyperdimensional Computing on Embedded Devices
- Author
-
Segura, Manuel E., Verges, Pere, Chen, Justin Tian Jin, Arangott, Ramesh, Garcia, Angela Kristine, Reynoso, Laura Garcia, Nicolau, Alexandru, Givargis, Tony, and Gago-Masague, Sergio
- Subjects
Computer Science - Machine Learning - Abstract
Alcohol consumption has a significant impact on individuals' health, with even more pronounced consequences when consumption becomes excessive. One approach to promoting healthier drinking habits is implementing just-in-time interventions, where timely notifications indicating intoxication are sent during heavy drinking episodes. However, the complexity or invasiveness of an intervention mechanism may deter an individual from using them in practice. Previous research tackled this challenge using collected motion data and conventional Machine Learning (ML) algorithms to classify heavy drinking episodes, but with impractical accuracy and computational efficiency for mobile devices. Consequently, we have elected to use Hyperdimensional Computing (HDC) to design a just-in-time intervention approach that is practical for smartphones, smart wearables, and IoT deployment. HDC is a framework that has proven results in processing real-time sensor data efficiently. This approach offers several advantages, including low latency, minimal power consumption, and high parallelism. We explore various HDC encoding designs and combine them with various HDC learning models to create an optimal and feasible approach for mobile devices. Our findings indicate an accuracy rate of 89\%, which represents a substantial 12\% improvement over the current state-of-the-art.
- Published
- 2024
5. Molecular Classification Using Hyperdimensional Graph Classification
- Author
-
Verges, Pere, Nunes, Igor, Heddes, Mike, Givargis, Tony, and Nicolau, Alexandru
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Neural and Evolutionary Computing ,Quantitative Biology - Quantitative Methods - Abstract
Our work introduces an innovative approach to graph learning by leveraging Hyperdimensional Computing. Graphs serve as a widely embraced method for conveying information, and their utilization in learning has gained significant attention. This is notable in the field of chemoinformatics, where learning from graph representations plays a pivotal role. An important application within this domain involves the identification of cancerous cells across diverse molecular structures. We propose an HDC-based model that demonstrates comparable Area Under the Curve results when compared to state-of-the-art models like Graph Neural Networks (GNNs) or the Weisfieler-Lehman graph kernel (WL). Moreover, it outperforms previously proposed hyperdimensional computing graph learning methods. Furthermore, it achieves noteworthy speed enhancements, boasting a 40x acceleration in the training phase and a 15x improvement in inference time compared to GNN and WL models. This not only underscores the efficacy of the HDC-based method, but also highlights its potential for expedited and resource-efficient graph learning.
- Published
- 2024
6. Detecting Contextual Network Anomalies with Graph Neural Networks
- Author
-
Latif-Martínez, Hamid, Suárez-Varela, José, Cabellos-Aparicio, Albert, and Barlet-Ros, Pere
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Networking and Internet Architecture - Abstract
Detecting anomalies on network traffic is a complex task due to the massive amount of traffic flows in today's networks, as well as the highly-dynamic nature of traffic over time. In this paper, we propose the use of Graph Neural Networks (GNN) for network traffic anomaly detection. We formulate the problem as contextual anomaly detection on network traffic measurements, and propose a custom GNN-based solution that detects traffic anomalies on origin-destination flows. In our evaluation, we use real-world data from Abilene (6 months), and make a comparison with other widely used methods for the same task (PCA, EWMA, RNN). The results show that the anomalies detected by our solution are quite complementary to those captured by the baselines (with a max. of 36.33% overlapping anomalies for PCA). Moreover, we manually inspect the anomalies detected by our method, and find that a large portion of them can be visually validated by a network expert (64% with high confidence, 18% with mid confidence, 18% normal traffic). Lastly, we analyze the characteristics of the anomalies through two paradigmatic cases that are quite representative of the bulk of anomalies., Comment: 7 pages, 3 figures, 2nd International Workshop on Graph Neural Networking (GNNet '23)
- Published
- 2023
- Full Text
- View/download PDF
7. A Self-Commissioning Edge Computing Method for Data-Driven Anomaly Detection in Power Electronic Systems
- Author
-
Gomez, Pere Izquierdo, Gajardo, Miguel E. Lopez, Mijatovic, Nenad, and Dragicevic, Tomislav
- Subjects
Computer Science - Machine Learning ,Electrical Engineering and Systems Science - Signal Processing - Abstract
Ensuring the reliability of power electronic converters is a matter of great importance, and data-driven condition monitoring techniques are cementing themselves as an important tool for this purpose. However, translating methods that work well in controlled lab environments to field applications presents significant challenges, notably because of the limited diversity and accuracy of the lab training data. By enabling the use of field data, online machine learning can be a powerful tool to overcome this problem, but it introduces additional challenges in ensuring the stability and predictability of the training processes. This work presents an edge computing method that mitigates these shortcomings with minimal additional memory usage, by employing an autonomous algorithm that prioritizes the storage of training samples with larger prediction errors. The method is demonstrated on the use case of a self-commissioning condition monitoring system, in the form of a thermal anomaly detection scheme for a variable frequency motor drive, where the algorithm self-learned to distinguish normal and anomalous operation with minimal prior knowledge. The obtained results, based on experimental data, show a significant improvement in prediction accuracy and training speed, when compared to equivalent models trained online without the proposed data selection process.
- Published
- 2023
- Full Text
- View/download PDF
8. Building a Graph-based Deep Learning network model from captured traffic traces
- Author
-
Güemes-Palau, Carlos, Galmés, Miquel Ferriol, Cabellos-Aparicio, Albert, and Barlet-Ros, Pere
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Machine Learning - Abstract
Currently the state of the art network models are based or depend on Discrete Event Simulation (DES). While DES is highly accurate, it is also computationally costly and cumbersome to parallelize, making it unpractical to simulate high performance networks. Additionally, simulated scenarios fail to capture all of the complexities present in real network scenarios. While there exists network models based on Machine Learning (ML) techniques to minimize these issues, these models are also trained with simulated data and hence vulnerable to the same pitfalls. Consequently, the Graph Neural Networking Challenge 2023 introduces a dataset of captured traffic traces that can be used to build a ML-based network model without these limitations. In this paper we propose a Graph Neural Network (GNN)-based solution specifically designed to better capture the complexities of real network scenarios. This is done through a novel encoding method to capture information from the sequence of captured packets, and an improved message passing algorithm to better represent the dependencies present in physical networks. We show that the proposed solution it is able to learn and generalize to unseen captured network scenarios., Comment: 8 pages, 4 figures
- Published
- 2023
9. Topological Graph Signal Compression
- Author
-
Bernárdez, Guillermo, Telyatnikov, Lev, Alarcón, Eduard, Cabellos-Aparicio, Albert, Barlet-Ros, Pere, and Liò, Pietro
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Networking and Internet Architecture - Abstract
Recently emerged Topological Deep Learning (TDL) methods aim to extend current Graph Neural Networks (GNN) by naturally processing higher-order interactions, going beyond the pairwise relations and local neighborhoods defined by graph representations. In this paper we propose a novel TDL-based method for compressing signals over graphs, consisting in two main steps: first, disjoint sets of higher-order structures are inferred based on the original signal --by clustering $N$ datapoints into $K\ll N$ collections; then, a topological-inspired message passing gets a compressed representation of the signal within those multi-element sets. Our results show that our framework improves both standard GNN and feed-forward architectures in compressing temporal link-based signals from two real-word Internet Service Provider Networks' datasets --from $30\%$ up to $90\%$ better reconstruction errors across all evaluation scenarios--, suggesting that it better captures and exploits spatial and temporal correlations over the whole graph-based network structure., Comment: Accepted as Oral at the Second Learning on Graphs Conference (LoG 2023). The recording of the talk can be found in https://www.youtube.com/watch?v=OcruIkiRkiU
- Published
- 2023
10. GraphCC: A Practical Graph Learning-based Approach to Congestion Control in Datacenters
- Author
-
Bernárdez, Guillermo, Suárez-Varela, José, Shi, Xiang, Xiao, Shihan, Cheng, Xiangle, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning ,Computer Science - Multiagent Systems - Abstract
Congestion Control (CC) plays a fundamental role in optimizing traffic in Data Center Networks (DCN). Currently, DCNs mainly implement two main CC protocols: DCTCP and DCQCN. Both protocols -- and their main variants -- are based on Explicit Congestion Notification (ECN), where intermediate switches mark packets when they detect congestion. The ECN configuration is thus a crucial aspect on the performance of CC protocols. Nowadays, network experts set static ECN parameters carefully selected to optimize the average network performance. However, today's high-speed DCNs experience quick and abrupt changes that severely change the network state (e.g., dynamic traffic workloads, incast events, failures). This leads to under-utilization and sub-optimal performance. This paper presents GraphCC, a novel Machine Learning-based framework for in-network CC optimization. Our distributed solution relies on a novel combination of Multi-agent Reinforcement Learning (MARL) and Graph Neural Networks (GNN), and it is compatible with widely deployed ECN-based CC protocols. GraphCC deploys distributed agents on switches that communicate with their neighbors to cooperate and optimize the global ECN configuration. In our evaluation, we test the performance of GraphCC under a wide variety of scenarios, focusing on the capability of this solution to adapt to new scenarios unseen during training (e.g., new traffic workloads, failures, upgrades). We compare GraphCC with a state-of-the-art MARL-based solution for ECN tuning -- ACC -- and observe that our proposed solution outperforms the state-of-the-art baseline in all of the evaluation scenarios, showing improvements up to $20\%$ in Flow Completion Time as well as significant reductions in buffer occupancy ($38.0-85.7\%$)., Comment: 11 pages, 7 figures, 2 tables
- Published
- 2023
11. HDCC: A Hyperdimensional Computing compiler for classification on embedded systems and high-performance computing
- Author
-
Vergés, Pere, Heddes, Mike, Nunes, Igor, Givargis, Tony, and Nicolau, Alexandru
- Subjects
Computer Science - Machine Learning ,Computer Science - Distributed, Parallel, and Cluster Computing - Abstract
Hyperdimensional Computing (HDC) is a bio-inspired computing framework that has gained increasing attention, especially as a more efficient approach to machine learning (ML). This work introduces the \name{} compiler, the first open-source compiler that translates high-level descriptions of HDC classification methods into optimized C code. The code generated by the proposed compiler has three main features for embedded systems and High-Performance Computing: (1) it is self-contained and has no library or platform dependencies; (2) it supports multithreading and single instruction multiple data (SIMD) instructions using C intrinsics; (3) it is optimized for maximum performance and minimal memory usage. \name{} is designed like a modern compiler, featuring an intuitive and descriptive input language, an intermediate representation (IR), and a retargetable backend. This makes \name{} a valuable tool for research and applications exploring HDC for classification tasks on embedded systems and High-Performance Computing. To substantiate these claims, we conducted experiments with HDCC on several of the most popular datasets in the HDC literature. The experiments were run on four different machines, including different hyperparameter configurations, and the results were compared to a popular prototyping library built on PyTorch. The results show a training and inference speedup of up to 132x, averaging 25x across all datasets and machines. Regarding memory usage, using 10240-dimensional hypervectors, the average reduction was 5x, reaching up to 14x. When considering vectors of 64 dimensions, the average reduction was 85x, with a maximum of 158x less memory utilization., Comment: 8 pages, 3 figures
- Published
- 2023
12. MAGNNETO: A Graph Neural Network-based Multi-Agent system for Traffic Engineering
- Author
-
Bernárdez, Guillermo, Suárez-Varela, José, López, Albert, Shi, Xiang, Xiao, Shihan, Cheng, Xiangle, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Machine Learning ,Computer Science - Multiagent Systems - Abstract
Current trends in networking propose the use of Machine Learning (ML) for a wide variety of network optimization tasks. As such, many efforts have been made to produce ML-based solutions for Traffic Engineering (TE), which is a fundamental problem in ISP networks. Nowadays, state-of-the-art TE optimizers rely on traditional optimization techniques, such as Local search, Constraint Programming, or Linear programming. In this paper, we present MAGNNETO, a distributed ML-based framework that leverages Multi-Agent Reinforcement Learning and Graph Neural Networks for distributed TE optimization. MAGNNETO deploys a set of agents across the network that learn and communicate in a distributed fashion via message exchanges between neighboring agents. Particularly, we apply this framework to optimize link weights in OSPF, with the goal of minimizing network congestion. In our evaluation, we compare MAGNNETO against several state-of-the-art TE optimizers in more than 75 topologies (up to 153 nodes and 354 links), including realistic traffic loads. Our experimental results show that, thanks to its distributed nature, MAGNNETO achieves comparable performance to state-of-the-art TE optimizers with significantly lower execution times. Moreover, our ML-based solution demonstrates a strong generalization capability to successfully operate in new networks unseen during training., Comment: IEEE Transactions on Cognitive Communications and Networking (2023). arXiv admin note: text overlap with arXiv:2109.01445
- Published
- 2023
- Full Text
- View/download PDF
13. A topological classifier to characterize brain states: When shape matters more than variance
- Author
-
Ferrà, Aina, Cecchini, Gloria, Fisas, Fritz-Pere Nobbe, Casacuberta, Carles, and Cos, Ignasi
- Subjects
Computer Science - Machine Learning ,Electrical Engineering and Systems Science - Signal Processing ,Mathematics - Algebraic Topology ,55N31, 92C20, 62R40, 68T09 - Abstract
Despite the remarkable accuracies attained by machine learning classifiers to separate complex datasets in a supervised fashion, most of their operation falls short to provide an informed intuition about the structure of data, and, what is more important, about the phenomena being characterized by the given datasets. By contrast, topological data analysis (TDA) is devoted to study the shape of data clouds by means of persistence descriptors and provides a quantitative characterization of specific topological features of the dataset under scrutiny. In this article we introduce a novel TDA-based classifier that works on the principle of assessing quantifiable changes on topological metrics caused by the addition of new input to a subset of data. We used this classifier with a high-dimensional electro-encephalographic (EEG) dataset recorded from eleven participants during a decision-making experiment in which three motivational states were induced through a manipulation of social pressure. After processing a band-pass filtered version of EEG signals, we calculated silhouettes from persistence diagrams associated with each motivated state, and classified unlabeled signals according to their impact on each reference silhouette. Our results show that in addition to providing accuracies within the range of those of a nearest neighbour classifier, the TDA classifier provides formal intuition of the structure of the dataset as well as an estimate of its intrinsic dimension. Towards this end, we incorporated dimensionality reduction methods to our procedure and found that the accuracy of our TDA classifier is generally not sensitive to explained variance but rather to shape, contrary to what happens with most machine learning classifiers., Comment: 21 pages, 13 figures
- Published
- 2023
14. Neural SDEs for Conditional Time Series Generation and the Signature-Wasserstein-1 metric
- Author
-
Lozano, Pere Díaz, Bagén, Toni Lozano, and Vives, Josep
- Subjects
Statistics - Machine Learning ,Computer Science - Machine Learning ,Mathematics - Probability - Abstract
(Conditional) Generative Adversarial Networks (GANs) have found great success in recent years, due to their ability to approximate (conditional) distributions over extremely high dimensional spaces. However, they are highly unstable and computationally expensive to train, especially in the time series setting. Recently, it has been proposed the use of a key object in rough path theory, called the signature of a path, which is able to convert the min-max formulation given by the (conditional) GAN framework into a classical minimization problem. However, this method is extremely expensive in terms of memory cost, sometimes even becoming prohibitive. To overcome this, we propose the use of \textit{Conditional Neural Stochastic Differential Equations}, which have a constant memory cost as a function of depth, being more memory efficient than traditional deep learning architectures. We empirically test that this proposed model is more efficient than other classical approaches, both in terms of memory cost and computational time, and that it usually outperforms them in terms of performance., Comment: 27 pages, 3 figures
- Published
- 2023
15. RouteNet-Fermi: Network Modeling with Graph Neural Networks
- Author
-
Ferriol-Galmés, Miquel, Paillisse, Jordi, Suárez-Varela, José, Rusek, Krzysztof, Xiao, Shihan, Shi, Xiang, Cheng, Xiangle, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Network models are an essential block of modern networks. For example, they are widely used in network planning and optimization. However, as networks increase in scale and complexity, some models present limitations, such as the assumption of Markovian traffic in queuing theory models, or the high computational cost of network simulators. Recent advances in machine learning, such as Graph Neural Networks (GNN), are enabling a new generation of network models that are data-driven and can learn complex non-linear behaviors. In this paper, we present RouteNet-Fermi, a custom GNN model that shares the same goals as Queuing Theory, while being considerably more accurate in the presence of realistic traffic models. The proposed model predicts accurately the delay, jitter, and packet loss of a network. We have tested RouteNet-Fermi in networks of increasing size (up to 300 nodes), including samples with mixed traffic profiles -- e.g., with complex non-Markovian models -- and arbitrary routing and queue scheduling configurations. Our experimental results show that RouteNet-Fermi achieves similar accuracy as computationally-expensive packet-level simulators and scales accurately to larger networks. Our model produces delay estimates with a mean relative error of 6.24% when applied to a test dataset of 1,000 samples, including network topologies one order of magnitude larger than those seen during training. Finally, we have also evaluated RouteNet-Fermi with measurements from a physical testbed and packet traces from a real-life network., Comment: This paper has been accepted for publication at IEEE/ACM Transactions on Networking 2023 (DOI: 10.1109/TNET.2023.3269983). \copyright 2023 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses
- Published
- 2022
- Full Text
- View/download PDF
16. Out-of-sample scoring and automatic selection of causal estimators
- Author
-
Kraev, Egor, Flesch, Timo, Lekunze, Hudson Taylor, Harley, Mark, and Morell, Pere Planell
- Subjects
Computer Science - Machine Learning ,Statistics - Methodology - Abstract
Recently, many causal estimators for Conditional Average Treatment Effect (CATE) and instrumental variable (IV) problems have been published and open sourced, allowing to estimate granular impact of both randomized treatments (such as A/B tests) and of user choices on the outcomes of interest. However, the practical application of such models has ben hampered by the lack of a valid way to score the performance of such models out of sample, in order to select the best one for a given application. We address that gap by proposing novel scoring approaches for both the CATE case and an important subset of instrumental variable problems, namely those where the instrumental variable is customer acces to a product feature, and the treatment is the customer's choice to use that feature. Being able to score model performance out of sample allows us to apply hyperparameter optimization methods to causal model selection and tuning. We implement that in an open source package that relies on DoWhy and EconML libraries for implementation of causal inference models (and also includes a Transformed Outcome model implementation), and on FLAML for hyperparameter optimization and for component models used in the causal models. We demonstrate on synthetic data that optimizing the proposed scores is a reliable method for choosing the model and its hyperparameter values, whose estimates are close to the true impact, in the randomized CATE and IV cases. Further, we provide examles of applying these methods to real customer data from Wise., Comment: 9 pages, 6 figures
- Published
- 2022
17. Predicting the power grid frequency of European islands
- Author
-
Onsaker, Thorbjørn Lund, Nygård, Heidi S., Gomila, Damià, Colet, Pere, Mikut, Ralf, Jumar, Richard, Maass, Heiko, Kühnapfel, Uwe, Hagenmeyer, Veit, and Schäfer, Benjamin
- Subjects
Statistics - Applications ,Computer Science - Machine Learning ,Electrical Engineering and Systems Science - Systems and Control ,Physics - Data Analysis, Statistics and Probability - Abstract
Modelling, forecasting and overall understanding of the dynamics of the power grid and its frequency are essential for the safe operation of existing and future power grids. Much previous research was focused on large continental areas, while small systems, such as islands are less well-studied. These natural island systems are ideal testing environments for microgrid proposals and artificially islanded grid operation. In the present paper, we utilize measurements of the power grid frequency obtained in European islands: the Faroe Islands, Ireland, the Balearic Islands and Iceland and investigate how their frequency can be predicted, compared to the Nordic power system, acting as a reference. The Balearic islands are found to be particularly deterministic and easy to predict in contrast to hard-to-predict Iceland. Furthermore, we show that typically 2-4 weeks of data are needed to improve prediction performance beyond simple benchmarks., Comment: 17 pages
- Published
- 2022
18. Torchhd: An Open Source Python Library to Support Research on Hyperdimensional Computing and Vector Symbolic Architectures
- Author
-
Heddes, Mike, Nunes, Igor, Vergés, Pere, Kleyko, Denis, Abraham, Danny, Givargis, Tony, Nicolau, Alexandru, and Veidenbaum, Alexander
- Subjects
Computer Science - Machine Learning - Abstract
Hyperdimensional computing (HD), also known as vector symbolic architectures (VSA), is a framework for computing with distributed representations by exploiting properties of random high-dimensional vector spaces. The commitment of the scientific community to aggregate and disseminate research in this particularly multidisciplinary area has been fundamental for its advancement. Joining these efforts, we present Torchhd, a high-performance open source Python library for HD/VSA. Torchhd seeks to make HD/VSA more accessible and serves as an efficient foundation for further research and application development. The easy-to-use library builds on top of PyTorch and features state-of-the-art HD/VSA functionality, clear documentation, and implementation examples from well-known publications. Comparing publicly available code with their corresponding Torchhd implementation shows that experiments can run up to 100x faster. Torchhd is available at: https://github.com/hyperdimensional-computing/torchhd.
- Published
- 2022
19. A multimodal approach for Parkinson disease analysis
- Author
-
Faundez-Zanuy, Marcos, Satue-Villar, Antonio, Mekyska, Jiri, Arreola, Viridiana, Sanz, Pilar, Paul, Carles, Guirao, Luis, Serra, Mateu, Rofes, Laia, Clavé, Pere, Sesa-Nogueras, Enric, and Roure, Josep
- Subjects
Quantitative Biology - Neurons and Cognition ,Computer Science - Machine Learning - Abstract
Parkinson's disease (PD) is the second most frequent neurodegenerative disease with prevalence among general population reaching 0.1-1 %, and an annual incidence between 1.3-2.0/10000 inhabitants. The mean age at diagnosis of PD is 55 and most patients are between 50 and 80 years old. The most obvious symptoms are movement-related; these include tremor, rigidity, slowness of movement and walking difficulties. Frequently these are the symptoms that lead to the PD diagnoses. Later, thinking and behavioral problems may arise, and other symptoms include cognitive impairment and sensory, sleep and emotional problems. In this paper we will present an ongoing project that will evaluate if voice and handwriting analysis can be reliable predictors/indicators of swallowing and balance impairments in PD. An important advantage of voice and handwritten analysis is its low intrusiveness and easy implementation in clinical practice. Thus, if a significant correlation between these simple analyses and the gold standard video-fluoroscopic analysis will imply simpler and less stressing diagnostic test for the patients as well as the use of cheaper analysis systems., Comment: 10 pages
- Published
- 2022
- Full Text
- View/download PDF
20. RouteNet-Erlang: A Graph Neural Network for Network Performance Evaluation
- Author
-
Ferriol-Galmés, Miquel, Rusek, Krzysztof, Suárez-Varela, José, Xiao, Shihan, Cheng, Xiangle, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Machine Learning - Abstract
Network modeling is a fundamental tool in network research, design, and operation. Arguably the most popular method for modeling is Queuing Theory (QT). Its main limitation is that it imposes strong assumptions on the packet arrival process, which typically do not hold in real networks. In the field of Deep Learning, Graph Neural Networks (GNN) have emerged as a new technique to build data-driven models that can learn complex and non-linear behavior. In this paper, we present \emph{RouteNet-Erlang}, a pioneering GNN architecture designed to model computer networks. RouteNet-Erlang supports complex traffic models, multi-queue scheduling policies, routing policies and can provide accurate estimates in networks not seen in the training phase. We benchmark RouteNet-Erlang against a state-of-the-art QT model, and our results show that it outperforms QT in all the network scenarios., Comment: arXiv admin note: text overlap with arXiv:2110.01261
- Published
- 2022
21. Accelerating Deep Reinforcement Learning for Digital Twin Network Optimization with Evolutionary Strategies
- Author
-
Güemes-Palau, Carlos, Almasan, Paul, Xiao, Shihan, Cheng, Xiangle, Shi, Xiang, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Machine Learning - Abstract
The recent growth of emergent network applications (e.g., satellite networks, vehicular networks) is increasing the complexity of managing modern communication networks. As a result, the community proposed the Digital Twin Networks (DTN) as a key enabler of efficient network management. Network operators can leverage the DTN to perform different optimization tasks (e.g., Traffic Engineering, Network Planning). Deep Reinforcement Learning (DRL) showed a high performance when applied to solve network optimization problems. In the context of DTN, DRL can be leveraged to solve optimization problems without directly impacting the real-world network behavior. However, DRL scales poorly with the problem size and complexity. In this paper, we explore the use of Evolutionary Strategies (ES) to train DRL agents for solving a routing optimization problem. The experimental results show that ES achieved a training time speed-up of 128 and 6 for the NSFNET and GEANT2 topologies respectively., Comment: 5 pages, 5 figures
- Published
- 2022
22. Graph Neural Networks for Communication Networks: Context, Use Cases and Opportunities
- Author
-
Suárez-Varela, José, Almasan, Paul, Ferriol-Galmés, Miquel, Rusek, Krzysztof, Geyer, Fabien, Cheng, Xiangle, Shi, Xiang, Xiao, Shihan, Scarselli, Franco, Cabellos-Aparicio, Albert, and Barlet-Ros, Pere
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Machine Learning ,Electrical Engineering and Systems Science - Signal Processing - Abstract
Graph neural networks (GNN) have shown outstanding applications in many fields where data is fundamentally represented as graphs (e.g., chemistry, biology, recommendation systems). In this vein, communication networks comprise many fundamental components that are naturally represented in a graph-structured manner (e.g., topology, configurations, traffic flows). This position article presents GNNs as a fundamental tool for modeling, control and management of communication networks. GNNs represent a new generation of data-driven models that can accurately learn and reproduce the complex behaviors behind real networks. As a result, such models can be applied to a wide variety of networking use cases, such as planning, online optimization, or troubleshooting. The main advantage of GNNs over traditional neural networks lies in its unprecedented generalization capabilities when applied to other networks and configurations unseen during training, which is a critical feature for achieving practical data-driven solutions for networking. This article comprises a brief tutorial on GNNs and their possible applications to communication networks. To showcase the potential of this technology, we present two use cases with state-of-the-art GNN models respectively applied to wired and wireless networks. Lastly, we delve into the key open challenges and opportunities yet to be explored in this novel research area.
- Published
- 2021
23. Scaling Graph-based Deep Learning models to larger networks
- Author
-
Ferriol-Galmés, Miquel, Suárez-Varela, José, Rusek, Krzysztof, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Machine Learning - Abstract
Graph Neural Networks (GNN) have shown a strong potential to be integrated into commercial products for network control and management. Early works using GNN have demonstrated an unprecedented capability to learn from different network characteristics that are fundamentally represented as graphs, such as the topology, the routing configuration, or the traffic that flows along a series of nodes in the network. In contrast to previous solutions based on Machine Learning (ML), GNN enables to produce accurate predictions even in other networks unseen during the training phase. Nowadays, GNN is a hot topic in the Machine Learning field and, as such, we are witnessing great efforts to leverage its potential in many different fields (e.g., chemistry, physics, social networks). In this context, the Graph Neural Networking challenge 2021 brings a practical limitation of existing GNN-based solutions for networking: the lack of generalization to larger networks. This paper approaches the scalability problem by presenting a GNN-based solution that can effectively scale to larger networks including higher link capacities and aggregated traffic on links.
- Published
- 2021
24. ENERO: Efficient Real-Time WAN Routing Optimization with Deep Reinforcement Learning
- Author
-
Almasan, Paul, Xiao, Shihan, Cheng, Xiangle, Shi, Xiang, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Machine Learning - Abstract
Wide Area Networks (WAN) are a key infrastructure in today's society. During the last years, WANs have seen a considerable increase in network's traffic and network applications, imposing new requirements on existing network technologies (e.g., low latency and high throughput). Consequently, Internet Service Providers (ISP) are under pressure to ensure the customer's Quality of Service and fulfill Service Level Agreements. Network operators leverage Traffic Engineering (TE) techniques to efficiently manage network's resources. However, WAN's traffic can drastically change during time and the connectivity can be affected due to external factors (e.g., link failures). Therefore, TE solutions must be able to adapt to dynamic scenarios in real-time. In this paper we propose Enero, an efficient real-time TE solution based on a two-stage optimization process. In the first one, Enero leverages Deep Reinforcement Learning (DRL) to optimize the routing configuration by generating a long-term TE strategy. To enable efficient operation over dynamic network scenarios (e.g., when link failures occur), we integrated a Graph Neural Network into the DRL agent. In the second stage, Enero uses a Local Search algorithm to improve DRL's solution without adding computational overhead to the optimization process. The experimental results indicate that Enero is able to operate in real-world dynamic network topologies in 4.5 seconds on average for topologies up to 100 edges., Comment: 12 pages, 9 figures
- Published
- 2021
- Full Text
- View/download PDF
25. IGNNITION: Bridging the Gap Between Graph Neural Networks and Networking Systems
- Author
-
Pujol-Perich, David, Suárez-Varela, José, Ferriol, Miquel, Xiao, Shihan, Wu, Bo, Cabellos-Aparicio, Albert, and Barlet-Ros, Pere
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Networking and Internet Architecture ,Electrical Engineering and Systems Science - Signal Processing - Abstract
Recent years have seen the vast potential of Graph Neural Networks (GNN) in many fields where data is structured as graphs (e.g., chemistry, recommender systems). In particular, GNNs are becoming increasingly popular in the field of networking, as graphs are intrinsically present at many levels (e.g., topology, routing). The main novelty of GNNs is their ability to generalize to other networks unseen during training, which is an essential feature for developing practical Machine Learning (ML) solutions for networking. However, implementing a functional GNN prototype is currently a cumbersome task that requires strong skills in neural network programming. This poses an important barrier to network engineers that often do not have the necessary ML expertise. In this article, we present IGNNITION, a novel open-source framework that enables fast prototyping of GNNs for networking systems. IGNNITION is based on an intuitive high-level abstraction that hides the complexity behind GNNs, while still offering great flexibility to build custom GNN architectures. To showcase the versatility and performance of this framework, we implement two state-of-the-art GNN models applied to different networking use cases. Our results show that the GNN models produced by IGNNITION are equivalent in terms of accuracy and performance to their native implementations in TensorFlow.
- Published
- 2021
- Full Text
- View/download PDF
26. Is Machine Learning Ready for Traffic Engineering Optimization?
- Author
-
Bernárdez, Guillermo, Suárez-Varela, José, López, Albert, Wu, Bo, Xiao, Shihan, Cheng, Xiangle, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Traffic Engineering (TE) is a basic building block of the Internet. In this paper, we analyze whether modern Machine Learning (ML) methods are ready to be used for TE optimization. We address this open question through a comparative analysis between the state of the art in ML and the state of the art in TE. To this end, we first present a novel distributed system for TE that leverages the latest advancements in ML. Our system implements a novel architecture that combines Multi-Agent Reinforcement Learning (MARL) and Graph Neural Networks (GNN) to minimize network congestion. In our evaluation, we compare our MARL+GNN system with DEFO, a network optimizer based on Constraint Programming that represents the state of the art in TE. Our experimental results show that the proposed MARL+GNN solution achieves equivalent performance to DEFO in a wide variety of network scenarios including three real-world network topologies. At the same time, we show that MARL+GNN can achieve significant reductions in execution time (from the scale of minutes with DEFO to a few seconds with our solution)., Comment: To appear at IEEE ICNP 2021
- Published
- 2021
27. Unveiling the potential of Graph Neural Networks for robust Intrusion Detection
- Author
-
Pujol-Perich, David, Suárez-Varela, José, Cabellos-Aparicio, Albert, and Barlet-Ros, Pere
- Subjects
Computer Science - Cryptography and Security ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning ,Computer Science - Networking and Internet Architecture - Abstract
The last few years have seen an increasing wave of attacks with serious economic and privacy damages, which evinces the need for accurate Network Intrusion Detection Systems (NIDS). Recent works propose the use of Machine Learning (ML) techniques for building such systems (e.g., decision trees, neural networks). However, existing ML-based NIDS are barely robust to common adversarial attacks, which limits their applicability to real networks. A fundamental problem of these solutions is that they treat and classify flows independently. In contrast, in this paper we argue the importance of focusing on the structural patterns of attacks, by capturing not only the individual flow features, but also the relations between different flows (e.g., the source/destination hosts they share). To this end, we use a graph representation that keeps flow records and their relationships, and propose a novel Graph Neural Network (GNN) model tailored to process and learn from such graph-structured information. In our evaluation, we first show that the proposed GNN model achieves state-of-the-art results in the well-known CIC-IDS2017 dataset. Moreover, we assess the robustness of our solution under two common adversarial attacks, that intentionally modify the packet size and inter-arrival times to avoid detection. The results show that our model is able to maintain the same level of accuracy as in previous experiments, while state-of-the-art ML techniques degrade up to 50% their accuracy (F1-score) under these attacks. This unprecedented level of robustness is mainly induced by the capability of our GNN model to learn flow patterns of attacks structured as graphs., Comment: 7 pages, 4 figures, 1 table
- Published
- 2021
28. The Graph Neural Networking Challenge: A Worldwide Competition for Education in AI/ML for Networks
- Author
-
Suárez-Varela, José, Ferriol-Galmés, Miquel, López, Albert, Almasan, Paul, Bernárdez, Guillermo, Pujol-Perich, David, Rusek, Krzysztof, Bonniot, Loïck, Neumann, Christoph, Schnitzler, François, Taïani, François, Happ, Martin, Maier, Christian, Du, Jia Lei, Herlich, Matthias, Dorfinger, Peter, Hainke, Nick Vincent, Venz, Stefan, Wegener, Johannes, Wissing, Henrike, Wu, Bo, Xiao, Shihan, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Artificial Intelligence ,Computer Science - General Literature ,Computer Science - Machine Learning - Abstract
During the last decade, Machine Learning (ML) has increasingly become a hot topic in the field of Computer Networks and is expected to be gradually adopted for a plethora of control, monitoring and management tasks in real-world deployments. This poses the need to count on new generations of students, researchers and practitioners with a solid background in ML applied to networks. During 2020, the International Telecommunication Union (ITU) has organized the "ITU AI/ML in 5G challenge'', an open global competition that has introduced to a broad audience some of the current main challenges in ML for networks. This large-scale initiative has gathered 23 different challenges proposed by network operators, equipment manufacturers and academia, and has attracted a total of 1300+ participants from 60+ countries. This paper narrates our experience organizing one of the proposed challenges: the "Graph Neural Networking Challenge 2020''. We describe the problem presented to participants, the tools and resources provided, some organization aspects and participation statistics, an outline of the top-3 awarded solutions, and a summary with some lessons learned during all this journey. As a result, this challenge leaves a curated set of educational resources openly available to anyone interested in the topic.
- Published
- 2021
- Full Text
- View/download PDF
29. Applying Graph-based Deep Learning To Realistic Network Scenarios
- Author
-
Ferriol-Galmés, Miquel, Suárez-Varela, José, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Machine Learning - Abstract
Recent advances in Machine Learning (ML) have shown a great potential to build data-driven solutions for a plethora of network-related problems. In this context, building fast and accurate network models is essential to achieve functional optimization tools for networking. However, state-of-the-art ML-based techniques for network modelling are not able to provide accurate estimates of important performance metrics such as delay or jitter in realistic network scenarios with sophisticated queue scheduling configurations. This paper presents a new Graph-based deep learning model able to estimate accurately the per-path mean delay in networks. The proposed model can generalize successfully over topologies, routing configurations, queue scheduling policies and traffic matrices unseen during the training phase.
- Published
- 2020
30. Intrinsic-Extrinsic Convolution and Pooling for Learning on 3D Protein Structures
- Author
-
Hermosilla, Pedro, Schäfer, Marco, Lang, Matěj, Fackelmann, Gloria, Vázquez, Pere Pau, Kozlíková, Barbora, Krone, Michael, Ritschel, Tobias, and Ropinski, Timo
- Subjects
Computer Science - Machine Learning ,Quantitative Biology - Biomolecules ,Statistics - Machine Learning - Abstract
Proteins perform a large variety of functions in living organisms, thus playing a key role in biology. As of now, available learning algorithms to process protein data do not consider several particularities of such data and/or do not scale well for large protein conformations. To fill this gap, we propose two new learning operations enabling deep 3D analysis of large-scale protein data. First, we introduce a novel convolution operator which considers both, the intrinsic (invariant under protein folding) as well as extrinsic (invariant under bonding) structure, by using $n$-D convolutions defined on both the Euclidean distance, as well as multiple geodesic distances between atoms in a multi-graph. Second, we enable a multi-scale protein analysis by introducing hierarchical pooling operators, exploiting the fact that proteins are a recombination of a finite set of amino acids, which can be pooled using shared pooling matrices. Lastly, we evaluate the accuracy of our algorithms on several large-scale data sets for common protein analysis tasks, where we outperform state-of-the-art methods., Comment: International Conference on Learning Representations (ICLR) 2021
- Published
- 2020
31. Deep Reinforcement Learning meets Graph Neural Networks: exploring a routing optimization use case
- Author
-
Almasan, Paul, Suárez-Varela, José, Rusek, Krzysztof, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Machine Learning - Abstract
Deep Reinforcement Learning (DRL) has shown a dramatic improvement in decision-making and automated control problems. Consequently, DRL represents a promising technique to efficiently solve many relevant optimization problems (e.g., routing) in self-driving networks. However, existing DRL-based solutions applied to networking fail to generalize, which means that they are not able to operate properly when applied to network topologies not observed during training. This lack of generalization capability significantly hinders the deployment of DRL technologies in production networks. This is because state-of-the-art DRL-based networking solutions use standard neural networks (e.g., fully connected, convolutional), which are not suited to learn from information structured as graphs. In this paper, we integrate Graph Neural Networks (GNN) into DRL agents and we design a problem specific action space to enable generalization. GNNs are Deep Learning models inherently designed to generalize over graphs of different sizes and structures. This allows the proposed GNN-based DRL agent to learn and generalize over arbitrary network topologies. We test our DRL+GNN agent in a routing optimization use case in optical networks and evaluate it on 180 and 232 unseen synthetic and real-world network topologies respectively. The results show that the DRL+GNN agent is able to outperform state-of-the-art solutions in topologies never seen during training., Comment: 12 pages
- Published
- 2019
- Full Text
- View/download PDF
32. RouteNet: Leveraging Graph Neural Networks for network modeling and optimization in SDN
- Author
-
Rusek, Krzysztof, Suárez-Varela, José, Almasan, Paul, Barlet-Ros, Pere, and Cabellos-Aparicio, Albert
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Network modeling is a key enabler to achieve efficient network operation in future self-driving Software-Defined Networks. However, we still lack functional network models able to produce accurate predictions of Key Performance Indicators (KPI) such as delay, jitter or loss at limited cost. In this paper we propose RouteNet, a novel network model based on Graph Neural Network (GNN) that is able to understand the complex relationship between topology, routing, and input traffic to produce accurate estimates of the per-source/destination per-packet delay distribution and loss. RouteNet leverages the ability of GNNs to learn and model graph-structured information and as a result, our model is able to generalize over arbitrary topologies, routing schemes and traffic intensity. In our evaluation, we show that RouteNet is able to predict accurately the delay distribution (mean delay and jitter) and loss even in topologies, routing and traffic unseen in the training (worst case MRE=15.4%). Also, we present several use cases where we leverage the KPI predictions of our GNN model to achieve efficient routing optimization and network planning., Comment: 12 pages
- Published
- 2019
- Full Text
- View/download PDF
33. Generalization error bounds for kernel matrix completion and extrapolation
- Author
-
Giménez-Febrer, Pere, Pagès-Zamora, Alba, and Giannakis, Georgios B.
- Subjects
Statistics - Machine Learning ,Computer Science - Machine Learning - Abstract
Prior information can be incorporated in matrix completion to improve estimation accuracy and extrapolate the missing entries. Reproducing kernel Hilbert spaces provide tools to leverage the said prior information, and derive more reliable algorithms. This paper analyzes the generalization error of such approaches, and presents numerical tests confirming the theoretical results.
- Published
- 2019
- Full Text
- View/download PDF
34. Autonomous Goal Exploration using Learned Goal Spaces for Visuomotor Skill Acquisition in Robots
- Author
-
Laversanne-Finot, Adrien, Péré, Alexandre, and Oudeyer, Pierre-Yves
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Neural and Evolutionary Computing ,Computer Science - Robotics ,Statistics - Machine Learning - Abstract
The automatic and efficient discovery of skills, without supervision, for long-living autonomous agents, remains a challenge of Artificial Intelligence. Intrinsically Motivated Goal Exploration Processes give learning agents a human-inspired mechanism to sequentially select goals to achieve. This approach gives a new perspective on the lifelong learning problem, with promising results on both simulated and real-world experiments. Until recently, those algorithms were restricted to domains with experimenter-knowledge, since the Goal Space used by the agents was built on engineered feature extractors. The recent advances of deep representation learning, enables new ways of designing those feature extractors, using directly the agent experience. Recent work has shown the potential of those methods on simple yet challenging simulated domains. In this paper, we present recent results showing the applicability of those principles on a real-world robotic setup, where a 6-joint robotic arm learns to manipulate a ball inside an arena, by choosing goals in a space learned from its past experience.
- Published
- 2019
35. Disease Knowledge Transfer across Neurodegenerative Diseases
- Author
-
Marinescu, Razvan V., Lorenzi, Marco, Blumberg, Stefano B., Young, Alexandra L., Morell, Pere P., Oxtoby, Neil P., Eshaghi, Arman, Yong, Keir X., Crutch, Sebastian J., Golland, Polina, and Alexander, Daniel C.
- Subjects
Computer Science - Machine Learning ,Computer Science - Computer Vision and Pattern Recognition ,Quantitative Biology - Quantitative Methods ,Statistics - Machine Learning - Abstract
We introduce Disease Knowledge Transfer (DKT), a novel technique for transferring biomarker information between related neurodegenerative diseases. DKT infers robust multimodal biomarker trajectories in rare neurodegenerative diseases even when only limited, unimodal data is available, by transferring information from larger multimodal datasets from common neurodegenerative diseases. DKT is a joint-disease generative model of biomarker progressions, which exploits biomarker relationships that are shared across diseases. Our proposed method allows, for the first time, the estimation of plausible, multimodal biomarker trajectories in Posterior Cortical Atrophy (PCA), a rare neurodegenerative disease where only unimodal MRI data is available. For this we train DKT on a combined dataset containing subjects with two distinct diseases and sizes of data available: 1) a larger, multimodal typical AD (tAD) dataset from the TADPOLE Challenge, and 2) a smaller unimodal Posterior Cortical Atrophy (PCA) dataset from the Dementia Research Centre (DRC), for which only a limited number of Magnetic Resonance Imaging (MRI) scans are available. Although validation is challenging due to lack of data in PCA, we validate DKT on synthetic data and two patient datasets (TADPOLE and PCA cohorts), showing it can estimate the ground truth parameters in the simulation and predict unseen biomarkers on the two patient datasets. While we demonstrated DKT on Alzheimer's variants, we note DKT is generalisable to other forms of related neurodegenerative diseases. Source code for DKT is available online: https://github.com/mrazvan22/dkt., Comment: accepted at MICCAI 2019, 13 pages, 5 figures, 2 tables
- Published
- 2019
36. Identification and Visualization of the Underlying Independent Causes of the Diagnostic of Diabetic Retinopathy made by a Deep Learning Classifier
- Author
-
de la Torre, Jordi, Valls, Aida, Puig, Domenec, and Romero-Aroca, Pere
- Subjects
Statistics - Machine Learning ,Computer Science - Machine Learning - Abstract
Interpretability is a key factor in the design of automatic classifiers for medical diagnosis. Deep learning models have been proven to be a very effective classification algorithm when trained in a supervised way with enough data. The main concern is the difficulty of inferring rationale interpretations from them. Different attempts have been done in last years in order to convert deep learning classifiers from high confidence statistical black box machines into self-explanatory models. In this paper we go forward into the generation of explanations by identifying the independent causes that use a deep learning model for classifying an image into a certain class. We use a combination of Independent Component Analysis with a Score Visualization technique. In this paper we study the medical problem of classifying an eye fundus image into 5 levels of Diabetic Retinopathy. We conclude that only 3 independent components are enough for the differentiation and correct classification between the 5 disease standard classes. We propose a method for visualizing them and detecting lesions from the generated visual maps.
- Published
- 2018
37. Matrix completion and extrapolation via kernel regression
- Author
-
Giménez-Febrer, Pere, Pagès-Zamora, Alba, and Giannakis, Georgios B.
- Subjects
Statistics - Machine Learning ,Computer Science - Machine Learning - Abstract
Matrix completion and extrapolation (MCEX) are dealt with here over reproducing kernel Hilbert spaces (RKHSs) in order to account for prior information present in the available data. Aiming at a faster and low-complexity solver, the task is formulated as a kernel ridge regression. The resultant MCEX algorithm can also afford online implementation, while the class of kernel functions also encompasses several existing approaches to MC with prior information. Numerical tests on synthetic and real datasets show that the novel approach performs faster than widespread methods such as alternating least squares (ALS) or stochastic gradient descent (SGD), and that the recovery error is reduced, especially when dealing with noisy data.
- Published
- 2018
- Full Text
- View/download PDF
38. Machine Learning in High Energy Physics Community White Paper
- Author
-
Albertsson, Kim, Altoe, Piero, Anderson, Dustin, Anderson, John, Andrews, Michael, Espinosa, Juan Pedro Araque, Aurisano, Adam, Basara, Laurent, Bevan, Adrian, Bhimji, Wahid, Bonacorsi, Daniele, Burkle, Bjorn, Calafiura, Paolo, Campanelli, Mario, Capps, Louis, Carminati, Federico, Carrazza, Stefano, Chen, Yi-fan, Childers, Taylor, Coadou, Yann, Coniavitis, Elias, Cranmer, Kyle, David, Claire, Davis, Douglas, De Simone, Andrea, Duarte, Javier, Erdmann, Martin, Eschle, Jonas, Farbin, Amir, Feickert, Matthew, Castro, Nuno Filipe, Fitzpatrick, Conor, Floris, Michele, Forti, Alessandra, Garra-Tico, Jordi, Gemmler, Jochen, Girone, Maria, Glaysher, Paul, Gleyzer, Sergei, Gligorov, Vladimir, Golling, Tobias, Graw, Jonas, Gray, Lindsey, Greenwood, Dick, Hacker, Thomas, Harvey, John, Hegner, Benedikt, Heinrich, Lukas, Heintz, Ulrich, Hooberman, Ben, Junggeburth, Johannes, Kagan, Michael, Kane, Meghan, Kanishchev, Konstantin, Karpiński, Przemysław, Kassabov, Zahari, Kaul, Gaurav, Kcira, Dorian, Keck, Thomas, Klimentov, Alexei, Kowalkowski, Jim, Kreczko, Luke, Kurepin, Alexander, Kutschke, Rob, Kuznetsov, Valentin, Köhler, Nicolas, Lakomov, Igor, Lannon, Kevin, Lassnig, Mario, Limosani, Antonio, Louppe, Gilles, Mangu, Aashrita, Mato, Pere, Meenakshi, Narain, Meinhard, Helge, Menasce, Dario, Moneta, Lorenzo, Moortgat, Seth, Neubauer, Mark, Newman, Harvey, Otten, Sydney, Pabst, Hans, Paganini, Michela, Paulini, Manfred, Perdue, Gabriel, Perez, Uzziel, Picazio, Attilio, Pivarski, Jim, Prosper, Harrison, Psihas, Fernanda, Radovic, Alexander, Reece, Ryan, Rinkevicius, Aurelius, Rodrigues, Eduardo, Rorie, Jamal, Rousseau, David, Sauers, Aaron, Schramm, Steven, Schwartzman, Ariel, Severini, Horst, Seyfert, Paul, Siroky, Filip, Skazytkin, Konstantin, Sokoloff, Mike, Stewart, Graeme, Stienen, Bob, Stockdale, Ian, Strong, Giles, Sun, Wei, Thais, Savannah, Tomko, Karen, Upfal, Eli, Usai, Emanuele, Ustyuzhanin, Andrey, Vala, Martin, Vasel, Justin, Vallecorsa, Sofia, Verzetti, Mauro, Vilasís-Cardona, Xavier, Vlimant, Jean-Roch, Vukotic, Ilija, Wang, Sean-Jiun, Watts, Gordon, Williams, Michael, Wu, Wenjing, Wunsch, Stefan, Yang, Kun, and Zapata, Omar
- Subjects
Physics - Computational Physics ,Computer Science - Machine Learning ,High Energy Physics - Experiment ,Statistics - Machine Learning - Abstract
Machine learning has been applied to several problems in particle physics research, beginning with applications to high-level physics analysis in the 1990s and 2000s, followed by an explosion of applications in particle and event identification and reconstruction in the 2010s. In this document we discuss promising future research and development areas for machine learning in particle physics. We detail a roadmap for their implementation, software and hardware resource requirements, collaborative initiatives with the data science community, academia and industry, and training the particle physics community in data science. The main objective of the document is to connect and motivate these areas of research and development with the physics drivers of the High-Luminosity Large Hadron Collider and future neutrino experiments and identify the resource needs for their implementation. Additionally we identify areas where collaboration with external communities will be of great benefit., Comment: Editors: Sergei Gleyzer, Paul Seyfert and Steven Schramm
- Published
- 2018
39. Curiosity Driven Exploration of Learned Disentangled Goal Spaces
- Author
-
Laversanne-Finot, Adrien, Péré, Alexandre, and Oudeyer, Pierre-Yves
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Neural and Evolutionary Computing ,Computer Science - Robotics ,Statistics - Machine Learning - Abstract
Intrinsically motivated goal exploration processes enable agents to autonomously sample goals to explore efficiently complex environments with high-dimensional continuous actions. They have been applied successfully to real world robots to discover repertoires of policies producing a wide diversity of effects. Often these algorithms relied on engineered goal spaces but it was recently shown that one can use deep representation learning algorithms to learn an adequate goal space in simple environments. However, in the case of more complex environments containing multiple objects or distractors, an efficient exploration requires that the structure of the goal space reflects the one of the environment. In this paper we show that using a disentangled goal space leads to better exploration performances than an entangled goal space. We further show that when the representation is disentangled, one can leverage it by sampling goals that maximize learning progress in a modular manner. Finally, we show that the measure of learning progress, used to drive curiosity-driven exploration, can be used simultaneously to discover abstract independently controllable features of the environment., Comment: The code used in the experiments is available at https://github.com/flowersteam/Curiosity_Driven_Goal_Exploration
- Published
- 2018
40. Unsupervised Learning of Goal Spaces for Intrinsically Motivated Goal Exploration
- Author
-
Péré, Alexandre, Forestier, Sébastien, Sigaud, Olivier, and Oudeyer, Pierre-Yves
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Intrinsically motivated goal exploration algorithms enable machines to discover repertoires of policies that produce a diversity of effects in complex environments. These exploration algorithms have been shown to allow real world robots to acquire skills such as tool use in high-dimensional continuous state and action spaces. However, they have so far assumed that self-generated goals are sampled in a specifically engineered feature space, limiting their autonomy. In this work, we propose to use deep representation learning algorithms to learn an adequate goal space. This is a developmental 2-stage approach: first, in a perceptual learning stage, deep learning algorithms use passive raw sensor observations of world changes to learn a corresponding latent space; then goal exploration happens in a second stage by sampling goals in this latent space. We present experiments where a simulated robot arm interacts with an object, and we show that exploration algorithms using such learned representations can match the performance obtained using engineered representations.
- Published
- 2018
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.