8 results on '"Pashami, Sepideh"'
Search Results
2. Mode tracking using multiple data streams
- Author
-
Bouguelia, Mohamed-Rafik, Karlsson, Alexander, Pashami, Sepideh, Nowaczyk, Sławomir, and Holst, Anders
- Published
- 2018
- Full Text
- View/download PDF
3. Parallel orthogonal deep neural network.
- Author
-
Mashhadi, Peyman Sheikholharam, Nowaczyk, Sławomir, and Pashami, Sepideh
- Subjects
- *
DEEP learning , *ORTHOGONALIZATION , *ARTIFICIAL neural networks , *DECISION making - Abstract
Ensemble learning methods combine multiple models to improve performance by exploiting their diversity. The success of these approaches relies heavily on the dissimilarity of the base models forming the ensemble. This diversity can be achieved in many ways, with well-known examples including bagging and boosting. It is the diversity of the models within an ensemble that allows the ensemble to correct the errors made by its members, and consequently leads to higher classification or regression performance. A mistake made by a base model can only be rectified if other members behave differently on that particular instance, and provide the aggregator with enough information to make an informed decision. On the contrary, lack of diversity not only lowers model performance, but also wastes computational resources. Nevertheless, in the current state of the art ensemble approaches, there is no guarantee on the level of diversity achieved, and no mechanism ensuring that each member will learn a different decision boundary from the others. In this paper, we propose a parallel orthogonal deep learning architecture in which diversity is enforced by design, through imposing an orthogonality constraint. Multiple deep neural networks are created, parallel to each other. At each parallel layer, the outputs of different base models are subject to Gram–Schmidt orthogonalization. We demonstrate that this approach leads to a high level of diversity from two perspectives. First, the models make different errors on different parts of feature space, and second, they exhibit different levels of uncertainty in their decisions. Experimental results confirm the benefits of the proposed method, compared to standard deep learning models and well-known ensemble methods, in terms of diversity and, as a result, classification performance. • We propose a novel deep learning ensemble method with the enforced diversity mechanism. • We introduce a new approach for analyzing ensemble diversity, taking into account model confidence. • We perform experimental evaluation and show improved classification performance. • We demonstrate an efficient way of building deep learning ensembles with end-to-end architecture. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
4. Rolling the dice for better deep learning performance: A study of randomness techniques in deep neural networks.
- Author
-
Altarabichi, Mohammed Ghaith, Nowaczyk, Sławomir, Pashami, Sepideh, Sheikholharam Mashhadi, Peyman, and Handl, Julia
- Subjects
- *
ARTIFICIAL neural networks , *DEEP learning , *COMPUTER vision , *DATA augmentation , *PERFORMANCE theory , *CLASSIFICATION - Abstract
This paper presents a comprehensive empirical investigation into the interactions between various randomization techniques in Deep Neural Networks (DNNs) and their impact on learning performance. It is well-established that injecting randomness into the training process of DNNs, through various approaches, at different stages, is often beneficial for reducing overfitting and improving generalization. Nonetheless, the interactions between randomness techniques such as weight noise, dropout, and many others remain poorly understood. Consequently, it is challenging to determine which methods can be effectively combined to optimize DNN performance. To address this issue, we categorize the existing randomness techniques into four key types: injection of noise/randomness at the data, model structure, optimization or learning stage. We use this classification to identify gaps in the current coverage of potential mechanisms for the introduction of randomness, leading to proposing two new techniques: adding noise to the loss function and random masking of the gradient updates. In our empirical study, we employ a Particle Swarm Optimizer (PSO) for hyperparameter optimization (HPO) to explore the space of possible configurations to determine where and how much randomness should be injected to maximize DNN performance. We assess the impact of various types and levels of randomness for DNN architectures across standard computer vision benchmarks: MNIST, FASHION-MNIST, CIFAR10, and CIFAR100. Across more than 30 000 evaluated configurations, we perform a detailed examination of the interactions between randomness techniques and their combined impact on DNN performance. Our findings reveal that randomness through data augmentation and in weight initialization are the main contributors to performance improvement. Additionally, correlation analysis demonstrates that different optimizers, such as Adam and Gradient Descent with Momentum, prefer distinct types of randomization during the training process. A GitHub repository with the complete implementation and generated dataset is available.2 [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Multi-domain adaptation for regression under conditional distribution shift.
- Author
-
Taghiyarrenani, Zahra, Nowaczyk, Sławomir, Pashami, Sepideh, and Bouguelia, Mohamed-Rafik
- Subjects
- *
ARTIFICIAL neural networks , *VECTOR spaces , *PROBLEM solving - Abstract
Domain adaptation (DA) methods facilitate cross-domain learning by minimizing the marginal or conditional distribution shift between domains. However, the conditional distribution shift is not well addressed by existing DA techniques for the cross-domain regression learning task. In this paper, we propose Multi-Domain Adaptation for Regression under Conditional shift (DARC) method. DARC constructs a shared feature space such that linear regression on top of that space generalizes to all domains. In other words, DARC aligns different domains according to the task-related information encoded in the values of the dependent variable. It is achieved using a novel Pairwise Similarity Preserver (PSP) loss function. PSP incentivizes the differences between the outcomes of any two samples, regardless of their domain(s), to match the distance between these samples in the constructed space. We perform experiments in both two-domain and multi-domain settings. The two-domain setting is helpful, especially when one domain contains few available labeled samples and can benefit from adaptation to a domain with many labeled samples. The multi-domain setting allows several domains, each with limited data, to be adapted collectively; thus, multiple domains compensate for each other's lack of data. The results from all the experiments conducted both on synthetic and real-world datasets confirm the effectiveness of DARC. • We propose a new pairwise similarity preserver loss function for regression tasks. • We propose a new domain adaptation method for regression tasks. • Our proposed method adapts domains with any arbitrary shift, concept, or covariate. • Our proposed method is capable of adapting multiple domains. • The results prove our method's efficiency in solving real-world problems. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
6. Fast Genetic Algorithm for feature selection — A qualitative approximation approach.
- Author
-
Altarabichi, Mohammed Ghaith, Nowaczyk, Sławomir, Pashami, Sepideh, and Mashhadi, Peyman Sheikholharam
- Subjects
- *
GENETIC algorithms , *PARTICLE swarm optimization , *FEATURE selection , *SWARM intelligence , *EVOLUTIONARY computation , *EVOLUTIONARY algorithms - Abstract
Evolutionary Algorithms (EAs) are often challenging to apply in real-world settings since evolutionary computations involve a large number of evaluations of a typically expensive fitness function. For example, an evaluation could involve training a new machine learning model. An approximation (also known as meta-model or a surrogate) of the true function can be used in such applications to alleviate the computation cost. In this paper, we propose a two-stage surrogate-assisted evolutionary approach to address the computational issues arising from using Genetic Algorithm (GA) for feature selection in a wrapper setting for large datasets. We define "Approximation Usefulness" to capture the necessary conditions to ensure correctness of the EA computations when an approximation is used. Based on this definition, we propose a procedure to construct a lightweight qualitative meta-model by the active selection of data instances. We then use a meta-model to carry out the feature selection task. We apply this procedure to the GA-based algorithm CHC (Cross generational elitist selection, Heterogeneous recombination and Cataclysmic mutation) to create a Qualitative approXimations variant, CHC Q X. We show that CHC Q X converges faster to feature subset solutions of significantly higher accuracy (as compared to CHC), particularly for large datasets with over 100K instances. We also demonstrate the applicability of the thinking behind our approach more broadly to Swarm Intelligence (SI), another branch of the Evolutionary Computation (EC) paradigm with results of PSO Q X , a qualitative approximation adaptation of the Particle Swarm Optimization (PSO) method. A GitHub repository with the complete implementation is available. 2 2 https://github.com/Ghaith81/Fast-Genetic-Algorithm-For-Feature-Selection. • A method to construct computationally efficient meta-models for feature selection. • A definition of the necessary conditions to ensure correctness of EA computation. • Our algorithm converges faster than a wrapper GA, particularly for large datasets. • The approach is useful independently of the used metaheuristic as shown with PSO. • We demonstrate the HUX crossover averaging effect when used in binary optimization. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
7. Material handling machine activity recognition by context ensemble with gated recurrent units.
- Author
-
Chen, Kunru, Rögnvaldsson, Thorsteinn, Nowaczyk, Sławomir, Pashami, Sepideh, Klang, Jonas, and Sternelöv, Gustav
- Subjects
- *
MATERIALS handling , *MACHINING , *MACHINE learning , *FORKLIFT trucks , *PRODUCT improvement , *PRODUCT design - Abstract
Research on machine activity recognition (MAR) is drawing more attention because MAR can provide productivity monitoring for efficiency optimization, better maintenance scheduling, product design improvement, and potential material savings. A particular challenge of MAR for human-operated machines is the overlap when transiting from one activity to another: during transitions, operators often perform two activities simultaneously, e.g., lifting the fork already while approaching a rack, so the exact time when one activity ends and another begins is uncertain. Machine learning models are often uncertain during such activity transitions, and we propose a novel ensemble-based method adapted to fuzzy transitions in a forklift MAR problem. Unlike traditional ensembles, where models in the ensemble are trained on different subsets of data, or with costs that force them to be diverse in their responses, our approach is to train a single model that predicts several activity labels, each under a different context. These individual predictions are not made by independent networks but are made using a structure that allows for sharing important features, i.e., a context ensemble. The results show that the gated recurrent unit network can provide medium or strong confident context ensembles for 95% of the cases in the test set, and the final forklift MAR result achieves accuracies of 97% for driving and 90% for load-handling activities. This study is the first to highlight the overlapping activity issue in MAR problems and to demonstrate that the recognition results can be significantly improved by designing a machine learning framework that addresses this issue. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
8. Pitfalls of assessing extracted hierarchies for multi-class classification.
- Author
-
del Moral, Pablo, Nowaczyk, Sławomir, Sant'Anna, Anita, and Pashami, Sepideh
- Subjects
- *
CLASSIFICATION - Abstract
• We identify several pitfalls in the process of extracting and evaluating methods to extract hierarchies in the context of HMC. • We propose using a random hierarchy as a necessary benchmark to evaluate the relevance of a good hierarchy, in this way we can establish whether our extracted hierarchy is capturing an existing structure between classes or not. • We demonstrate how the relevance of the hierarchy quality depends on the complexity of the dataset and the complexity of the classifiers used. • We perform experimental evaluation of different approaches for extracting hierarchies to showcase the previous contributions. Using hierarchies of classes is one of the standard methods to solve multi-class classification problems. In the literature, selecting the right hierarchy is considered to play a key role in improving classification performance. Although different methods have been proposed, there is still a lack of understanding of what makes a hierarchy good and what makes a method to extract hierarchies perform better or worse. To this effect, we analyze and compare some of the most popular approaches to extracting hierarchies. We identify some common pitfalls that may lead practitioners to make misleading conclusions about their methods. To address some of these problems, we demonstrate that using random hierarchies is an appropriate benchmark to assess how the hierarchy's quality affects the classification performance. In particular, we show how the hierarchy's quality can become irrelevant depending on the experimental setup: when using powerful enough classifiers, the final performance is not affected by the quality of the hierarchy. We also show how comparing the effect of the hierarchies against non-hierarchical approaches might incorrectly indicate their superiority. Our results confirm that datasets with a high number of classes generally present complex structures in how these classes relate to each other. In these datasets, the right hierarchy can dramatically improve classification performance. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.