34 results
Search Results
2. Survey on the research direction of EEG-based signal processing.
- Author
-
Congzhong Sun and Chaozhou Mou
- Subjects
ARTIFICIAL neural networks ,SIGNAL processing ,DATA augmentation ,GENERATIVE adversarial networks ,MACHINE learning - Abstract
Electroencephalography (EEG) is increasingly important in Brain-Computer Interface (BCI) systems due to its portability and simplicity. In this paper, we provide a comprehensive review of research on EEG signal processing techniques since 2021, with a focus on preprocessing, feature extraction, and classification methods. We analyzed 61 research articles retrieved from academic search engines, including CNKI, PubMed, Nature, IEEE Xplore, and Science Direct. For preprocessing, we focus on innovatively proposed preprocessing methods, channel selection, and data augmentation. Data augmentation is classified into conventional methods (sliding windows, segmentation and recombination, and noise injection) and deep learning methods [Generative Adversarial Networks (GAN) and Variation AutoEncoder (VAE)]. We also pay attention to the application of deep learning, and multi-method fusion approaches, including both conventional algorithm fusion and fusion between conventional algorithms and deep learning. Our analysis identifies 35 (57.4%), 18 (29.5%), and 37 (60.7%) studies in the directions of preprocessing, feature extraction, and classification, respectively. We find that preprocessing methods have become widely used in EEG classification (96.7% of reviewed papers) and comparative experiments have been conducted in some studies to validate preprocessing. We also discussed the adoption of channel selection and data augmentation and concluded several mentionable matters about data augmentation. Furthermore, deep learning methods have shown great promise in EEG classification, with Convolutional Neural Networks (CNNs) being the main structure of deep neural networks (92.3% of deep learning papers). We summarize and analyze several innovative neural networks, including CNNs and multi-structure fusion. However, we also identified several problems and limitations of current deep learning techniques in EEG classification, including inappropriate input, low cross-subject accuracy, unbalanced between parameters and time costs, and a lack of interpretability. Finally, we highlight the emerging trend of multi-method fusion approaches (49.2% of reviewed papers) and analyze the data and some examples. We also provide insights into some challenges of multi-method fusion. Our review lays a foundation for future studies to improve EEG classification performance. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
3. Early dementia detection with speech analysis and machine learning techniques.
- Author
-
Jahan, Zerin, Khan, Surbhi Bhatia, and Saraee, Mo
- Subjects
NATURAL language processing ,MACHINE learning ,SPEECH ,ARTIFICIAL neural networks ,DEMENTIA - Abstract
This in-depth study journey explores the context of natural language processing and text analysis in dementia detection, revealing their importance in a variety of fields. Beginning with an examination of the widespread and influence of text data. The dataset utilised in this study is from TalkBank's DementiaBank, which is basically a vast database of multimedia interactions built with the goal of examining communication patterns in the context of dementia. The various communication styles dementia patients exhibit when communicating with others are seen from a unique perspective by this specific dataset. Thorough data preprocessing procedures, including cleansing, tokenization, and structuring, are undertaken, with a focus on improving prediction capabilities through the combination of textual and non-textual information in the field of feature engineering. In the subsequent phase, the precision, recall, and F1-score metrics of Support Vector Machines (SVM), K-Nearest Neighbours (KNN), Random Forest, and Artificial Neural Networks (ANN) are assessed. Empirical facts are synthesized using text analysis methods and models to formulate a coherent conclusion. The significance of text data analysis, the revolutionary potential of natural language processing, and the direction for future research are highlighted in this synthesis. Throughout this paper, readers are encouraged to leverage text data to embark on their own adventures in the evolving, data-centric world of dementia detection. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Mental arithmetic task detection using geometric features extraction of EEG signal based on machine learning.
- Author
-
Hoda Edris, ABADI, Mohammad Karimi, MORIDANI, and Mahshid, MIRZAKHANI
- Subjects
MENTAL arithmetic ,FEATURE extraction ,MACHINE learning ,ATTENTION-deficit hyperactivity disorder ,ARTIFICIAL neural networks - Abstract
BACKGROUND: Mental arithmetic analysis based on electroencephalogram (EEG) signals can help to understand some disorders such as attention deficit hyperactivity disorder, arithmetic disorder, or autism spectrum disorder in which learning is difficult. Most mental computation detection and classification systems rely on the characteristics of a single channel, however, the understanding of the connections between EEG channels, which certainly contains valuable information, is still evolving. The methods presented in this paper are the result of a research project that introduces an alternative method for better and faster receipt of information from the EEG signals of individuals, which are generally complex and nonlinear. METHODS: The EEGs of 66 healthy individuals were recorded in two rest modes and mental task a designed, with a sampling frequency of 500 Hz. To classify these two modes, we extracted features from our recordings to differentiate the EEG signals of these two groups in a single channel as well as combine possible channels. The new method that was proposed was the extraction of several geometric features from Poincaré design analysis, which used the necessary comparison t-test to determine brain differences, with a significance level of less than 0.05 in the state of mental calculations and facial rest. Also, an artificial neural network (ANN) has been used for automatic learning and diagnosis in the two mentioned modes. RESULTS: The results of this paper show that by using a combination of geometric properties (sides, angles, shortest distance, slope, and coefficients of the third-degree equation) using selected channels (FP1, F7, C4, O1) can achieve 100 % accuracy. The sensitivity reached 100 %. As well as 100 % feature. CONCLUSIONS: With the help of mental calculation, it is possible to diagnose, treat, rehabilitate and rehabilitation people who have lost the function of a part of their brain due to a disease in this field (Tab. 6, Fig. 15, Ref. 45). [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
5. Improved Siamese Palmprint Authentication Using Pre-Trained VGG16-Palmprint and Element-Wise Absolute Difference.
- Author
-
Ezz, Mohamed, Alanazi, Waad, Mostafa, Ayman Mohamed, Hamouda, Eslam, Elbashir, Murtada K., and Alruily, Meshrif
- Subjects
PALMPRINT recognition ,BIOMETRIC identification ,DEEP learning ,MACHINE learning ,ARTIFICIAL neural networks - Abstract
Palmprint identification has been conducted over the last two decades in many biometric systems. High-dimensional data with many uncorrelated and duplicated features remains difficult due to several computational complexity issues. This paper presents an interactive authentication approach based on deep learning and feature selection that supports Palmprint authentication. The proposed model has two stages of learning; the first stage is to transfer pre-trained VGG-16 of ImageNet to specific features based on the extraction model. The second stage involves the VGG-16 Palmprint feature extraction in the Siamese network to learn Palmprint similarity. The proposed model achieves robust and reliable end-to-end Palmprint authentication by extracting the convolutional features using VGG-16 Palmprint and the similarity of two input Palmprint using the Siamese network. The second stage uses the CASIA dataset to train and test the Siamese network. The suggested model outperforms comparable studies based on the deep learning approach achieving accuracy and EER of 91.8% and 0.082%, respectively, on the CASIA left-hand images and accuracy and EER of 91.7% and 0.084, respectively, on the CASIA right-hand images. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
6. Multiscale Convolutional Neural Networks for Fault Diagnosis of Wind Turbine Gearbox.
- Author
-
Jiang, Guoqian, He, Haibo, Yan, Jun, and Xie, Ping
- Subjects
ARTIFICIAL neural networks ,SIGNAL convolution ,MULTISCALE modeling ,ARTIFICIAL intelligence ,FAULT diagnosis ,DEEP learning ,WIND turbines ,GEARBOXES - Abstract
This paper proposes a novel intelligent fault diagnosis method to automatically identify different health conditions of wind turbine (WT) gearbox. Unlike traditional approaches, where feature extraction and classification are separately designed and performed, this paper aims to automatically learn effective fault features directly from raw vibration signals while classify the type of faults in a single framework, thus providing an end-to-end learning-based fault diagnosis system for WT gearbox without additional signal processing and diagnostic expertise. Considering the multiscale characteristics inherent in vibration signals of a gearbox, a new multiscale convolutional neural network (MSCNN) architecture is proposed to perform multiscale feature extraction and classification simultaneously. The proposed MSCNN incorporates multiscale learning into the traditional CNN architecture, which has two merits: 1) high-level fault features can be effectively learned by the hierarchical learning structure with multiple pairs of convolutional and pooling layers; and 2) multiscale learning scheme can capture complementary and rich diagnosis information at different scales. This greatly improves the feature learning ability and enables better diagnosis performance. The proposed MSCNN approach is evaluated through experiments on a WT gearbox test rig. Experimental results and comprehensive comparison analysis with respect to the traditional CNN and traditional multiscale feature extractors have demonstrated the superiority of the proposed method. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
7. Survey and Evaluation of Neural 3D Shape Classification Approaches.
- Author
-
Mirbauer, Martin, Krabec, Miroslav, Krivanek, Jaroslav, and Sikudova, Elena
- Subjects
ARTIFICIAL neural networks ,MACHINE learning ,OBJECT recognition (Computer vision) ,CLASSIFICATION ,GEOMETRIC shapes ,COMPUTER graphics ,CLASSIFICATION algorithms - Abstract
Classification of 3D objects – the selection of a category in which each object belongs – is of great interest in the field of machine learning. Numerous researchers use deep neural networks to address this problem, altering the network architecture and representation of the 3D shape used as an input. To investigate the effectiveness of their approaches, we conduct an extensive survey of existing methods and identify common ideas by which we categorize them into a taxonomy. Second, we evaluate 11 selected classification networks on two 3D object datasets, extending the evaluation to a larger dataset on which most of the selected approaches have not been tested yet. For this, we provide a framework for converting shapes from common 3D mesh formats into formats native to each network, and for training and evaluating different classification approaches on this data. Despite being partially unable to reach the accuracies reported in the original papers, we compare the relative performance of the approaches as well as their performance when changing datasets as the only variable to provide valuable insights into performance on different kinds of data. We make our code available to simplify running training experiments with multiple neural networks with different prerequisites. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
8. Learning Subspace-Based RBFNN Using Coevolutionary Algorithm for Complex Classification Tasks.
- Author
-
Tian, Jin, Li, Minqiang, Chen, Fuzan, and Feng, Nan
- Subjects
MACHINE learning ,SUBSPACES (Mathematics) ,ARTIFICIAL neural networks ,EVOLUTIONARY algorithms ,FEATURE extraction - Abstract
Many real-world classification problems are characterized by samples of a complex distribution in the input space. The classification accuracy is determined by intrinsic properties of all samples in subspaces of features. This paper proposes a novel algorithm for the construction of radial basis function neural network (RBFNN) classifier based on subspace learning. In this paper, feature subspaces are obtained for every hidden node of the RBFNN during the learning process. The connection weights between the input layer and the hidden layer are adjusted to produce various subspaces with dominative features for different hidden nodes. The network structure and dominative features are encoded in two subpopulations that are cooperatively coevolved using the coevolutionary algorithm to achieve a better global optimality for the estimated RBFNN. Experimental results illustrate that the proposed algorithm is able to obtain RBFNN models with both better classification accuracy and simpler network structure when compared with other learning algorithms. Thus, the proposed model provides a more flexible and efficient approach to complex classification tasks by employing the local characteristics of samples in subspaces. [ABSTRACT FROM PUBLISHER]
- Published
- 2016
- Full Text
- View/download PDF
9. A Survey on Artificial Intelligence in Chinese Sign Language Recognition.
- Author
-
Jiang, Xianwei, Satapathy, Suresh Chandra, Yang, Longxiang, Wang, Shui-Hua, and Zhang, Yu-Dong
- Subjects
SIGN language ,ARTIFICIAL intelligence ,CONVOLUTIONAL neural networks ,CHINESE language ,ARTIFICIAL neural networks ,FEATURE extraction - Abstract
Chinese Sign Language (CSL) offers the main means of communication for the hearing impaired in China. Sign Language Recognition (SLR) can shorten the distance between the hearing-impaired and healthy people and help them integrate into the society. Therefore, SLR has become the focus of sign language application research. Over the years, the continuous development of new technologies provides a source and motivation for SLR. This paper aims to cover the most recent approaches in Chinese Sign Language Recognition (CSLR). With a thorough review of superior methods from 2000 to 2019 in CSLR researches, various techniques and algorithms such as scale-invariant feature transform, histogram of oriented gradients, wavelet entropy, Hu moment invariant, Fourier descriptor, gray-level co-occurrence matrix, dynamic time warping, principal component analysis, autoencoder, hidden Markov model (HMM), support vector machine (SVM), random forest, skin color modeling method, k-NN, artificial neural network, convolutional neural network (CNN), and transfer learning are discussed in detail, which are based on several major stages, that is, data acquisition, preprocessing, feature extraction, and classification. CSLR was summarized from some aspect as follows: methods of classification and feature extraction, accuracy/performance evaluation, and sample size/datasets. The advantages and limitations of different CSLR approaches were compared. It was found that data acquisition is mainly through Kinect and camera, and the feature extraction focuses on hand's shape and spatiotemporal factors, but ignoring facial expressions. HMM and SVM are used most in the classification. CNN is becoming more and more popular, and a deep neural network-based recognition approach will be the future trend. However, due to the complexity of the contemporary Chinese language, CSLR generally has a lower accuracy than other SLR. It is necessary to establish an appropriate dataset to conduct comparable experiments. The issue of decreasing accuracy as the dataset increases needs to resolve. Overall, our study is hoped to give a comprehensive presentation for those people who are interested in CSLR and SLR and to further contribute to the future research. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
10. Data augmentation for handwritten digit recognition using generative adversarial networks.
- Author
-
Jha, Ganesh and Cecotti, Hubert
- Subjects
HANDWRITING recognition (Computer science) ,ARTIFICIAL neural networks ,COMPUTER vision ,SUPERVISED learning ,DEEP learning ,FEATURE extraction - Abstract
Supervised learning techniques require labeled examples that can be time consuming to obtain. In particular, deep learning approaches, where all the feature extraction stages are learned within the artificial neural network, require a large number of labeled examples to train the model. Various data augmentation techniques can be performed to overcome this issue by taking advantage of known variations that have no impact on the label of an example. Typical solutions in computer vision and document analysis and recognition are based on geometric transformations (e.g. shift and rotation) and random elastic deformations of the original training examples. In this paper, we consider Generative Adversarial Networks (GAN), a technique that does not require prior knowledge of the possible variabilities that exist across examples to create novel artificial examples. In the case of a training dataset with a low number of labeled examples, which are described in a high dimensional space, the classifier may generalize poorly. Therefore, we aim at enriching databases of images or signals for improving the classifier performance by designing a GAN for creating artificial images. While adding more images through a GAN can help, the extent to which it will help is unknown, and it may degrade the performance if too many artificial images are added. The approach is tested on four datasets on handwritten digits (Latin, Bangla, Devanagri, and Oriya). The accuracy for each dataset shows that the addition of GAN generated images in the training dataset provides an improvement of the accuracy. However, the results suggest that the addition of too many GAN generated images deteriorates the performance. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
11. Learning Sensor-Specific Spatial-Spectral Features of Hyperspectral Images via Convolutional Neural Networks.
- Author
-
Mei, Shaohui, Ji, Jingyu, Hou, Junhui, Li, Xu, and Du, Qian
- Subjects
ARTIFICIAL neural networks ,HYPERSPECTRAL imaging systems ,IMAGING systems ,GEOGRAPHIC spatial analysis ,SPECTRUM analysis - Abstract
Convolutional neural network (CNN) is well known for its capability of feature learning and has made revolutionary achievements in many applications, such as scene recognition and target detection. In this paper, its capability of feature learning in hyperspectral images is explored by constructing a five-layer CNN for classification (C-CNN). The proposed C-CNN is constructed by including recent advances in deep learning area, such as batch normalization, dropout, and parametric rectified linear unit (PReLU) activation function. In addition, both spatial context and spectral information are elegantly integrated into the C-CNN such that spatial-spectral features are learned for hyperspectral images. A companion feature-learning CNN (FL-CNN) is constructed by extracting fully connected feature layers in this C-CNN. Both supervised and unsupervised modes are designed for the proposed FL-CNN to learn sensor-specific spatial-spectral features. Extensive experimental results on four benchmark data sets from two well-known hyperspectral sensors, namely airborne visible/infrared imaging spectrometer (AVIRIS) and reflective optics system imaging spectrometer (ROSIS) sensors, demonstrate that our proposed C-CNN outperforms the state-of-the-art CNN-based classification methods, and its corresponding FL-CNN is very effective to extract sensor-specific spatial-spectral features for hyperspectral applications under both supervised and unsupervised modes. [ABSTRACT FROM PUBLISHER]
- Published
- 2017
- Full Text
- View/download PDF
12. Motor Imagery Classification via Temporal Attention Cues of Graph Embedded EEG Signals.
- Author
-
Zhang, Dalin, Chen, Kaixuan, Jian, Debao, and Yao, Lina
- Subjects
BRAIN-computer interfaces ,ELECTROENCEPHALOGRAPHY ,MOTOR imagery (Cognition) ,ARTIFICIAL neural networks ,CLASSIFICATION ,SYNCHRONOUS electric motors ,MACHINE learning - Abstract
Motor imagery classification from EEG signals is essential for motor rehabilitation with a Brain-Computer Interface (BCI). Most current works on this issue require a subject-specific adaptation step before applied to a new user. Thus the research of directly extending a pre-trained model to new users is particularly desired and indispensable. As brain dynamics fluctuate considerably across different subjects, it is challenging to design practical hand-crafted features based on prior knowledge. Regarding this gap, this paper proposes a Graph-based Convolutional Recurrent Attention Model (G-CRAM) to explore EEG features across different subjects for motor imagery classification. A graph structure is first developed to represent the positioning information of EEG nodes. Then a convolutional recurrent attention model learns EEG features from both spatial and temporal dimensions and emphasizes on the most distinguishable temporal periods. We evaluate the proposed approach on two benchmark EEG datasets of motor imagery classification on the subject-independent testing. The results show that the G-CRAM achieves superior performance to state-of-the-art methods regarding recognition accuracy and ROC-AUC. Furthermore, model interpretation studies reveal the learning process of different neural network components and demonstrate that the proposed model can extract detailed features efficiently. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
13. Cluster Ensemble Method and Convolution Neural Network Model for Predicting Mental Illness.
- Author
-
M. V., Ananthapadmanabha, A. C., Dhanesh Kumar, S., Sabariraju, M., Eswar, and Mathi, Senthilkumar
- Subjects
ARTIFICIAL neural networks ,MENTAL illness ,MACHINE learning ,EMOTIONS ,PSYCHOMETRICS - Abstract
One in every four individuals has a diagnosable mental disease in a year. Around 20% of children and adolescents have a mental health condition and often ignore it. It is found that 93% of youth use social media to communicate and engage, as it reflects their emotions, moods, and thoughts. As a result, machine learning algorithms may anticipate people's moods and emotions based on their postings and comments. On the other hand, psychometric tests use questions to determine how individuals think, feel, behave, and react. It is necessary to investigate a hybrid approach for identifying people's mental illness by combining social media inputs and psychometric tests, especially during a pandemic. The hybrid approach can combine the results from both the models to reflect on the user's digital & non-digital reactions to certain sensitive situations to determine their mental state. Hence, the present paper aims to develop a web framework that can forecast the emergence of mental illness in the future based on data from social media comments and real-time data from psychometric tests using machine learning algorithms. The proposed work includes the cluster ensemble method for social media posts and a convolution neural network model for psychometric tests. This model predicts mental illness with an accuracy of 87.05 percent. The individual can use this result to take the required precautions by visiting a psychologist. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
14. Visualization and classification of mushroom species with multi-feature fusion of metaheuristics-based convolutional neural network model.
- Author
-
Özbay, Erdal, Özbay, Feyza Altunbey, and Gharehchopogh, Farhad Soleimanian
- Subjects
ARTIFICIAL neural networks ,CONVOLUTIONAL neural networks ,K-nearest neighbor classification ,FEATURE selection ,FEATURE extraction - Abstract
Determining the correct mushroom species with the necessary ecological characteristics is critical to continue mushroom production, which is essential in gastronomy. The mushroom farmers and collectors technique may help identify toxic mushrooms by detecting poisonous mushrooms using images of different mushroom species with distinctive morphological features. However, it can be not easy to distinguish between species. This paper used a dataset of 6714 mushroom images obtained from nine different mushroom species to classify the mushroom species. For a more straightforward comprehension of mushroom images and feature extraction by reanalysis of data sets, data visualization was performed using Grad-CAM, LIME, and Heatmap methods. Residual block-based Convolutional Neural Network (CNN) architectures are trained to automatically classify the concatenated feature map obtained from the Grad-CAM, LIME, and Heatmap methods. After extracting the deep features of the images from each architecture, the Atom Search Optimization (ASO) algorithm has been used to select the most distinctive features. The 6714×9000 size of the concatenated feature map was reduced to 6714×600 using the ASO algorithm. Classification results were evaluated using six different classifiers based on the feature map obtained to determine the mushroom species. The nine classes of mushroom species were classified successfully with 95.45 % accuracy using the proposed model with the ASO algorithm and KNN classifier. The methodology introduces novel visualization techniques for interpreting CNN-based models' decisions in mushroom species classification tasks. Using metaheuristics-based CNN models with multi-feature fusion techniques allows the model to leverage diverse sources of information, potentially enhancing its ability to discriminate between mushroom species and achieve higher classification accuracy than existing methods. This study can advance the mushroom species classification field by introducing new methodologies, improving classification accuracy, providing insights into model interpretability, and facilitating knowledge transfer to related fields. • The proposed model automatically detects mushroom species. • Grad-CAM, LIME, and Heatmap methods based on CNN architectures are used to visualize mushroom images. • The ASO algorithm has been successfully applied to residual block-based CNN models for feature selection. • The hybrid model classification accuracy for different species of mushrooms is 95.45 %. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. Texture Classification of Machined Surfaces Using Image Processing and Machine Learning Techniques.
- Author
-
Patel, Dhiren R., Vakharia, Vinay, and Kiran, Mysore B.
- Subjects
IMAGE processing ,CHARGE coupled devices ,ARTIFICIAL neural networks ,RANDOM forest algorithms ,SURFACE texture ,MACHINE learning ,DIGITAL image processing - Abstract
Copyright of FME Transactions is the property of University of Belgrade, Faculty of Mechanical Engineering and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2019
- Full Text
- View/download PDF
16. Enhanced quantum-based neural network learning and its application to signature verification.
- Author
-
Patel, Om Prakash, Tiwari, Aruna, Chaudhary, Rishabh, Nuthalapati, Sai Vidyaranya, Bharill, Neha, Prasad, Mukesh, Hussain, Farookh Khadeer, and Hussain, Omar Khadeer
- Subjects
ARTIFICIAL neural networks ,QUANTUM computing ,SUPPORT vector machines ,BACK propagation ,MACHINE learning - Abstract
In this paper, an enhanced quantum-based neural network learning algorithm (EQNN-S) which constructs a neural network architecture using the quantum computing concept is proposed for signature verification. The quantum computing concept is used to decide the connection weights and threshold of neurons. A boundary threshold parameter is introduced to optimally determine the neuron threshold. This parameter uses min, max function to decide threshold, which assists efficient learning. A manually prepared signature dataset is used to test the performance of the proposed algorithm. To uniquely identify the signature, several novel features are selected such as the number of loops present in the signature, the boundary calculation, the number of vertical and horizontal dense patches, and the angle measurement. A total of 45 features are extracted from each signature. The performance of the proposed algorithm is evaluated by rigorous training and testing with these signatures using partitions of 60–40 and 70–30%, and a tenfold cross-validation. To compare the results derived from the proposed quantum neural network, the same dataset is tested on support vector machine, multilayer perceptron, back propagation neural network, and Naive Bayes. The performance of the proposed algorithm is found better when compared with the above methods, and the results verify the effectiveness of the proposed algorithm. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
17. Dense Semantic Labeling of Subdecimeter Resolution Images With Convolutional Neural Networks.
- Author
-
Volpi, Michele and Tuia, Devis
- Subjects
LABELING theory ,ARTIFICIAL neural networks ,DEEP learning ,IMAGE processing ,DESCRIPTOR systems - Abstract
Semantic labeling (or pixel-level land-cover classification) in ultrahigh-resolution imagery (<10 cm) requires statistical models able to learn high-level concepts from spatial data, with large appearance variations. Convolutional neural networks (CNNs) achieve this goal by learning discriminatively a hierarchy of representations of increasing abstraction. In this paper, we present a CNN-based system relying on a downsample-then-upsample architecture. Specifically, it first learns a rough spatial map of high-level representations by means of convolutions and then learns to upsample them back to the original resolution by deconvolutions. By doing so, the CNN learns to densely label every pixel at the original resolution of the image. This results in many advantages, including: 1) the state-of-the-art numerical accuracy; 2) the improved geometric accuracy of predictions; and 3) high efficiency at inference time. We test the proposed system on the Vaihingen and Potsdam subdecimeter resolution data sets, involving the semantic labeling of aerial images of 9- and 5-cm resolution, respectively. These data sets are composed by many large and fully annotated tiles, allowing an unbiased evaluation of models making use of spatial information. We do so by comparing two standard CNN architectures with the proposed one: standard patch classification, prediction of local label patches by employing only convolutions, and full patch labeling by employing deconvolutions. All the systems compare favorably or outperform a state-of-the-art baseline relying on superpixels and powerful appearance descriptors. The proposed full patch labeling CNN outperforms these models by a large margin, also showing a very appealing inference time. [ABSTRACT FROM PUBLISHER]
- Published
- 2017
- Full Text
- View/download PDF
18. Classification of Non-Conventional Ships Using a Neural Bag-Of-Words Mechanism.
- Author
-
Polap, Dawid and Wlodarczyk-Sielicka, Marta
- Subjects
ARTIFICIAL neural networks ,AUTOMATIC identification ,CAMCORDERS ,SYSTEM identification ,CLASSIFICATION ,DYNAMIC positioning systems ,RADARSAT satellites - Abstract
The existing methods for monitoring vessels are mainly based on radar and automatic identification systems. Additional sensors that are used include video cameras. Such systems feature cameras that capture images and software that analyzes the selected video frames. Methods for the classification of non-conventional vessels are not widely known. These methods, based on image samples, can be considered difficult. This paper is intended to show an alternative way to approach image classification problems; not by classifying the entire input data, but smaller parts. The described solution is based on splitting the image of a ship into smaller parts and classifying them into vectors that can be identified as features using a convolutional neural network (CNN). This idea is a representation of a bag-of-words mechanism, where created feature vectors might be called words, and by using them a solution can assign images a specific class. As part of the experiment, the authors performed two tests. In the first, two classes were analyzed and the results obtained show great potential for application. In the second, the authors used much larger sets of images belonging to five vessel types. The proposed method indeed improved the results of classic approaches by 5%. The paper shows an alternative approach for the classification of non-conventional vessels to increase accuracy. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
19. Multichannel Classifier for Recognizing Acoustic Impacts Recorded with a phi-OTDR.
- Author
-
Barantsov, Ivan Alekseevich, Pnev, Alexey Borisovich, Koshelev, Kirill Igorevich, Garin, Egor Olegovich, Pozhar, Nickolai Olegovich, and Khan, Roman Igorevich
- Subjects
MACHINE learning ,ARTIFICIAL neural networks ,FEATURE extraction - Abstract
The purpose of this work is to increase the security of the perimeter of an area from unauthorized intrusions by creating an improved algorithm for classifying acoustic impacts recorded with a sensor system based on a phase-sensitive optical time reflectometer (phi-OTDR). The algorithm includes machine learning, so a dataset consisting of two classes was assembled. The dataset consists of two classes. The first class is the data of the steps, and the second class is other non-stepping influences (engine noise, a passing car, a passing cyclist, etc.). As an intrusion signal, a human walking signal is analyzed and recorded in frames of 5 s, which passed the threshold condition. Since, in most cases, the intruder moves on foot to overcome the perimeter, the analysis of the acoustic effects generated during the step will increase the efficiency of the perimeter detection tools. When walking quietly, step signals can be quite weak, and background signals can contain high energy and visually resemble the signals you are looking for. Therefore, an algorithm was created that processes space–time diagrams developed in real time, which are grayscale images. At the same time, during the processing of one image, two more images are calculated, which are the result of processing the denoised autoencoder and the created mathematical model of the adaptive correlation. Then, the three obtained images are fed to the input of the created three-channel neural network classifier, which includes convolutional layers for the automatic extraction of spatial features. The probability of correctly detecting steps is 98.3% and that of background actions is 97.93%. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
20. Coconut trees classification based on height, inclination, and orientation using MIN-SVM algorithm.
- Author
-
Megalingam, Rajesh Kannan, Kuttankulangara Manoharan, Sakthiprasad, Babu, Dasari Hema Teja Anirudh, Sriram, Ghali, Lokesh, Karanam, and Kariparambil Sudheesh, Sankardas
- Subjects
COCONUT palm ,DEEP learning ,MACHINE learning ,CONVOLUTIONAL neural networks ,ARTIFICIAL neural networks ,SUPPORT vector machines - Abstract
A computerized coconut tree detection system can help dendrologists and laypersons in identifying coconut trees based on three morphological parameters including height, inclination, and orientation. These three parameters help to determine the health and the nature of growth of coconut trees which influences the design and use of robots for harvesting coconuts. Deep learning is a powerful tool used for feature extraction as it is better in extracting deeper details (features) in an image. In this research work, a new Modified Inception Net based Hyper Tuning Support Vector Machine classification method named MIN-SVM is proposed for coconut tree classification based on three morphological parameters including height, inclination and orientation. The features from the pre-processed coconut tree images were extracted using four distinct Convolutional Neural Network models including Visual Geometry Group, Inception Net, ResNet, and MIN-SVM. These extracted features were then classified using a Machine Learning model named Support Vector Machine (SVM). The MIN-SVM have achieved a remarkable accuracy of 95.35 percent as contrasted to Visual Geometry Group (91.90%), Inception Net (81.66%), and ResNet (71.95%). The features extracted from Modified Inception Net fitted good with SVM classifier. Experimental results show that MIN-SVM can be powerful computerized automated system to identify coconut trees based on height, inclination, and orientation. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
21. Classification of brain hemorrhage computed tomography images using OzNet hybrid algorithm.
- Author
-
Ozaltin, Oznur, Coskun, Orhan, Yeniay, Ozgur, and Subasi, Abdulhamit
- Subjects
INTRACRANIAL hemorrhage ,FISHER discriminant analysis ,ARTIFICIAL neural networks ,CONVOLUTIONAL neural networks ,NAIVE Bayes classification ,FEATURE extraction - Abstract
Classification of brain hemorrhage computed tomography (CT) images provides a better diagnostic implementation for emergency patients. Attentively, each brain CT image must be examined by doctors. This situation is time‐consuming, exhausting, and sometimes leads to making errors. Hence, we aim to find the best algorithm owing to a requirement for automatic classification of CT images to detect brain hemorrhage. In this study, we developed OzNet hybrid algorithm, which is a novel convolution neural networks (CNN) algorithm. Although OzNet achieves high classification performance, we combine it with Neighborhood Component Analysis (NCA) and many classifiers: Artificial neural networks (ANN), Adaboost, Bagging, Decision Tree, K‐Nearest Neighbor (K‐NN), Linear Discriminant Analysis (LDA), Naïve Bayes and Support Vector Machines (SVM). In addition, Oznet is utilized for feature extraction, where 4096 features are extracted from the fully connected layer. These features are reduced to have significant and informative features with minimum loss by NCA. Eventually, we use these classifiers to classify these significant features. Finally, experimental results display that OzNet‐NCA‐ANN excellent classifier model and achieves 100% accuracy with created Dataset 2 from Brain Hemorrhage CT images. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
22. Advanced artificial neural network classification for detecting preterm births using EHG records.
- Author
-
Fergus, Paul, Idowu, Ibrahim, Hussain, Abir, and Dobbins, Chelsea
- Subjects
- *
PREMATURE labor , *DELIVERY (Obstetrics) , *ARTIFICIAL neural networks , *ELECTROHYSTEROGRAPHY , *SIGNAL processing , *MACHINE learning - Abstract
Globally, the rate of preterm births are increasing, thus resulting in significant health, development and economic problems. Current methods for the early detection of such births are inadequate. Nevertheless, there has been some evidence that the analysis of uterine electrical signals, collected from the abdominal surface, could provide an independent and easier way to diagnose true labour and detect the onset of preterm delivery. Using advanced machine learning algorithms, in conjunction with Electrohysterography signal processing, numerous studies have focused on detecting true labour several days prior to the event. However, in this paper, the Electrohysterography signals have been used to detect preterm births. This has been achieved using an open dataset, which contains 262 records for women who delivered at term and 38 who delivered prematurely. Several new features from Electromyography studies have been utilised, as well as feature-ranking techniques to determine their discriminative capabilities in detecting term and preterm records. Seven different artificial neural networks were then used to identify these records. The results illustrate that the combination of the Levenberg–Marquardt trained Feed-Forward Neural Network, Radial Basis Function Neural Network and the Random Neural Network classifiers performed the best, with 91% for sensitivity, 84% for specificity, 94% for the area under the curve and 12% for the mean error rate. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
23. Extracting classification rules from modified fuzzy min–max neural network for data with mixed attributes.
- Author
-
Shinde, Swati and Kulkarni, Uday
- Subjects
FEATURE extraction ,FUZZY systems ,ARTIFICIAL neural networks ,SUPERVISED learning ,MACHINE learning - Abstract
This paper proposes the modified fuzzy min–max neural network (MFMMN) classification model to perform the supervised classification of data. The basic fuzzy min–max neural network (FMMN) can only be applied to the continuous attribute values and cannot handle the discrete values. Also justification of the classification results given by FMMN required to be obtained to make it more applicable to real world applications. These both issues are solved in the proposed MFMMN. In the MFMMN, each hyperbox have min–max values defined in terms of continuous attributes and a set of binary strings defined for discrete attributes. Bitwise ‘ and ’ and ‘ or ’ operators are used to update the discrete values associated with each hyperbox. The trained network is pruned to remove the less useful hyperboxes based on their confidence factor. The proposed model is applied to nine different datasets taken from the University of California, Irvine (UCI) machine learning repository. Finally the case study of a real time weather data is evaluated using MFMMN. The experimental results show that the proposed model has given very good accuracy. In addition to accuracy, the number of hyperboxes obtained after pruning are very less which lead to less number of concise rules and reduced computational complexity. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
24. Stress Detection System for Working Pregnant Women Using an Improved Deep Recurrent Neural Network.
- Author
-
Sharma, Sameer Dev, Sharma, Sonal, Singh, Rajesh, Gehlot, Anita, Priyadarshi, Neeraj, and Twala, Bhekisipho
- Subjects
ARTIFICIAL neural networks ,PREGNANT women ,RECURRENT neural networks ,WOMEN employees ,FEATURE selection ,MACHINE learning ,RECURRENT miscarriage - Abstract
Stress is a concerning issue in today's world. Stress in pregnancy harms both the development of children and the health of pregnant women. As a result, assessing the stress levels of working pregnant women is crucial to aid them in developing and growing professionally and personally. In the past, many machine-learning (ML) and deep-learning (DL) algorithms have been made to predict the stress of women. It does, however, have some problems, such as a more complicated design, a high chance of misclassification, a high chance of making mistakes, and less efficiency. With these considerations in mind, our article will use a deep-learning model known as the deep recurrent neural network (DRNN) to predict the stress levels of working pregnant women. Dataset preparation, feature extraction, optimal feature selection, and classification with DRNNs are all included in this framework. Duplicate attributes are removed, and missing values are filled in during the preprocessing of the dataset. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
25. Multimodal biometric authentication based on deep fusion of electrocardiogram (ECG) and finger vein.
- Author
-
El-Rahiem, Basma Abd, El-Samie, Fathi E. Abd, and Amin, Mohamed
- Subjects
FINGERS ,MULTIMODAL user interfaces ,BIOMETRIC identification ,ARTIFICIAL neural networks ,FEATURE extraction ,CONVOLUTIONAL neural networks ,MACHINE learning - Abstract
Biometric identification depends on the statistical analysis of the unique physical and behavioral characteristics of individuals. However, a unimodal biometric system is susceptible to different attacks such as spoof attacks. To overcome these limitations, we propose a multimodal biometric authentication system based on deep fusion of electrocardiogram (ECG) and finger vein. The proposed system has three main components, which are biometric pre-processing, deep feature extraction, and authentication. During the pre-processing, normalization and filtering techniques are adapted for each biometric. In the feature extraction process, the features are extracted using a proposed deep Convolutional Neural Network (CNN) model. Then, the authentication process is performed on the extracted features using five well-known machine learning classifiers: Support Vector Machine (SVM), K-Nearest Neighbors (KNNs), Random Forest (RF), Naive Bayes (NB), and Artificial Neural Network (ANN). In addition, to represent the deep features in a low-dimensional feature space and speed up the authentication task, we adopt Multi-Canonical Correlation Analysis (MCCA). We combine the two biometric systems based on ECG and finger vein into a single multimodal biometric system using feature and score fusion. The performance of the proposed system is tested on two finger vein (TW finger vein and VeinPolyU finger vein) databases and two ECG (MWM-HIT and ECG-ID) databases. Experimental results reveal improvement in terms of authentication performance with Equal Error Rates (EERs) of 0.12% and 1.40% using feature fusion and score fusion, respectively. Furthermore, the authentication with the proposed multimodal system using MCCA feature fusion with a KNN classifier shows an increase of accuracy by an average of 10% compared with those of other machine learning algorithms. Therefore, the proposed biometric system is effective in performing secure authentication and assisting the stakeholders in making accurate authentication of users. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
26. Supervised Learning for Nonsequential Data: A Canonical Polyadic Decomposition Approach.
- Author
-
Haliassos, Alexandros, Konstantinidis, Kriton, and Mandic, Danilo P.
- Subjects
SUPERVISED learning ,MACHINE learning ,ARTIFICIAL neural networks - Abstract
Efficient modeling of feature interactions underpins supervised learning for nonsequential tasks, characterized by a lack of inherent ordering of features (variables). The brute force approach of learning a parameter for each interaction of every order comes at an exponential computational and memory cost (curse of dimensionality). To alleviate this issue, it has been proposed to implicitly represent the model parameters as a tensor, the order of which is equal to the number of features; for efficiency, it can be further factorized into a compact tensor train (TT) format. However, both TT and other tensor networks (TNs), such as tensor ring and hierarchical Tucker, are sensitive to the ordering of their indices (and hence to the features). To establish the desired invariance to feature ordering, we propose to represent the weight tensor through the canonical polyadic (CP) decomposition (CPD) and introduce the associated inference and learning algorithms, including suitable regularization and initialization schemes. It is demonstrated that the proposed CP-based predictor significantly outperforms other TN-based predictors on sparse data while exhibiting comparable performance on dense nonsequential tasks. Furthermore, for enhanced expressiveness, we generalize the framework to allow feature mapping to arbitrarily high-dimensional feature vectors. In conjunction with feature vector normalization, this is shown to yield dramatic improvements in performance for dense nonsequential tasks, matching models such as fully connected neural networks. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
27. Deep Learning-Based Forecasting Approach in Smart Grids With Microclustering and Bidirectional LSTM Network.
- Author
-
Jahangir, Hamidreza, Tayarani, Hanif, Gougheri, Saleh Sadeghi, Golkar, Masoud Aliakbar, Ahmadian, Ali, and Elkamel, Ali
- Subjects
SMART power grids ,LOAD forecasting (Electric power systems) ,DEEP learning ,RECURRENT neural networks ,ARTIFICIAL neural networks ,FORECASTING ,RENEWABLE energy sources - Abstract
Uncertainty modeling of renewable energy sources, load demand, electricity price, etc. create a high volume of data in smart grids. Accordingly, in this article, a precise forecasting method based on a deep learning concept with microclustering (MC) task is presented. The MC method is structured based on hybrid unsupervised and supervised clustering tasks by K-means and Gaussian support vector machine, respectively. In the proposed method, the input data sequence is clustered by the MC task, and then the forecasting process is employed. By applying the MC, input data in each hour are categorized into different groups, and a distinctive forecasting unit is allocated to each one. In this way, more clusters and forecasting networks are earmarked for the hours with higher fluctuation rates. The bi-directional long short-term memory (B-LSTM), which is one of the newest recurrent artificial neural networks, is proposed as the forecasting unit. The B-LSTM has bidirectional memory—feedforward and feedback loops—that helps us to investigate both previous and future hidden layers data. The optimal number of clusters in each hour is determined based on the Davies–Bouldin index. To evaluate the performance of the proposed method, in this study, three forecasting tasks including the wind speed, load demand, and electricity price are studied in different periods using the Ontario province, Canada, data set. The results are compared with other benchmarking methods to verify the robustness and effectiveness of the proposed method. In fact, the proposed method, which is equipped with the MC technique and B-LSTM networks, significantly promotes the forecasting results, especially in spike points. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
28. Multiscale Context-Aware Ensemble Deep KELM for Efficient Hyperspectral Image Classification.
- Author
-
Xi, Bobo, Li, Jiaojiao, Li, Yunsong, Song, Rui, Sun, Weiwei, and Du, Qian
- Subjects
MACHINE learning ,CLASSIFICATION ,ARTIFICIAL neural networks ,SAMPLE size (Statistics) ,FEATURE extraction - Abstract
Recently, multiscale spatial features have been widely utilized to improve the hyperspectral image (HSI) classification performance. However, fixed-size neighborhood involving the contextual information probably leads to misclassifications, especially for the boundary pixels. Additionally, it has been demonstrated that deep neural network (DNN) is practical to extract representative features for the classification tasks. Nevertheless, under the condition of high dimensionality versus small sample sizes, DNN tends to be over-fitting and it is generally time-consuming due to the deep-level feature learning process. To alleviate the aforementioned issues, we propose a multiscale context-aware ensemble deep kernel extreme learning machine (MSC-EDKELM) for efficient HSI classification. First, the scene of the HSI data set is over-segmented in multiscale via using the adaptive superpixel segmentation technique. Second, superpixel pattern (SP) and attentional neighboring superpixel pattern (ANSP) are generated by leveraging the superpixel maps, which can automatically comprise local and global contextual information, respectively. Afterward, an ensemble deep kernel extreme learning machine (EDKELM) is presented to investigate the deep-level characteristics in the SP and ANSP. Finally, the category of each pixel is accurately determined by the decision fusion and weighted output layer fusion strategy. Experimental results on four real-world HSI data sets demonstrate that the proposed frameworks outperform some classic and state-of-the-art methods with high computational efficiency, which can be employed to serve real-time applications. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
29. More Diverse Means Better: Multimodal Deep Learning Meets Remote-Sensing Imagery Classification.
- Author
-
Hong, Danfeng, Gao, Lianru, Yokoya, Naoto, Yao, Jing, Chanussot, Jocelyn, Du, Qian, and Zhang, Bing
- Subjects
REMOTE sensing ,DEEP learning ,ARTIFICIAL neural networks ,CONVOLUTIONAL neural networks ,MULTIMODAL user interfaces ,SURFACE of the earth - Abstract
Classification and identification of the materials lying over or beneath the earth’s surface have long been a fundamental but challenging research topic in geoscience and remote sensing (RS), and have garnered a growing concern owing to the recent advancements of deep learning techniques. Although deep networks have been successfully applied in single-modality-dominated classification tasks, yet their performance inevitably meets the bottleneck in complex scenes that need to be finely classified, due to the limitation of information diversity. In this work, we provide a baseline solution to the aforementioned difficulty by developing a general multimodal deep learning (MDL) framework. In particular, we also investigate a special case of multi-modality learning (MML)—cross-modality learning (CML) that exists widely in RS image classification applications. By focusing on “what,” “where,” and “how” to fuse, we show different fusion strategies as well as how to train deep networks and build the network architecture. Specifically, five fusion architectures are introduced and developed, further being unified in our MDL framework. More significantly, our framework is not only limited to pixel-wise classification tasks but also applicable to spatial information modeling with convolutional neural networks (CNNs). To validate the effectiveness and superiority of the MDL framework, extensive experiments related to the settings of MML and CML are conducted on two different multimodal RS data sets. Furthermore, the codes and data sets will be available at https://github.com/danfenghong/IEEE%5fTGRS%5fMDL-RS , contributing to the RS community. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
30. Transfer learning features for predicting aesthetics through a novel hybrid machine learning method.
- Author
-
Carballal, Adrian, Fernandez-Lozano, Carlos, Heras, Jonathan, and Romero, Juan
- Subjects
BLENDED learning ,ARTIFICIAL neural networks ,SIGNAL convolution ,MACHINE learning ,AESTHETICS ,SUPERVISED learning ,GENETIC correlations - Abstract
The automatic assessment of the aesthetic value of an image is a task with many applications but really complex and challenging, due to the subjective component of the aesthetics for humans. The computational systems that carry out this task are usually composed of a set of ad hoc metrics proposed by the researchers and a machine learning system. We propose a new approach that fully automates the metrics creation process, its filtering and adjustment without human subjectivity. Thus, it does not depend on the authors' human aesthetic intuitions. Our proposal is therefore based on the integration of two machine learning algorithms: CNN, which works as a feature extractor, and Correlation by Genetic Search (CGS)—a novel regression method, working as a supervised learning method. CGS is based on the creation of an adjusted linear regression model using Pearson's correlation as a measure of performance in an evolutionary process. Experiments were conducted on a very well-known aesthetics database called "Photo.net" with more than a million images from over 400,000 users. The comparison of results with other approaches using the same dataset demonstrates that the fusion of CNN transfer learning features with this specific machine learning method has achieved robust and significantly better results than other state-of-the-art methods and hybrid approaches in terms of AUROC (0.93), accuracy (0.93) and Pearson's correlation value (0.94). [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
31. CLR-based deep convolutional spiking neural network with validation based stopping for time series classification.
- Author
-
Gautam, Anjali and Singh, Vrijendra
- Subjects
ARTIFICIAL neural networks ,TIME series analysis ,BIOLOGICALLY inspired computing ,MACHINE learning ,FEATURE extraction ,OPTIMAL stopping (Mathematical statistics) ,CLASSIFICATION - Abstract
Huge amount of time series data over several domains such as engineering, biomedical and finance, demands the development of efficient methods for the problem of time series classification. The classification of univariate and multivariate time series together using a single architecture is a very difficult task. In this work, a bio-inspired convolutional spiking neural network (CSNN) is proposed for both univariate and multivariate time series. For this, first we develop a simple transformation to convert raw time series sequences into matrices. The CSNN is a three staged framework which include convolutional feature extraction, spike encoding using soft leaky integrate and fire (Soft-LIf) and classification. As spikes generated are differentiable, thus the learning algorithm for CSNN uses error-backpropagation with cyclical learning rates (CLR) and RMSprop optimizer. Additionally, validation based stopping rules are employed to overcome the overfitting which also provides a set of parameters associated with low validation set loss. Thereafter, to demonstrate the accuracy and robustness of proposed CSNN model, we have used University of California (UCR) univariate as well as University of East Anglia (UEA) multivariate datasets to perform the experiments. Moreover, we conduct comparative empirical performance evaluation with benchmark methods and also with recent deep networks proposed for time series classification. Our results reveal that proposed CSNN advances the baseline methods by achieving higher performance accuracy for both univariate and multivariate datasets. It is shown that the CLR with RMSprop optimizer is able to achieve faster convergence, however CLR and adaptive rates are considered competitive to each other. In addition, we also address the optimal model selection and study the effects of different factors on the performance of proposed CSNN. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
32. Dictionary Learning for Adaptive GPR Landmine Classification.
- Author
-
Giovanneschi, Fabio, Mishra, Kumar Vijay, Gonzalez-Huici, Maria Antonia, Eldar, Yonina C., and Ender, Joachim H. G.
- Subjects
ARTIFICIAL neural networks ,SUPPORT vector machines ,CLASSIFICATION algorithms ,GROUND penetrating radar ,SINGULAR value decomposition ,ONLINE algorithms ,CLASSIFICATION ,FEATURE extraction - Abstract
Ground-penetrating radar (GPR) target detection and classification is a challenging task. Here, we consider online dictionary learning (DL) methods to obtain sparse representations (SR) of the GPR data to enhance feature extraction for target classification via support vector machines. Online methods are preferred because traditional batch DL like K-times singular value decomposition (K-SVD) is not scalable to high-dimensional training sets and infeasible for real-time operation. We also develop Drop-Off MINi-batch Online Dictionary Learning (DOMINODL), which exploits the fact that a lot of the training data may be correlated. The DOMINODL algorithm iteratively considers elements of the training set in small batches and drops off samples which become less relevant. For the case of abandoned anti-personnel landmines classification, we compare the performance of K-SVD with three online algorithms: classical online dictionary learning (ODL), its correlation-based variant, and DOMINODL. Our experiments with real data from L-band GPR show that online DL methods reduce learning time by 36%–93% and increase mine detection by 4%–28% over K-SVD. Our DOMINODL is the fastest and retains similar classification performance as the other two online DL approaches. We use a Kolmogorov–Smirnoff test distance and the Dvoretzky–Kiefer–Wolfowitz inequality for the selection of DL input parameters leading to enhanced classification results. To further compare with the state-of-the-art classification approaches, we evaluate a convolutional neural network (CNN) classifier, which performs worse than the proposed approach. Moreover, when the acquired samples are randomly reduced by 25%, 50%, and 75%, sparse decomposition-based classification with DL remains robust while the CNN accuracy is drastically compromised. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
33. Two-Stream Deep Architecture for Hyperspectral Image Classification.
- Author
-
Hao, Siyuan, Wang, Wei, Ye, Yuanxin, Nie, Tingyuan, and Bruzzone, Lorenzo
- Subjects
HYPERSPECTRAL imaging systems ,CLASSIFICATION ,ARTIFICIAL neural networks ,DEEP learning ,REMOTE sensing ,PIXELS - Abstract
Most traditional approaches classify hyperspectral image (HSI) pixels relying only on the spectral values of the input channels. However, the spatial context around a pixel is also very important and can enhance the classification performance. In order to effectively exploit and fuse both the spatial context and spectral structure, we propose a novel two-stream deep architecture for HSI classification. The proposed method consists of a two-stream architecture and a novel fusion scheme. In the two-stream architecture, one stream employs the stacked denoising autoencoder to encode the spectral values of each input pixel, and the other stream takes as input the corresponding image patch and deep convolutional neural networks are employed to process the image patch. In the fusion scheme, the prediction probabilities from two streams are fused by adaptive class-specific weights, which can be obtained by a fully connected layer. Finally, a weight regularizer is added to the loss function to alleviate the overfitting of the class-specific fusion weights. Experimental results on real HSIs demonstrate that the proposed two-stream deep architecture can achieve competitive performance compared with the state-of-the-art methods. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
34. Electromyogram-Based Classification of Hand and Finger Gestures Using Artificial Neural Networks.
- Author
-
Lee, Kyung Hyun, Min, Ji Young, and Byun, Sangwon
- Subjects
ARTIFICIAL neural networks ,FINGERS ,GESTURE ,FEATURE extraction ,ONE-way analysis of variance ,SUPPORT vector machines - Abstract
Electromyogram (EMG) signals have been increasingly used for hand and finger gesture recognition. However, most studies have focused on the wrist and whole-hand gestures and not on individual finger (IF) gestures, which are considered more challenging. In this study, we develop EMG-based hand/finger gesture classifiers based on fixed electrode placement using machine learning methods. Ten healthy subjects performed ten hand/finger gestures, including seven IF gestures. EMG signals were measured from three channels, and six time-domain (TD) features were extracted from each channel. A total of 18 features was used to build personalized classifiers for ten gestures with an artificial neural network (ANN), a support vector machine (SVM), a random forest (RF), and a logistic regression (LR). The ANN, SVM, RF, and LR achieved mean accuracies of 0.940, 0.876, 0.831, and 0.539, respectively. One-way analyses of variance and F-tests showed that the ANN achieved the highest mean accuracy and the lowest inter-subject variance in the accuracy, respectively, suggesting that it was the least affected by individual variability in EMG signals. Using only TD features, we achieved a higher ratio of gestures to channels than other similar studies, suggesting that the proposed method can improve the system usability and reduce the computational burden. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.