10 results on '"Zhang, Chongsheng"'
Search Results
2. A parameter-free label propagation algorithm for person identification in stereo videos
- Author
-
Zhang, Chongsheng, Bi, Jingjun, Liu, Changchang, and Chen, Ke
- Published
- 2016
- Full Text
- View/download PDF
3. Multi-Imbalance: An open-source software for multi-class imbalance learning.
- Author
-
Zhang, Chongsheng, Bi, Jingjun, Xu, Shixin, Ramentol, Enislay, Fan, Gaojuan, Qiao, Baojun, and Fujita, Hamido
- Subjects
- *
OPEN source software , *SOURCE code , *COMPUTER software , *SOFTWARE development tools , *MACHINE learning - Abstract
Abstract Imbalance classification is one of the most challenging research problems in machine learning. Techniques for two-class imbalance classification are relatively mature nowadays, yet multi-class imbalance learning is still an open problem. Moreover, the community lacks a suitable software tool that can integrate the major works in the field. In this paper, we present Multi-Imbalance, an open source software package for multi-class imbalanced data classification. It provides users with seven different categories of multi-class imbalance learning algorithms, including the latest advances in the field. The source codes and documentations for Multi-Imbalance are publicly available at https://github.com/chongshengzhang/Multi_Imbalance. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
4. An empirical evaluation of high utility itemset mining algorithms.
- Author
-
Zhang, Chongsheng, Almpanidis, George, Wang, Wanwan, and Liu, Changchang
- Subjects
- *
DATA mining , *STOCK exchanges , *DATA analysis , *BIOINFORMATICS , *RECOMMENDER systems - Abstract
High utility itemset mining (HUIM) has emerged as an important research topic in data mining, with applications to retail-market data analysis, stock market prediction, and recommender systems, etc. However, there are very few empirical studies that systematically compare the performance of state-of-the-art HUIM algorithms. In this paper, we present an experimental evaluation on 10 major HUIM algorithms, using 9 real world and 27 synthetic datasets to evaluate their performance. Our experiments show that EFIM and d2HUP are generally the top two performers in running time, while EFIM also consumes the least memory in most cases. In order to compare these two algorithms in depth, we use another 45 synthetic datasets with varying parameters so as to study the influence of the related parameters, in particular the number of transactions, the number of distinct items and average transaction length, on the running time and memory consumption of EFIM and d2HUP. In this work, we demonstrate that, d2HUP is more efficient than EFIM under low minimum utility values and with large sparse datasets, in terms of running time; although EFIM is the fastest in dense real datasets, it is among the slowest algorithms in sparse datasets. We suggest that, when a dataset is very sparse or the average transaction length is large, and running time is favoured over memory consumption, d2HUP should be chosen. Finally, we compare d2HUP and EFIM with two newest algorithms, mHUIMiner and ULB-Miner, and find these two algorithms have moderate performance. This work has reference value for researchers and practitioners when choosing the most appropriate HUIM algorithm for their specific applications. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
5. An up-to-date comparison of state-of-the-art classification algorithms.
- Author
-
Zhang, Chongsheng, Liu, Changchang, Zhang, Xiangliang, and Almpanidis, George
- Subjects
- *
CENTRAL processing units , *CLASSIFICATION algorithms , *COMPARATIVE studies , *DEEP learning , *RANDOM forest algorithms , *SUPPORT vector machines - Abstract
Current benchmark reports of classification algorithms generally concern common classifiers and their variants but do not include many algorithms that have been introduced in recent years. Moreover, important properties such as the dependency on number of classes and features and CPU running time are typically not examined. In this paper, we carry out a comparative empirical study on both established classifiers and more recently proposed ones on 71 data sets originating from different domains, publicly available at UCI and KEEL repositories. The list of 11 algorithms studied includes Extreme Learning Machine (ELM), Sparse Representation based Classification (SRC), and Deep Learning (DL), which have not been thoroughly investigated in existing comparative studies. It is found that Stochastic Gradient Boosting Trees (GBDT) matches or exceeds the prediction performance of Support Vector Machines (SVM) and Random Forests (RF), while being the fastest algorithm in terms of prediction efficiency. ELM also yields good accuracy results, ranking in the top-5 , alongside GBDT, RF, SVM, and C4.5 but this performance varies widely across all data sets. Unsurprisingly, top accuracy performers have average or slow training time efficiency. DL is the worst performer in terms of accuracy but second fastest in prediction efficiency. SRC shows good accuracy performance but it is the slowest classifier in both training and testing. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
6. The anti-bouncing data stream model for web usage streams with intralinkings.
- Author
-
Zhang, Chongsheng, Masseglia, Florent, and Lechevallier, Yves
- Subjects
- *
WEB analytics , *STREAMING technology , *DATA mining , *ELECTRONIC records , *DATA analysis - Abstract
Abstract: Web usage mining is a significant research area with applications in various fields. However, Web usage data is usually considered streaming, due to its high volumes and rates. Because of these characteristics, we only have access, at any point in time, to a small fraction of the stream. When the data is observed through such a limited window, it is challenging to give a reliable description of the recent usage data. We show that data intralinkings, i.e. a usage record (event) may be associated with other records (events) in the same dataset, are common for Web usage streams. Therefore, in order to have a more authentic grasp of Web usage behaviors, the corresponding data stream models for Web usage streams should be able to process such intralinkings. We study the important consequences of the constraints and intralinkings, through the “bounce rate” problem and the clustering of usage streams. Then we propose the user-centric ABS (the Anti-Bouncing Stream) model which combines the advantages of previous models but avoids their drawbacks. First, ABS is the first data stream model that is able to seize the intralinkings between the Web usage records. It is also the first user-centric data stream model that can associate the usage records for the users in the Web usage streams. Second, owing to its simple but effective management principle, the data in ABS is available at any time for analysis. Under the same resource constraints as existing models in the literature, ABS can better model the recent data. Third, ABS can better measure the bounce rates for Web usage streams. We demonstrate its superiority through a theoretical study and experiments on two real-world data sets. [Copyright &y& Elsevier]
- Published
- 2014
- Full Text
- View/download PDF
7. A unified deep semi-supervised graph learning scheme based on nodes re-weighting and manifold regularization.
- Author
-
Dornaika, Fadi, Bi, Jingjun, and Zhang, Chongsheng
- Subjects
- *
SUPERVISED learning , *LEARNING problems - Abstract
In recent years, semi-supervised learning on graphs has gained importance in many fields and applications. The goal is to use both partially labeled data (labeled examples) and a large amount of unlabeled data to build more effective predictive models. Deep Graph Neural Networks (GNNs) are very useful in both unsupervised and semi-supervised learning problems. As a special class of GNNs, Graph Convolutional Networks (GCNs) aim to obtain data representation through graph-based node smoothing and layer-wise neural network transformations. However, GCNs have some weaknesses when applied to semi-supervised graph learning: (1) it ignores the manifold structure implicitly encoded by the graph; (2) it uses a fixed neighborhood graph and focuses only on the convolution of a graph, but pays little attention to graph construction; (3) it rarely considers the problem of topological imbalance. To overcome the above shortcomings, in this paper, we propose a novel semi-supervised learning method called Re-weight Nodes and Graph Learning Convolutional Network with Manifold Regularization (ReNode-GLCNMR). Our proposed method simultaneously integrates graph learning and graph convolution into a unified network architecture, which also enforces label smoothing through an unsupervised loss term. At the same time, it addresses the problem of imbalance in graph topology by adaptively reweighting the influence of labeled nodes based on their distances to the class boundaries. Experiments on 8 benchmark datasets show that ReNode-GLCNMR significantly outperforms the state-of-the-art semi-supervised GNN methods. 1 1 The code is available at https://github.com/BiJingjun/ReNode-GLCNMR. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
8. Open-set long-tailed recognition via orthogonal prototype learning and false rejection correction.
- Author
-
Deng, Binquan, Kamel, Aouaidjia, and Zhang, Chongsheng
- Abstract
Learning from data with long-tailed and open-ended distributions is highly challenging. In this work, we propose OLPR , which is a new dual-stream O pen-set L ong-tailed recognition framework based on orthogonal P rototype learning and false R ejection correction. It consists of a Probabilistic Prediction Learning (PPL) branch and a Distance Metric Learning (DML) branch. The former is used to generate prediction probability for image classification. The latter learns orthogonal prototypes for each class by computing three distance losses, which are the orthogonal prototype loss among all the prototypes, the balanced Softmin distance based cross-entropy loss between each prototype and its corresponding input sample, and the adversarial loss for making the open-set space more compact. Furthermore, for open-set learning, instead of merely relying on binary decisions, we propose an Iterative Clustering Module (ICM) to categorize similar open-set samples and correct the false rejected closed-set samples simultaneously. If a sample is detected as a false rejection, i.e., a sample of the known classes is incorrectly identified as belonging to the unknown classes, we will re-classify the sample to the closest known/closed-set class. We conduct extensive experiments on ImageNet-LT, Places-LT, CIFAR-10/100-LT benchmark datasets, as well as a new long-tailed open-ended dataset that we build. Experimental results demonstrate that OLPR improves over the best competitors by up to 2.2% in terms of overall classification accuracy in closed-set settings, and up to 4% in terms of F-measure in open-set settings, which are very remarkable. [ABSTRACT FROM AUTHOR]
- Published
- 2025
- Full Text
- View/download PDF
9. imFTP: Deep imbalance learning via fuzzy transition and prototypical learning.
- Author
-
Hou, Yaxin, Ding, Weiping, and Zhang, Chongsheng
- Subjects
- *
DEEP learning , *SET theory , *FUZZY sets , *LEARNING modules , *MULTIPLE criteria decision making - Abstract
Although many methods have been proposed for tackling the class-imbalance problem, they still suffer from the insufficient feature representative capability and the overfitting problem. This paper proposes a new deep im balance learning approach based on the F uzzy set T heory and the P rototypical learning mechanism, abbreviated as imFTP for short, which consists of an adaptive smooth sampling module, a self-learnable prototypical learning module, and a fuzzy transition module. The adaptive smooth sampling module adaptively adjusts the sampling frequency of different classes to ensure their adequate opportunity to participate in the training process, which can mitigate the overfitting problem. The self-learnable prototypical learning module devises a clustering distance based Softmax cross-entropy loss and an intra-class clustering loss to improve the feature representation and discrimination capability of the model. The fuzzy transition module utilizes the fuzzy set theory to transform sample features effectively, which further enhances the feature representation capability of the model, meanwhile alleviates the overfitting problem. Experimental results on 15 benchmark datasets demonstrate that our method outperforms the best competitor by more than 3% in terms of the Macro-F1 metric, which is very significant. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. RCAR-UNet: Retinal vessel segmentation network algorithm via novel rough attention mechanism.
- Author
-
Ding, Weiping, Sun, Ying, Huang, Jiashuang, Ju, Hengrong, Zhang, Chongsheng, Yang, Guang, and Lin, Chin-Teng
- Subjects
- *
RETINAL blood vessels , *ARTIFICIAL neural networks , *NEUTROSOPHIC logic , *BLOOD vessels , *ROUGH sets , *SIMILARITY (Psychology) , *FEATURE extraction - Abstract
The health status of the retinal blood vessels is a significant reference for rapid and non-invasive diagnosis of various ophthalmological, diabetic, and cardio-cerebrovascular diseases. However, retinal vessels are characterized by ambiguous boundaries, with multiple thicknesses and obscured lesion areas. These phenomena cause deep neural networks to face the characteristic channel uncertainty when segmenting retinal blood vessels. The uncertainty in feature channels will affect the channel attention coefficient, making the deep neural network incapable of paying attention to the detailed features of retinal vessels. This study proposes a retinal vessel segmentation via a rough channel attention mechanism. First, the method integrates deep neural networks to learn complex features and rough sets to handle uncertainty for designing rough neurons. Second, a rough channel attention mechanism module is constructed based on rough neurons, and embedded in U-Net skip connection for the integration of high-level and low-level features. Then, the residual connections are added to transmit low-level features to high-level to enrich network feature extraction and help back-propagate the gradient when training the model. Finally, multiple comparison experiments were carried out on three public fundus retinal image datasets to verify the validity of Rough Channel Attention Residual U-Net (RCAR-UNet) model. The results show that the RCAR-UNet model offers high superiority in accuracy, sensitivity, F1, and Jaccard similarity, especially for the precise segmentation of fragile blood vessels, guaranteeing blood vessels' continuity. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.