15 results on '"Bian, Zekang"'
Search Results
2. Shared style linear k nearest neighbor classification method
- Author
-
Zhang, Jin, Bian, Zekang, and Wang, Shitong
- Published
- 2024
- Full Text
- View/download PDF
3. Weighted adaptively ensemble clustering method based on fuzzy Co-association matrix
- Author
-
Bian, Zekang, Qu, Jia, Zhou, Jie, Jiang, Zhibin, and Wang, Shitong
- Published
- 2024
- Full Text
- View/download PDF
4. Style linear k-nearest neighbor classification method
- Author
-
Zhang, Jin, Bian, Zekang, and Wang, Shitong
- Published
- 2024
- Full Text
- View/download PDF
5. Hybrid-ensemble-based interpretable TSK fuzzy classifier for imbalanced data
- Author
-
Bian, Zekang, Zhang, Jin, Nojima, Yusuke, Chung, Fu-lai, and Wang, Shitong
- Published
- 2023
- Full Text
- View/download PDF
6. Self-paced and Bayes-decision-rule linear KNN prediction
- Author
-
Zhang, Jin, Bian, Zekang, and Wang, Shitong
- Published
- 2022
- Full Text
- View/download PDF
7. Multi-view local linear KNN classification: theoretical and experimental studies on image classification
- Author
-
Jiang, Zhibin, Bian, Zekang, and Wang, Shitong
- Published
- 2020
- Full Text
- View/download PDF
8. Bayes-Decisive Linear KNN with Adaptive Nearest Neighbors.
- Author
-
Zhang, Jin, Bian, Zekang, and Wang, Shitong
- Abstract
While the classical KNN (k nearest neighbor) shares its avoidance of the consistent distribution assumption between training and testing samples to achieve fast prediction, it still faces two challenges: (a) its generalization ability heavily depends on an appropriate number k of nearest neighbors; (b) its prediction behavior lacks interpretability. In order to address the two challenges, a novel Bayes-decisive linear KNN with adaptive nearest neighbors (i.e., BLA-KNN) is proposed to obtain the following three merits: (a) a diagonal matrix is introduced to adaptively select the nearest neighbors and simultaneously improve the generalization capability of the proposed BLA-KNN method; (b) the proposed BLA-KNN method owns the group effect, which inherits and extends the group property of the sum of squares for total deviations by reflecting the training sample class-aware information in the group effect regularization term; (c) the prediction behavior of the proposed BLA-KNN method can be interpreted from the Bayes-decision-rule perspective. In order to do so, we first use a diagonal matrix to weigh each training sample so as to obtain the importance of the sample, while constraining the importance weights to ensure that the adaptive k value is carried out efficiently. Second, we introduce a class-aware information regularization term in the objective function to obtain the nearest neighbor group effect of the samples. Finally, we introduce linear expression weights related to the distance measure between the testing and training samples in the regularization term to ensure that the interpretation of Bayes-decision-rule can be performed smoothly. We also optimize the proposed objective function using an alternating optimization strategy. We experimentally demonstrate the effectiveness of the proposed BLA-KNN method by comparing it with 7 comparative methods on 15 benchmark datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Enhanced Fuzzy Random Forest by Using Doubly Randomness and Copying From Dynamic Dictionary Attributes.
- Author
-
Bian, Zekang, Chung, Fu-Lai, and Wang, Shitong
- Subjects
RANDOM forest algorithms ,RUNNING speed ,COMPARATIVE method ,DECISION trees ,RADIO frequency - Abstract
While fuzzy random forest (FRF) as a fuzzy implementation of random forest has earned its strong ambiguity/uncertainty handling capability on a rich variety of considerably low dimensional datasets, this article revisits FRF and attempts to enhance its generalization capability and computational speed on high dimensional datasets. For the first issue, in addition to the original use of randomness in FRF, a doubly randomness is newly introduced into the generation of both the candidate attributes and the best splitting attributes in FRF. For the computational speed issue, while the proposed new fuzzy information gain (NFG) measure does not apply to all candidate attributes, the remaining NFG values can be quickly retrieved by copying from the dynamically generated dictionary. As a result, a new method called enhanced fuzzy random forest (E-FRF) is proposed and justified theoretically from the consistency perspective. Our extensive experimental results indicate that the proposed method E-FRF has at least comparable performance to the comparative methods, and is an advantageous alternative to FRF in terms of both testing accuracy and running speed in most of the adopted high dimensional datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
10. Fuzzy KNN Method With Adaptive Nearest Neighbors.
- Author
-
Bian, Zekang, Vong, Chi Man, Wong, Pak Kin, and Wang, Shitong
- Abstract
Due to its strong performance in handling uncertain and ambiguous data, the fuzzy ${k}$ -nearest-neighbor method (FKNN) has realized substantial success in a wide variety of applications. However, its classification performance would be heavily deteriorated if the number k of nearest neighbors was unsuitably fixed for each testing sample. This study examines the feasibility of using only one fixed k value for FKNN on each testing sample. A novel FKNN-based classification method, namely, fuzzy KNN method with adaptive nearest neighbors (A-FKNN), is devised for learning a distinct optimal k value for each testing sample. In the training stage, after applying a sparse representation method on all training samples for reconstruction, A-FKNN learns the optimal ${k}$ value for each training sample and builds a decision tree (namely, A-FKNN tree) from all training samples with new labels (the learned optimal ${k}$ values instead of the original labels), in which each leaf node stores the corresponding optimal ${k}$ value. In the testing stage, A-FKNN identifies the optimal ${k}$ value for each testing sample by searching the A-FKNN tree and runs FKNN with the optimal ${k}$ value for each testing sample. Moreover, a fast version of A-FKNN, namely, FA-FKNN, is designed by building the FA-FKNN decision tree, which stores the optimal k value with only a subset of training samples in each leaf node. Experimental results on 32 UCI datasets demonstrate that both A-FKNN and FA-FKNN outperform the compared methods in terms of classification accuracy, and FA-FKNN has a shorter running time. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
11. Fuzzy Density Peaks Clustering.
- Author
-
Bian, Zekang, Chung, Fu-Lai, and Wang, Shitong
- Subjects
DENSITY ,AMBIGUITY ,COMPARATIVE method ,SOFT sets ,PARALLEL algorithms ,FUZZY sets ,TASK analysis - Abstract
As an exemplar-based clustering method, the well-known density peaks clustering (DPC) heavily depends on the computation of kernel-based density peaks, which incurs two issues: first, whether kernel-based density can facilitate a large variety of data well, including cases where ambiguity and uncertainty of the assignment of the data points to their clusters may exist, and second, whether the concept of density peaks can be interpreted and manipulated from the perspective of soft partitions (e.g., fuzzy partitions) to achieve enhanced clustering performance. In this article, in order to provide flexible adaptability for tackling ambiguity and uncertainty in clustering, a new concept of fuzzy peaks is proposed to express the density of a data point as the fuzzy-operator-based coupling of the fuzzy distances between a data point and its neighbors. As a fuzzy variant of DPC, a novel fuzzy density peaks clustering (FDPC) method FDPC based on fuzzy operators (especially S-norm operators) is accordingly devised along with the same algorithmic framework of DPC. With an appropriate choice of a fuzzy operator with its associated tunable parameter for a clustering task, FDPC can indeed inherit the advantage of fuzzy partitions and simultaneously provide flexibility in enhancing clustering performance. The experimental results on both synthetic and real data sets demonstrate that the proposed method outperforms or at least remains comparable to the comparative methods in clustering performance by choosing appropriate parameters in most cases. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
12. Joint Learning of Spectral Clustering Structure and Fuzzy Similarity Matrix of Data.
- Author
-
Bian, Zekang, Ishibuchi, Hisao, and Wang, Shitong
- Subjects
FUZZY algorithms ,CLUSTER analysis (Statistics) - Abstract
When spectral clustering analysis is applied, a similarity matrix of data plays a vital role in both clustering performance and stability of clustering results. In order to enhance the clustering performance and maintain the stability of the clustering results, a new method to jointly learn the similarity matrix and the clustering structure, called the joint learning method (FSCM) of spectral clustering structure and fuzzy similarity matrix of data, is proposed in this paper. In FSCM, the capability of a double-index fuzzy C-means clustering algorithm is used to determine an appropriate fuzzy similarity between any pair of data points. A fuzzy similarity matrix of data is also determined by adaptively assigning fuzzy neighbors of data points so the spectral clustering structure of data can be found and the clustering stability of FSCM can be assured. Experimental results on synthetic and real datasets demonstrate the effectiveness of the proposed method. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
13. Residual Sketch Learning for a Feature-Importance-Based and Linguistically Interpretable Ensemble Classifier.
- Author
-
Bian Z, Zhang J, Chung FL, and Wang S
- Abstract
Motivated by both the commonly used "from wholly coarse to locally fine" cognitive behavior and the recent finding that simple yet interpretable linear regression model should be a basic component of a classifier, a novel hybrid ensemble classifier called hybrid Takagi-Sugeno-Kang fuzzy classifier (H-TSK-FC) and its residual sketch learning (RSL) method are proposed. H-TSK-FC essentially shares the virtues of both deep and wide interpretable fuzzy classifiers and simultaneously has both feature-importance-based and linguistic-based interpretabilities. RSL method is featured as follows: 1) a global linear regression subclassifier on all original features of all training samples is generated quickly by the sparse representation-based linear regression subclassifier training procedure to identify/understand the importance of each feature and partition the output residuals of the incorrectly classified training samples into several residual sketches; 2) by using both the enhanced soft subspace clustering method (ESSC) for the linguistically interpretable antecedents of fuzzy rules and the least learning machine (LLM) for the consequents of fuzzy rules on residual sketches, several interpretable Takagi-Sugeno-Kang (TSK) fuzzy subclassifiers are stacked in parallel through residual sketches and accordingly generated to achieve local refinements; and 3) the final predictions are made to further enhance H-TSK-FC's generalization capability and decide which interpretable prediction route should be used by taking the minimal-distance-based priority for all the constructed subclassifiers. In contrast to existing deep or wide interpretable TSK fuzzy classifiers, benefiting from the use of feature-importance-based interpretability, H-TSK-FC has been experimentally witnessed to have faster running speed and better linguistic interpretability (i.e., fewer rules and/or TSK fuzzy subclassifiers and smaller model complexities) yet keep at least comparable generalization capability.
- Published
- 2024
- Full Text
- View/download PDF
14. Three-layer heterogeneous network based on the integration of CircRNA information for MiRNA-disease association prediction.
- Author
-
Qu J, Liu S, Li H, Zhou J, Bian Z, Song Z, and Jiang Z
- Abstract
Increasing research has shown that the abnormal expression of microRNA (miRNA) is associated with many complex diseases. However, biological experiments have many limitations in identifying the potential disease-miRNA associations. Therefore, we developed a computational model of Three-Layer Heterogeneous Network based on the Integration of CircRNA information for MiRNA-Disease Association prediction (TLHNICMDA). In the model, a disease-miRNA-circRNA heterogeneous network is built by known disease-miRNA associations, known miRNA-circRNA interactions, disease similarity, miRNA similarity, and circRNA similarity. Then, the potential disease-miRNA associations are identified by an update algorithm based on the global network. Finally, based on global and local leave-one-out cross validation (LOOCV), the values of AUCs in TLHNICMDA are 0.8795 and 0.7774. Moreover, the mean and standard deviation of AUC in 5-fold cross-validations is 0.8777+/-0.0010. Especially, the two types of case studies illustrated the usefulness of TLHNICMDA in predicting disease-miRNA interactions., Competing Interests: The authors declare that they have no competing interests., (© 2024 Qu et al.)
- Published
- 2024
- Full Text
- View/download PDF
15. Neighborhood-based inference and restricted Boltzmann machine for microbe and drug associations prediction.
- Author
-
Cheng X, Qu J, Song S, and Bian Z
- Subjects
- Humans, Algorithms, Machine Learning, Area Under Curve, SARS-CoV-2, COVID-19
- Abstract
Background: Efficient identification of microbe-drug associations is critical for drug development and solving problem of antimicrobial resistance. Traditional wet-lab method requires a lot of money and labor in identifying potential microbe-drug associations. With development of machine learning and publication of large amounts of biological data, computational methods become feasible., Methods: In this article, we proposed a computational model of neighborhood-based inference (NI) and restricted Boltzmann machine (RBM) to predict potential microbe-drug association (NIRBMMDA) by using integrated microbe similarity, integrated drug similarity and known microbe-drug associations. First, NI was used to obtain a score matrix of potential microbe-drug associations by using different thresholds to find similar neighbors for drug or microbe. Second, RBM was employed to obtain another score matrix of potential microbe-drug associations based on contrastive divergence algorithm and sigmoid function. Because generalization ability of individual method is poor, we used an ensemble learning to integrate two score matrices for predicting potential microbe-drug associations more accurately. In particular, NI can fully utilize similar (neighbor) information of drug or microbe and RBM can learn potential probability distribution hid in known microbe-drug associations. Moreover, ensemble learning was used to integrate individual predictor for obtaining a stronger predictor., Results: In global leave-one-out cross validation (LOOCV), NIRBMMDA gained the area under the receiver operating characteristics curve (AUC) of 0.8666, 0.9413 and 0.9557 for datasets of DrugVirus, MDAD and aBiofilm, respectively. In local LOOCV, AUCs of 0.8512, 0.9204 and 0.9414 were obtained for NIRBMMDA based on datasets of DrugVirus, MDAD and aBiofilm, respectively. For five-fold cross validation, NIRBMMDA acquired AUC and standard deviation of 0.8569 ± -0.0027, 0.9248 ± -0.0014 and 0.9369 ± -0.0020 on the basis of datasets of DrugVirus, MDAD and aBiofilm, respectively. Moreover, case study for severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) showed that 13 out of the top 20 predicted drugs were verified by searching literature. The other two case studies indicated that 17 and 17 out of the top 20 predicted microbes for the drug of ciprofloxacin and minocycline were confirmed by identifying published literature, respectively., Competing Interests: The authors declare there are no competing interests., (©2022 Cheng et al.)
- Published
- 2022
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.