13 results on '"classifier integration"'
Search Results
2. An effective fine grading method of BI-RADS classification in mammography.
- Author
-
Lin, Fei, Sun, Hang, Han, Lu, Li, Jing, Bao, Nan, Li, Hong, Chen, Jing, Zhou, Shi, and Yu, Tao
- Abstract
Purpose: Mammography is an important imaging technique for the detection of early breast cancer. Doctors classify mammograms as Breast Imaging Reporting and Data Systems (BI-RADS). This study aims to provide an intelligent BI-RADS grading prediction method, which can help radiologists and clinicians to distinguish the most challenging 4A, 4B, and 4C cases in mammography. Methods: Firstly, the breast region, the lesion region, and the corresponding region in the contralateral breast were extracted. Four categories of features were extracted from the original images and the images after the wavelet transform. Secondly, an optimized sequential forward floating selection (SFFS) was used for feature selection. Finally, a two-layer classifier integration was employed for fine grading prediction. 45 cases from the hospital and 500 cases from Digital Database for Screening Mammography (DDSM) database were used for evaluation. Results: The classification performance of the support vector machine (SVM), Bayes, and random forest is very close on the 45 testing set, with the area under the receiver operating characteristic curve (AUC) of 0.978, 0.967, and 0.968. On the DDSM set, the AUC achieves 0.931, 0.938, and 0.874. Using the mean probability prediction, the AUC on the two datasets reaches 0.998 and 0.916. However, they are all significantly higher than the doctors' diagnosis, with the AUC of 0.807 and 0.725. Conclusions: A BI-RADS fine grading (2, 3, 4A, 4B, 4C, 5) prediction model was proposed. Through the evaluation from different datasets, the performance is proved higher than that of the doctors, which may provide great help for clinical BI-RADS classification diagnosis. Therefore, our method can produce more effective and reliable results. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
3. Integration and Selection of Linear SVM Classifiers in Geometric Space
- Author
-
Robert Burduk and Jedrzej Biedrzycki
- Subjects
classifier integration ,ensemble of classifiers ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
Integration or fusion of the base classifiers is the final stage of creating multiple classifiers system. Known methods in this step use base classifier outputs, which are class labels or values of the confidence (predicted probabilities) for each class label. In this paper we propose an integration process which takes place in the geometric space. It means that the fusion of base classifiers is done using their decision boundaries. In order to obtain one decision boundary from boundaries defined by base classifiers the median or weighted average method will be used. In addition, the proposed algorithm uses the division of the entire feature space into disjoint regions of competence as well as the process of selection of base classifiers is carried out. The aim of the experiments was to compare the proposed algorithms with the majority voting method and assessment which of the analyzed approaches to integration of the base classifiers creates a more effective ensemble.
- Published
- 2019
- Full Text
- View/download PDF
4. Decision Tree Integration Using Dynamic Regions of Competence
- Author
-
Jędrzej Biedrzycki and Robert Burduk
- Subjects
decision tree ,random forest ,majority voting ,classifier ensemble ,classifier integration ,Science ,Astrophysics ,QB460-466 ,Physics ,QC1-999 - Abstract
A vital aspect of the Multiple Classifier Systems construction process is the base model integration. For example, the Random Forest approach used the majority voting rule to fuse the base classifiers obtained by bagging the training dataset. In this paper we propose the algorithm that uses partitioning the feature space whose split is determined by the decision rules of each decision tree node which is the base classification model. After dividing the feature space, the centroid of each new subspace is determined. This centroids are used in order to determine the weights needed in the integration phase based on the weighted majority voting rule. The proposal was compared with other Multiple Classifier Systems approaches. The experiments regarding multiple open-source benchmarking datasets demonstrate the effectiveness of our method. To discuss the results of our experiments, we use micro and macro-average classification performance measures.
- Published
- 2020
- Full Text
- View/download PDF
5. Integrating Heterogeneous Prediction Models in the Cloud
- Author
-
Chen, Hung-Chen, Wei, Chih-Ping, Chen, Yu-Cheng, Lan, Ci-Wei, van der Aalst, Wil, editor, Mylopoulos, John, editor, Rosemann, Michael, editor, Shaw, Michael J., editor, Szyperski, Clemens, editor, Zhang, Dongsong, editor, and Yue, Wei T., editor
- Published
- 2012
- Full Text
- View/download PDF
6. Integration of Global and Local Feature for Age Estimation of Facial Images
- Author
-
Kou, Jie, Du, Ji-Xiang, Zhai, Chuan-Min, Hutchison, David, editor, Kanade, Takeo, editor, Kittler, Josef, editor, Kleinberg, Jon M., editor, Mattern, Friedemann, editor, Mitchell, John C., editor, Naor, Moni, editor, Nierstrasz, Oscar, editor, Pandu Rangan, C., editor, Steffen, Bernhard, editor, Sudan, Madhu, editor, Terzopoulos, Demetri, editor, Tygar, Doug, editor, Vardi, Moshe Y., editor, Weikum, Gerhard, editor, Goebel, Randy, editor, Siekmann, Jörg, editor, Wahlster, Wolfgang, editor, Huang, De-Shuang, editor, Ma, Jianhua, editor, Jo, Kang-Hyun, editor, and Gromiha, M. Michael, editor
- Published
- 2012
- Full Text
- View/download PDF
7. Integration of Decision Trees Using Distance to Centroid and to Decision Boundary
- Author
-
Robert Burduk and Jedrzej Biedrzycki
- Subjects
classifier integration ,General Computer Science ,Computer science ,Decision tree ,Centroid ,QA75.5-76.95 ,computer.software_genre ,classifier integrat ,Theoretical Computer Science ,ComputingMethodologies_PATTERNRECOGNITION ,ensemble of classifiers ,Electronic computers. Computer science ,Decision boundary ,Data mining ,computer ,distance to decision boundary - Abstract
Plethora of ensemble techniques have been implemented and studied in order to achieve better classification results than base classifiers. In this paper an algorithm for integration of decision trees is proposed, which means that homogeneous base classifiers will be used. The novelty of the presented approach is the usage of the simultaneous distance of the object from the decision boundary and the center of mass of objects belonging to one class label in order to determine the score functions of base classifiers. This means that the score function assigned to the class label by each classifier depends on the distance of the classified object from the decision boundary and from the centroid. The algorithm was evaluated using an open-source benchmarking dataset. The results indicate an improvement in the classification quality in comparison to the referential method - majority voting method.
- Published
- 2020
8. Protein function prediction based on data fusion and functional interrelationship.
- Author
-
Meng, Jun, Wekesa, Jael-Sanyanda, Shi, Guan-Li, and Luan, Yu-Shi
- Subjects
- *
PROTEIN analysis , *PREDICTION theory , *DATA fusion (Statistics) , *FUNCTIONAL analysis , *BIOINFORMATICS , *PROTEOMICS - Abstract
One of the challenging tasks of bioinformatics is to predict more accurate and confident protein functions from genomics and proteomics datasets. Computational approaches use a variety of high throughput experimental data, such as protein-protein interaction (PPI), protein sequences and phylogenetic profiles, to predict protein functions. This paper presents a method that uses transductive multi-label learning algorithm by integrating multiple data sources for classification. Multiple proteomics datasets are integrated to make inferences about functions of unknown proteins and use a directed bi-relational graph to assign labels to unannotated proteins. Our method, bi-relational graph based transductive multi-label function annotation (Bi-TMF) uses functional correlation and topological PPI network properties on both the training and testing datasets to predict protein functions through data fusion of the individual kernel result. The main purpose of our proposed method is to enhance the performance of classifier integration for protein function prediction algorithms. Experimental results demonstrate the effectiveness and efficiency of Bi-TMF on multi-sources datasets in yeast, human and mouse benchmarks. Bi-TMF outperforms other recently proposed methods. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
9. Integration and Selection of Linear SVM Classifiers in Geometric Space
- Author
-
Jedrzej Biedrzycki and Robert Burduk
- Subjects
ComputingMethodologies_PATTERNRECOGNITION ,classifier integration ,svm ,Electronic computers. Computer science ,ensemble of classifiers ,QA75.5-76.95 - Abstract
Integration or fusion of the base classifiers is the final stage of creating multiple classifiers system. Known methods in this step use base classifier outputs, which are class labels or values of the confidence (predicted probabilities) for each class label. In this paper we propose an integration process which takes place in the geometric space. It means that the fusion of base classifiers is done using their decision boundaries. In order to obtain one decision boundary from boundaries defined by base classifiers the median or weighted average method will be used. In addition, the proposed algorithm uses the division of the entire feature space into disjoint regions of competence as well as the process of selection of base classifiers is carried out. The aim of the experiments was to compare the proposed algorithms with the majority voting method and assessment which of the analyzed approaches to integration of the base classifiers creates a more effective ensemble.
- Published
- 2019
10. Decision Tree Integration Using Dynamic Regions of Competence
- Author
-
Robert Burduk and Jedrzej Biedrzycki
- Subjects
Majority rule ,Computer science ,Feature vector ,Decision tree ,General Physics and Astronomy ,lcsh:Astrophysics ,0102 computer and information sciences ,02 engineering and technology ,computer.software_genre ,01 natural sciences ,Article ,lcsh:QB460-466 ,decision tree ,0202 electrical engineering, electronic engineering, information engineering ,lcsh:Science ,classifier integration ,majority voting ,Centroid ,Decision rule ,Benchmarking ,lcsh:QC1-999 ,Random forest ,classifier ensemble ,ComputingMethodologies_PATTERNRECOGNITION ,010201 computation theory & mathematics ,lcsh:Q ,020201 artificial intelligence & image processing ,Data mining ,computer ,lcsh:Physics ,random forest ,Subspace topology - Abstract
A vital aspect of the Multiple Classifier Systems construction process is the base model integration. For example, the Random Forest approach used the majority voting rule to fuse the base classifiers obtained by bagging the training dataset. In this paper we propose the algorithm that uses partitioning the feature space whose split is determined by the decision rules of each decision tree node which is the base classification model. After dividing the feature space, the centroid of each new subspace is determined. This centroids are used in order to determine the weights needed in the integration phase based on the weighted majority voting rule. The proposal was compared with other Multiple Classifier Systems approaches. The experiments regarding multiple open-source benchmarking datasets demonstrate the effectiveness of our method. To discuss the results of our experiments, we use micro and macro-average classification performance measures.
- Published
- 2020
- Full Text
- View/download PDF
11. Decision Tree Integration Using Dynamic Regions of Competence.
- Author
-
Biedrzycki, Jędrzej and Burduk, Robert
- Subjects
- *
DECISION trees , *PLURALITY voting , *ALGORITHMS , *PERFORMANCE , *CENTROID - Abstract
A vital aspect of the Multiple Classifier Systems construction process is the base model integration. For example, the Random Forest approach used the majority voting rule to fuse the base classifiers obtained by bagging the training dataset. In this paper we propose the algorithm that uses partitioning the feature space whose split is determined by the decision rules of each decision tree node which is the base classification model. After dividing the feature space, the centroid of each new subspace is determined. This centroids are used in order to determine the weights needed in the integration phase based on the weighted majority voting rule. The proposal was compared with other Multiple Classifier Systems approaches. The experiments regarding multiple open-source benchmarking datasets demonstrate the effectiveness of our method. To discuss the results of our experiments, we use micro and macro-average classification performance measures. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
12. Training set selection and swarm intelligence for enhanced integration in multiple classifier systems.
- Author
-
Mohammed, Amgad M., Onieva, Enrique, and Woźniak, Michał
- Subjects
SWARM intelligence ,MACHINE learning ,FEATURE selection ,PROCESS optimization ,MULTIPLE intelligences ,FORECASTING - Abstract
Multiple classifier systems (MCS s) constitute one of the most competitive paradigms for obtaining more accurate predictions in the field of machine learning. Systems of this type should be designed efficiently in all of their stages, from data preprocessing to multioutput decision fusion. In this article, we present a framework for utilizing the power of instance selection methods and the search capabilities of swarm intelligence to train learning models and to aggregate their decisions. The process consists of three steps: First, the essence of the complete training data set is captured in a reduced set via the application of intelligent data sampling. Second, the reduced set is used to train a group of heterogeneous classifiers using bagging and distance-based feature sampling. Finally, swarm intelligence techniques are applied to identify a pattern among multiple decisions to enhance the fusion process by assigning class-specific weights for each classifier. The proposed methodology yielded competitive results in experiments that were conducted on 25 benchmark datasets. The Matthews correlation coefficient (MCC) is regarded as the objective to be maximized by various nature-inspired metaheuristics, which include the moth-flame optimization algorithm (MFO), the grey wolf optimizer (GWO) and the whale optimization algorithm (WOA). • Proposing a framework for discussing the intersection between instance selection, ensemble learning and swarm intelligence. • Heterogeneous ensemble system considering two additional steps to increase the diversity of the set of classifiers: Bagging, Distance-based feature selection. • Swarm intelligence for combining multiple decisions. • Matthews Correlation Coefficient (MCC) to evaluate ensemble under the search capabilities of swarm intelligence. • Getting out more accuracy from the reduced data. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
13. Cross-modality image feature fusion diagnosis in breast cancer.
- Author
-
Jiang M, Han L, Sun H, Li J, Bao N, Li H, Zhou S, and Yu T
- Subjects
- Bayes Theorem, Female, Humans, Mammography, ROC Curve, Retrospective Studies, Breast Neoplasms diagnostic imaging
- Abstract
Considering the complementarity of mammography and breast MRI, the research of feature fusion diagnosis based on cross-modality images was explored to improve the accuracy of breast cancer diagnosis. 201 patients with both mammography and breast MRI were collected retrospectively, including 117 cases of benign lesions and 84 cases of malignant ones. Two feature optimization strategies of sequential floating forward selection (SFFS), SFFS-1 and SFFS-2, were defined based on the sequential floating forward selection method. Each strategy was used to analyze the diagnostic performance of single-modality images and then to study the feature fusion diagnosis of cross-modality images. Three feature fusion approaches were compared: optimizing MRI features and then fusing those of mammography; optimizing mammography features and then fusing those of MRI; selecting the effective features from the whole feature set (mammography and MRI). Support vector machine, Naive Bayes, and K-nearest neighbor were employed as the classifiers and were finally integrated to get better performance. The average accuracy and area under the ROC curve (AUC) of MRI (88.56%, 0.9 for SFFS-1, 88.39%, 0.89 for SFFS-2) were better than mammography (84.25%, 0.84 for SFFS-1, 80.43%, 0.80 for SFFS-2). Furthermore, compared with a single modality, the average accuracy and AUC of cross-modality feature fusion can improve from 85.40% and 0.86 to 89.66% and 0.91. Classifier integration improved the accuracy and AUC from 90.49%, 0.92 to 92.37%, and 0.97. Cross-modality image feature fusion can achieve better diagnosis performance than a single modality. Feature selection strategy SFFS-1 has better efficiency than SFFS-2. Classifier integration can further improve diagnostic accuracy., (© 2021 Institute of Physics and Engineering in Medicine.)
- Published
- 2021
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.