339 results on '"INFORMATION ENTROPY"'
Search Results
2. Structural analysis of bond information entropy and HOMO-LUMO gap in CLO and KFI zeolites
- Author
-
Daniel, Paul, Arockiaraj, Micheal, Peter, Pancras, and Clement, Joseph
- Published
- 2025
- Full Text
- View/download PDF
3. Exploring the interconnections between total cloud water content and water vapor mixing ratio with other cloud microphysical variables in northward-moving typhoon precipitation via information entropy: A hybrid causal analysis approach using wavelet coherence and Liang–Kleeman information flow
- Author
-
Wu, Xianghua, Ren, Miaomiao, Zhou, Linyi, Li, Yashao, Chen, Jinghua, Li, Wanting, Yang, Kai, and Wang, Weiwei
- Published
- 2025
- Full Text
- View/download PDF
4. Assessing geological structure uncertainties in groundwater models using transition probability-based realizations
- Author
-
Huang, Shiqi, Hu, Litang, Li, Binghua, Wu, Xia, Gan, Lin, Sun, Jianchong, and Zhang, Menglin
- Published
- 2025
- Full Text
- View/download PDF
5. TIEOD: Three-way concept-based information entropy for outlier detection
- Author
-
Hu, Qian, Zhang, Jun, Mi, Jusheng, Yuan, Zhong, and Li, Meizheng
- Published
- 2025
- Full Text
- View/download PDF
6. Information Entropy Evaluation Method for the Disassemblability of Smartphones.
- Author
-
Huang, Haihong, Xue, Yuhao, Zhu, Libin, and Cui, Chuangchuang
- Abstract
Smartphones are a vast category of consumer electronic products. Design For Disassembly (DFD) is a methodology that can effectively reduce the disassembly, recycling, and maintenance costs of smartphones. However, due to the diverse constraints posed by the connections among smartphone components, the variability of disassembling tools makes it challenging to objectively characterize the disassemblability of the smartphone. Therefore, information entropy is introduced to characterize the complex state of the system. Disassemblability can be expressed by calculating the operating time of the disassembling tool through information entropy. After multiplying by a constrained quantity factor, the Improved Disassembling Tool Entropy (IDTE) responds to changes in disassemblability as the structural level changes. According to the structure and recycling direction, the disassembling level of smartphones can be divided into module level, part level, and hybrid part-module level. Based on the Maynard Operational Sequencing Technique (MOST), the basic unit of operating time of the disassembling tool is calculated. For the hybrid selective Disassembly Sequence Planning (DSP) of the part and module levels, the Improved Double Genetic Algorithm (IDGA) model is established to compute the optimal disassembly sequence corresponding to each program level. This model calculates the IDTE corresponding to the optimal disassembly sequences at each disassembling level. The validity of IDTE was verified by a coupled comparison test between IDTE and theoretical disassembly time. Finally, an analysis of the disassemblability variations was conducted for two different models of smartphones based on their structure. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Energy management strategy with mutation protection for fuel cell electric vehicles.
- Author
-
Wang, Da, Mei, Lei, Song, Chuanxue, Jin, Liqiang, Xiao, Feng, and Song, Shixin
- Subjects
- *
FUEL cell vehicles , *ENERGY management , *FUEL cells , *HYDROGEN economy , *HYBRID electric vehicles , *HYDROGEN as fuel - Abstract
For hydrogen fuel cell vehicles, energy management strategies (EMS) are vital for balancing fuel cell and battery power, limiting fuel cell power, maintaining state of charge (SOC) fluctuation range and mitigating degradation. Reinforcement learning-based EMS, especially using deep Q-network (DQN) and deep deterministic policy gradient (DDPG), have demonstrated potential for enhancing the fuel economy of hybrid electric vehicles (HEV) and fuel cell electric vehicles (FCEV). This research proposes mutation protection DQN (MPD)-based EMS to improve hydrogen fuel economy and reduce fuel cell degradation under driving cycles with plenty of mutations. By quantifying mutation, exploring its relationship with driving conditions and integrating a mutation protection module with DQN, MPD-based EMS achieves approximately 11% and 6% better fuel economy compared to the other two learning-based EMS. Additionally, it also reduces fuel cell degradation by approximately 21% and 13%. • Quantifying the mutation of driving conditions through variance. • Utilizing information entropy to process mutations. • Introducing a dynamic mutation threshold value. • Designing mutation protection module and integrating it into the DQN. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Initial selection of configurations based on information entropy for multi-level design optimization.
- Author
-
Saad, Mohammed H. and Xue, Deyi
- Abstract
An efficient multi-level design optimization method is developed in this research based on information entropy to select the configurations with high potential to achieve the optimal design solution. The generic design is modeled by an AND-OR tree based on design requirements. A node in the AND-OR tree is used to describe partial design solution such as a component or an assembly with design parameters. The different design configurations are created from the AND-OR tree through tree-based search, and each design configuration is modeled by design parameters. Optimization is conducted at two levels. Parameter optimization is carried out to identify the optimal parameter values for each design configuration, while configuration optimization is carried out to identify the optimal configuration. Information entropy is employed to evaluate the partial configuration candidates modeled as branches in the AND-OR tree to eliminate the branches that are unlikely to lead to the optimal solution. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
9. Magnetic anomaly detection via a combination approach of minimum entropy and gradient orthogonal functions.
- Author
-
Zhou, Jiaqi, Wang, Chengdong, Peng, Genzhai, Yan, Huan, Zhang, Zhihong, and Chen, Yong
- Subjects
MAGNETIC anomalies ,ORTHOGONAL functions ,ENTROPY ,ORTHONORMAL basis ,MAGNETIC entropy - Abstract
Aiming at the problem of magnetic anomaly detection under low signal-to-noise ratio, a full magnetic gradient detection via minimum entropy and gradient orthonormal basis function is proposed in this paper. Firstly, the mathematical model of magnetic dipole gradient tensor and the experimental method of collecting magnetic gradient tensor signals are introduced. Secondly, a new minimum entropy detector is designed, the basic functions which are processed by new minimum entropy detector can be orthogonalized easily. Thirdly, a group of new standard orthogonal basis functions are constructed according to the characteristics of magnetic gradient anomaly signals. Finally, the detector and orthogonal basis functions are used to detect the magnetic anomaly. Both numerical simulation signals and experiment signals are used to prove the effectiveness of the methods. • A new magnetic gradient minimum entropy detector is designed on the orthonormal basic function. • A group of new standard orthogonal basis functions are constructed on the characteristics of magnetic gradient anomaly signals. • The detector and orthogonal basis functions are used to detect the magnetic anomaly. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
10. Two-strand ladder network variants: Localization, multifractality, and quantum dynamics under an Aubry-André-Harper kind of quasiperiodicity.
- Author
-
Biswas, Sougata
- Subjects
- *
PHASE transitions , *WAVE packets , *QUANTUM theory , *ELECTRONIC spectra , *ENTROPY (Information theory) , *TRANSITION metals - Abstract
In this paper we demonstrate, using a couple of variants of a two-strand ladder network that, a quasiperiodic Aubry-André-Harper (AAH) modulation applied to the vertical strands, mimicking a deterministic distortion in the network, can give rise to certain exotic features in the electronic spectrum of such systems. While, for the simplest ladder network all the eigenstates become localized as the modulation strength crosses a threshold, for the second variant, modeling an ultrathin graphene nano-ribbon, the central part of the energy spectrum remains populated by extended wavefunctions. The multifractal character in the energy spectrum is observed for both these networks close to the critical values of the modulation. We substantiate our findings also by studying the quantum dynamics of a wave packet on such decorated lattices. Interestingly, while the mean square displacement (MSD) changes in the usual manner in a pure two-strand ladder network as the modulation strength varies, for the ultrathin graphene nanoribbon the temporal behavior of the MSD remains unaltered only up to a strong modulation strength. This, we argue, is due to the extendedness of the wavefunction at the central part of the energy spectrum. Other measurements like the return probability, temporal autocorrelation function, the time dependence of the inverse participation ratio, and the information entropy are calculated for both networks with different modulation strengths and corroborate our analytical findings. • We have investigated in detail the behavior of eigenstates and multifractality in two decorated lattices. • The ladder network shows the metal–insulator transition. • A perfect metal-to-insulator phase transition is not possible in the ultrathin graphene nano-ribbon network. • We have explored that both the distorted networks exhibit multifractality in the neighborhood of their critical points. • We substantiate our findings also by studying the quantum dynamics of a wave packet on such decorated lattices. [ABSTRACT FROM AUTHOR]
- Published
- 2025
- Full Text
- View/download PDF
11. An integrated framework of deep learning and entropy theory for enhanced high-dimensional permeability field identification in heterogeneous aquifers.
- Author
-
Cao, Mingxu, Dai, Zhenxue, Chen, Junjun, Yin, Huichao, Zhang, Xiaoying, Wu, Jichun, Thanh, Hung Vo, and Soltanian, Mohamad Reza
- Subjects
- *
STOCHASTIC systems , *GROUNDWATER flow , *ENTROPY (Information theory) , *PERMEABILITY , *AQUIFERS , *DEEP learning - Abstract
• Uncertainty of stochastic system responses may lead to non-unique inversion results of high-dimensional parameters when adopting entropy-based monitoring network design criteria. • An ensemble monitoring network is developed to enhance the robustness of the inversion framework for heterogeneous permeability field estimation. • Study evaluates the influence of observation information in various identification scenarios on the inversion results of high-dimensional parameters. Accurately estimating high-dimensional permeability ( k ) fields through data assimilation is critical for minimizing uncertainties in groundwater flow and solute transport simulations. However, designing an effective monitoring network to obtain diverse system responses in heterogeneous aquifers for data assimilation presents significant challenges. To investigate the influence of different measurement types (hydraulic heads, solute concentrations, and permeability) and monitoring strategies on the accuracy of permeability characterization, this study integrates a deep learning-based surrogate modeling approach and the entropy-based maximum information minimum redundancy (MIMR) monitoring design criterion into a data assimilation framework. An ensemble MIMR-optimized method is developed to provide more comprehensive monitoring information and avoid missing key information due to the randomness of stochastic response datasets in entropy analysis. A numerical case of solute transport with log-Gaussian permeability fields is presented, with twelve scenarios designed by combining different measurement types and monitoring strategies. The results demonstrated that the proposed ensemble MIMR-optimized method significantly improved the k -field estimates compared to the conventional MIMR method. Additionally, high prediction accuracy in forward modeling is essential for ensuring reliable inversion results, especially for observation data with strong nonlinearity. The findings of this study enhance our understanding and management of k -field estimation in heterogeneous aquifers, contributing to the development of more robust inversion frameworks for general data assimilation tasks. [Display omitted] [ABSTRACT FROM AUTHOR]
- Published
- 2025
- Full Text
- View/download PDF
12. Chaos based image encryption scheme to secure sensitive multimedia content in cloud storage.
- Author
-
Umar, Talha, Nadeem, Mohammad, and Anwer, Faisal
- Subjects
- *
DATA encryption , *CLOUD storage , *SINE function , *TRIGONOMETRIC functions , *ENTROPY (Information theory) , *IMAGE encryption - Abstract
Cloud computing offers a variety of on-demand services to users and has gained significant prominence in the contemporary era. The security of information stored in cloud data centers has become a central concern, especially for sensitive data like medical images, videos, and multimedia content that require heightened protection when stored in data centers. Cloud users have a responsibility to ensure the security of their data through strategic measures. Focusing on maintaining the privacy of images stored in the cloud, this research introduces an innovative image encryption technique that is based on a modified Skew Tent chaotic map. The suggested modification to the chaotic map includes combining the Skew Tent map with both the sine trigonometric function and perturbation technology. This results in enhanced randomness, more intricate dynamical behavior, a broader chaotic range, and increased sensitivity to initial values in contrast to alternative chaotic maps. The incorporation of this adapted map into a stepwise procedure, involving two rounds of permutation followed by diffusion, effectively accomplishes image data encryption for cloud storage systems. These consecutive operations collectively enhance the encryption method's randomness and robustness. Through simulations conducted using Cloudsim, the cipher-image exhibits a uniform distribution and achieves a commendable information entropy score of 7.996749. The encryption algorithm significantly reduces correlation coefficients from 1 in the original image to 0 in the encrypted image, while maintaining NPCR (Number of Pixel Change Rate) and UACI (Unified Average Changing Intensity) values within the critical range. Additionally, both theoretical analysis and practical evaluations confirm the algorithm's resilience against exhaustive, occlusion, and classical attacks. Moreover, extending this encryption framework to video data, a novel video encryption approach is proposed. • Novel image encryption with modified Skew Tent map for cloud privacy. • Method merges Skew Tent map, sine function, and perturbation technology. • The technique boosts chaos, complexity, range, and initial value sensitivity. • Cloudsim simulations show cipher-image uniformity & high entropy of 7.996749. • Encryption lowers correlation from 1 to 0, keeps NPCR and UACI in critical range. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. Optimising the acquisition conditions of high information quality low-field NMR signals based on a cutting-edge approach applying information theory and Taguchi's experimental designs – Virgin olive oil as an application example.
- Author
-
Arroyo-Cerezo, Alejandra, Jiménez-Carvelo, Ana M., López-Ruíz, Rosalía, Tello-Liébana, María, and Cuadros-Rodríguez, Luis
- Subjects
- *
ENTROPY (Information theory) , *NMR spectrometers , *TAGUCHI methods , *OLIVE oil , *ANALYTICAL chemistry - Abstract
Developing a new spectrometric analytical method based on a fingerprinting approach requires optimisation of the experimental stage, particularly with novel instruments like benchtop low-field NMR spectrometers. To ensure high-quality LF-NMR spectra before developing the multivariate model, an experimental design to optimise instrument conditions is essential. However, difficult-to-control factors may be critical for optimisation. Taguchi methodology addresses these factors to obtain a system robust to variation. This study uses the Taguchi methodology to optimise instrument settings for acquiring high-quality 1H and 13C LF-NMR signals in a short time from virgin olive oil (VOO). Two experimental trials (for 1H and 13C signals, respectively) were carried out and analysed to find an optimal and robust combination of instrument settings against changes in two difficult-to-control factors: ambient temperature and small deviations of the NMR tube volume (700 ± 50 μL). The responses to be optimised, run time and spectral information quality, were analysed separately and jointly, as some factors showed opposite behaviour in the effect on the responses. Multiple response analysis based on suitable desirability functions yielded a combination of factors resulting in desirability values above 0.8 for 1H LF-NMR signals and almost 1.0 for 13C LF-NMR signals. In addition, a novel approach to assess the information quality of an analytical signal was proposed, addressing a major challenge in analytical chemistry. By applying information theory and calculating information entropy, this approach demonstrated its potential for selecting the highest quality (i.e. most informative) analytical signals. The acquisition instrument conditions of LF-NMR were successfully optimised using Taguchi methodology to acquire highly informative 1H and 13C spectra in a minimum run time. The importance lies in the future development of non-targeted analytical applications for VOO quality control. In addition, the innovative use of information entropy to a priori assess the signal quality represents a significant advance and proposes a solution to a long-standing challenge in analytical chemistry. [Display omitted] • Instrument settings for 1H and 13C LF-NMR signals acquisition were optimised. • Taguchi experimental designs were performed to optimise a robust system. • A novel proposal to a priori assess the information quality of an analytical signal is presented. • Information theory was applied to select the most informative fingerprint. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
14. Similarity measure method of near-infrared spectrum combined with multi-attribute information.
- Author
-
Zhang, Jinfeng, Qin, Yuhua, Tian, Rongkun, Bai, Xiaoli, and Liu, Jing
- Subjects
- *
NEAR infrared spectroscopy , *DISTRIBUTION (Probability theory) , *ENTROPY (Information theory) , *PRODUCT design , *DATA distribution , *TOBACCO - Abstract
[Display omitted] • This paper uses NIR spectroscopy combined with intelligent analysis algorithms to improve the accuracy of similarity measurement. • The method effectively extracts sample information of NIR spectra in high-dimensional space, retaining the original data features. • The Sinkhorn distance is introduced to improve the t-SNE algorithm, which can focus more on the global probability distributions. • To mitigate multi-attribute impacts on similarity measure, a multi-attribute distance matrix is constructed by information entropy. • The method can assist in the maintenance and design of the product formula. Due to the high-dimensionality, redundancy, and non-linearity of the near-infrared (NIR) spectra data, as well as the influence of attributes such as producing area and grade of the sample, which can all affect the similarity measure between samples. This paper proposed a t-distributed stochastic neighbor embedding algorithm based on Sinkhorn distance (St-SNE) combined with multi-attribute data information. Firstly, the Sinkhorn distance was introduced which can solve problems such as KL divergence asymmetry and sparse data distribution in high-dimensional space, thereby constructing probability distributions that make low-dimensional space similar to high-dimensional space. In addition, to address the impact of multi-attribute features of samples on similarity measure, a multi-attribute distance matrix was constructed using information entropy, and then combined with the numerical matrix of spectral data to obtain a mixed data matrix. In order to validate the effectiveness of the St-SNE algorithm, dimensionality reduction projection was performed on NIR spectral data and compared with PCA, LPP, and t-SNE algorithms. The results demonstrated that the St-SNE algorithm effectively distinguishes samples with different attribute information, and produced more distinct projection boundaries of sample category in low-dimensional space. Then we tested the classification performance of St-SNE for different attributes by using the tobacco and mango datasets, and compared it with LPP, t-SNE, UMAP, and Fisher t-SNE algorithms. The results showed that St-SNE algorithm had the highest classification accuracy for different attributes. Finally, we compared the results of searching the most similar sample with the target tobacco for cigarette formulas, and experiments showed that the St-SNE had the highest consistency with the recommendation of the experts than that of the other algorithms. It can provide strong support for the maintenance and design of the product formula. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. Adaptive fuzzy neighborhood decision tree.
- Author
-
Cui, Xinyu, Wang, Changzhong, An, Shuang, and Qian, Yuhua
- Subjects
DATA structures ,FEATURE selection ,FUZZY sets ,ROUGH sets ,ENTROPY (Information theory) ,DECISION trees - Abstract
Decision tree algorithms have gained widespread acceptance in machine learning, with the central challenge lying in devising an optimal splitting strategy for node sample subspaces. In the context of continuous data, conventional approaches typically involve fuzzifying data or adopting a dichotomous scheme akin to the CART tree. Nevertheless, fuzzifying continuous features often entails information loss, whereas the dichotomous approach can generate an excessive number of classification rules, potentially leading to overfitting. To address these limitations, this study introduces an adaptive growth decision tree framework, termed the fuzzy neighborhood decision tree (FNDT). Initially, we establish a fuzzy neighborhood decision model by leveraging the concept of fuzzy inclusion degree. Furthermore, we delve into the topological structure of misclassified samples under the proposed decision model, providing a theoretical foundation for the construction of FNDT. Subsequently, we utilize conditional information entropy to sift through original features, prioritizing those that offer the maximum information gain for decision tree nodes. By leveraging the conditional decision partitions derived from the fuzzy neighborhood decision model, we achieve an adaptive splitting method for optimal features, culminating in an adaptive growth decision tree algorithm that relies solely on the inherent structure of real-valued data. Experimental evaluations reveal that, compared with advanced decision tree algorithms, FNDT exhibits a simple tree structure, stronger generalization capabilities, and superior performance in classifying continuous data. • Existing methods often split node samples either by preset fuzzy sets or dichotomy. • We propose a novel node-cutting method which only depends on data structure itself. • We introduce a fuzzy neighborhood decision model for constructing adaptive decision trees. • The proposed decision tree has a minimal rule set and robust generalization ability. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. An adaptive interval many-objective evolutionary algorithm with information entropy dominance.
- Author
-
Cui, Zhihua, Qu, Conghong, Zhang, Zhixia, Jin, Yaqing, Cai, Jianghui, Zhang, Wensheng, and Chen, Jinjun
- Subjects
OPTIMIZATION algorithms ,EVOLUTIONARY algorithms ,CONFIDENCE intervals ,ENTROPY (Information theory) ,PROBLEM solving - Abstract
Interval many-objective optimization problems (IMaOPs) involve more than three conflicting objectives with interval parameters. Various real-world applications under uncertainty can be modeled as IMaOPs to solve, so effectively handling IMaOPs is crucial for solving practical problems. This paper proposes an adaptive interval many-objective evolutionary algorithm with information entropy dominance (IMEA-IED) to tackle IMaOPs. Firstly, an interval dominance method based on information entropy is proposed to adaptively compare intervals. This method constructs convergence entropy and uncertainty entropy related to interval features and innovatively introduces the idea of using global information to regulate the direction of local interval comparison. Corresponding interval confidence levels are designed for different directions. Additionally, a novel niche strategy is designed through interval population partitioning. This strategy introduces a crowding distance increment for improved subpopulation comparison and employs an updated reference vector method to adjust the search regions for empty subpopulations. The IMEA-IED is compared with seven interval optimization algorithms on 60 interval test problems and a practical application. Empirical results affirm the superior performance of our proposed algorithm in tackling IMaOPs. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
17. Crucial time of emergency monitoring for reliable numerical pollution source identification.
- Author
-
Yang, Ruiyi, Jiang, Jiping, Pang, Tianrui, Yang, Zhonghua, Han, Feng, Li, Hailong, Wang, Hongjie, and Zheng, Yi
- Subjects
- *
WATER quality management , *CHEMICAL spills , *PECLET number , *MUNICIPAL water supply , *ENTROPY (Information theory) - Abstract
• Crucial time exists in the accumulation of monitoring data for Bayesian inversion. • Section number and location impact crucial time, but frequency and error level do not. • Relative crucial time determined by Peclet number and mainly controlled by dispersion. • Spatial structure of crucial time uncovered and explained by information entropy theory. • Novel design method on emergency monitoring largely improve PSI practicality. The Pollution source identification (PSI) is an important issue on river water quality management especially for urban receiving water. Numerical inversion method is theoretically an effective PSI technique, which employs monitored downstream pollutant breakthrough curves to identify the pollution source. In practice, it is important to know how much monitoring data should be accumulated to provide PSI results with acceptable accuracy and uncertainty. However, no literature reports on this key point and it seriously handers the numerical PSI technology to mature practical applications. To seek a monitoring guideline for PSI, we conducted extensively numerical experiments for single-point source instantaneous release taking Bayesian-MCMC method as the baseline inversion technique. The crucial time (T c) phenomenon was found during the data accumulation process for Bayesian source inversion. After T c , estimated source parameters subsequent sustained low error levels and uncertainty convergence. Results shown the presence of T c impacted by the number and location of monitoring sections, while monitoring frequency and data error do not. Under different river hydrodynamic conditions, relative crucial time (Λ) is determined by the river's Peclet number, and minimum effective Λ was controlled by dispersion coefficient (D x). Analytic spatial structure of Λ(U, D x) was uncovered and this relationship successfully explained by the information entropy theory. Based on these findings, a novel design method of PSI emergency monitoring network for preparedness plan and a practical framework of PSI for emergency response were established. These findings fill the important knowledge gap in PSI applications and the guidelines provide valuable references for river water quality management. [Display omitted] [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. IEA-GNN: Anchor-aware graph neural network fused with information entropy for node classification and link prediction.
- Author
-
Zhang, Peiliang, Chen, Jiatao, Che, Chao, Zhang, Liang, Jin, Bo, and Zhu, Yongjun
- Subjects
- *
ENTROPY (Information theory) , *INFORMATION networks , *REPRESENTATIONS of graphs , *LEARNING strategies , *POINT set theory , *CLASSIFICATION - Abstract
Graph neural networks are essential in mining complex relationships in graphs. However, most methods ignore the global location information of nodes and the discrepancy between symmetrically located nodes, resulting in the inability to distinguish between nodes with homogeneous network neighborhoods. We propose an Anchor-aware Graph Neural Network fused with Information Entropy (IEA-GNN) to capture the global location information of nodes in the graph. IEA-GNN first calculates the information entropy of nodes and constructs candidate sets of anchors. We define the calculation method of the distance from any node to the anchor points and incorporate the relative distance information between nodes at initialization. The nonlinear distance-weighted aggregation learning strategy based on the anchor points of candidate sets is used to obtain the nodes' feature information, which can be captured more effectively by fusing the global location information to the node representation with the selected anchor points. Selecting anchor points based on information entropy avoid the aggregation of anchor points in the graph, highlighting the positional differences between nodes and making it easier to distinguish homogeneous neighborhood nodes. Experimental results of node classification and link prediction on five datasets show that IEA-GNN outperforms the baseline model. • Enhancing graph representation learning with information entropy-based anchor selection. • Calculating nodes entropy with information contribution to avoid anchor point aggregation. • Introducing global location information to highlight the differences between nodes. • The performance of IEA-GNN is more stable than stochastic anchor point selection models. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
19. Information entropy approach to design adaptability evaluation.
- Author
-
Sun, Zhilin, Wang, Kaifeng, and Gu, Peihua
- Subjects
ENTROPY ,SOFT sets ,QUALITY function deployment ,ENTROPY (Information theory) ,MACHINE tools - Abstract
Adaptable design is an increasingly utilized design paradigm that can effectively and efficiently create a new design by adapting an existing design to maintain, upgrade and extend the functionality of a product by enhancing its adaptability. This paper presents an approach integrating soft sets and entropy theories to develop a mapping model between design requirements and design solutions to evaluate the adaptability when the design or product has insufficient design information. The design adaptabilities of cylindrical gear machine tools were used to validate the feasibility of the proposed approach and a modified design with improved adaptability is realized. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
20. Analysis of multimodal data fusion from an information theory perspective.
- Author
-
Dai, Yinglong, Yan, Zheng, Cheng, Jiangchang, Duan, Xiaojun, and Wang, Guojun
- Subjects
- *
MULTISENSOR data fusion , *INFORMATION theory , *PROCESS capability , *DATA analysis , *DEEP learning , *SPEECH perception - Abstract
[Display omitted] • Reduce the information fusion cases into an essential process of decreasing information entropy. • Provide basic definitions for multimodal data analysis and model uncertainty measurement. • Introduced a latent code embedding method to constrain the latent representation space. Inspired by the McGurk effect, studies on multimodal data fusion start with audio-visual speech recognition tasks. Multimodal data fusion research was not popular for a period of time because the capacity of traditional machine learning is limited. Recently, advances in deep learning techniques have provided new opportunities for multimodal data fusion. Powerful deep learning models have the capacity to process high-dimensional and complex multimodal data, and multimodal deep learning has the potential to process multimodal data at the human level. However, there is still a lack of theoretical analytical methods relating data information with model performance. In this work, we propose basic concepts and principles to gain insight into the process of multimodal data fusion from an information theory perspective. We analyze different multimodal data fusion cases, such as redundant, noisy, consistent, and contradictory data fusion. We define the model accuracy upper bound for multimodal tasks and prove that a multimodal model with an extra modal channel can perform better in theory when extra modal data provide more effective information for prediction. We explicitly inspect the latent representation space and analyze the information loss of the representation space transformation in deep learning for the first time. From a naive example to a multimodal deep learning example, we demonstrate the theoretical analysis method for evaluating a multimodal data fusion model, and the experimental results validate the definitions and principles. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
21. An access control model for medical big data based on clustering and risk.
- Author
-
Jiang, Rong, Han, Shanshan, Yu, Yimin, and Ding, Weiping
- Subjects
- *
ACCESS control , *BIG data , *ENTROPY (Information theory) , *QUALITY of service , *AT-risk behavior - Abstract
• Proposes a new manner of spectral clustering to decrease the sensitivity • Gives an algorithm to calculate information entropy so as to value the risk ranks. • Evaluates SC-RBAC by the datasets and it is robust and efficient Access control has been widely adopted by distributed platforms, and its effectiveness is of great importance to the quality of services provided by such platforms. However, traditional access control is difficult to apply to scenarios where authorization changes frequently and to extremely large-scale datasets with limited resources. This paper proposes an access control model based on spectral clustering (SC) and risk (SC-RBAC), which is more suitable for big data medical scenarios. Based on user history access data, an improved SC algorithm is used to cluster doctor users. Then, the user classification is introduced as a parameter into the information entropy to improve the accuracy of quantifying the user's access behavior risk. Finally, based on the accurate risk value of access behavior, we assign access rights to users through the access control function constructed in the paper. Experimental results show that in three different situations, the model proposed in this paper can distinguish the two types of doctors well, the accuracy of the model can reach more than 90%, and it outperforms other access control models. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
22. CSF: Closed-mask-guided semantic fusion method for semantic perception of unknown scenes.
- Author
-
Chen, Minghong, Shu, Ruijun, Zhu, Dongchen, Li, Jiamao, and Zhang, Xiaolin
- Subjects
- *
ENTROPY (Information theory) , *CONFIDENCE , *AUDITORY masking - Abstract
• A mask-guided semantic fusion method for optimizing the results of semantic segmentation. • A closed mask generation (CEG) module for post-processing of the edge detection algorithm. • A semantic confidence fusion (SCF) module to fuse the semantic segmentation results. • This method does not need to obtain the target data in advance. • This method has been verified in two scenarios. Though many high-precision semantic segmentation models have been proposed, how to improve the generalization ability of these models is still an urgent problem. Recently, a great number of unsupervised domain adaptation (UDA) algorithms around domain adaptation problems have been studied in semantic segmentation. These methods require labeled source domain data and unlabeled target domain data. In this paper, we propose a closed-mask-guided semantic fusion method (CSF) to improve the semantic segmentation results of unknown scenes, where the target domain data is not obtained in advance. First, a Closed Mask Generation (CMG) module is designed to convert the edge detection result into a mask that can segment the image into several image blocks. Then, a Semantic Confidence Fusion (SCF) module based on information entropy and voting method is introduced, which can select reliable semantic segmentation results for each image block by comparing the confidence of several semantic segmentation networks. In addition, the experimental results on both KITTI and COCO Stuff datasets validate the effectiveness of this method. The code is publicly available at https://github.com/tryhere/CSF. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
23. Probabilistic summarization via importance-driven sampling for large-scale patch-based scientific data visualization.
- Author
-
Yang, Yang, Wu, Yu, and Cao, Yi
- Subjects
- *
SCIENTIFIC visualization , *DATA visualization , *ENTROPY (Information theory) , *DATA reduction - Abstract
Probabilistic summarization is the process of creating compact statistical representations of the original data. It is used for data reduction, and to facilitate efficient post-hoc visualization for large-scale patch-based data generated in parallel numerical simulation. To ensure high reconstruction accuracy, existing methods typically merge and repartition data patches stored across multiple processor cores, which introduces time-consuming processing. Therefore, this paper proposes a novel probabilistic summarization method for large-scale patch-based scientific data. It considers neighborhood statistical properties by importance-driven sampling guided by the information entropy, thus eliminating the requirement of patch merging and repartitioning. In addition, the reconstruction value of a given spatial location is estimated by coupling the statistical representations of each data patch and the sampling results, thereby maintaining high reconstruction accuracy. We demonstrate the effectiveness of our method using five datasets, with a maximum grid size of one billion. The experimental results show that the method presented in this paper reduced the amount of data by about one order of magnitude. Compared with the current state-of-the-art methods, our method had higher reconstruction accuracy and lower computational cost. [Display omitted] • A novel probabilistic summarization method for large-scale patch-based scientific data. • The neighborhood statistical properties are considered by importance-driven sampling guided by the information entropy. • Compared with the current state-of-the-art methods, our method had higher reconstruction accuracy and lower computational cost. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
24. Image based information hiding via minimization of entropy and randomization.
- Author
-
Zhao, Xuemeng and Song, Yinglei
- Subjects
- *
BINARY sequences , *ENTROPY (Information theory) , *DYNAMIC programming , *IMAGING systems , *PIXELS , *INPAINTING - Abstract
In this paper, a new approach that can effectively and securely hide information into color images with significantly improved security and hiding capacity is proposed. The proposed approach performs information hiding in three major steps. As the first step, two binary sequences are constructed from the least significant bits in the pixels of a cover image and the information that needs to be embedded, the information entropies of both sequences are minimized with a dynamic programming method. In the second step, the resulting sequences are randomly reshuffled into randomized sequences with mappings based on a set of one-dimensional chaotic systems, a single binary sequence can be obtained by a matching operation performed between the two randomized sequences. Finally, an inverse mapping is applied to the sequence obtained in the second step, and the transformed sequence is embedded into the least significant bits in the pixels of the cover image. Both analysis and experiments show that the proposed approach can achieve guaranteed performance in both security and capacity for long binary sequences. In addition, a comparison with other state-of-the-art methods for image-based information hiding suggests that the proposed approach can achieve significantly improved performance and is promising for practical applications. [ABSTRACT FROM AUTHOR]
- Published
- 2025
- Full Text
- View/download PDF
25. Eigen-Entropy: A metric for multivariate sampling decisions.
- Author
-
Huang, Jiajing, Yoon, Hyunsoo, Wu, Teresa, Candan, Kasim Selcuk, Pradhan, Ojas, Wen, Jin, and O'Neill, Zheng
- Subjects
- *
ENTROPY (Information theory) , *DATA distribution , *STATISTICAL correlation , *EIGENVALUES , *SAMPLING (Process) - Abstract
Sampling is a technique to help identify a representative data subset that captures the characteristics of the whole dataset. Most existing sampling algorithms require distribution assumptions of the multivariate data, which may not be available beforehand. This study proposes a new metric called Eigen-Entropy (EE), which is based on information entropy for the multivariate dataset. EE is a model-free metric because it is derived based on eigenvalues extracted from the correlation coefficient matrix without any assumptions on data distributions. We prove that EE measures the composition of the dataset, such as its heterogeneity or homogeneity. As a result, EE can be used to support sampling decisions, such as which samples and how many samples to consider with respect to the application of interest. To demonstrate the utility of the EE metric, two sets of use cases are considered. The first use case focuses on classification problems with an imbalanced dataset, and EE is used to guide the rendering of homogeneous samples from minority classes. Using 10 public datasets, it is demonstrated that two oversampling techniques using the proposed EE method outperform reported methods from the literature in terms of precision, recall, F-measure, and G-mean. In the second experiment, building fault detection is investigated where EE is used to sample heterogeneous data to support fault detection. Historical normal datasets collected from real building systems are used to construct the baselines by EE for 14 test cases, and experimental results indicate that the EE method outperforms benchmark methods in terms of recall. We conclude that EE is a viable metric to support sampling decisions. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
26. Optimization to identify the adapted product design and product adaptation process with initial evaluation of information quality in branches of AND-OR tree based on information entropy.
- Author
-
Saad, Mohammed and Xue, Deyi
- Subjects
- *
TREE branches , *ENTROPY (Information theory) , *PRODUCT design , *PROCESS optimization , *PRODUCTION planning - Abstract
Adaptable products are used to satisfy new customer requirements through changes of products such as upgrading of modules during their utilization stages. Compared with the creation of a new product from scratch, the process to adapt an existing product to a new one can result in savings for the customer. In this research, a new optimization approach to identify the adapted product design and product adaptation process is developed with initial evaluation of information quality in branches of AND-OR tree based on information entropy. In this approach, various candidates of design configurations and operation processes for adaptation of a product are modelled by an AND-OR tree with design and process nodes. Design and process nodes are further associated with design and process parameters. Optimization to identify the best adapted design and adaptation process is conducted at two levels: configuration/process optimization level and parameter optimization level. Information entropy is employed to evaluate the information quality of partial configuration/process candidates modelled as branches in the AND-OR tree to eliminate the branches that are unlikely to lead to the optimal solution for improvement of the optimization efficiency. A numerical example and an engineering application are developed to demonstrate the effectiveness of the newly introduced approach. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
27. A real-time measurement and analysis method for gas holdup in a wet scrubber with the use of image information entropy.
- Author
-
Zhao, Lei, You, Ruoyu, Liu, Junjie, and Chen, Qingyan
- Subjects
- *
SCRUBBER (Chemical technology) , *ENTROPY (Information theory) , *GAS analysis , *NON-uniform flows (Fluid dynamics) , *CHEMICAL purification - Abstract
• A fast real-time method for measurement of gas holdup characteristics with image information entropy. • The global and local profiles of gas holdup could be collected simultaneously. • Sieve plate resistance and bubbling state were analyzed with the use of image information entropy. Wet scrubbers are often used in various types of industrial workshops to separate dust particles from the air. Non-uniform gas–liquid flow patterns above the sieve plate can result in significant losses in purification efficiency for particles. To improve the flow pattern via sieve plate design and to more accurately predict purification efficiency, a better understanding of the gas holdup is required. Various methods have been developed to measure and analyze the gas holdup from local and global perspectives. However, for guiding the operation and design of reactors, it is obviously insufficient to obtain the phase holdup characteristics under a steady state. The spatial and temporal characteristics of gas holdup in a wet scrubber are rather difficult to measure in real time with the current techniques. In this study, a fast real-time method for measurement of gas holdup characteristics was proposed, based on image information entropy analysis. The axial and radial profiles of gas holdup under different bubble washing operation statuses were obtained and are discussed here. More detailed information about the gas holdup above the sieve plate under various operation conditions has been attained. Sieve plate resistance and bubbling state were analyzed with the use of image information entropy. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
28. A self-adaptive density-based clustering algorithm for varying densities datasets with strong disturbance factor.
- Author
-
Cai, Zihao, Gu, Zhaodong, and He, Kejing
- Subjects
- *
TEXT mining , *ENTROPY (Information theory) , *DATA mining , *DATA distribution , *IMAGE analysis - Abstract
Clustering is a fundamental task in data mining, aiming to group similar objects together based on their features or attributes. With the rapid increase in data analysis volume and the growing complexity of high-dimensional data distribution, clustering has become increasingly important in numerous applications, including image analysis, text mining, and anomaly detection. DBSCAN is a powerful tool for clustering analysis and is widely used in density-based clustering algorithms. However, DBSCAN and its variants encounter challenges when confronted with datasets exhibiting clusters of varying densities in intricate high-dimensional spaces affected by significant disturbance factors. A typical example is multi-density clustering connected by a few data points with strong internal correlations, a scenario commonly encountered in the analysis of crowd mobility. To address these challenges, we propose a Self-adaptive Density-Based Clustering Algorithm for Varying Densities Datasets with Strong Disturbance Factor (SADBSCAN). This algorithm comprises a data block splitter, a local clustering module, a global clustering module, and a data block merger to obtain adaptive clustering results. We conduct extensive experiments on both artificial and real-world datasets to evaluate the effectiveness of SADBSCAN. The experimental results indicate that SADBSCAN significantly outperforms several strong baselines across different metrics, demonstrating the high adaptability and scalability of our algorithm. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
29. An entropy-based causality framework for cross-level faults diagnosis and isolation in building HVAC systems.
- Author
-
Huang, Jiajing, Ghalamsiah, Naghmeh, Patharkar, Abhidnya, Pradhan, Ojas, Chu, Mengyuan, Wu, Teresa, Wen, Jin, O'Neill, Zheng, and Selcuk Candan, Kasim
- Subjects
- *
FAULT diagnosis , *BAYESIAN analysis , *CAUSAL inference , *BUILDING performance , *ENTROPY (Information theory) - Abstract
• Entropy-based causality learning framework is proposed for building fault diagnosis. • The proposed synchronicity concept is employed to measure building symptom correlations. • Causal learning aims to determine Bayesian network structure. • Modelica test cases demonstrate the framework's effectiveness. Faults, such as malfunctioning sensors, equipment, and control systems, significantly affect a building's performance. Automatic fault detection and diagnosis (AFDD) tools have shown great potential in improving building performances, including both energy efficiency and indoor environment quality. Since modern buildings have integrated systems where multiple subsystems and equipment are coupled, many faults in a building are cross-level faults, i.e., faults occurring in one component that trigger operational abnormalities in other subsystems. Compared with non-cross-level faults, it is more challenging to isolate the root cause of a cross-level faults due to the system coupling effects. Bayesian networks (BNs) have been studied for the root cause isolation for building faults. While promising, existing BN-based diagnosis methods highly rely on expert domain knowledge, which is time-consuming and labor expensive, especially for cross-level faults. To address this challenge, we propose an entropy-based causality learning framework, termed Eigen-Entropy Causal Learning (EECL), to learn BN structures. The proposed method is data-driven without the use of expert domain knowledge; it utilizes causal inference to determine the causal mechanisms between faults status and symptoms to construct the BN model. To demonstrate the effectiveness of the proposed framework, three fault test cases are used for evaluation in this study. Experimental results show that the BN constructed by the proposed framework is able to conduct building cross-level faults diagnosis with a comparable isolation accuracy to those by domain knowledge while maintaining less complexed BN structure. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
30. Engineering complete delocalization of single particle states in a class of one dimensional aperiodic lattices: A quantum dynamical study.
- Author
-
Biswas, Sougata and Chakrabarti, Arunava
- Subjects
- *
QUANTUM theory , *FLUX pinning , *WAVE packets , *TRANSITION metals , *MAGNETIC flux , *METAL-insulator transitions - Abstract
We study quantum dynamics of a wave packet on a class of one dimensional decorated aperiodic lattices, described within a tight binding formalism. We look for the possibility of finding extended single particle states even in the absence of any translational periodicity. The chosen lattices are stubbed with one or more atoms, tunnel coupled to the backbone, thereby introducing a minimal quasi-one dimensionality. It is seen that, for a group of such lattices a certain correlation between the numerical values of the hopping amplitudes leads to a complete delocalization of single particle states. In some other cases, a special value of a magnetic flux trapped in the loops present in the geometries delocalize the states, leading to a flux driven insulator to metal transition. The mean square displacement, temporal autocorrelation function, the time dependence of the inverse participation ratio, or the information entropy – the so-called hallmarks of studying localization based on dynamics – all of them indicate such a complete turnover in the nature of the single particle states and the character of the energy spectrum under suitable conditions. The results shown in this work using quasiperiodic lattices of the Fibonacci family are much more general and hold good even for a randomly disordered arrangement of the building blocks of the systems considered, and indicate a subtle universality class under which these lattices can be grouped. • We examine the quantum dynamics of a wave packet on a family of one-dimensional aperiodic lattices. • It is observed that a specific connection between the hopping amplitudes for a set of such lattices results in a total delocalization of single particle states. • In other instances, the states are delocalized due to a unique value of a magnetic flux that is trapped in the loops found in the geometries. • We analyze the mean square displacement, temporal autocorrelation function, time dependence of the inverse participation ratio, and information entropy. • The findings presented here with quasiperiodic lattices from the Fibonacci family are far more broadly applicable. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
31. A new active-learning function for adaptive Polynomial-Chaos Kriging probability density evolution method.
- Author
-
Zhou, Tong and Peng, Yongbo
- Subjects
- *
ACTIVE learning , *PROBABILITY density function , *ENTROPY (Information theory) , *KRIGING , *STRUCTURAL reliability , *INFORMATION theory , *GEOLOGICAL statistics - Abstract
• Notation of region of interest with critical contribution to failure probability is proposed. • A new active-learning function is devised based on the theory of information entropy. • Sufficient PC-Kriging accuracy in the range of the region of interest can be secured. • Computational burden can be significantly alleviated by the new active-learning function. • The proposed approach outperforms other adaptive surrogate-assisted simulation methods. A new active-learning function is developed to integrate Polynomial-Chaos Kriging with probability density evolution method. First, the relative importance of each representative point to the probability of failure is separately measured, thereby the region of interest is defined for the probability density evolution method, which covers the representative points exerting vital contributions to the probability of failure. Then, a new learning function called the probability density evolution method-oriented information entropy is readily devised based on the information theory, and the stopping condition is defined by specifying a threshold for the learning function values. Two examples are studied to show the efficacy of the adaptive Polynomial-Chaos Kriging probability density evolution method, and the recommended values of two key parameters associated with this new learning function are provided. Moreover, comprehensive comparisons are conducted against several existing reliability methods. The results highlight the advantage of the proposed active-learning function for structural reliability analysis in terms of both computational accuracy and efficiency. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
32. Information mining of customers preferences for product specifications determination using big sales data.
- Author
-
Zhang, Jian, Lin, Peihuang, and Simeone, Alessandro
- Abstract
Product competitiveness is highly influenced by its related design specifications. Information retrieval of customers preferences for the specification determination is essential to product design and development. Big sales data is an emerging resource for mining customers preferences on product specifications. In this work, information entropy is used for customers preferences information quantification on product specifications firstly. Then, a method of information mining for customers preferences estimation is developed by using big sales data. On this basis, a density-based clustering analysis is carried out on customers preferences as a decision support tool for the determination and selection of product design specifications. A case study related to electric bicycle specifications determination using big sales data is reported to illustrate and validate the proposed method. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
33. Dynamic information fusion in multi-source incomplete interval-valued information system with variation of information sources and attributes.
- Author
-
Zhang, Xiaoyan, Chen, Xiuwei, Xu, Weihua, and Ding, Weiping
- Subjects
- *
INFORMATION resources , *INFORMATION storage & retrieval systems , *ENTROPY (Information theory) , *BIG data , *HUFFMAN codes - Abstract
Interval-valued data describe the random phenomenon that abounds in the real world, a pivotal research orientation in uncertainty processing. With the rapid development of big data, we may gather information from multiple information sources. To effectively acquire knowledge from multiple information sources, information fusion is commonly used to get a unified representation. However, sometimes data gathered from multiple sources may be lost; it is meaningful and necessary to study the fusion of multi-source incomplete interval-valued data. We propose a novel information fusion method based on information entropy for multi-source incomplete interval-valued data and four incremental fusion mechanisms characterized by the change in information sources and attributes. The corresponding static and dynamic fusion algorithms are designed, and their time complexities are analyzed. Experimental results show that the proposed method outperforms the mean, max, and min fusion methods. Furthermore, the four incremental fusion mechanisms reduced the runtime compared with the static fusion mechanism. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
34. Optimal scale combination selection based on genetic algorithm in generalized multi-scale decision systems for classification.
- Author
-
Yang, Ying, Zhang, Qinghua, Zhao, Fan, Cheng, Yunlong, Xie, Qin, and Wang, Guoyin
- Subjects
- *
GENETIC algorithms , *SUBSET selection , *ENTROPY (Information theory) , *ROUGH sets , *MACHINE learning - Abstract
Optimal scale combination (OSC) selection plays a crucial role in multi-scale decision systems for data mining and knowledge discovery, and its aim is to select an appropriate subsystem for classification or decision-making while keeping a certain consistency criterion. Selecting the OSC with existing methods requires judging the consistency of all multi-scale attributes; however, judging consistency and selecting scales for unimportant multi-scale attributes increases the selection cost in vain. Moreover, the existing definitions of OSC are only applicable to rough set classifiers (RSCs), which makes the selected OSC perform poorly on other machine learning classifiers. To this end, the main objective of this paper is to investigate multi-scale attribute subset selection and OSC selection applicable to any classifier in generalized multi-scale decision systems. First, a novel consistency criterion based on the multi-scale attribute subset is proposed, which is called p -consistency criterion. Second, the relevance and redundancy among multi-scale attributes are measured based on the information entropy, and an algorithm for selecting the multi-scale attribute subset is given based on this. Third, an extended definition of OSC, called the accuracy OSC, is proposed, which can be widely applied to classification tasks using any classifier. On this basis, an OSC selection algorithm based on genetic algorithm is proposed. Finally, the results of many experiments show that the proposed method can significantly improve the classification accuracy and selection efficiency. [ABSTRACT FROM AUTHOR]
- Published
- 2025
- Full Text
- View/download PDF
35. Strategic node identification in complex network dynamics.
- Author
-
Nikougoftar, Elaheh
- Subjects
- *
ENTROPY (Information theory) , *ENTROPY , *RUMOR , *SCALABILITY , *ADVERTISING - Abstract
Detecting significant nodes in intricate networks is essential for various purposes, including market advertising, rumor management, and predicting scientific publications. Existing algorithms, from basic degree methods to more complex approaches, have been developed, but there is a need for a more robust solution. Traditional methods often focus on local network details, neglecting global aspects. This study introduces a network structure entropy-based node importance ranking method that considers both local and global information. The method's efficacy is validated through comparisons with three benchmarks, showcasing strong performance on two real-world datasets. Further work could explore scalability and applicability in dynamic scenarios. • Introducing a node importance ranking method based on network structure entropy. • Integrating local and global structural information to develop a comprehensive scoring mechanism. • The construction of the score function is oriented toward the perspective of node removal. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
36. A novel method to attribute reduction based on weighted neighborhood probabilistic rough sets.
- Author
-
Xie, Jingjing, Hu, Bao Qing, and Jiang, Haibo
- Subjects
- *
ROUGH sets , *DATA binning , *INFORMATION theory , *NEIGHBORHOODS , *INFORMATION storage & retrieval systems - Abstract
Attribute reduction is an important application of rough set theory. Most existing rough set models do not consider the weight information of attributes in information systems. In this paper, we first study the weight of each attribute in information systems by using data binning method and information entropy theory. Then, we propose a new rough set model of weighted neighborhood probabilistic rough sets (WNPRSs) and investigate its basic properties. Meanwhile, the dependency degree formula of an attribute relative to an attribute subset is defined based on WNPRSs. Subsequently, we design a novel attribute reduction method by using WNPRSs and the corresponding algorithm is also given. Finally, to evaluate the performance of the proposed algorithm, we conduct a data experiment and compare it with other existing attribute reduction algorithms on eight public datasets. Experimental result demonstrates that the proposed attribute reduction algorithm is effective and performs better than some of the existing algorithms. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
37. Effects of unsaturated fatty acid methyl esters on the oxidation stability of biodiesel determined by gas chromatography-mass spectrometry and information entropy methods.
- Author
-
Wang, Wenchao, Liu, Huili, Li, Fashe, Wang, Hua, Ma, Xin, Li, Jingjing, Zhou, Li, and Xiao, Quan
- Subjects
- *
FATTY acid methyl esters , *UNSATURATED fatty acids , *BIODIESEL fuels , *OXIDATION , *ENTROPY (Information theory) - Abstract
To explore the trend of fatty acid methyl esters (FAMEs) during the oxidation process of biodiesel, this study uses gas chromatography - mass spectrometry (GC-MS) combined with information entropy theory to propose a new method to analyze the effect of FAMEs on the stability of biodiesel (especially with different amounts of unsaturated FAMEs). The induction periods of the main FAMEs in biodiesel were determined by the Rancimat method. The results showed that the longest induction period of methyl stearate was 248 h, while the induction of methyl linoleate was only 0.12 h. Accelerated oxidation of biodiesel, using GC-MS analysis showed that the oxidation trends of FAMEs was different. The oxidation rate of methyl linoleate was the fastest. A mathematical equation was proposed to calculate the trend of FAMEs. The results showed that during the 5 h accelerated an oxidation process, the content of methyl linoleate decreased from 28.35% to 2.10%, while the content of methyl oleate decreased by 18.21%. Using information entropy, the weighting coefficients of FAMEs in the oxidation process of biodiesel were calculated. The weighting coefficient of methyl linoleate was as high as 0.6797. Methyl linoleate was the main reason why biodiesel was easily oxidized. • GC-MS and information entropy methods was employed for FAMEs oxidation evaluation. •The weight coefficients of different FAMEs in the oxidation process are obtained. •Methyl linoleate is the main internal reason that affects the oxidation of biodiesel. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
38. Attribute granules-based object entropy for outlier detection in nominal data.
- Author
-
Liu, Chang, Peng, Dezhong, Chen, Hongmei, and Yuan, Zhong
- Subjects
- *
OUTLIER detection , *ASSOCIATION rule mining , *LATTICE theory , *ENTROPY , *GRANULAR computing , *ENTROPY (Information theory) - Abstract
Concept lattice theory, which is one of the key mathematical models of granular computing, is capable of successfully dealing with uncertain information in nominal data. It has been applied to machine learning tasks such as data reduction, classification, and association rule mining. For the problem of outlier detection in nominal data, this paper presents a concept lattice theory-based approach for detecting outliers in nominal data. First, subcontexts and concept lattices based on subsets of objects are discussed. Then, information entropy is introduced into the formal context, and an object entropy based on attribute granules is proposed. Finally, a nominal data-oriented outlier detection method is explored based on the proposed object entropy. The experimental results show that the proposed detection method can effectively detect outliers in nominal data. Besides, the results of the hypothesis testing indicate that the proposed method is statistically significantly different from the other methods. The code is publicly available online at https://github.com/from-china-to/OEOD. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
39. CEDR: Contrastive Embedding Distribution Refinement for 3D point cloud representation.
- Author
-
Yang, Feng, Cao, Yichao, Xue, Qifan, Jin, Shuai, Li, Xuanpeng, and Zhang, Weigong
- Subjects
- *
POINT cloud , *CLOUD computing - Abstract
The distinguishable deep features are essential for the 3D point cloud recognition as they influence the search for the optimal classifier. Most existing point cloud classification methods mainly focus on local information aggregation while ignoring the feature distribution of the whole dataset that indicates more informative and intrinsic semantic relationships of labeled data, if better exploited, which could learn more distinguishing inter-class features. Our work attempts to construct a more distinguishable feature space through performing feature distribution refinement inspired by contrastive learning and sample mining strategies, without modifying the model architecture. To explore the full potential of feature distribution refinement, two modules are involved to boost exceptionally distributed samples distinguishability in an adaptive manner: (i) Confusion-Prone Classes Mining (CPCM) module is aimed at hard-to-distinct classes, which alleviates the massive category-level confusion by generating class-level soft labels; (ii) Entropy-Aware Attention (EAA) mechanism is proposed to remove influence of the trivial cases which could substantially weaken model performance. Our method achieves competitive results on multiple applications of point cloud. In particular, our method gets 85.8% accuracy on ScanObjectNN, and substantial performance gains up to 2.7% in DCGNN, 3.1% in PointNet++, and 2.4% in GBNet. Our code is available at https://github.com/YangFengSEU/CEDR. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
40. AK-SYS-IE: A novel adaptive Kriging-based method for system reliability assessment combining information entropy.
- Author
-
Yuan, Kai, Sui, Xi, Zhang, Shijie, Xiao, Ning-cong, and Hu, Jinghan
- Subjects
- *
STRUCTURAL reliability , *INFORMATION theory , *RELIABILITY in engineering , *ENTROPY (Information theory) , *FINITE element method , *KRIGING - Abstract
• Information entropy is first used for Kriging-based system reliability assessment. • The prediction uncertainty of the system state is quantified by information entropy. • A learning function is proposed using the prediction uncertainty of the system state. • Sample with the largest uncertainty get more attention for model reconstruction. Structural reliability assessment is a popular topic in engineering problems, particularly for the larger and more complex systems with implicit performance functions, also called black-box problems. The reliability assessment for a black-box problem must continuously be computed using simulation models, such as the finite element model, a highly time-consuming process with a high computational cost. The adaptive Kriging has gained considerable attention over the past decade. The Kriging-based reliability assessment method reduces the computational cost to a great extent, on the premise of ensuring the accuracy of reliability assessment. However, many of the currently published system reliability assessment methods, construct adaptive Kriging models by reducing the probability of incorrect prediction of the system state, and do not make full use of the uncertainty of the system state prediction information. To this end, a new Kriging-based method for structural system reliability assessment is proposed in this study. First, the probabilities of incorrect or correct system state predictions were understood from the perspective of information entropy. Second, an active learning strategy is proposed based on information entropy theory. Finally, the advantages of the proposed method are demonstrated and highlighted through several numerical examples. The results show that the proposed method achieves a good balance between the accuracy and computational cost, and the numerical magnitude effect does not affect the computational cost. Moreover, this is an effective method for assessing the reliability of complex systems. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
41. Big data-assisted urban governance: A comprehensive system for business documents classification of the government hotline.
- Author
-
Zhang, Zicheng, Li, Anguo, Wang, Li, Cao, Wei, and Yang, Jianlin
- Subjects
- *
BIG data , *NETWORK governance , *GOVERNMENT publications , *NEW words , *DATA structures - Abstract
The government service platform, exemplified by the government hotline, has to handle extensive volumes of business documents that contain rich and timely public opinion information and citizens' demands. However, manual processing struggles to process large-scale text data, adversely impacting operating costs and the quality of government services. This study proposes a comprehensive system for business document classification of the government hotline (BDCGHS) in China to address these challenges. BDCGHS leverages information entropy fused with term frequency-inverse document frequency (TF-IDF) weight to mine new words from business documents of the government hotline, and store them in a new word repository. These new words optimize Chinese word segmentation and text representation for text classification. We introduce a novel data structure called nested balanced binary tree to expedite new word mining, yielding a computational speed of almost five times than the Trie trees. Comparative experiments on the THUNews and government hotline datasets validate our proposed improvement BDCGHS algorithm's superior performance 3 % over text classification algorithms. Compared to the latest bidirectional encoder representations from the transformers (BERT) model, BDCGHS enhances the accuracy of order dispatch based on business documents by almost 3 %. It has also demonstrated stable operations in two Chinese cities for over a year, yielding favorable results. [Display omitted] • An embedded balanced binary tree structure is proposed for new word discovery. • A new word database is constructed for the government hotline. • The effects of mainstream classification algorithms are compared based on a new word database. • An intelligent text classification system for the government hotline is constructed and the results have promising results. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
42. Learning bayesian network parameters from limited data by integrating entropy and monotonicity.
- Author
-
Fan, Zhiping, Zhou, Liang, Komolafe, Temitope Emmanuel, Ren, Zhengyun, Tong, Yinghao, and Feng, Xue
- Subjects
- *
MACHINE learning , *TIME complexity , *BAYESIAN analysis , *ENTROPY (Information theory) , *INFORMATION networks , *UNCERTAIN systems - Abstract
High-accuracy parameter learning in Bayesian Networks (BNs) is a key challenge in real-time decision support applications, particularly when the available data are limited. Prior/Expert knowledge was introduced to eliminate the drawbacks of insufficient information; however, this method is subjective. In this study, we explored the use of monotonicity constraints to control the causal relationships between the nodes and their parents-nodes in BNs and proposed a new learning algorithm called global domain monotonicity based on maximum information entropy (GDM-MIE), which was designed for parameter learning in uncertain discrete BNs with nonlinear equality constraints when only finite data are available. In the proposed algorithm, a class of monotonicity is encoded as a constraint on information entropy and the parameter learning problem is transformed into constraints among the node parameters based on a known network. Furthermore, we considered the parameters as uncertain entropy information and discussed the monotonicity among parameters in the global spatial domain, proving the accuracy of the logical relationships of the model, system reliability, and time complexity. Finally, the proposed method was validated using standard BNs, and its performance was analyzed by comparing the proposed method with the existing learning algorithms. The results showed that the proposed method is more accurate and has better Kullback–Leibler divergence. To revalidate the rationality of the proposed method, the Alarm and Asia networks were employed as special cases. The GDM-MIE was found to achieve the intended goal of observing the estimated parameters by closely approximating the original real parameters with a small sample size, indicating that the proposed algorithm can served as an efficient and feasible method for learning Bayesian parameter. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
43. A two-way accelerator for feature selection using a monotonic fuzzy conditional entropy.
- Author
-
Yang, Yanyan, Chen, Degang, Ji, Zhenyan, Zhang, Xiao, and Dong, Lianjie
- Subjects
- *
FEATURE selection , *ENTROPY , *ROUGH sets , *FUZZY sets - Abstract
Fuzzy rough set is a highly effective mathematical method for feature selection, which offers clear interpretability without expert knowledge. However, most of fuzzy-rough feature selection methods are to rely on all samples and candidate features during the selection of a best feature at each iteration. This often shares high computation complexity and is inefficient for large datasets. Therefore, a two-way accelerator for feature selection is presented by stepwise narrowing the search space for both samples and features. As the foundation for our accelerator, a monotonic fuzzy conditional entropy, named parameterized fuzzy granule-based conditional entropy, is first proposed to guide the feature selection process. After identifying a best feature, a sample-based accelerator is then designed to disregard redundant samples for the calculation of the newly defined entropy. A feature-based accelerator is further proposed to eliminate redundant candidate features, of which the inclusion cannot change the proposed entropy of a currently selected feature subset. Our accelerator is finally developed by integrating the sample-based accelerator with the feature-based accelerator. Experimental comparisons demonstrate the effectiveness and efficiency of the proposed two-way accelerator for feature selection. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
44. Sparse measure of bearing fault features based on Legendre wavelet multi-scale multi-mode Entropy.
- Author
-
Zheng, Xiaoyang, Huang, Yan, Xin, Yu, Zhang, Zhiyu, Liu, Weishuo, and Liu, Dezhi
- Subjects
- *
ROTATING machinery , *ENTROPY , *FAULT diagnosis , *GENETIC algorithms , *MAINTENANCE costs , *MACHINERY safety , *ENTROPY (Information theory) - Abstract
The efficient diagnosis of the bearing fault categories is vitally significant to enhance the overall safety of rotating machinery and decrease the maintenance cost. This paper introduces an innovative sparse measure approach for the bearing fault characteristics using Legendre wavelet multi-scale multi-mode (LWMSMM) frequency domain. The motivation behind this approach lies in the need for an effective method that overcomes complexities associated with extracting transient characteristics and frequency-related bearing fault features. The core principle of the proposed method is to leverage information Entropy (IE) to pick up the most sensitive fault features and utilize the genetic algorithm (GA) to optimize the control parameters of LWMSMM frame, including the decomposition level and the number of wavelet base, so as to thoroughly match the more salient fault characteristics of the bearing. The essence of the effectiveness of the method lies in two aspects. One is that the rich properties especial diverse regularities of LW can effectively approximate the complex fault characteristics. The other is that IE can effectively represent the dynamic characteristics of the faults at each resolution scale and each mode frequency domain. Finally, experiments are conducted on three datasets to validate the effectiveness and robustness of the presented method. The experimental results demonstrate that the developed method can accurately identify different fault categories with the simpler classifiers and achieves a diagnosis accuracy of 100 %, surpassing cutting-edge approaches, and it provides a new and promising method for rotating machinery in real industrial applications. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
45. Information entropy analysis of the relation between climate and thermal adaptation: A case study in hot summer and cold winter region of China.
- Author
-
Jing, Shenglan, Lei, Yonggang, Song, Chongfang, and Wang, Fei
- Abstract
The shift of paradigm from reductionist to holism in the field of thermal adaptation promoted a boom in thermal adaptation theory and provided a systemic perspective on this issue. The thermal adaptation theory confronts twin difficulties of scale and complexity. The complexity and entropy of occupants' thermal adaptation emerging. This paper aims to investigate the information entropy exchange in the coupled climate-building-occupants system. First, we justified that the building-occupants system is an open system and exchanges energy, matter, and entropy with its surrounding climate system. Then, correlation analysis results suggested that climate entropy could influence thermal adaptation entropy. Thermal adaptation entropy fluctuated with climate entropy and the difference varied between −0.5 and 0.5 below 31 °C. A generalized entropy balance model was proposed for quantifying the entropy exchange among coupled climate-building-occupants systems for both free-running and air-conditioning buildings. The quantitative relation between the entropy of thermal adaptation and H_DTR reveals the dynamic occupant's thermal adaptation under the process of entropy exchange with the climate system. The H_DTR and entropy relation models provide a new perspective on thermal adaptation coupled with climate beyond outdoor daily mean temperatures. Control strategies that integrate the dynamic characteristics of DTR can bring more economical, energy-efficient and comfortable effects to the operation of air conditioning or heating systems. Our findings shed some light on the dynamical and complex thermal adaptation. In further studies, more results and methods from meteorological research could be used for thermal adaptation research and interdisciplinary theories could be adopted to help us better understand occupants' thermal adaptation. • The building-occupants system is an open system and exchange energy, matter, and entropy with the climate system. • The entropy of SCV and Clo show a strong positive relationship with DTR entropy. • Thermal adaptation entropy fluctuated with DTR entropy and the difference varied between −0.5 and 0.5 below 31 °C. • The entropy balance model was proposed for quantifying the entropy exchange. • Different H_DTR may lead to the difference in acceptable temperature ranges. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
46. A design of fuzzy rule-based classifier optimized through softmax function and information entropy.
- Author
-
Han, Xiaoyu, Zhu, Xiubin, Pedrycz, Witold, Mostafa, Almetwally M., and Li, Zhiwu
- Subjects
ENTROPY (Information theory) ,FUZZY systems - Abstract
Takagi–Sugeno–Kang (TSK) classifiers have achieved great success in many applications due to their interpretability and transparent model reliability for users. At present, however, how to evaluate classification results is still an unsolved issue for TSK classifiers. This study designs a fuzzy rule-based classifier based on TSK classifiers, the outputs of which for an instance can be considered as the membership grades that the instance belongs to all classes. Then, an information entropy-based method is proposed to estimate the certainty of the outputs, which facilitates the further evaluation of the classification results of the instance for users. If the confidence level is not high, users can reject the classification results, and use other more advanced classifiers or collect more information about the instance. Moreover, the developed mechanism is suitable for handling large data since the adaptive moment estimation algorithm is used to identify the parameters of it. Experimental results demonstrate that the developed mechanism outperforms several rule-based classifiers. • The aim is to form a fuzzy rule-based classifier through information entropy. • The developed mechanism can provide the confidence of classification results. • The proposed classifier can handle large data due to the adaptive moment estimation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
47. Three-way approximate reduct based on information-theoretic measure.
- Author
-
Gao, Can, Wang, Zhicheng, and Zhou, Jie
- Subjects
- *
APPROXIMATE reasoning , *ENTROPY (Information theory) - Abstract
Three-way decision is a typical and popular methodology for decision-making and approximate reasoning, while attribute reduction is an important research topic in three-way decision. However, most attribute reduction methods based on three-way decision strictly rely on the preservation of measure criterion, which not only explicitly limits the efficiency of attribute reduction and also implicitly confines the generalization ability of the resulting reduct. In this study, we present a new three-way approximate attribute reduction method based on information-theoretic measure. More specifically, a unified framework for approximate attribute reduction is first provided. Then, the process of attribute reduction is considered to determine each attribute to be the positive region, boundary region, or negative region in terms of its correlation to the decision attribute. The negative attributes can be removed by the preservation of information-theoretic measure, while some boundary attributes are further iteratively eliminated by relaxing the measure criterion. An approximate reduct is finally formed by the positive attributes and the remaining boundary attributes. On several public UCI data sets, the proposed method achieves a much better attribute reduction rate and simultaneously gains an improvement in performance when comparing with other attribute reduction methods. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
48. Generalized information entropy and generalized information dimension.
- Author
-
Zhan, Tianxiang, Zhou, Jiefeng, Li, Zhen, and Deng, Yong
- Subjects
- *
ENTROPY (Information theory) , *UNCERTAINTY (Information theory) , *INFORMATION theory , *FRACTAL dimensions , *INFORMATION measurement , *ENTROPY - Abstract
The concept of entropy has played a significant role in thermodynamics and information theory, and is also a current research hotspot. Information entropy, as a measure of information, has many different forms, such as Shannon entropy and Deng entropy, but there is no unified interpretation of information from a measurement perspective. To address this issue, this article proposes Generalized Information Entropy (GIE) that unifies entropies based on mass function. Meanwhile, GIE establishes the relationship between entropy, fractal dimension, and number of events. Therefore, Generalized Information Dimension (GID) has been proposed, which extends the definition of information dimension from probability to mass fusion. GIE plays a role in approximation calculation and coding systems. In the application of coding, information from the perspective of GIE exhibits a certain degree of particle nature that the same event can have different representational states, similar to the number of microscopic states in Boltzmann entropy. • Generalized Information Entropy is compatible with Shannon entropy, Deng entropy, and RPS entropy. • Generalized Information Dimension achieves approximate calculation of large number entropy. • The form of Generalized Information Entropy is closer to Boltzmann entropy. • The linearity of Deng entropy is further explained (Zhao et al., 2024). [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
49. DNF: A differential network flow method to identify rewiring drivers for gene regulatory networks.
- Author
-
Xie, Jiang, Yang, Fuzhang, Wang, Jiao, Karikomi, Mathew, Yin, Yiting, Sun, Jiamin, Wen, Tieqiao, and Nie, Qing
- Subjects
- *
NEURONAL differentiation , *NEURAL stem cells , *GENE regulatory networks , *BIOLOGICAL networks , *CELL differentiation , *DISEASE progression - Abstract
Differential network analysis has become an important approach in identifying driver genes in development and disease. However, most studies capture only local features of the underlying gene-regulatory network topology. These approaches are vulnerable to noise and other changes which mask driver-gene activity. Therefore, methods are urgently needed which can separate the impact of true regulatory elements from stochastic changes and downstream effects. We propose the differential network flow (DNF) method to identify key regulators of progression in development or disease. Given the network representation of consecutive biological states, DNF quantifies the essentiality of each node by differences in the distribution of network flow, which are capable of capturing comprehensive topological differences from local to global feature domains. DNF achieves more accurate driver-gene identification than other state-of-the-art methods when applied to four human datasets from The Cancer Genome Atlas and three single-cell RNA-seq datasets of murine neural and hematopoietic differentiation. Furthermore, we predict key regulators of crosstalk between separate networks underlying both neuronal differentiation and the progression of neurodegenerative disease, among which App is predicted as a driver gene of neural stem cell differentiation. Our method is a new approach for quantifying the essentiality of genes across networks of different biological states. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
50. Entropy-based concept drift detection in information systems.
- Author
-
Sun, Yingying, Mi, Jusheng, and Jin, Chenxia
- Subjects
- *
ENTROPY (Information theory) , *INFORMATION storage & retrieval systems , *DECISION making , *ENTROPY - Abstract
As time passes, the data within information systems may continuously evolve, causing the target concept to drift. To ensure the effectiveness of data-driven decision making, it is crucial to detect drift in a timely manner and gather relevant information. In this paper, we introduce two methods that can directly detect concept drift in the provided information system, by considering a new perspective on uncertainty. First, using entropy under a single attribute constraint, we define the uncertainty of the target concept in an information system. By integrating the uncertainty of each attribute, the overall uncertainty of the target concept in the information system is obtained. Subsequently, two concept drift detection methods are proposed, namely EBTBM (Entropy-Based Threshold-Based Method) and EBSBM (Entropy-Based Sampling-Based Method). These methods utilize the defined uncertainty of the target concept as a statistical measure of the difference between two data blocks. Finally, extensive experiments on artificial and real-world data sets are conducted to validate the effectiveness of the proposed concept drift detection methods. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.