Submitted:
02 October 2024
Posted:
03 October 2024
You are already at the latest version
Abstract
Keywords:
1. Introduction
1.1. Aims and Scope
1.2. Classical Entropies and Generalized Entropies
- SK1
- Continuity. depends continuously on all variables for each W.
- SK2
- Maximality. For all W,
- SK3
- Expansibility. For all W and ,
- SK4
- Strong additivity (or separability). For all ,where .
1.3. Methodology and Organization of this Review
- G1
- Entropies based on Shannon’s entropy. These are entropies that use Shannon’s function (1) on probability distributions obtained from the data by a number of techniques. This group comprises: Dispersion entropy and fluctuation-based dispersion entropy, energy entropy and empirical mode decomposition energy entropy, Fourier entropy and fractional Fourier entropy, graph entropy, permutation entropy, spectral entropy, and wavelet entropy.
- G2
- Entropies based on other information-theoretical concepts, such as the correlation integral, divergences, unconditional or conditional mutual information, etc. This group comprises: Approximate entropy, cross entropy and categorical cross entropy, excess entropy, kernel entropy, relative entropy or Kullback-Leibler divergence, and transfer entropy.
- G3
- Entropies tailored to specific needs or inspired by other entropies. This group comprises: Bubble entropy, entanglement entropy, fuzzy entropy, intrinsic mode entropy, Kaniadakis entropy, Rao’s quadratic entropy, rank-based entropy, sample entropy, and tone entropy.
2. Applications in Data Analysis and Machine Learning
2.1. Approximate Entropy
-
Applications
- Anesthetic drug effects. Another field of applications is the quantification of anesthetic drug effects on the brain activity as measured by EEGs, including comparative testing of different anesthetics [24].
- Emotion recognition. Along with other entropies, approximate entropy has been used for EEG-based human emotion recognition [25].
- Physiological time series. See [29] for an overview of applications of approximate entropy to the analysis of physiological time series.
- Sleep research. The applications of approximate entropy include sleep research, in particular, the separation of sleep stages based on EEG data [30].
2.2. Bubble Entropy
-
Applications
- Fault bearing detection. Bubble entropy is used to reinforce the accuracy of fault bearing diagnosis through the Gorilla Troops Optimization (GTO) algorithm for classification [33]. A similar application can be found for the so-called Improved Hierarchical Refined Composite Multiscale Multichannel Bubble Entropy [34].
- Feature extraction. Bubble entropy is compared with dispersion entropy (Section 2.6) in the extraction of single and double features in [35].
2.3. Categorical cross entropy
-
Applications
- Deep Learning. CCE is used in deep neural networks when dealing with noisy labels. An improved categorical cross entropy (ICCE) is also used in this case [36].
- Multi-class classification. CCE is used in the analysis of multi-channel time-series multi-class classification. It is the standard loss function for tasks such as image classification, text classification, and speech recognition [37].
- Reinforcement learning. CCE is used as an improvement of value function training (using classification instead of regression) mainly in games [38].
- Semi-supervised learning. CCE is used in pseudo-labelling to optimize convolutional neural networks parameters [39].
2.4. Cross Entropy
-
Applications
- Deep learning. Cross entropy is a standard loss function for training deep neural networks, particularly those involving softmax activation functions. It is very useful for applications such as object detection, language translation, and sentiment analysis [40]. In this regard, empirical evidence with limited and noisy data suggests that to measure the top- error (a common measure of performance in machine learning performed with deep neural networks trained with the cross entropy loss), the loss function must be smooth, meaning that it should incorporate a smoothing parameter to handle small probability events [41].
- Feature selection. Cross entropy is used to select significant features of binary values from highly imbalanced large datasets via a framework called FMC Selector [42].
- Image analysis. Wavelet analysis together with cross entropy are used in image segmentation, object recognition, texture analysis (e.g., fabric defect detection) and pattern classification [43].
- Learning-to-rank methods. In [44] the author proposes a learning-to-rank loss function that is based on cross entropy. Learning-to-rank methods form a class of ranking algorithms that are widely applied in information retrieval.
- Multiclass classification. Cross entropy is used to enhance the efficiency of solving support vector machines for multi-class classification problems [45].
- Semi-supervised clustering. Cross entropy is employed along with the information bottleneck method in semi-supervised clustering. It is robust to noisy labels and automatically determines the optimal number of clusters under mild conditions [46].
2.5. Differential Entropy
-
Applications
- Anomaly detection. Differential entropy can measure changes in the probability density function of an analog signal which reveals an anomaly in the source, whether it is a mechanical system or a patient [47].
- Emotion recognition. Differential entropy has been used in [25] to extract features in EEG-based human emotion recognition.
- Feature selection. A feature selection algorithm based on differential entropy to evaluate feature subsets has been proposed in [48]. This algorithm effectively represents uncertainty in the boundary region of a fuzzy rough model and demonstrates improved performance in selecting optimal feature subsets, thereby enhancing classification accuracy; see [49] for an implementation.
- Generative models. Variational Autoencoders and other generative models leverage differential entropy to model the latent space of continuous data distributions. These models can learn better representations of the input data, thus improving the performance [50].
- Mutual information. Differential entropy is instrumental to compute the mutual information of continuous-valued random variables and processes, e.g., autoregressive processes. Seen in speech processing (linear prediction), seismic signal processing and biological signal processing [51].
- Probabilistic models. Differential entropy is utilized in probabilistic models such as Gaussian Mixture Models to describe the uncertainty and distribution of continuous variables. This approach is applicable to image processing and network inference as well [52].
2.6. Dispersion Entropy
-
Applications
- Feature extraction. Multiscale fuzzy dispersion entropy is applied in fault diagnosis of rotating machinery to capture the dynamical variability of time series across various scales of complexity [54].
- Image classification. A multiscale version of dispersion entropy called MDispEn2D has been used with biomedical data to measure the impact of key parameters that may greatly influence the entropy values obtained in image classification [55].
- Signal classification. Other generalization of dispersion entropy, namely, fractional fuzzy dispersion entropy, has been proposed as a fuzzy membership function for signal classification tasks [56].
- Time series analysis. Multiscale graph-based dispersion entropy is a generalization of dispersion entropy used to analyze multivariate time series data in graph and complex network frameworks, e.g., weather and two-phase flow data; it combines temporal dynamics with topological relationships [59].
2.7. Energy Entropy and Empirical Mode Decomposition Energy Entropy
-
Applications
- Chatter detection. This application involves detecting vibrations and noise in machining operations that can indicate chattering. In this regard, energy entropy can detect chatter in robotic milling [61].
- Feature extraction. Energy entropy is calculated via the empirical decomposition of the signal into intrinsic mode functions and serves as a feature for machine learning models used in chatter detection. The chatter feature extraction method is based on the largest energy entropy [64].
- Time-series forecasting. EMD energy entropy was used in [65] to predict short-term electricity consumption by taking into account the data variability, i.e., that power consumption data is non-stationary, nonlinear, and influenced by the season, holidays, and other factors. In [66], this entropy was the tool to distinguish two kinds of financial markets.
2.8. Entanglement Entropy
-
Applications
- Feature extraction. In quantum machine learning, entanglement entropy is used for feature extraction by representing data in a form that highlights quantum correlations and thus, leveraging the quantum properties of the data [69].
- Quantum models. Entanglement entropy is used in quantum models to quantify unknown entanglement by using neural networks to predict entanglement measures of unknown quantum states based on experimentally measurable data: moments or correlation data produced by local measurements [70].
2.9. Excess Entropy
-
Applications
- Image segmentation. Excess entropy is used to measure the structural information of a 2D or 3D image and then determine the optimal threshold in a segmentation algorithm proposed in [73]. The working hypothesis of this thresholding-based segmentation algorithm is that the optimal threshold corresponds to the maximum excess entropy (i.e., to a segmentation with maximum structure).
- Machine learning. In [74], the authors present a method called machine-learning iterative calculation of entropy, for calculating the entropy of physical systems by iteratively dividing the system into smaller subsystems and estimating the mutual information between each pair of halves.
- Neural estimation in adversarial generative models. Mutual Information Neural Estimator is a scalable estimator used in high dimensional continuous data analysis that optimizes mutual information. The authors apply this estimator to Generative Adversarial Networks (GANs) [75].
- Time series analysis. In this application, total excess entropy is used for classifying stationary time series into long-term and short-term memory. A stationary sequence with finite block entropy is long-term memory if its excess entropy is infinite [76].
2.10. Fluctuation-Based Dispersion Entropy
-
Applications
-
Fault diagnosis. The so-called refined composite moving average FDispEn is used in machinery fault diagnosis by analysing vibration signals [78].Refined composite multiscale FDispEn and supervised manifold mapping are used in fault diagnosis for feature extraction in planetary gearboxes [79].Multivariate hierarchical multiscale FDispEn along with multi-cluster feature selection and Gray-Wolf Optimization-based Kernel Extreme Learning Machine helps diagnose faults in rotating machinery. It also captures the high-dimensional fault features hidden in multichannel vibration signals [80].
- Feature extraction. FDispEn gives rise to hierarchical refined multi-scale fluctuation-based dispersion entropy, used to extract underwater target features in marine environments and weak target echo signals, thereby improving the detection performance of active sonars [81].
- Robustness in spectrum sensing. Reference [82] proposes a machine learning implementation of spectrum sensing using an improved version of the FDispEnt as a feature vector. This improved version shows enhanced robustness to noise.
- Signal classification. FDispEn helps distinguish various physiological states of biomedical time series and it is commonly used in biomedicine. It is also used to estimate the dynamical variability of the fluctuations of signals applied to neurological diseases [83]. Fluctuation-based reverse dispersion entropy is applied to signal classification combined with k-nearest neighbor [84].
- Time series analysis. FDispEn is used to quantify the uncertainty of time series to account for knowledge on parameters sensitivity and studying the effects of linear and nonlinear mapping on the defined entropy in [77]. FDispEn is defined as a measure for dealing with fluctuations in time series. Then, the performance is compared to complexity measures such as permutation entropy (Section 2.20), sample entropy (Section 2.25), and Lempel-Ziv complexity [9,85].
-
2.11. Fourier Entropy
-
Applications
- Decision trees: The Fourier Entropy-Influence Conjecture, made by Friedgut and Kalai [86], says that the Fourier entropy of any Boolean function f is upper bounded, up to a constant factor, by the total influence (or average sensitivity) of f. This conjecture, applied to decision trees, gives interesting results that boil down to , meaning that , where denotes the minimum number of leaves in a decision tree that computes f and is independent of f and [87]. Another similar application to decision trees can be found in [88].
- Learning theory. The Fourier Entropy-Influence Conjecture is closely related to the problem of learning functions in the membership model. It is said that if a function has small Fourier entropy it means that its Fourier transform is concentrated on a few characters, i.e. the function can be approximated by a sparse polynomial, a class that is very important in the context of learning theory [89]. Learning theory provides the mathematical foundation for understanding how algorithms learn from data, guiding the development of machine learning models.
2.12. Fractional Fourier Entropy
-
Applications
- Artificial intelligence. Two-dimensional fractional Fourier entropy helps to diagnose COVID-19 by extracting features from chest CT images [93].
- Biomedical image classification. Fractional Fourier entropy has proven helpful in detecting pathological brain conditions. By using it as a new feature in Magnetic Resonance Imaging, the classification of images is improved in time and cost [94].
- Deep learning. Fractional Fourier entropy is used in the detection of gingivitis via feed-forward neural networks. It reduces the complexity of image extraction before classification and can obtain better image eigenvalues [95].
- Emotion recognition. Fractional Fourier entropy, along with two binary support vector machines, helps improve the accuracy of emotion recognition from physiological signals in electrocardiogram and galvanic skin responses [96].
- Multilabel classification. Fractional Fourier entropy has been used in a tea-category identification system, which can automatically determine tea category from images captured by a 3 charge-coupled device digital camera [97].
2.13. Fuzzy Entropy
-
Applications
- Clustering and time-series analysis. Fuzzy entropy is used in problems of robustness against outliers in clustering techniques in [101].
- Data analysis. Fuzzy entropy is proposed in [102] to assess the strength of fuzzy rules with respect to a dataset, based on the greatest energy and smallest entropy of a fuzzy relation.
- Fault detection. Fuzzy entropy (along with dispersion entropy, Section 2.6) was the best performer in a comparative study of entropy-based methods for detecting motor faults [103]. Multiscale fuzzy entropy is used to measure complexity in time series in rolling bearing fault diagnosis [104].
- Feature selection and mathematical modelling. Fuzzy entropy is used in feature selection to evaluate the relevance and contribution of each feature in Picture Fuzzy Sets [105].
- Image classification. Fuzzy entropy, in the form of multivariate multiscale fuzzy entropy, is proposed and tested in [106] for the study of texture in color images and their classification.
- Image segmentation. Fuzzy entropy is the objective function of a colour image segmentation technique based on an improved cuckoo search algorithm [107].
2.14. Graph Entropy
-
Applications
- Dimension reduction and feature selection. Graph entropy gives rise to the Conditional Graph Entropy that helps in the alternating minimization problem [112].
- Graph structure. Graph entropy is used to measure the information content of graphs, as well as to evaluate the complexity of the hierarchical structure of a graph [113].
- Graph-based time series analysis. Graph entropy can be used in time series analysis in conjunction with any method that transforms time series into graphs. An example is the HV graph entropy presented above; see [114] and references therein.
- Node embedding dimension selection. Graph entropy is applied in Graph Neural Networks through the Minimum Graph Entropy algorithm. It calculates the ideal node embedding dimension of any graph [115].
2.15. Havrda–Charvát Entropy
-
Applications
- Computer vision. An HC entropy-based technique for group-wise registration of point sets with unknown correspondence is used in graphics, medical imaging and pattern recognition. By defining the HC entropy for cumulative distribution functions (CDFs), the corresponding CDF-HC divergence quantifies the dissimilarity between CDFs estimated from each point-set in the given population of point sets [118].
- Financial time series. Weighted HC entropy outperforms regular HC entropy when used as a complexity measure in financial time series. The weights turn out to be useful for showing amplitude differences between series with the same order mode (i.e. similarities in patterns or specific states) and robust to noise [119].
- Image segmentation and classification. HC entropy is applied as loss function in image segmentation and classification tasks using convolutional neural networks in [120].
- Loss functions in deep learning. HC entropy can be used to design loss functions in deep learning models. These loss functions are particularly useful in scenarios with small datasets, common in medical applications [121].
2.16. Intrinsic Mode Entropy
-
Applications
- Language gesture recognition. IME is used in [123] to analyse data from a 3-dimensional accelerometer and a five-channel surface electromyogram of the user’s dominant forearm for automated recognition of Greek sign language gestures.
- Neural data analysis. An IME version with improved discriminatory capacity in the analysis of neural data is proposed in [124].
- Time series analysis. IME is used in nonlinear time series analysis to efficiently characterize the underlying dynamics [122]. As any multiscale entropy, IME is particularly useful for the analysis of physiological time series [125]. See also [126] for an application to the analysis of postural steadiness.
2.17. Kaniadakis Entropy
-
Applications
- Image segmentation. Kaniadakis entropy is used in image thresholding to segment images with long-tailed distribution histograms; the parameter is selected via a swarm optimization search algorithm [129].
- Images threshold selection. Kaniadakis entropy can be used to construct an objective function for image thresholding. By using the energy curve and the Black Widow optimization algorithm with Gaussian mutation, this approach can be performed on both grayscale and colour images of different modalities and dimensions [130].
- Seismic imaging. Application of the Maximum Entropy Principle with leads to the Kaniadakis distribution, a deformation of the Gaussian distribution that has application, e.g., in seismic imaging [131].
2.18. Kernel Entropy
-
Applications
- Complexity of time-series. The authors of [133] present experimental evidence that Gaussian kernel entropy outperforms approximate entropy when it comes to analyze the complexity of time series.
- Fetal heart rate discrimination. In [134] the authors compare the performance of several kernel entropies on fetal heart rates discrimination, with the result that the circular and Cauchy kernels outperform other, more popular kernels, such as the Gaussian or the spherical ones.
- Parkinson’s disease. Gaussian kernel entropy, along with other nonlinear features, is used in [135] in the task of automatic classification of speech signals from subjects with Parkinson’s disease and a control set.
- Pathological speech signal analysis. Reference [132] is a study of several approaches in the field of pathological speech signal analysis. Among the new pathological voice measures, the authors include different kernel-based approximate and sample entropies.
2.19. Kolmogorov-Sinai Entropy
-
Applications
- Time series analysis. The perhaps main practical application of the Kolmogorov-Sinai entropy is the analysis of nonlinear, real-valued time series, where it is used to characterize the underlying dynamical system, in particular, its chaotic behavior. Recent practical examples include short-term heart rate variability [140], physical models of the vocal membranes [139], autonomous driving [141], and EEG-based human emotion recognition [25,142].
2.20. Permutation Entropy
-
Applications
- Analysis of EEGs. One of the first applications of permutation entropy was the analysis of EEGs of subjects with epilepsy because normal and abnormal signals (during epileptic seizures) have different complexities [156]. Furthermore, since permutation entropy can be computed in virtually real time, it has been used to predict seizures in epilepsy patients by tracking dynamical changes in EEGs [147]. Further examples can be found in the reviews [18,147]. Results can be improved using permutation Rényi and Tsallis entropy due to their additional, fine-tunable parameter [150,157].
- Determinism detection. Time series generated by one-dimensional maps have necessarily forbidden ordinal patterns of all sufficiently large lengths L[146]. Theoretical results under some provisos and numerical results in other cases show that the same happens with higher dimensional maps [158,159]. Therefore, the scaling of permutation entropy with L can distinguish noisy deterministic signals from random signals [146,160].
- Emotion recognition. Permutation entropy is used to help in tasks of feature extraction in EEGs [25].
- Obstructive sleep apnea. A combination of permutation entropy-based indices and other entropic metrics was used in [161] to distinguish subjets with obstructive sleep apnea from a control group. The data consisted of heart rate and beat-to-beat blood pressure recordings.
- Prediction. Permutation entropy has been used along with variational modal decomposition to predict wind power [162].
- Speech signals. In their seminal paper [143], Bandt and Pompe used precisely permutation entropy to analyze speech signals and showed that it is robust with respect to the window length, sampling frequency and observational noise.
- The causality-complexity plane. Permutation entropy together with the so-called statistical complexity builds the causality-complexity plane, that has proven to be a powerful tool to discriminate and classify time series [163]. By using variants of the permutation entropy and the statistical complexity, the corresponding variants of the causality-complexity plane are obtained, possibly with enhanced discriminatory abilities for the data at hand [149].
- Unstructured data. Nearest-neighbor permutation entropy is an innovative extension of permutation entropy tailored for unstructured data, irrespective of their spatial or temporal configuration and dimensionality, including, e.g., liquid crystal textures [164].
2.21. Rao’s Quadratic Entropy
-
Applications
- Environmental monitoring. RQE helps calculate the environmental heterogeneity index and assist prioritization schemes [166].
- Genetic diversity metrics. RQE is used to measure diversity for a whole collection of alleles to accommodate different genetic distance coding schemes and computational tractability in case of large datasets [167].
- Unsupervised classification. RQE is used as a framework in the support vector data description algorithm for risk management, enhancing knowledge in terms of interpretation, optimization, among others [168].
2.22. Rank-Based Entropy
-
Applications
- Anomaly detection. RbE is applied in mixed data analysis to check the influence of categorical features, using Jaccard index for anomaly ranking and classification [170].
- Feature selection. RbE is used in the Entropy-and-Rank-based-Correlation framework to select features, e.g., in the detection of fruit diseases [171].
- Mutual information. RbE is used to rank mutual information in decision trees for monotonic classification [172].
- Node importance. RbE is employed in the analysis of graphs to rank nodes taking into account the local and global structure of the information [173].
- QSAR models. RbE is employed in Quantitative Structure-Activity Relationship models (QSAR) to analyse their stability via “rank order entropy”, suggesting that certain models typically used should be discarded [174].
- Time series classification. RbE helps classify order of earliness in time series to generate probability distributions in different stages [176].
2.23. Relative Entropy
-
Applications
- Anomaly detection. KL divergence has been used for plane control in Software-Defined Networking as a method to detect Denial of Service attacks in [177].
- Bayesian networks. The efficient computation of the KL divergence of two probability distributions, each one coming from a different Bayesian network (with possibly different structures), has been considered in [178].
- Feature selection. The authors of [179] show that the KL divergence is useful in information-theoretic feature selection due to the fact that maximising conditional likelihood corresponds to minimising KL-divergence between the true and predicted class posterior probabilities.
- Parameter minimization in ML. Parameters that minimize the KL divergence minimize also the cross entropy and the negative log likelihood. So, the KL divergence is useful in optimization problems where the loss function is a cross-entropy [181].
2.24. Rényi Entropy
-
Applications
- Automated identification. Average Renyi entropy, along with other entropic measures, have been used as inputs for SVM algorithms to classify focal or non-focal EEGs of subjects affected by partial epilepsy [185].
- Clustering. Rényi entropy can provide robust similarity measures that are less sensitive to outliers [182].
- Extreme entropy machines. Rényi’s quadratic entropy is used in the construction of extreme entropy machines to improve classification problems [186].
- Feature selection and character recognition. Adjustment of the parameter can help to emphasize different parts of the underlying probability distribution and hence the selection of the most informative features. Rényi entropy is used for feature selection in [182,187]. Max-entropy is used in [188] for convolutional feature extraction and improvement of image perception.
2.25. Sample Entropy
-
Applications
- Automated identification. Average sample entropy and other entropy measures are used as input for an SVM algorithm to classify focal and non-focal EEG signals of subjects with epilepsy [185].
- Fault diagnosis. Sample entropy has been used for multi-fault diagnosis in lithium batteries [192].
- Image classification. Sample entropy, in the form of multivariate multiscale sample entropy, is used for classifying RGB colour images to compare textures, based on a threshold to measure similarity [106].
- Image texture analysis. Two-dimensional sample entropy has shown to be a useful texture feature quantifier for the analysis of biomedical images [193].
- Mutual information. Modified sample entropy has been used in skin blood flow signals to analyse mutual information and, hence, study the association of microvascular dysfunction in different age groups [194].
- Neurodegenerative disease classification. Sample entropy is used to classify neurodegenerative diseases. Gait signals, support vector machines and nearest neighbours are employed to process the features extracted using sample entropy [195].
- Time-series analysis. Sample entropy, often in the form of multiscale sample entropy, is a popular tool in time series analysis, in particular with biomedical data [198]. For example, it is used for the fast diagnosis and monitoring of Parkinson’s disease [199] and human emotion recognition [25] using EEGs. A modified version of multiscale sample entropy has recently been used for diagnosing epilepsy [200]. See [29] for an overview of applications of sample entropy to the analysis of physiological time series.
- Weather forecasting. Sample entropy is applied in weather forecasting by using transductive feature selection methods based on clustering-based sample entropy [201].
2.26. Shannon Entropy
-
Applications
- Accurate prediction. Shannon entropy is employed in machine learning models to improve the accuracy of predictions of molecular properties in the screening and development of drug molecules and other functional materials [202].
- Anomaly detection. Shannon entropy is employed in sensors (Internet of Things) to identify anomalies using the CorrAUC algorithm [203].
- Artificial intelligence. Shannon entropy contributes to the creation of the Kolmogorov Learning Cycle, which acts as a framework to optimize "Entropy Economy", helped by the intersection of Algorithmic Information Theory (AIT) and Machine Learning (ML). This framework enhances the performance of the Kolmogorov Structure Function, leading to the development of "Additive AI". By integrating principles from both AIT and ML, this approach aims to improve algorithmic efficiency and effectiveness, driving innovation in AI by balancing information theory with practical machine learning applications [204].
- Automated identification.Average Shannon entropy and other entropy measures are used as inputs of an SVM algorithm to classify focal or non-focal EEG signals of subjects with epilepsy [185].
- Fault bearing diagnosis. Multi-scale stationary wavelet packet analysis and the Fourier amplitude spectrum are combined to obtain a new discriminative Shannon entropy feature that is called stationary wavelet packet Fourier entropy in [205]. Features extracted by this method are then used to diagnose bearing failure.
- Feature selection. Shannon’s mutual information entropy is used to design an entire information-theoretic framework that improves the selection of features in [179]. Shannon entropy is employed in biological science to improve classification of data in clustering of genes using microarray data [206].
- Hard clustering. Shannon entropy is used as a criterion to measure the confidence in unsupervised clustering tasks [207].
- Natural language processing.Shannon entropy quantifies the predictability (or redundancy) of a text. Therefore, it is instrumental in language modelling, text compression, information retrieval, among others [2,9]. For example, it is used in [208] for keyword extraction, i.e., to rank the relevance of words.
- Policy learning.Shannon entropy acts as a regularization inside of an iterative policy optimization method for certain quadratic linear control scenarios [209].
- Signal analysis. Shannon entropy is used as the cost functional of compression algorithms in sound and image processing [210].
- Statistical inference. According to the Maximum Entropy Principle of Jaynes [211], "in making inferences on the basis of partial information we must use the probability distribution which has maximum entropy subject to whatever is known". This principle has been traditionally applied with the Shannon entropy and several moment constraints of a probability distribution to infer the actual distribution [9,18].
2.27. Spectral Entropy
-
Applications
- Audio analysis. Spectral entropy has been applied for robust audio content classification in noisy signals [213]. Specifically, spectral entropy is used to segment input signals into noisy audio and noise. Also, spectral entropy (in the form of Multiband Spectral Entropy Signature) has been shown to outperform other approaches in the task of sound recognition [214].
- Damage event detection. Spectral entropy detects damage in vibration recordings from a wind turbine gearbox [215].
- Data time compression. Spectral entropy has been successfully applied to identify important segments in speech, enabling time-compression of speech for skimming [216].
- Deep learning synchronization. Spectral entropy evaluates synchronization in neuronal networks, providing analysis of possibly noisy recordings collected with microelectrode arrays [217].
- Hyperspectral anomaly detection. Hyperspectral Conditional Entropy enters into the Entropy Rate Superpixel Algorithm, which is used in hyper spectral-spatial data to recognize unusual patterns [219].
- Signal detection. Spectral entropy has been used to detect cetacean vocalization in marine audio data [220]. The time frequency decomposition was done with the short time Fourier transform and the continuous wavelet transform.
2.28. Tone Entropy
-
Applications
- Biomedical analysis. Tone entropy has been employed to study the autonomic nervous system in age groups at high-risk of cardiovascular diseases [222]. In [223], tone entropy was used to study the influence of gestational ages on the development of the foetal autonomic nervous system by analyzing the foetal heart rate variability.
- Time series. Tone entropy has been used in time series analysis to differentiate between physiologic and synthetic interbeat time series [224].
2.29. Topological and Topology-Based Entropies
-
Applications
- Cardiac dynamics. Given a (finite) time series, the out-link entropy is derived from the adjacency matrix of its ordinal network. This entropy has been used to classify cardiac dynamics in [228].
- Convolutional neural networks. The authors of [229] propose a method for quantitatively clarifying the status of single unit in convolutional neural networks using algebraic topological tools. Unit status is indicated via the calculation of a topology-based entropy, called feature entropy.
- Damage detection. Persistent entropy can also address the damage detection problem in civil engineering structures. In particular, to solve the supervised classification damage detection problem [230].
- Detection of determinism. Permutation topological entropy (i.e., the topological entropy of the distribution of ordinal patterns of length obtained from a time series) can be used to detect determinism in continuous-valued time series. Actually, it suffices to check the growth of ordinal patterns with increasing L’s, since this growth is exponential for deterministic signals ( converges to a finite number) and factorial for random ones ( diverges) [146,158].
- Financial time series. Topological entropy has been applied to horizontal visibility graphs of financial time series in [231].
- Similarity of piecewise linear functions. Piecewise linear functions are a useful mathematical tool in different areas of applied mathematics, including signal processing and machine learning methods. In this regard, persistent entropy (a topological entropy based on persistent homology) can be used to measure their similarity [232].
2.30. Transfer Entropy
-
Applications
- Accelerated training in Convolutional Neural Networks (CNN). The authors of [235] propose a training mechanism for CNN architectures that integrates transfer entropy feedback connections. In this way, the training process is accelerated as fewer epochs are needed. Furthermore, it generates stability, hence, it can be considered a smoothing factor.
- Improving accuracy in Graph Convolutional Neural Networks (GCN). The accuracy of a GCN can be improved by using node relational characteristics (such as heterophily), degree information, and feature-based transfer entropy calculations. However, depending on the number of grah nodes, the computation of the transfer entropy can significantly increase the computational load [236].
- Improving neural network performance. A small, few-layer artificial neural network that employs feedback can reach top level performance on standard benchmark tasks, otherwise only obtained by large feed-forward structures. To show this, the authors of [237] use feed-forward transfer entropy between neurons to structure feedback connectivity.
- Multivariate time series forecasting. Transfer entropy is used to establish causal relationships in multivariate time series converted into graph neural networks, each node corresponding to a variable and edges representing the casual relationships between the variables. Such neural networks are then used for prediction [238].
- Time series analysis. The main application of transfer entropy since its formulation has been the analysis of multivariate time series (whether biomedical, physical, economical, financial, ...) for revealing causal relationships via information directionality. See [239] and the references therein for the conceptual underpinnings and practical applications.
2.31. Tsallis Entropy
-
Applications
- Anomaly detection. Tsallis entropy is used in network intrusion detection by detecting botnet-like malware based on anomalous patterns in the network [184].
- Clustering. A Tsallis entropy based categorical data clustering algorithm is proposed in [241]. It is shown there, that when the attributes have a power law behavior the proposed algorithm outperforms existing Shannon entropy-based clustering algorithms.
- Feature selection. Tsallis-entropy-based feature selection is used in [242] to identify significant features, which boosts the classification performance in machine learning. The authors propose an algorithm to optimize both the classifier (a Support Vector Machine) and Tsallis entropy parameters, so improving the classification accuracy.
- Image segmentation. Tsallis Entropy is maximized to use it for segmenting images by maximizing the entropy within different regions of the image [243].
- Pre-seismic signals. Tsallis entropy has been used in [244] to analyze pre-seismic electromagnetic signals.
2.32. Von Neumann Entropy
-
Applications
- Feature selection and dimensionality reduction. Von Neumann entropy is employed in the case of kernelized relevance vector machines to asses dimensionality reduction for better model performances [245].
- Graph-based learning. In [246] the authors propose a method to identify vital nodes in hypergraphs that is based on von Neumann entropy. More precisely, this method is based on the high-order line graph structure of hypergraphs and measures changes in network complexity using von Neumann entropy.
- Graph similarity and anomaly detection. The von Neumann graph entropy (VNGE) is used to measure the information divergence and distance between graphs in a sequence. This is used for various learning tasks involving network-based data. The Fast Incremental von Neumann Graph Entropy algorithm reduces the computation time of the VNGE, making it feasible for real-time applications and large datasets [247].
- Network analysis. Von Neumann entropy is used in [248] to build visualization histograms from the edges of networks and then component analysis is performed on a sample for different networks.
- Pattern recognition in neurological time series. Von Neumann entropy has been used (together with other entropies) in [249] for automated pattern recognition in neurological conditions, a crucial task in patient monitoring and medical diagnosis.
2.33. Wavelet Entropy
-
Applications
- Emotion recognition. Wavelet entropy can detect little variations in signals and it has been used in [25] to develop an automatic EEG classifier.
- Fault detection. Wavelet entropy is applied to monitor the condition of machinery and detect faults by analysing vibration signals [251].
- Feature extraction. Wavelet entropy is used to extract features from biomedical signals such as EEG and ECG to identify different physiological states or detect abnormalities [250].
3. Discussion
4. Conclusions
- The choice of a particular entropy-based method depends in general on the application. Some methods may be more popular than others because they are designed for the purpose or dataset at hand, or simply because they have some computational advantage. In this regard, Section 2 presented possible candidates for different applications but no performance comparison between them was discussed. In fact, such a comparison would require a case-by-case test as, for example, in Reference [103], where the authors study motor fault detection with the approximate, dispersion, energy, fuzzy, permutation, sample and Shannon entropies (see Section 2.13 for the best performer). Along with the selection of the “right” entropy, a common concern among practitioners is the choice of parameters and hyperparameters. A combination of methods and parameter settings may be also a good approach in practice [161] [249].
- We have not included applications of entropy to cryptography in this review because they belong to the general field of Data Science (through data security) rather than to Data Analysis. Entropy (mainly Shannon’s and Rényi’s entropies) have been applied to measure the “randomness” of encrypted messages in the so-called chaotic cryptography, which applies ideas from Chaos Theory and chaos synchronization to the masking of analog signals [252]. To deal with digital signals, new tools such as discrete entropy [253] and discrete Lyapunov exponents [254] have also been developed for application to chaotic cryptography, inspired by their conventional counterparts.
- We have not delved into the numerical methods to compute entropies from the data. Being functionals of probability distributions, most methods for computing entropy are based on the estimation of data probabilities. In the case of discrete-valued data, the probabilities are usually estimated via relative frequencies (the maximum likelihood estimator) and possibly extrapolation methods in case of undersampling [146]. In the case of continuous-valued data, the probability densities are usually estimated via kernel density estimation (also called Parzen-Rosenblatt windowing) [255]. Furthermore, there are methods that do not rely on probability estimation like, e.g., Lempel-Ziv complexity, which resorts to pattern matching [85]. Also, in some particular cases, the entropy can be estimated via spectral information. For example, the quadratic Rényi entropy can be directly estimated via inner product matrices and principal component analysis [256]. See [257] for a general review on the estimation of entropy.
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- von Neumann, J. Mathematische Grundlagen der Quantenmechanik; Springer Verlag: Berlin, Heidelberg, New York, 1971; p. 150. [Google Scholar] [CrossRef]
- Shannon, C.E. A mathematical theory of communication. The Bell System Technical Journal 1948, 27, 379–423. [Google Scholar] [CrossRef]
- Gibbs, J.W. Elementary Principles in Statistical Mechanics: Developed with Especial Reference to the Rational Foundation of Thermodynamics; Charles Scribner’s Sons: Farmington Hills, MI, USA, 1902. [Google Scholar]
- Kolmogorov, A.N. Entropy per unit time as a metric invariant of automorphisms. Doklady of the Russian Academy of Sciences 1959, 124, 754–755. [Google Scholar]
- Sinai, Y.G. On the Notion of Entropy of a Dynamical System. Doklady of the Russian Academy of Sciences 1959, 124, 768–771. [Google Scholar] [CrossRef]
- Adler, R.L.; Konheim, A.G.; McAndrew, M.H. Topological entropy. Transactions of the American Mathematical Society 1965, 114, 309–319. [Google Scholar] [CrossRef]
- Walters, P. An Introduction to Ergodic Theory; Springer Verlag: New York, 2000. [Google Scholar]
- Khinchin, A.I. Mathematical Foundations of Information Theory; Dover: New York, 1957. [Google Scholar]
- Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons, Inc.: Hoboken, NJ, 2005; pp. 243–244. [Google Scholar] [CrossRef]
- Planck, M. Zur Theorie des Gesetzes der Energieverteilung im Normalspectrum. Verhandlungen der Deutschen Physikalischen Gesellschaft 1900, 2, 237–245. [Google Scholar]
- Csiszár, I. Axiomatic Characterization of Information Measures. Entropy 2008, 10, 261–273. [Google Scholar] [CrossRef]
- Rényi, A. On measures of entropy and information. Proceedings of the 4th Berkeley Symposium on Mathematical Statistics and Probability, 1961, Vol. 1, pp. 547–561.
- Amigó, J.M.; Balogh, S.G.; Hernández, S. A Brief Review of Generalized Entropies. Entropy 2018, 20. [Google Scholar] [CrossRef]
- Havrda, J.; Charvát, F. Quantification method of classification processes. Kybernetika 1967, 3, 30–35. [Google Scholar]
- Tsallis, C. Possible generalization of Boltzmann-Gibbs statistics. Journal of Statistical Physics 1988, 52, 479–487. [Google Scholar] [CrossRef]
- Ribeiro, M.; Henriques, T.; Castro, L.; Souto, A.; Antunes, L.; Costa-Santos, C.; Teixeira, A. The Entropy Universe. Entropy 2021, 23. [Google Scholar] [CrossRef]
- Katok, A. Fifty years of entropy in dynamics: 1958–2007. Journal of Modern Dynamics 2007, 1, 545–596. [Google Scholar] [CrossRef]
- Amigó, J.M.; Keller, K.; Unakafova, V. On entropy, entropy-like quantities, and applications. Frontiers in Entropy Across the Disciplines; Freeden, W.; Nashed, M., Eds. World Scientific, 2022, pp. 197–231. [CrossRef]
- Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy analysis of biological signals. Physical Review E 2005, 71, 021906. [Google Scholar] [CrossRef]
- Pincus, J.M. Approximate entropy as a measure of system complexity. Proceedings of the National Academy of Sciences 1991, 88, 2297–2301. [Google Scholar] [CrossRef]
- Delgado-Bonal, A.; Marshak, A. Approximate Entropy and Sample Entropy: A Comprehensive Tutorial. Entropy 2019, 21, 541. [Google Scholar] [CrossRef]
- Hornero, R.; Abásolo, D.; Escudero, J.; Gómez, C. Nonlinear analysis of electroencephalogram and magnetoencephalogram recordings in patients with Alzheimer’s disease. Philosophical Transactions of the Royal Society of London, Series A 2009, 367, 317–336. [Google Scholar] [CrossRef]
- Morabito, F.C.; Labate, D.; La Foresta, F.; Bramanti, A.; Morabito, G.; Palamara, I. Multivariate Multi-Scale Permutation Entropy for Complexity Analysis of Alzheimer’s Disease EEG. Entropy 2012, 14, 1186–1202. [Google Scholar] [CrossRef]
- Liang, Z.; Wang, Y.; Sun, X.; Li, D.; Voss, L.J.; Sleigh, J.W.; Hagihira, S.; Li, X. EEG Entropy Measures in Anesthesia. Frontiers in Computational Neuroscience 2015, 9, 16. [Google Scholar] [CrossRef]
- Patel, P.R.; Annavarapu, R.N. EEG-based human emotion recognition using entropy as a feature extraction measure. Brain Informatics 2021, 8, 20. [Google Scholar] [CrossRef]
- Kannathal, N.; Choo, M.L.; Acharya, U.R.; Sadasivan, P.K. Entropies for Detection of Epilepsy in EEG. Computer Methods and Programs in Biomedicine 2005, 80, 187–194. [Google Scholar] [CrossRef]
- Srinivasan, V.; Eswaran, C.; Sriraam, N. Approximate Entropy-Based Epileptic EEG Detection Using Artificial Neural Networks. IEEE Transactions on Information Technology in Biomedicine 2007, 11, 288–295. [Google Scholar] [CrossRef]
- Jouny, C.C.; Bergey, G.K. Characterization of Early Partial Seizure Onset: Frequency, Complexity and Entropy. Clinical Neurophysiology 2012, 123, 658–669. [Google Scholar] [CrossRef]
- Richman, J.; Moorman, J. Physiological Time-Series Analysis Using Approximate Entropy and Sample Entropy. American Journal of Physiology - Heart and Circulatory Physiology 2000, 278, H2039–H2049. [Google Scholar] [CrossRef]
- Acharya, U.R.; Faust, O.; Kannathal, N.; Chua, T.; Laxminarayan, S. Non-linear Analysis of EEG Signals at Various Sleep Stages. Computer Methods and Programs in Biomedicine 2005, 80, 37–45. [Google Scholar] [CrossRef]
- Manis, G.; Aktaruzzaman, M.; Sassi, R. Bubble Entropy: An Entropy Almost Free of Parameters. IEEE Transactions on Biomedical Engineering 2017, 64, 2711–2718. [Google Scholar] [CrossRef]
- Manis, G.; Bodini, M.; Rivolta, M.W.; Sassi, R. A Two-Steps-Ahead Estimator for Bubble Entropy. Entropy 2021, 23. [Google Scholar] [CrossRef]
- Gong, J.; Yang, X.; Wang, H.; Shen, J.; Liu, W.; Zhou, F. Coordinated method fusing improved bubble entropy and artificial Gorilla Troops Optimizer optimized KELM for rolling bearing fault diagnosis. Applied Acoustics 2022, 195, 108844. [Google Scholar] [CrossRef]
- Gong, J.; Yang, X.; Qian, K.; Chen, Z.; Han, T. Application of improved bubble entropy and machine learning in the adaptive diagnosis of rotating machinery faults. Alexandria Engineering Journal 2023, 80, 22–40. [Google Scholar] [CrossRef]
- Jiang, X.; Yi, Y.; Wu, J. Analysis of the synergistic complementarity between bubble entropy and dispersion entropy in the application of feature extraction. Frontiers in Physics 2023, 11. [Google Scholar] [CrossRef]
- Li, P.; He, X.; Song, D.; Ding, Z.; Qiao, M.; Cheng, X.; Li, R. Improved Categorical Cross-Entropy Loss for Training Deep Neural Networks with Noisy Labels. Pattern Recognition and Computer Vision; Ma, H.; Wang, L.; Zhang, C.; Wu, F.; Tan, T.; Wang, Y.; Lai, J.; Zhao, Y., Eds. Springer International Publishing, 2021, pp. 78–89. [CrossRef]
- Spindelböck, T.; Ranftl, S.; von der Linden, W. Cross-Entropy Learning for Aortic Pathology Classification of Artificial Multi-Sensor Impedance Cardiography Signals. Entropy 2021, 23, 1661. [Google Scholar] [CrossRef]
- Farebrother, J.; Orbay, J.; Vuong, Q.; Taïga, A.A.; Chebotar, Y.; Xiao, T.; Irpan, A.; Levine, S.; Castro, P.S.; Faust, A.; Kumar, A.; Agarwal, R. Stop Regressing: Training Value Functions via Classification for Scalable Deep RL. arXiv:2403.03950, arXiv:2403.03950 2024, [2403.03950]. [CrossRef]
- Arazo, E.; Ortego, D.; Albert, P.; O’Connor, N.E.; McGuinness, K. Pseudo-Labeling and Confirmation Bias in Deep Semi-Supervised Learning. arXiv:1908.02983, arXiv:1908.02983 2020. [CrossRef]
- Mao, A.; Mohri, M.; Zhong, Y. Cross-entropy loss functions: theoretical analysis and applications. arXiv:2304.07288, arXiv:2304.07288. [CrossRef]
- Berrada, L.; Zisserman, A.; Kumar, M.P. Smooth Loss Functions for Deep Top-k Classification. arXiv:1802.07595, arXiv:1802.07595 2018, [1802.07595]. [CrossRef]
- Wang, Z.; Zhu, Q. A Cross-Entropy Based Feature Selection Method for Binary Valued Data Classification; Springer International Publishing, 2022; pp. 1406–1416. [CrossRef]
- Kim, S.C.; Kang, T.J. Texture classification and segmentation using wavelet packet frame and Gaussian mixture model. Pattern Recognition 2007, 40, 1207–1221. [Google Scholar] [CrossRef]
- Bruch, S. An Alternative Cross Entropy Loss for Learning-to-Rank. arXiv:1911.09798, arXiv:1911.09798 2021. [CrossRef]
- Santosa, B. Multiclass Classification with Cross Entropy-Support Vector Machines. Procedia Computer Science 2015, 72, 345–352. [Google Scholar] [CrossRef]
- Śmieja, M.; Geiger, B.C. Semi-supervised cross-entropy clustering with information bottleneck constraint. Information Sciences 2017, 421, 254–271. [Google Scholar] [CrossRef]
- Orchard, M.E.; Olivares, B.; Cerda, M.; Silva, J.F. Anomaly Detection based on Information-Theoretic Measures and Particle Filtering Algorithms. Annual Conference of the Prognostics and Health Management (PHM) Society 2012, 4. [Google Scholar] [CrossRef]
- Bishop, C.M. Pattern Recognition and Machine Learning; Springer Science + Business Media, 2006.
- Qu, Y.; Li, R.; Deng, A.; Shang, C.; Shen, Q. Non-unique Decision Differential Entropy-Based Feature Selection. Neurocomputing 2020, 393, 187–193. [Google Scholar] [CrossRef]
- Grassucci, E.; Comminiello, D.; Uncini, A. An Information-Theoretic Perspective on Proper Quaternion Variational Autoencoders. Entropy 2021, 23, 856. [Google Scholar] [CrossRef]
- Gibson, J. Entropy Power, Autoregressive Models, and Mutual Information. Entropy 2018, 20, 750. [Google Scholar] [CrossRef]
- Robin, S.; Scrucca, L. Mixture-based estimation of entropy. Computational Statistics & Data Analysis 2023, 177, 107582. [Google Scholar] [CrossRef]
- Rostaghi, M.; Azami, H. Dispersion Entropy: A Measure for Time-Series Analysis. IEEE Signal Processing Letters 2016, 23, 610–614. [Google Scholar] [CrossRef]
- Rostaghi, M.; Khatibi, M.M.; Ashory, M.R.; Azami, H. Refined Composite Multiscale Fuzzy Dispersion Entropy and Its Applications to Bearing Fault Diagnosis. Entropy 2023, 25, 1494. [Google Scholar] [CrossRef]
- Furlong, R.; Hilal, M.; O’Brien, V.; Humeau-Heurtier, A. Parameter Analysis of Multiscale Two-Dimensional Fuzzy and Dispersion Entropy Measures Using Machine Learning Classification. Entropy 2021, 23, 1303. [Google Scholar] [CrossRef]
- Hu, B.; Wang, Y.; Mu, J. A new fractional fuzzy dispersion entropy and its application in muscle fatigue detection. Mathematical Biosciences and Engineering 2024, 21, 144–169. [Google Scholar] [CrossRef]
- Dhandapani, R.; Mitiche, I.; McMeekin, S.; Mallela, V.S.; Morison, G. Enhanced Partial Discharge Signal Denoising Using Dispersion Entropy Optimized Variational Mode Decomposition. Entropy 2021, 23, 1567. [Google Scholar] [CrossRef]
- Li, G.; Yang, Z.; Yang, H. A Denoising Method of Ship Radiated Noise Signal Based on Modified CEEMDAN, Dispersion Entropy, and Interval Thresholding. Electronics 2019, 8, 597. [Google Scholar] [CrossRef]
- Fabila-Carrasco, J.S.; Tan, C.; Escudero, J. Graph-Based Multivariate Multiscale Dispersion Entropy: Efficient Implementation and Applications to Real-World Network Data. arXiv:2405.00518, arXiv:2405.00518 2024, [2405.00518]. [CrossRef]
- Ge, H.; Chen, G.; Yu, H.; Chen, H.; An, F. Theoretical Analysis of Empirical Mode Decomposition. Symmetry 2018, 10, 623. [Google Scholar] [CrossRef]
- Liu, C.; Zhu, L.; Ni, C. Chatter detection in milling process based on VMD and energy entropy. Mechanical Systems and Signal Processing 2018, 105, 169–182. [Google Scholar] [CrossRef]
- Gao, Z.; Liu, Y.; Wang, Q.; Wang, J.; Luo, Y. Ensemble empirical mode decomposition energy moment entropy and enhanced long short-term memory for early fault prediction of bearing. Measurement 2022, 188, 110417. [Google Scholar] [CrossRef]
- Yu, Y.; Dejie, Y.; Junsheng, C. A roller bearing fault diagnosis method based on EMD energy entropy and ANN. Journal of Sound and Vibration 2006, 294, 269–277. [Google Scholar] [CrossRef]
- Yang, Z.; Luo, S.; Zhong, P.; Chen, R.; Pan, C.; Li, K. An EMD and IMF Energy Entropy-Based Optimized Feature Extraction and Classification Scheme for Single Trial EEG Signal. Journal of Mechanics in Medicine and Biology 2023, 23, 2340063. [Google Scholar] [CrossRef]
- Zhu, G.; Peng, S.; Lao, Y.; Su, Q.; Sun, Q. Short-Term Electricity Consumption Forecasting Based on the EMD-Fbprophet-LSTM Method. Mathematical Problems in Engineering 2021, 2021, 6613604. [Google Scholar] [CrossRef]
- Gao, J.; Shang, P. Analysis of complex time series based on EMD energy entropy plane. Nonlinear Dynamics 2019, 96, 465–482. [Google Scholar] [CrossRef]
- Headrick, M. Lectures on entanglement entropy in field theory and holography. arXiv:1907.08126, arXiv:1907.08126 2019, [1907.08126]. [CrossRef]
- Rieger, M.; Reh, M.; Gärtner, M. Sample-efficient estimation of entanglement entropy through supervised learning. Physical Review A 2024, 109, 012403. [Google Scholar] [CrossRef]
- Liu, Y.; Li, W.J.; Zhang, X.; Lewenstein, M.; Su, G.; Ran, S.J. Entanglement-Based Feature Extraction by Tensor Network Machine Learning. Frontiers in Applied Mathematics and Statistics 2021, 7, 716044. [Google Scholar] [CrossRef]
- Lin, X.; Chen, Z.; Wei, Z. Quantifying Unknown Quantum Entanglement via a Hybrid Quantum-Classical Machine Learning Framework. Physical Review A 2023, 107. [Google Scholar] [CrossRef]
- Abdallah, S.A.; Plumbley, M.D. A measure of statistical complexity based on predictive information with application to finite spin systems. Physics Letters A 2012, 376, 275–281. [Google Scholar] [CrossRef]
- Crutchfield, J.P.; Packard, N.H. Symbolic dynamics of noisy chaos. Physica D: Nonlinear Phenomena 1983, 7, 201–223. [Google Scholar] [CrossRef]
- Bardera, A.; Boada, I.; Feixas, M., S. M. Image Segmentation Using Excess Entropy. Journal of Signal Processing Systems 2009, 54, 205–214. [Google Scholar] [CrossRef]
- Nir, A.; Sela, E.; Beck, R.; Bar-Sinai, Y. Machine-Learning Iterative Calculation of Entropy for Physical Systems. Proceedings of the National Academy of Sciences 2020, 117, 30234–30240. [Google Scholar] [CrossRef]
- Belghazi, M.I.; Baratin, A.; Rajeswar, S.; Ozair, S.; Bengio, Y.; Courville, A.; Hjelm, R.D. MINE: Mutual Information Neural Estimation. arXiv:1801.04062, arXiv:1801.04062 2021.
- Xiang, X.; Zhou, J. An Excess Entropy Approach to Classify Long-Term and Short-Term Memory Stationary Time Series. Mathematics 2023, 11, 2448. [Google Scholar] [CrossRef]
- Azami, H.; Escudero, J. Amplitude- and Fluctuation-Based Dispersion Entropy. Entropy 2018, 20. [Google Scholar] [CrossRef]
- Chen, Y.; Chen, J.; Qiang, Y.; Yuan, Z.; Yang, J. Refined composite moving average fluctuation dispersion entropy and its application on rolling bearing fault diagnosis. Review of Scientific Instruments 2023, 94, 105110. [Google Scholar] [CrossRef]
- Su, H.; Wang, Z.; Cai, Y.; Ding, J.; Wang, X.; Yao, L. Refined Composite Multiscale Fluctuation Dispersion Entropy and Supervised Manifold Mapping for Planetary Gearbox Fault Diagnosis. Machines 2023, 11, 47. [Google Scholar] [CrossRef]
- Zhou, F.; Han, J.; Yang, X. Multivariate hierarchical multiscale fluctuation dispersion entropy: Applications to fault diagnosis of rotating machinery. Applied Acoustics 2021, 182, 108271. [Google Scholar] [CrossRef]
- Li, Z.; Lan, T.; Li, Z.; Gao, P. Exploring Relationships between Boltzmann Entropy of Images and Building Classification Accuracy in Land Cover Mapping. Entropy 2023, 25, 1182. [Google Scholar] [CrossRef]
- Baldini, G.; Chareau, J.M.; Bonavitacola, F. Spectrum Sensing Implemented with Improved Fluctuation-Based Dispersion Entropy and Machine Learning. Entropy 2021, 23, 1611. [Google Scholar] [CrossRef]
- Azami, H.; Arnold, S.E.; Sanei, S.; Chang, Z.; Sapiro, G.; Escudero, J.; Gupta, A.S. Multiscale Fluctuation-based Dispersion Entropy and its Applications to Neurological Diseases. ArXiv:1902.10825, 1902. [Google Scholar] [CrossRef]
- Jiao, S.; Geng, B.; Li, Y.; Zhang, Q.; Wang, Q. Fluctuation-based reverse dispersion entropy and its applications to signal classification. Applied Acoustics 2021, 175, 107857. [Google Scholar] [CrossRef]
- Amigó, J.M.; Szczepanski, J.; Wajnryb, E.; Sanchez-Vives, M.V. Estimating the entropy of spike trains via Lempel-Ziv complexity. Neural Computation 2004, 16, 717–736. [Google Scholar] [CrossRef]
- Friedgut, E.; Kalai, G. Every Monotone Graph Property Has A Sharp Threshold. Proceedings of the American Mathematical Society 1999, 124, 2993–3002. [Google Scholar] [CrossRef]
- Chakraborty, S.; Kulkarni, R.; Lokam, S.; Saurabh, N. Upper bounds on Fourier entropy. Theoretical Computer Science 2016, 654, 92–112. [Google Scholar] [CrossRef]
- O’Donnell, R.; Wright, J.; Zhou, Y. The Fourier Entropy–Influence Conjecture for Certain Classes of Boolean Functions. Automata, Languages and Programming; Aceto, L.; Henzinger, M.; Sgall, J., Eds. Springer Berlin Heidelberg, 2011, pp. 330–341. [CrossRef]
- Kelman, E.; Kindler, G.; Lifshitz, N.; Minzer, D.; Safra, M. Towards a Proof of the Fourier–Entropy Conjecture? 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS), 2020, pp. 247–258. [CrossRef]
- Almeida, L.B. The fractional Fourier transform and time-frequency representations. IEEE Transactions on Signal Processing 1994, 42, 3084–3091. [Google Scholar] [CrossRef]
- Tao, R.; Zhao, X.; Li, W.; Li, H.C.; Du, Q. Hyperspectral Anomaly Detection by Fractional Fourier Entropy. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 2019, 12, 4920–4929. [Google Scholar] [CrossRef]
- Zhang, L.; Ma, J.; Cheng, B.; Lin, F. Fractional Fourier Transform-Based Tensor RX for Hyperspectral Anomaly Detection. Remote Sensing 2022, 14, 797. [Google Scholar] [CrossRef]
- Wang, S.H.; Zhang, X.; Zhang, Y.D. DSSAE: Deep Stacked Sparse Autoencoder Analytical Model for COVID-19 Diagnosis by Fractional Fourier Entropy. ACM Transactions on Management Information System (TMIS) 2021, 13. [Google Scholar] [CrossRef]
- Wang, S.; Zhang, Y.; Yang, X.; Sun, P.; Dong, Z.; Liu, A.; Yuan, T.F. Pathological Brain Detection by a Novel Image Feature—Fractional Fourier Entropy. Entropy 2015, 17, 8278–8296. [Google Scholar] [CrossRef]
- Yan, Y. Gingivitis detection by fractional Fourier entropy with optimization of hidden neurons. International Journal of Cognitive Computing in Engineering 2020, 1, 36–44. [Google Scholar] [CrossRef]
- Panahi, F.; Rashidi, S.; Sheikhani, A. Application of fractional Fourier transform in feature extraction from Electrocardiogram and Galvanic Skin RESPONSE for emotion recognition. Biomedical Signal Processing and Control 2021, 69, 102863. [Google Scholar] [CrossRef]
- Zhang, Y.; Yang, X.; Cattani, C.; Rao, R.V.; Wang, S.; Phillips, P. Tea Category Identification Using a Novel Fractional Fourier Entropy and Jaya Algorithm. Entropy 2016, 18, 77. [Google Scholar] [CrossRef]
- Ishikawa, A.; Mieno, H. The fuzzy entropy concept and its application. Fuzzy Sets and Systems 1979, 2, 113–123. [Google Scholar] [CrossRef]
- Zheng, J.; Cheng, J.; Yang, Y.; Luo, S. A rolling bearing fault diagnosis method based on multi-scale fuzzy entropy and variable predictive model-based class discrimination. Mechanism and Machine Theory 2014, 78, 187–200. [Google Scholar] [CrossRef]
- Markechová, D.; Riečan, B. Entropy of Fuzzy Partitions and Entropy of Fuzzy Dynamical Systems. Entropy 2016, 18, 19. [Google Scholar] [CrossRef]
- D’Urso, P.; De Giovanni, L.; Vitale, V. Robust DTW-based entropy fuzzy clustering of time series. Annals of Operations Research 2023. [Google Scholar] [CrossRef]
- Di Martino, F.; Sessa, S. Energy and Entropy Measures of Fuzzy Relations for Data Analysis. Entropy 2018, 20, 424. [Google Scholar] [CrossRef]
- Aguayo-Tapia, S.; Avalos-Almazan, G.; Rangel-Magdaleno, J.J. Entropy-Based Methods for Motor Fault Detection: A Review. Entropy 2024, 26, 1–21. [Google Scholar] [CrossRef]
- Jinde, Z.; Chen, M.J.; Junsheng, C.; Yang, Y. Multiscale fuzzy entropy and its application in rolling bearing fault diagnosis. Zhendong Gongcheng Xuebao/Journal of Vibration Engineering 2014, 27, 145–151. [Google Scholar]
- Kumar, R.; Bisht, D.C.S. Picture fuzzy entropy: A novel measure for managing uncertainty in multi-criteria decision-making. Decision Analytics Journal 2023, 9, 100351. [Google Scholar] [CrossRef]
- Lhermitte, E.; Hilal, M.; Furlong, R.; O’Brien, V.; Humeau-Heurtier, A. Deep Learning and Entropy-Based Texture Features for Color Image Classification. Entropy 2022, 24, 1577. [Google Scholar] [CrossRef]
- Tan, Z.; Li, K.; Wang, Y. An improved cuckoo search algorithm for multilevel color image thresholding based on modified fuzzy entropy. Journal of Ambient Intelligence and Humanized Computing 2021. [Google Scholar] [CrossRef]
- Korner, J. Coding of an information source having ambiguous alphabet and the entropy of graphs. Transactions of the 6th Prague conference on Information Theory.
- Simonyi, G. Graph Entropy: A Survey. Combinatorial Optimization 1995, 20, 399–444. [Google Scholar]
- Luque, B.; Lacasa, L.; Ballesteros, F.; Luque, J. Horizontal visibility graphs: exact results for random time series. Physical Review E 2009, 80, 046103. [Google Scholar] [CrossRef]
- Lacasa, L.; Just, W. Visibility graphs and symbolic dynamics. Physica D: Nonlinear Phenomena 2018, 374-375, 35–44. [Google Scholar] [CrossRef]
- Harangi, V.; Niu, X.; Bai, B. Conditional graph entropy as an alternating minimization problem. arXiv:2209.00283, arXiv:2209.00283 2023, [2209.00283]. [CrossRef]
- Wu, J.; Chen, X.; Xu, K.; Li, S. Structural Entropy Guided Graph Hierarchical Pooling. arXiv:2206.13510, arXiv:2206.13510 2022, [2206.13510]. [CrossRef]
- Juhnke-Kubitzke, M.; Köhne, D.; Schmidt, J. Counting Horizontal Visibility Graphs. arXiv:2111.02723, arXiv:2111.02723 2021, [2111.02723]. [CrossRef]
- Luo, G.; Li, J.; Su, J.; Peng, H.; Yang, C.; Sun, L.; Yu, P.S.; He, L. Graph Entropy Guided Node Embedding Dimension Selection for Graph Neural Networks. arXiv:2105.03178, arXiv:2105.03178 2021, [2105.03178]. [CrossRef]
- Zhu, G.; Li, Y.; Wen, P. Analysis of alcoholic EEG signals based on horizontal visibility graph entropy. Brain Informatics 2014, 1, 19–25. [Google Scholar] [CrossRef]
- Yu, M.; Hillebrand, A.; Gouw, A.; Stam, C.J. Horizontal visibility graph transfer entropy (HVG-TE): A novel metric to characterize directed connectivity in large-scale brain networks. NeuroImage 2017, 156, 249–264. [Google Scholar] [CrossRef]
- Chen, T.; Vemuri, B.; Rangarajan, A.; Eisenschenk, S.J. Group-Wise Point-Set Registration Using a Novel CDF-Based Havrda-Charvát Divergence. International Journal of Computer Vision 2010, 86, 111–124. [Google Scholar] [CrossRef]
- Shi, Y.; Wu, Y.; Shang, P. Research on weighted Havrda–Charvat’s entropy in financial time series. Physica A: Statistical Mechanics and its Applications 2021, 572, 125914. [Google Scholar] [CrossRef]
- Brochet, T.; Lapuyade-Lahorgue, J.; Bougleux, S.; Salaun, M.; Ruan, S. Deep learning using Havrda-Charvat entropy for classification of pulmonary endomicroscopy. arXiv:2104.05450, arXiv:2104.05450 2021, [2104.05450]. [CrossRef]
- Brochet, T.; Lapuyade-Lahorgue, J.; Huat, A.; Thureau, S.; Pasquier, D.; Gardin, I.; Modzelewski, R.; Gibon, D.; Thariat, J.; Grégoire, V.; Vera, P.; Ruan, S. A Quantitative Comparison between Shannon and Tsallis–Havrda–Charvat Entropies Applied to Cancer Outcome Prediction. Entropy 2022, 24, 436. [Google Scholar] [CrossRef]
- Amoud, H.; Snoussi, H.; Hewson, D.; Doussot, M.; Duchene, J. Intrinsic Mode Entropy for Nonlinear Discriminant Analysis. IEEE Signal Processing Letters 2007, 14, 297–300. [Google Scholar] [CrossRef]
- E. , K.V.; Hadjileontiadis, L.J. Intrinsic mode entropy: An enhanced classification means for automated Greek Sign Language gesture recognition. 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 2008, pp. 5057–5060. [Google Scholar] [CrossRef]
- Hu, M.; Liang, H. Intrinsic mode entropy based on multivariate empirical mode decomposition and its application to neural data analysis. Cognitive Neurodynamics 2011, 5, 277–284. [Google Scholar] [CrossRef]
- Hu, M.; Liang, H. , Multiscale Entropy: Recent Advances. In Complexity and Nonlinearity in Cardiovascular Signals; Barbieri, R.; Scilingo, E.; Valenza, G., Eds.; Springer International Publishing, 2017; pp. 115–138. [CrossRef]
- Amoud, H.; Snoussi, H.; Hewson, D.; Duchêne, J. Intrinsic Mode Entropy for postural steadiness analysis. 4th European Conference of the International Federation for Medical and Biological Engineering. Springer, Berlin, Heidelberg, 2009, Vol. 22, IFMBE Proceedings, pp. 212–215. [CrossRef]
- Kaniadakis, G. Statistical mechanics in the context of special relativity. Physical Review E 2002, 66, 056125. [Google Scholar] [CrossRef]
- Kaniadakis, G. Relativistic Roots of κ-Entropy. Entropy 2024, 26, 406. [Google Scholar] [CrossRef]
- Lei, B.; Fan, J.I. Adaptive Kaniadakis entropy thresholding segmentation algorithm based on particle swarm optimization. Soft Computing 2020, 24, 7305–7318. [Google Scholar] [CrossRef]
- Jena, B.; Naik, M.K.; Panda, R. A novel Kaniadakis entropy-based multilevel thresholding using energy curve and Black Widow optimization algorithm with Gaussian mutation. 2023 International Conference in Advances in Power, Signal, and Information Technology (APSIT), 2023, pp. 86–91. [CrossRef]
- da Silva, S.L.E.F.; de Araújo, J.M.; de la Barra, E.; Corso, G. A Graph-Space Optimal Transport Approach Based on Kaniadakis κ-Gaussian Distribution for Inverse Problems Related to Wave Propagation. Entropy 2023, 25, 990. [Google Scholar] [CrossRef]
- Mekyska, J.; Janousova, E.; Gomez-Vilda, P.; Smekal, Z.; Rektorova, I.; Eliasova, I.; Kostalova, M.; Mrackova, M.; Alonso-Hernandez, J.B.; Faundez-Zanuy, M.; López-de Ipiña, K. Robust and complex approach of pathological speech signal analysis. Neurocomputing 2015, 167, 94–111. [Google Scholar] [CrossRef]
- Xu, L.S.; Wang, K.Q.; Wang, L. Gaussian kernel approximate entropy algorithm for analyzing irregularity of time-series. 2005 International Conference on Machine Learning and Cybernetics, 2005, Vol. 9, pp. 5605–5608. [CrossRef]
- Zaylaa, A.; Saleh, S.; Karameh, F.; Nahas, Z. andBouakaz, A. Cascade of nonlinear entropy and statistics to discriminate fetal heart rates. Proceedings of the 2016 3rd International Conference on Advances in Computational Tools for Engineering Applications (ACTEA). IEEE Xplore, 2016, pp. 152–157. [CrossRef]
- Orozco-Arroyave, J.; Arias-Londono, J.; Vargas-Bonilla, J.; Nöth, E. Analysis of speech from people with Parkinson’s disease through nonlinear dynamics. Advances in Nonlinear Speech Processing. NOLISP 2013. Lecture Notes in Computer Science, vol 7911; Drugman, T.; Dutoit, T., Eds. Springer, 2013, pp. 112–119. [CrossRef]
- Kolmogorov, A.N. A New Metric Invariant of Transitive Dynamical Systems and Automorphisms of Lebesgue Spaces. Proceedings of the Steklov Institute of Mathematics 1986, 169, 97–102. [Google Scholar]
- Pesin, Y. Characteristic Lyapunov exponents and smooth ergodic theory. Russian Mathematical Surveys 1977, 32, 55. [Google Scholar] [CrossRef]
- Stolz, I.; Keller, K. A General Symbolic Approach to Kolmogorov-Sinai Entropy. Entropy 2017, 19, 675. [Google Scholar] [CrossRef]
- Shiozawa, K.; Tokuda, I. Estimating Kolmogorov-Sinai entropy from time series of high-dimensional complex systems. Physics Letters A 2024, 510, 129531. [Google Scholar] [CrossRef]
- Karmakar, C.; Udhayakumar, R.; Palaniswami, M. Entropy Profiling: A Reduced-Parametric Measure of Kolmogorov-Sinai Entropy from Short-Term HRV Signal. Entropy 2020, 22, 1396. [Google Scholar] [CrossRef]
- Kiss, G.; Bakucz, P. Using Kolmogorov Entropy to Verify the Description Completeness of Traffic Dynamics of Highly Autonomous Driving. Applied Sciences 2024, 14, 2261. [Google Scholar] [CrossRef]
- Aftanas, L.I.; Lotova, N.V.; Koshkarov, V.I.; Pokrovskaja, V.L.; Popov, S.A.; Makhnev, V.P. Non-linear analysis of emotion EEG: calculation of Kolmogorov entropy and the principal Lyapunov exponent. Neuroscience Letters 1997, 226, 13–16. [Google Scholar] [CrossRef]
- Bandt, C.; Pompe, B. Permutation Entropy: A Natural Complexity Measure for Time Series. Physical Review Letters 2002, 88, 174102. [Google Scholar] [CrossRef]
- Bandt, C.; Keller, G.; Pompe, B. Entropy of interval maps via permutations. Nonlinearity 2002, 15, 1595. [Google Scholar] [CrossRef]
- Riedl, M.; Müller, A.; Wessel, N. Practical considerations of permutation entropy. The European Physical Journal Special Topics 2013, 222, 249–262. [Google Scholar] [CrossRef]
- Amigó, J.M. Permutation Complexity in Dynamical Systems; Springer Verlag, 2010. [CrossRef]
- Zanin, M.; Zunino, L.; Rosso, O.A.; Papo, D. Permutation Entropy and Its Main Biomedical and Econophysics Applications: A Review. Entropy 2012, 14, 1553–1577. [Google Scholar] [CrossRef]
- Amigó, J.M.; Keller, K.; Kurths, J. Recent Progress in Symbolic Dynamics and Permutation Complexity —Ten Years of Permutation Entropy. The European Physical Journal Special Topics 2013, 222, 241–247. [Google Scholar] [CrossRef]
- Amigó, J.M.; Rosso, O.A. Ordinal methods: Concepts, applications, new developments, and challenges—In memory of Karsten Keller (1961–2022). Chaos 2023, 33, 080401. [Google Scholar] [CrossRef]
- Mammone, N.; Duun-Henriksen, J.; Kjaer, T.W.; Morabito, F.C. Differentiating Interictal and Ictal States in Childhood Absence Epilepsy through Permutation Rényi Entropy. Entropy 2015, 17, 4627–4643. [Google Scholar] [CrossRef]
- Zunino, L.; Pérez, D.G.; Kowalski, A.; Martin, M.T.; Garavaglia, M.; Plastino, A.; Rosso, O.A. Fractional Brownian motion, fractional Gaussian noise, and Tsallis permutation entropy. Physica A: Statistical Mechanics and its Applications 2008, 387, 6057–6068. [Google Scholar] [CrossRef]
- Stosic, D.; Stosic, D.; Stosic, T.; Stosic, B. Generalized weighted permutation entropy. Chaos 2022, 32. [Google Scholar] [CrossRef]
- Yin, Y.; Sun, K.; He, S. Multiscale permutation Rényi entropy and its application for EEG signals. PLOS ONE 2018, 13, 1–15. [Google Scholar] [CrossRef]
- Li, C.; Shang, P. Multiscale Tsallis permutation entropy analysis for complex physiological time series. Physica A 2019, 523, 10–20. [Google Scholar] [CrossRef]
- Azami, H.; Escudero, J. Improved multiscale permutation entropy for biomedical signal analysis: Interpretation and application to electroencephalogram recordings. Biomedical Signal Processing and Control 2016, 23, 28–41. [Google Scholar] [CrossRef]
- Keller, K.; Lauffer, H. Symbolic Analysis of High-Dimensional Time Series. International Journal of Bifurcation and Chaos 2003, 13, 2657–2668. [Google Scholar] [CrossRef]
- Keller, K.; Mangold, T.; Stolz, L.; Werner, J. Permutation Entropy: New Ideas and Challenges. Entropy 2017, 19, 134. [Google Scholar] [CrossRef]
- Amigó, J.M.; Kennel, M.B. Forbidden ordinal patterns in higher dimensional dynamics. Physica D 2008, 237, 2893–2899. [Google Scholar] [CrossRef]
- Amigó, J.M.; Keller, K. Permutation entropy: One concept, two approaches. The European Physical Journal Special Topics 2013, 222, 263–273. [Google Scholar] [CrossRef]
- Carpi, L.C.; Saco, P.M.; Rosso, O.A. Missing ordinal patterns in correlated noises. Physica A 2010, 389, 2020–2029. [Google Scholar] [CrossRef]
- Pilarczyk, P.; Graff, G.; Amigó, J.M.; Tessmer. ; Narkiewicz, K.; Graff, B. Differentiating patients with obstructive sleep apnea from healthy controls based on heart rate-blood pressure coupling quantified by entropy-based indices. Chaos 2023, 33, 103140. [Google Scholar] [CrossRef]
- Qu, Z.; Hou, X.; Hu, W.; Yang, R.; Ju, C. Wind power forecasting based on improved variational mode decomposition and permutation entropy. Clean Energy 2023, 7, 1032–1045. [Google Scholar] [CrossRef]
- Rosso, O.A.; Larrondo, H.A.; Martin, M.T.; Plastino, A.; Fuentes, M.A. Distinguishing Noise from Chaos. Physical Review Letters 2007, 99, 154102. [Google Scholar] [CrossRef]
- Voltarelli, L.G.J.M.; Pessa, A.A.B.; Zunino, L.; Zola, R.S.; Lenzi, E.K.; Perc, M.; Ribeiro, H.V. Characterizing unstructured data with the nearest neighbor permutation entropy. Chaos 2024, 34, 053130. [Google Scholar] [CrossRef]
- Rao, C.R. Diversity and Dissimilarity Coefficients: A Unified Approach. Theoretical Population Biology 1982, 21, 24–43. [Google Scholar] [CrossRef]
- Doxa, A.; Prastacos, P. Using Rao’s quadratic entropy to define environmental heterogeneity priority areas in the European Mediterranean biome. Biological Conservation 2020, 241, 108366. [Google Scholar] [CrossRef]
- Smouse, P.E.; Banks, S.C.; Peakall, R. Converting quadratic entropy to diversity: Both animals and alleles are diverse, but some are more diverse than others. PLOS ONE 2017, 12, 1–19. [Google Scholar] [CrossRef]
- Dionne, G.; Koumou, G. Machine Learning and Risk Management: SVDD Meets RQE. Technical Report 18-6, HEC Montreal, Canada Research Chair in Risk Management, 2018.
- Citi, L. ; G. , G.; Mainardi, L. Rank-based Multi-Scale Entropy analysis of heart rate variability. Proceedings of the Computing in Cardiology 2014, 2014, pp. 597–600. [Google Scholar]
- Garchery, M.; Granitzer, M. On the influence of categorical features in ranking anomalies using mixed data. Procedia Computer Science 2018, 126, 77–86. [Google Scholar] [CrossRef]
- Khan, M.A.; Akram, T.; Sharif, M.; Alhaisoni, M.; Saba, T.; Nawaz, N. A probabilistic segmentation and entropy-rank correlation-based feature selection approach for the recognition of fruit diseases. EURASIP Journal on Image and Video Processing 2021, 2021, 14. [Google Scholar] [CrossRef]
- Hu, Q.; Che, X.; Zhang, L.; Zhang, D.; Guo, M.; Yu, D. Rank Entropy-Based Decision Trees for Monotonic Classification. IEEE Transactions on Knowledge and Data Engineering 2012, 24, 2052–2064. [Google Scholar] [CrossRef]
- Liu, S.; Gao, H. The Structure Entropy-Based Node Importance Ranking Method for Graph Data. Entropy 2023, 25, 941. [Google Scholar] [CrossRef]
- McLellan, M.R.; Ryan, M.D.; Breneman, C.M. Rank order entropy: why one metric is not enough. Journal of Chemical Information and Modeling 2011, 51, 2302–2319. [Google Scholar] [CrossRef]
- Diks, C.; Panchenko, V. Rank-based Entropy Tests for Serial Independence. Studies in Nonlinear Dynamics & Econometrics 2008, 12. [Google Scholar] [CrossRef]
- C. , S.; Li, H.; M., S.; S., H. A Ranking-Based Cross-Entropy Loss for Early Classification of Time Series. IEEE Transactions on Neural Networks and Learning Systems 2024, 35, 11194–11203. [Google Scholar] [CrossRef]
- Niknami, N.; Wu, J. Entropy-KL-ML: Enhancing the Entropy-KL-Based Anomaly Detection on Software-Defined Networks. IEEE Transactions on Network Science and Engineering 2022, 9, 4458–4467. [Google Scholar] [CrossRef]
- Moral, S.; Cano, A.; Gómez-Olmedo, M. Computation of Kullback–Leibler Divergence in Bayesian Networks. Entropy 2021, 23, 1122. [Google Scholar] [CrossRef]
- Brown, G.; Pocock, A.; Zhao, M.J.; Luján, M. Conditional likelihood maximisation: a unifying framework for information theoretic feature selection. The Journal of Machine Learning Research 2012, 13, 27–66. [Google Scholar]
- Chaimovich, M.; Chaimovich, A. Relative Resolution: An Analysis with the Kullback–Leibler Entropy. Journal of Chemical Theory and Computation 2024, 20, 2074–2087. [Google Scholar] [CrossRef]
- Draelos, R. Connections: Log-Likelihood, Cross-Entropy, KL-Divergence, Logistic Regression, and Neural Networks. GlassBox Medicine, .: online, 7 December 2024. [Google Scholar]
- De La Pava Panche, I.; Alvarez-Meza, A.M.; Orozco-Gutierrez, A. A Data-Driven Measure of Effective Connectivity Based on Renyi’s α-Entropy. Frontiers in Neuroscience 2019, 13, 1277. [Google Scholar] [CrossRef]
- Rioul, O. The Interplay between Error, Total Variation, Alpha-Entropy and Guessing: Fano and Pinsker Direct and Reverse Inequalities. Entropy 2023, 25, 978. [Google Scholar] [CrossRef]
- Berezinski, P.; Jasiul, B.; Szpyrka, M. An Entropy-Based Network Anomaly Detection Method. Entropy 2015, 17, 2367–2408. [Google Scholar] [CrossRef]
- Sharma, R.; Pachori, R.B.; Acharya, U.R. Application of Entropy Measures on Intrinsic Mode Functions for the Automated Identification of Focal Electroencephalogram Signals. Entropy 2015, 17, 669–691. [Google Scholar] [CrossRef]
- Czarnecki, W.M.; Tabor, J. Extreme entropy machines: Robust information theoretic classification. Pattern Analysis and Applications 2017, 20, 383–400. [Google Scholar] [CrossRef]
- Sluga, D.; Lotrič, U. Quadratic Mutual Information Feature Selection. Entropy 2017, 19, 157. [Google Scholar] [CrossRef]
- Gowdra, N.; Sinha, R.; MacDonell, S. Examining convolutional feature extraction using Maximum Entropy (ME) and Signal-to-Noise Ratio (SNR) for image classification. IECON 2020 The 46th Annual Conference of the IEEE Industrial Electronics Society, 2020, pp. 471–476. [CrossRef]
- Mammone, N.; La Foresta, F.; Morabito, F.C. Automatic Artifact Rejection from Multichannel Scalp EEG by Wavelet ICA. IEEE Sensors Journal 2012, 12, 533–542. [Google Scholar] [CrossRef]
- Poza, J.; Escudero, J.; Hornero, R.; Fernandez, A.; Sanchez, C.I. Regional Analysis of Spontaneous MEG Rhythms in Patients with Alzheimer’s Disease Using Spectral Entropies. Annals of Biomedical Engineering 2008, 36, 141–152. [Google Scholar] [CrossRef]
- Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropY. Journal of Physiology – Heart and Circulatory Physiology 2000, 278, 2039–H2049. [Google Scholar] [CrossRef]
- Shang, Y.; Lu, G.; Kang, Y.; Zhou, Z.; Duan, B.; Zhang, C. A multi-fault diagnosis method based on modified Sample Entropy for lithium-ion battery strings. Journal of Power Sources 2020, 446, 227275. [Google Scholar] [CrossRef]
- Silva, L.E.V.; Senra Filho, A.C.S.; Fazan, V.P.S.; Felipe, J.C.; Murta Junior, L.O. Two-dimensional sample entropy: assessing image texture through irregularity. Biomedical Physics & Engineering Express 2016, 2, 045002. [Google Scholar] [CrossRef]
- Liao, F.; Jan, Y.K. Using Modified Sample Entropy to Characterize Aging-Associated Microvascular Dysfunction. Frontiers in Physiology 2016, 7. [Google Scholar] [CrossRef]
- Nam N., Q. D.; Liu, A.B.; Lin, C.W. Development of a Neurodegenerative Disease Gait Classification Algorithm Using Multiscale Sample Entropy and Machine Learning Classifiers. Entropy 2020, 22, 1340. [Google Scholar] [CrossRef]
- Lake, D.E. Renyi Entropy Measures of Heart Rate Gaussianity. IEEE Transactions on Biomedical Engineering 2005, 53, 21–27. [Google Scholar] [CrossRef]
- Lake, D.E. Accurate estimation of entropy in very short physiological time series: the problem of atrial fibrillation detection in implanted ventricular devices. American journal of Physiology - Heart and Circulatory Physiology 2011, 300, H319–H325. [Google Scholar] [CrossRef]
- Humeau-Heurtier, A. Evaluation of Systems’ Irregularity and Complexity: Sample Entropy, Its Derivatives, and Their Applications across Scales and Disciplines. Entropy 2018, 20, 794. [Google Scholar] [CrossRef]
- Belyaev, M.; Murugappan, M.; Velichko, A.; Korzun, D. Entropy-Based Machine Learning Model for Fast Diagnosis and Monitoring of Parkinson’s Disease. Sensors 2023, 23, 8609. [Google Scholar] [CrossRef]
- Lin, G.; Lin, A. Modified multiscale sample entropy and cross-sample entropy based on horizontal visibility graph. Chaos, Solitons & Fractals 2022, 165, 112802. [Google Scholar] [CrossRef]
- Karevan, Z.; Suykens, J.A.K. Transductive Feature Selection Using Clustering-Based Sample Entropy for Temperature Prediction in Weather Forecasting. Entropy 2018, 20, 264. [Google Scholar] [CrossRef]
- Guha, R.; Velegol, D. Harnessing Shannon entropy-based descriptors in machine learning models to enhance the prediction accuracy of molecular properties. Journal of Cheminformatics 2023, 15, 54. [Google Scholar] [CrossRef]
- DeMedeiros, K.; Hendawi, A.; Alvarez, M. A Survey of AI-Based Anomaly Detection in IoT and Sensor Networks. Sensors 2023, 23, 1352. [Google Scholar] [CrossRef]
- Evans, S.C.; Shah, T.; Huang, H.; Ekanayake, S.P. The Entropy Economy and the Kolmogorov Learning Cycle: Leveraging the intersection of Machine Learning and Algorithmic Information Theory to jointly optimize energy and learning. Physica D 2024, 461, 134051. [Google Scholar] [CrossRef]
- Rodriguez, N.; Barba, L.; Alvarez, P.; Cabrera-Guerrero, G. Stationary Wavelet-Fourier Entropy and Kernel Extreme Learning for Bearing Multi-Fault Diagnosis. Entropy 2019, 21, 540. [Google Scholar] [CrossRef]
- Soltanian, A.R.; Rabiei, N.; Bahreini, F. Feature Selection in Microarray Data Using Entropy Information. In Computational Biology; Husi, H., Ed.; Exon Publications, 2019; chapter 10. [CrossRef]
- Hoayek, A.; Rullière, D. Assessing Clustering Methods Using Shannon’s Entropy. hal-03812055v1.
- Yang, Z.; Lei, J.; Fan, K.; Lai, Y. Keyword extraction by entropy difference between the intrinsic and extrinsic mode. Physica A 2013, 392, 4523–4531. [Google Scholar] [CrossRef]
- Guo, X.; Li, X.; Xu, R. Fast Policy Learning for Linear Quadratic Control with Entropy Regularization. arXiv:2311.14168, arXiv:2311.14168 2023, [2311.14168]. [CrossRef]
- Coifman, R.; Wickerhauser, M. Entropy-Based Algorithms for Best Basis Selection. IEEE Transactions on Information Theory 1992, 38, 713–718. [Google Scholar] [CrossRef]
- Jaynes, E.T. Information Theory and Statistical Mechanics. Physical Review 1957, 106, 620–630. [Google Scholar] [CrossRef]
- Kapur, J.N.; Kesavan, H.K. Entropy Optimization Principles and Their Applications. In Entropy and Energy Dissipation in Water Resources; Springer: Berlin/Heidelberg, Germany, 1992; pp. 3–20. [Google Scholar] [CrossRef]
- Wang, K.C. Robust Audio Content Classification Using Hybrid-Based SMD and Entropy-Based VAD. Entropy 2020, 22, 183. [Google Scholar] [CrossRef]
- Manzo-Martínez, A.; Gaxiola, F.; Ramírez-Alonso, G.; Martínez-Reyes, F. A Comparative Study in Machine Learning and Audio Features for Kitchen Sounds Recognition. Computación y Sistemas 2022, 26, 4244. [Google Scholar] [CrossRef]
- Civera, M.; Surace, C. An Application of Instantaneous Spectral Entropy for the Condition Monitoring of Wind Turbines. Applied Sciences 2022, 12, 1059. [Google Scholar] [CrossRef]
- Ajmal, M.; Kushki, A.; Plataniotis, K.N. Time-Compression of Speech in Information Talks Using Spectral Entropy. Eighth International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS ’07), 2007, pp. 80–80. [CrossRef]
- Kapucu, F.E.; Välkki, I.; Mikkonen, J.E.; Leone, C.; Lenk, K.; Tanskanen, J.M.A.; Hyttinen, J.A.K. Spectral Entropy Based Neuronal Network Synchronization Analysis Based on Microelectrode Array Measurements. Frontiers in Computational Neuroscience 2016, 10, 112. [Google Scholar] [CrossRef]
- Ra, J.S.; Li, T.; Li, Y. A novel spectral entropy-based index for assessing the depth of anaesthesia. Brain Informatics 2021, 8, 1–12. [Google Scholar] [CrossRef]
- Liu, S.; Li, Z.; Wang, G.; Qiu, X.; Liu, T.; Cao, J.; Zhang, D. Spectral–Spatial Feature Fusion for Hyperspectral Anomaly Detection. Sensors 2024, 24, 1652. [Google Scholar] [CrossRef]
- Rademan, M.W.; Versfeld, D.J.J.; du Preez, J.A. Soft-Output Signal Detection for Cetacean Vocalizations Using Spectral Entropy, K-Means Clustering and the Continuous Wavelet Transform. Ecological Informatics 2023, 74, 101990. [Google Scholar] [CrossRef]
- Oida, E.; Moritani, T.; Yamori, Y. Tone-Entropy Analysis on Cardiac Recovery After Dynamic Exercise. Journal of Applied Physiology 1997, 82, 1794–1801. [Google Scholar] [CrossRef]
- Khandoker, A.H.; Al Zaabi, Y.; Jelinek, H.F. What Can Tone and Entropy Tell Us About Risk of Cardiovascular Diseases? 2019 Computing in Cardiology (CinC), 2019, pp. 1–4. [CrossRef]
- Khandoker, A.; Karmakar, C.; Kimura, Y.; Endo, M.; Oshio, S.; Palaniswami, M. Tone Entropy Analysis of Foetal Heart Rate Variability. Entropy 2015, 17, 1042–1053. [Google Scholar] [CrossRef]
- Karmakar, C.K.; Khandoker, A.H.; Palaniswami, M. Multi-scale Tone Entropy in differentiating physiologic and synthetic RR time series. 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), 2013, pp. 6135–6138. [CrossRef]
- Amigó, J.M.; Giménez, Á. A Simplified Algorithm for the Topological Entropy of Multimodal Maps. Entropy 2014, 16, 627–644. [Google Scholar] [CrossRef]
- Amigó, J.M.; Giménez, A. Formulas for the topological entropy of multimodal maps based on min-max symbols. Discrete and Continuous Dynamical Systems B 2015, 20, 3415–3434. [Google Scholar] [CrossRef]
- Lum, P.; Singh, G.; Lehman, A.; Ishkanov, T.; Vejdemo-Johansson, M.; Alagappan, M.; Carlsson, J.; Carlsson, G. Extracting insights from the shape of complex data using topology. Scientific Reports 2013, 3, 1236. [Google Scholar] [CrossRef]
- McCullough, M.; Small, M.; Iu, H.H.C.; Stemler, T. Multiscale Ordinal Network Analysis of Human Cardiac Dynamics. Philosophical Transactions of the Royal Society A 2017, 375, 20160292. [Google Scholar] [CrossRef]
- Zhao, Y.; Zhang, H. Quantitative Performance Assessment of CNN Units via Topological Entropy Calculation. arXiv:2103.09716, arXiv:2103.09716 2022, [2103.09716]. [CrossRef]
- Jiménez-Alonso, J.F.; López-Martínez, J.; Blanco-Claraco, J.L.; González-Díaz, R.; Sáez, A. A topological entropy-based approach for damage detection of civil engineering structures. 5th International Conference on Mechanical Models in Structural Engineering (CMMoST 2019), 2019, pp. 55–62.
- Rong, L.; Shang, P. Topological entropy and geometric entropy and their application to the horizontal visibility graph for financial time series. Nonlinear Dynamics 2018, 92, 41–58. [Google Scholar] [CrossRef]
- Rucco, M.; Gonzalez-Diaz, R.; Jimenez, M.; Atienza, N.; Cristalli, C.; Concettoni, E.; Ferrante, A.; Merelli, E. A new topological entropy-based approach for measuring similarities among piecewise linear functions. Signal Processing 2017, 134, 130–138. [Google Scholar] [CrossRef]
- Schreiber, T. Measuring Information Transfer. Physical Review Letters 2000, 85, 461–464. [Google Scholar] [CrossRef]
- Granger, C.W.J. Investigating Causal Relations by Econometric Models and Cross-Spectral Methods. Econometrica 1969, 37, 424–438. [Google Scholar] [CrossRef]
- Moldovan, A.; Caţaron, A.; Răzvan, A. Learning in Convolutional Neural Networks Accelerated by Transfer Entropy. Entropy 2021, 23, 1281. [Google Scholar] [CrossRef]
- Moldovan, A.; Caţaron, A.; Andonie, R. Transfer Entropy in Graph Convolutional Neural Networks. arXiv:2406.06632, arXiv:2406.06632 2024. [CrossRef]
- Herzog, S.; Tetzlaff, C.; Wörgötter, F. Transfer entropy-based feedback improves performance in artificial neural networks. arXiv:1706.04265, arXiv:1706.04265 2017, [arXiv:cs.LG/1706.04265]. [CrossRef]
- Duan, Z.; Xu, H.; Huang, Y.; Feng, J.; Wang, Y. Multivariate Time Series Forecasting with Transfer Entropy Graph. Tsinghua Science and Technology 2023, 28, 141–149. [Google Scholar] [CrossRef]
- Amblard, P.O.; Michel, O.J.J. The relation between Granger causality and directed information theory: A Review. Entropy 2013, 15, 113–143. [Google Scholar] [CrossRef]
- Alomani, G.; Kayid, M. Further Properties of Tsallis Entropy and Its Application. Entropy 2023, 25, 199. [Google Scholar] [CrossRef]
- Sharma, S.; Bassi, I. Efficacy of Tsallis Entropy in Clustering Categorical Data. 2019 IEEE Bombay Section Signature Conference (IBSSC), 2019, pp. 1–5. [CrossRef]
- Wu, D.; Jia, H.; Abualigah, L.; Xing, Z.; Zheng, R.; Wang, H.; Altalhi, M. Enhance Teaching-Learning-Based Optimization for Tsallis-Entropy-Based Feature Selection Classification Approach. Processes 2022, 10, 360. [Google Scholar] [CrossRef]
- Naidu, M.S.R.; Kumar, P.R. Tsallis Entropy Based Image Thresholding for Image Segmentation. In Computational Intelligence in Data Mining; Springer, Singapore, 2017; Vol. 556, pp. 371–379. [CrossRef]
- Kalimeri, M.; Papadimitriou, C.; Balasis, G.; Eftaxias, K. Dynamical complexity detection in pre-seismic emissions using nonadditive Tsallis entropy. Physica A 2008, 387, 1161–1172. [Google Scholar] [CrossRef]
- Belanche-Muñoz, L.A.; Wiejacha, M. Analysis of Kernel Matrices via the von Neumann Entropy and Its Relation to RVM Performances. Entropy 2023, 25, 154. [Google Scholar] [CrossRef]
- Hu, F.; Tian, K.; Zhang, Z.K. Identifying Vital Nodes in Hypergraphs Based on Von Neumann Entropy. Entropy 2023, 25, 1263. [Google Scholar] [CrossRef]
- Chen, P.Y.; Wu, L.; Liu, S.; Rajapakse, I. Fast Incremental von Neumann Graph Entropy Computation: Theory, Algorithm, and Applications. arXiv:1805.11769, arXiv:1805.11769 2019, [1805.11769]. [CrossRef]
- Ye, C.; Wilson, R.C.; Hancock, E.R. Network analysis using entropy component analysis. Journal of Complex Networks 2017, 6, 831. [Google Scholar] [CrossRef]
- Huang, Y.; Zhao, Y.; Capstick, A.; Palermo, F.; Haddadi, H.; Barnaghi, P. Analyzing entropy features in time-series data for pattern recognition in neurological conditions. Artificial Intelligence in Medicine 2024, 150, 102821. [Google Scholar] [CrossRef]
- Rosso, O.A.; Blanco, S.; Yordanova, J.; Kolev, V.; Figliola, A.; Schürmann, M.; Başar, E. Wavelet entropy: a new tool for analysis of short duration brain electrical signals. Journal of Neuroscience Methods 2001, 105, 65–75. [Google Scholar] [CrossRef]
- Hu, P.; Zhao, C.; Huang, J.; Song, T. Intelligent and Small Samples Gear Fault Detection Based on Wavelet Analysis and Improved CNN. Processes 2023, 11, 2969. [Google Scholar] [CrossRef]
- Cuomo, K.M.; Oppenheim, A.V.; Strogatz, S.H. Synchronization of Lorenz-Based Chaotic Circuits with Applications to Communications. IEEE Transactions on Circuits and Systems II 1993, 40, 626–633. [Google Scholar] [CrossRef]
- Amigó, J.M.; Kocarev, L.; Tomovski, I. Discrete entropy. Physica D 2007, 228, 77–85. [Google Scholar] [CrossRef]
- Amigó, J.M.; Kocarev, L.; Szczepanski, J. Discrete Lyapunov exponent and resistance to differential cryptanalysis. IEEE Transactions on Circuits and Systems II 2007, 54, 882–886. [Google Scholar] [CrossRef]
- Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning, second edition ed.; Springer: New York, 2009. [Google Scholar] [CrossRef]
- Jenssen, R. Kernel Entropy Component Analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence 2010, 32, 847–860. [Google Scholar] [CrossRef]
- Paninski, L. Estimation of Entropy and Mutual Information. Neural Computation 2003, 15, 1191–1253. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
