Abstract
About 160 years ago, the concept of entropy was introduced in thermodynamics by Rudolf Clausius. Since then, it has been continually extended, interpreted, and applied by researchers in many scientific fields, such as general physics, information theory, chaos theory, data mining, and mathematical linguistics. This paper presents The Entropy Universe, which aims to review the many variants of entropies applied to time-series. The purpose is to answer research questions such as: How did each entropy emerge? What is the mathematical definition of each variant of entropy? How are entropies related to each other? What are the most applied scientific fields for each entropy? We describe in-depth the relationship between the most applied entropies in time-series for different scientific fields, establishing bases for researchers to properly choose the variant of entropy most suitable for their data. The number of citations over the past sixteen years of each paper proposing a new entropy was also accessed. The Shannon/differential, the Tsallis, the sample, the permutation, and the approximate entropies were the most cited ones. Based on the ten research areas with the most significant number of records obtained in the Web of Science and Scopus, the areas in which the entropies are more applied are computer science, physics, mathematics, and engineering. The universe of entropies is growing each day, either due to the introducing new variants either due to novel applications. Knowing each entropy’s strengths and of limitations is essential to ensure the proper improvement of this research field.
Keywords: entropy measures, information theory, time-series, application areas
1. Introduction
Despite its long history, to many, the term entropy still appears not to be easily understood. Initially, the concept was applied to thermodynamics, but it is becoming more popular in other fields. The concept of entropy has a complex history. It has been the subject of diverse reconstructions and interpretations making it very confusing and difficult to understand, implement, and interpret.
Up to the present, many different types of entropy methods have emerged, with a large number of different purposes and possible application areas. Various descriptions and meanings of entropy are provided in the scientific community, bewildering researchers, students, and professors [1,2,3]. The miscellany in the research papers by the widespread use of entropy in many disciplines leads to many contradictions and misconceptions involving entropy, summarized in Von Neumann’s sentence, “Whoever uses the term ‘entropy’ in a discussion always wins since no one knows what entropy really is, so in a debate, one always has the advantage” [4,5].
Researchers have already studied entropy measurement problems, but there are still several questions to answer. In 1983, Batten [6] discussed entropy theoretical ideas that have led to the suggested nexus between the physicists’ entropy concept and measures of uncertainty or information. Amigó et al. [7] presented a review of only generalized entropies, which from a mathematical point of view, are non-negative functions defined on probability distributions that satisfy the first three Shannon–Khinchin axioms [8]. In 2019, Namdari and Zhaojun [9] reviewed the entropy concept for uncertainty quantification of stochastic processes of lithium-ion battery capacity data. However, those works do not present an in-depth analysis of how entropies are related to each other.
Several researchers, such as the ones of reference [10,11,12,13,14,15,16,17,18,19], consider that entropy is an essential tool for time-series analysis and apply this measure in several research areas. Nevertheless, the choice of a specific entropy is made in an isolated and unclear way. There is no complete study on the areas of entropies application in the literature to the best of our knowledge. We believe that a study of the importance and application of each entropy in time-series will help researchers understand and choose the most appropriate measure for their problem.
Hence, considering entropies applied to time-series, the focus of our work is to demystify and clarify the concept of entropy, describing:
How the different concepts of entropy arose.
The mathematical definitions of each entropy.
How the entropies are related to each other.
Which are the areas of application of each entropy and their impact in the scientific community.
2. Building the Universe of Entropies
In this section, we describe in detail how we built our universe of entropies. We describe the entropies, mathematical definitions, the respective origin, and the relationship between each other. We also address some issues with the concept of entropy and, its extension to the study of continuous random variables.
Figure 1 is the timeline (in logarithmic scale) of the universe of entropies covered in this paper. Although we have referred to some entropies in more than one section, in the Figure 1, the colors refer to the section where each entropy has been defined.
Boltzmann, Gibbs, Hartley, quantum, Shannon and Boltzmann-Gibbs-Shannon entropies, the first concepts of entropy are described in Section 2.1. Section 2.2 is dedicated to entropies derived from Shannon entropy such as differential entropy (Section 2.2.1), spectral entropy (Section 2.2.2), tone-entropy (Section 2.2.3), wavelet entropy (Section 2.2.4), empirical mode decomposition energy entropy (Section 2.2.5), and -entropy (Section 2.2.6).
Kolmogorov, topological, and geometric entropies are described in Section 2.3. Section 2.4 describes the particular cases of Rényi entropy (Section 2.4.1), smooth Rényi entropy (Section 2.4.2), and Rényi entropy for continuous random variable and the different definition of quadratic entropy (Section 2.4.3). Havrda–Charvát structural -entropy and Tsallis entropy are detailed in Section 2.5 while permutation entropy and related entropies are described in Section 2.6.
Section 2.7 is dedicated to rank-based and bubble entropies while topological information content, graph entropy and horizontal visibility graph entropy are detailed in Section 2.8. In Section 2.9,are described the approximate, the sample, and related entropies such as the quadratic sample, coefficient of the sample, and intrinsic mode entropies (Section 2.9.1), dispersion entropy and fluctuation-based dispersion entropy (Section 2.9.2), fuzzy entropy (Section 2.9.3), modified sample entropy (Section 2.9.4), fuzzy measure entropy (Section 2.9.5), and kernel entropies (Section 2.9.6).
2.1. Early Times of the Entropy Concept
In 1864, Rudolf Clausius introduced the term entropy in thermodynamics from the Greek word Entropein for transformation and change [20]. The concept of entropy arises, providing a statement of the second law of thermodynamics. Later, statistical mechanics provided a connection between thermodynamic entropy and the logarithm of the number of microstates in the system’s macrostate. This work is attributed to Ludwig Boltzmann and the Boltzmann entropy [21], S, was defined as
(1) |
where is the thermodynamic unit of the measurement of the entropy and is the Boltzmann constant, W is called the thermodynamic probability or statistical weight and is the total number of microscopic states or complexions compatible with the macroscopic state of the system. The Boltzmann entropy, also known as Boltzmann-Gibbs entropy [22] or thermodynamic entropy [23], is a function on phase space and is thus defined for an individual system. Equation (1) was originally formulated by Boltzmann between 1872 and 1875 [24] but later put into its current form by Max Planck in about 1900 [25].
In 1902, Gibbs published Elementary Principles in Statistical Mechanics [26], where he introduced Gibbs entropy. In contrast to Boltzmann entropy, the Gibbs entropy of a macroscopic classical system is a function of a probability distribution over phase space, i.e., an ensemble. The Gibbs entropy [27] in the phase space of system X, at time t, is defined by the quantity, :
(2) |
where is the Boltzmann’s constant and is the density of the temporal evolution of the probability distribution in the phase space.
Boltzmann entropy was defined for a macroscopic state of a system, while Gibbs entropy is the generalization of the Boltzmann entropy for over an ensemble, which is over the probability distribution of macrostates [21,28].
In 1928, the electrical engineer Ralph Hartley [29] proposed that the amount of information associated with any finite set of entities could be understood as a function of the set’s size. Hartley defined the amount of information associated with the finite set X as the logarithm to some base b of the size, as shown in Equation (3). This amount is known as Hartley entropy and is related to the particular cases of Rényi entropy (see Section 2.4.1).
(3) |
In 1932, Von Neumann generalized Gibbs entropy to quantum mechanics, and it is known as Von Neumann entropy (quantum Gibbs entropy or quantum entropy) [30]. The quantum entropy was initially defined as Shannon entropy associated with the density matrix’s eigenvalues, but sixteen years before Shannon entropy was defined [31]. Von Neumann entropy is the quantum generalization of Shannon entropy [32].
In 1948, the American Claude Shannon published “A mathematical theory of communication” in July and October issues of Bell System technical journal [33]. He proposed the notion of entropy to measure how the information within a signal can be quantified with absolute precision as the amount of unexpected data contained in the message. This measure is known as Shannon entropy (SE) and is defined as:
(4) |
where is a time-series, by convention and represents the probability of , . Therefore and .
In 1949, at the request of Shannon employer, Warren Weaver, the paper of Shannon is republished as a book [34], preceded by an introductory exhibition by Weaver. Weaver’s text [35] attempts to explain how Shannon’s ideas can extend far beyond their initial objectives to all sciences that address communication problems in the broad sense. Weaver is sometimes cited as the first author, if not the only author of information theory [36]. Nevertheless, as Weaver himself stated: “No one could realize more keenly than I do that my own contribution to this book is infinitesimal as compared with Shannon’s” [37]. Shannon and Weaver, in the book “The Mathematical Theory of Communication” [38], referred to Tolman (1938), who in turn attributes to Pauli (1933), the definition of entropy Shannon used [39].
Many authors [7,40,41] used the term Boltzmann-Gibbs-Shannon entropy (BGS), generalizing the entropy expressions of Boltzmann, Gibbs, and Shannon and following the ideas of Stratonovich [42] in 1955:
(5) |
where is the Boltzmann constant, W is the number of microstates consistent with the macroscopic constraints of a given thermodynamical system and is the probability that the system is in the microstate i.
In 1957, Khinchin [43] proposed an axiomatic definition of the Boltzmann entropy, based on four requirements, known as the Shannon-Khinchin axioms [8]. He also demonstrated that Shannon entropy is generalized by
(6) |
where k is a positive constant representing the desired unit of measurement. This property enables us to change the logarithm base in the definition, i.e., . Entropy can be changed from one base to another by multiplying by the appropriate factor [44]. Therefore, depending on the application area, instead of the natural logarithm in Equation (4), one can use logarithms from other bases.
There are different approaches to the derivation of Shannon entropy based on different postulates or axioms [44,45,46,47].
Figure 2 represents the origin of the universe of entropies that we propose in this paper, as well as the relationships between the entropies that we describe in this section.
2.2. Entropies Derived from Shannon Entropy
Based on Shannon entropy, many researchers have been devoted to enhancing the performance of Shannon entropy for more accurate complexity estimation, such as differential entropy, spectral entropy, tone-entropy, wavelet entropy, empirical mode decomposition energy entropy, and entropy. Throughout this paper, other entropies related to Shannon entropy have been analyzed, such as Rényi, Tsallis, permutation, and dispersion entropies. However, due to its context, we decided to include them in other sections.
In Figure 3, we represent how the entropies described below are related to each other.
2.2.1. Differential Entropy
Shannon entropy was formalized for discrete probability distributions. However, the concept of entropy can be extended to continuous distributions through a quantity known as differential entropy (DE) (also referred to as continuous entropy). The concept of DE was introduced by Shannon [33] like a continuous-case extension of Shannon entropy [44]. However, further analysis reveals several shortcomings that render it far less useful than it appears.
The information-theoretic quantity for continuous one-dimensional random variables is differential entropy. The DE for a continuous time-series X with probability density is defined as:
(7) |
where S is the support set of the random variable.
In the discrete case, we had a set of axioms from which we derived Shannon entropy and, therefore, a collection of properties that the measure must present. Charles Marsh [48] stated that “continuous entropy on its own proved problematic,” the differential entropy is not a limit of the Shannon entropy for . On the contrary, it differs from the limit of the Shannon entropy by an infinite displacement [48]. Among other problems, continuous entropy can be negative, while discrete entropy is always non-negative. For example, when the continuous random variable U is uniformly distributed over the interval , Equation (7) results in:
(8) |
The entropy value obtained in Equation (8) is negative when the length of the interval is .
2.2.2. Spectral Entropy
In 1992, Kapur et al. [49] proposed spectral entropy (SpEn) that uses the power spectral density obtained from the Fourier transformation method [50]. The power spectral density represents the distribution of power as a function of frequency. Thus, the normalization of yields a probability density function. Using the definition of Shannon entropy, SpEn can be defined as:
(9) |
where is the frequency band. Spectral entropy is usually normalized , where is the number of frequency components in the range .
2.2.3. Tone-Entropy
In 1997, Oida et al. [51] proposed the tone-entropy (T-E) analysis to characterize the time-series of percentage index (PI) of heart period variation. In this paper, the authors used the tone, and the entropy to characterize the time-series PI. Considered a time-series , PI is defined as:
(10) |
The Tone is defined as the first-order moment (arithmetic average) of this PI time-series as:
(11) |
Entropy is defined from the PI’s probability distribution using Shannon’s Equation (4) with instead of .
2.2.4. Wavelet Entropy
Wavelet entropy (WaEn) was introduced by Rosso and Blanco [52] in 2001. The WaEn is an indicator of the disorder degree associated with the multi-frequency signal response [11]. The definition of WaEn is given as follows:
(12) |
where denotes the probability distribution of time-series, and i represents different resolution levels.
2.2.5. Empirical Mode Decomposition Energy Entropy
In 2006, Yu et al. [53] proposed the Empirical Mode Decomposition Energy entropy (EMDEnergyEn) that quantifies the regularity of time-series with the help of the intrinsic mode functions (IMFs) [12] obtained by the empirical mode decomposition (EMD).
Assume that we have obtained n IMFs, three steps are required to obtain the EMDEnergyEn as follows:
- Calculate energy for each ith IMFs :
where m represents the length of IMF.(13) - Calculate the total energy of these n efficient IMFs:
(14) - Calculate the energy entropy of IMF:
where denotes the EMDEnergyEn in the whole of the original signal, and denotes the percentage of the energy of the IMF number i relative to the total energy entropy.(15)
2.2.6. Entropy
In 2011, the entropy was introduced by Chen et al. [54]. This measure is sensitive to the dynamic range of the time-series. The entropy contains two terms, in which the first term measures the probabilistic uncertainty obtained with Shannon entropy. The second term measures the dispersion in the error variable:
(16) |
where is the time-series and .
The entropy converges to DE when the scale () tends to zero. When the scale defaults to the natural numbers ( and therefore ) the entropy is indistinguishable from SE.
2.3. Kolmogorov, Topological and Geometric Entropies
In 1958, Kolmogorov [55] introduced the concept of entropy in the dynamical system as a measure-preserving transformation and studied the attendant property of completely positive entropy (K-property). In 1959, his student Sinai [56,57] formulated Kolmogorov-Sinai entropy (or metric entropy or Kolmogorov entropy [58]) that is suitable for arbitrary automorphisms of Lebesgue spaces. The Kolmogorov-Sinai entropy is equivalent to a generalized version of SE under certain plausible assumptions [59].
For the case of a state space, it can be partitioned into hypercubes of content in an m-dimensional dynamical system and observed at time intervals , defining the Kolmogorov-Sinai entropy as [60]:
(17) |
where denote the joint probability that the state of the system is in the hypercube at the time , at , …, in the hypercube at and . For stationary processes it can be shown that
(18) |
It is impossible to calculate Equation (17) for because different estimation methods have been proposed, such as approximate entropy [61] and Rényi entropy [62], and compression [63].
The Kolmogorov-Sinai entropy and conditional entropy coincide for a stochastic process X, where is the random variable obtained by sampling the process X at the present time n, and [64].
In this case, conditional entropy quantifies the amount of information in the current process that cannot be explained by its history. If the process is random, the system produces information at the maximum rate, producing the maximum conditional entropy. If, on the contrary, the process is completely predictable, the system does not produce new information, and conditional entropy is zero. When the process is stationary, the system produces new information at a constant rate, meaning that the conditional entropy does not change over time [65]. Note that conditional entropy is, more broadly, the entropy of a random variable conditioned to the knowledge of another random variable [44].
In 1965, inspired by Kolmogorov-Sinai entropy, the concept of the topological entropy (TopEn) was introduced by Adler et al. [66] to describe the complexity of a single map acting on a compact metric space. The notion of TopEn is similar to the notion of metric entropy: instead of a probability measure space, we have a metric space [67].
After 1965, many researchers proposed other notions of TopEn [68,69,70,71,72]. Most of the new notions extended the concept to more general functions or spaces, but the idea of measuring the complexity of the systems was preserved among all these new notions. In [56], the authors review the notions of topological entropy, give an overview of the relation between the notions and fundamental properties of topological entropy.
In 2018, Rong and Shang [10] introduced TopEn based on time-series to characterizes the total exponential complexity of a quantified system with a single number. The authors began by choosing a symbolic method to transform the time-series into a symbolic sequence . At the same time, consider k the number of different alphabets of Y and represents the number of different sets of words with length n. Note that and the TopEn of the time-series was defined as:
(19) |
The maximum of is , then TopEn can reach a maximum of 1, while TopEn’s minimum value is 0, which is reached when . Addabbo and Blackmore [73] showed that metric entropy, with the imposition of some additional properties, is a special case of topological entropy and Shannon entropy is a particular form of metric entropy.
In 1988, Walczak et al. [74] introduced the geometric entropy (or foliation entropy) to study a foliation dynamics, which can be considered as a generalization of TopEn of a single group [75,76,77]. Recently, Rong and Shang [10] proposed geometric entropy for time-series based on the multiscale method (see Section 2.10) and the original definition of geometric entropy provided by Walczak. To calculate the value of TopEn and geometric entropy in their work, the authors used horizontal visibility graphs [78] to transform the time-series into a symbolic series. More details about horizontal visibility graphs in Section 2.8.
In Figure 4, we show the relationships between the entropies described in this section.
2.4. Rényi Entropy
In 1961, Rényi entropy (RE) or q-entropy was introduced by Alfréd Rényi [79] and played a significant role in information theory. Rényi entropy, a generalization of Shannon entropy, is a family of functions of order q () for quantifying the diversity, uncertainty, or randomness of a given system defined as:
(20) |
In the following sections, we describe particular cases of Rényi entropy, smooth Rényi entropy, Rényi entropy for a continuous random variable, and discuss issues with the different definitions of quadratic entropy.
Figure 5 illustrates the relations between Shannon entropy, Rényi entropy, particular cases of Rényi entropy, and smooth Rényi entropy.
2.4.1. Particular Cases of Rényi Entropy
There are some particular cases of Rényi entropy [80,81,82], for example, if , is the Hartley entropy (see Equation (3)).
The Rényi entropy converges to the well-known Shannon entropy, eventually, with multiplication by a constant resulting from the base change property, i.e., when and is replaced by .
When , is called collision entropy [80] and is the negative logarithm of the likelihood of two independent random variables with the same probability distribution to have the same value. The collision entropy measures the probability for two elements drawn according to this distribution to collide. Many papers refers to Rényi entropy [81] when using , even when that choice it is not explicit.
The concept of maximum entropy (or Max-Entropy) arose in statistical mechanics in the nineteenth century and has been advocated for use in a broader context by Edwin Thompson Jaynes [83] in 1957. The Max-Entropy can be obtained from Rényi entropy when , and the limit exists, as:
(21) |
It is the largest value of , which justifies the name maximum entropy. The particular cases of Rényi entropy when , and the limit exists as
(22) |
is called minimum entropy (or Min-Entropy) because it is the smallest value of . The Min-Entropy was proposed by Edward Posner [84] in 1975. According to reference [85], the collision entropy and Min-Entropy are related by Equation by the following:
(23) |
2.4.2. Smooth Rényi Entropy
In 2004, Renner and Wolf [86] proposed the -smooth Rényi entropy for characterizing the fundamental properties of a random variable, such as the amount of uniform randomness that may be extracted from the random variable:
(24) |
where is the set of probability distributions which are -close to P; P is the probability distribution with range Z; and . For the particular case of a significant number of independent and identically distributed (i.i.d.) random variables, smooth Rényi entropy approaches the Rényi entropy. In this special case, if the smooth Rényi entropy approaches the Rényi entropy.
2.4.3. Rényi entropy for Continuous Random Variables and the Different Definition of Quadratic Entropy
According to Lake [87], if X is an absolutely continuous random variable with density f, the Rényi entropy of order q (or q-entropy) is defined as:
(25) |
where letting q tend to 1, and using L’Hospitals rule results in differential entropy, i.e., . Lake [87] uses the term quadratic entropy when .
In 1982, Rao [88,89] gave a different definition for quadratic entropy. He introduced quadratic entropy as a new measure of diversity in biological populations, which considers differences between categories of species. For the discrete and finite case, the quadratic entropy is defined as:
(26) |
where is the difference between the i-th and the j-th category and () are the theoretical probabilities corresponding to the s species in the multinomial model.
2.5. Havrda–Charvát Structural Entropy and Tsallis Entropy
The Havrda–Charvát Structural entropy was introduced in 1967 within information theory [90]. It may be considered as a new generalization of the Shannon entropy, different from the generalization given by Rényi. The Havrda–Charvát Structural entropy, , was defined as:
(27) |
for and . When , the Havrda–Charvát Structural -entropy converges for Shannon entropy minus multiplication by a constant ().
In 1988, the concept of Tsallis entropy (TE) was introduced by Constantino Tsallis as a possible generalization of Boltzmann-Gibbs entropy to nonextensive physical systems and prove that the Boltzmann-Gibbs entropy is recovered as [7,91]. TE is identical in form to Havrda–Charvát Structural entropy and Constantino Tsallis defined it as:
(28) |
where k is a positive constant [91]. In particular, when implies that
(29) |
When in Equation (29) we recover Shannon entropy [92]. In Equation (29) if we consider k to be the Boltzmann constant and , , we recover the Boltzmann entropy to .
Figure 6 summarizes the relationships between the entropies that we describe in this section.
2.6. Permutation Entropy and Related Entropies
Bandt and Pompe [13] introduced permutation entropy (PE) in 2002. PE is an ordinal analysis method, in which a given time-series is divided into a series of ordinal patterns for describing the order relations between the present and a fixed number of equidistant past values [93].
A time-series, as a set of N consecutive data points, is defined as . PE can be obtained by the following steps. From the original time-series X, let us define the vectors as:
(30) |
with , where m is the embedding dimension, and is the embedding lag or time delay. Reconstruct the phase space so that the time-series maps to a trajectory matrix ,
(31) |
Each state vector has an ordinal pattern. It is determined by a comparison of neighboring values. Sort in ascending order, and the ordinal pattern is the ranking of . The trajectory matrix is thus transformed into ordinal pattern matrix,
(32) |
It is called permutation because there is a transformation from to and for a given embedding dimension m, at most permutations exist in total.
Select all distinguishable permutations and number them , . For all the possible permutations , the relative frequency is calculated by
(33) |
where and # represents the number of elements.
According to information theory, the information contained in is measured as and PE is finally computed as:
(34) |
Since can maximally reach , PE is generally normalized as:
(35) |
This is the information contained in comparing m consecutive values of the time-series. Bandt and Pompe [13] defined the permutation entropy per symbol of order m (), dividing by since comparisons start with the second value:
(36) |
In reference [13], “Permutation entropy — a natural complexity measure for time-series” the authors also proposed the sorting entropy (SortEn) defined as: , . This entropy determines the information contained in sorting the mth value among the previous when their order is already known.
Based on the probability of the permutation , other entropy can be defined, e.g., Zhao et al. [14], in 2013, introduced Rényi permutation entropy (RPE) as:
(37) |
Zunino et al. [92], based on the definition of TE, proposed the normalized Tsallis permutation entropy (NTPE) and defined it as:
(38) |
In Figure 7, we showed how Tsallis, Rényi, Shannon and the entropies described in this section are related to each other. The rank-based and bubble entropies, described the Section 2.7, are related to permutation entropy and are also represented in the Figure 7.
2.7. Rank-Based Entropy and Bubble Entropy
Citi et al. [94] proposed the rank-based entropy (RbE), in 2014. The RbE consists of an alternative entropy metric based on the amount of shuffling required for ordering the mutual distances between m-long vectors when the next observation is considered, that is, when the corresponding -long vectors are considered [94]. Operationally, RbE can be defined by the following steps:
Compute, for , the mutual distances vectors: and where is the infinity norm of vector that is, and is the index assigned for each pair, with .
Consider vector and find the permutation such that the vector is sorted in ascending order. Now, if the system is deterministic, we expect that if the vectors and are close, then the new observation from each vector should be close too. In other words, should be almost sorted too. Compute inversion count which is a measure of a vector‘s disorder.
Determine the largest index satisfying and compute the number I of inversion pairs such that , and .
- Compute the RbE as:
(39)
The concept of bubble entropy (BEn) was introduced by Manis et al. [95] in 2017 and it is based on permutation entropy (see Section 2.6), where the vectors in the embedding space are ranked and this rank is inspired by rank-based entropy.
The computation of BEn is as follows:
Sort each vector , defined in Equation (30), of elements in ascending order, counting the number of swaps necessary. The number is obtained by bubble sort [96]. For more details about bubble sort see paper [97].
Compute an histogram from values and normalize it by , to obtain the probabilities (describing how likely a given number of swaps is).
Repeat steps 1 to 3 to compute .
- Compute BEn by:
(41)
2.8. Topological Information Content, Graph Entropy and Horizontal Visibility Graph Entropy
In the literature, there are variations in the definition of graph entropy [99]. Inspired by Shannon’s entropy, the first researchers to define and investigate the entropy of graphs were Rashevsky [100], Trucco [101], and Mowshowitz [102,103,104,105]. In 1955, Rashevsky introduced the information measures for graphs called topological information content and defined as:
(42) |
where denotes the number of topologically equivalent vertices in the ith vertex orbit of G and k is the number of different orbits. Vertices are considered as topologically equivalent if they belong to the same orbit of a graph G. In 1956, Trucco [101] introduced similar entropy measures applying the same principle to the edge automorphism group.
In 1968, Mowshowitz [102,103,104,105] introduced an information measure using chromatic decompositions of graphs and explored the properties of structural information measures relative to two different equivalence relations defined on the vertices of a graph. Mowshowitz [102] defined graph entropy as:
(43) |
where , denotes an arbitrary chromatic decomposition of a graph G, is the chromatic number of G.
In 1973, János Körne [106] introduced a different definition of graph entropy linked to problems in information and coding theory. The context was to determine the performance of the best possible encoding of messages emitted by an information source where the symbols belong to a finite vertex set V. According to Körne, the graph entropy of G, denoted by is defined as:
(44) |
where X is chosen uniformly from V, Y ranges over independent sets of G, the joint distribution of X and Y is such that with probability one, and is the mutual information of X and Y. We will not go deeper into the analysis of the graph entropy defined by Körner because it involves the concept of mutual information.
There is no unanimity in the research community regarding the first author that defined graph entropy. Some researchers consider that graph entropy was defined by Mowshowitz in 1968 [50,99] but other researchers consider that it was introduced by János Körne in 1973 [107,108]. New entropies have emerged based on both concepts of graph entropy. When the term graph entropy is mentioned, care must be taken and understand the underlying definition.
The concept of horizontal visibility graph entropy, a measure based on one definition of graph entropy and in the concept of horizontal visibility graph, was proposed by Zhu et al. [109], in 2014. The horizontal visibility graph is a type of complex network, based on a visibility algorithm [78], and was introduced by Luque et al. [110], in 2009. Let a time-series be mapped into a graph where a time point is mapped into a node . The relationship between any two points is represented by an edge , which are connected if and only if the maximal values between and are less than both of them. Therefore, each edge can be defined as [111]:
where implies that the edge does not exist, otherwise it does.
The graph entropy, based on either vertices or edges, can be computed as follows:
(45) |
where is a probability degree k over a degree sequence of the graph G. It is obtained by counting the number of nodes having degree k divided by the size of the degree sequence. Notice that, the more fluctuating the degree sequence is, the larger the graph entropy is. There are others graph entropy calculation methods based on either vertices or edges [99].
We summarize how the entropies described in this section are related to each other, in Figure 8.
2.9. Approximate and Sample Entropies
Pincus [112] introduced the approximate entropy (ApEn), in 1991. ApEn is derived from Kolmogorov entropy [113] and it is a technique used to quantify the amount of regularity and the unpredictability of fluctuations over time-series data [114]. In order to calculate the ApEn the new series of a vector of length m (embedding dimension) are constructed, , based on Equation (30). For each vector , the value , where r is referred as a tolerance value, is computed as:
(46) |
Here the distance between the vector and its neighbor is:
(47) |
The value can also be defined as:
(48) |
where is the Heaviside function
Next, the average of the natural logarithm of is computed for all i:
(49) |
Since in practice N is a finite number, the statistical estimate is computed as:
The disadvantages of ApEn are: it lacks relative consistency, its strong dependence on data length and is often lower than expected for short records [15]. To overcome the disadvantages of ApEn, in 2000 Richman and Moorman [15] proposed the sample entropy (SampEn) to replace ApEn by excluding self-matches and thereby reducing the computing time by one-half in comparison with ApEn. For the SampEn [15] calculation, the same parameters defined for the ApEn, m, and r are required. Considering A as the number of vector pairs of length having , with and B as the total number of template matches of length m also with , the SampEn is defined as:
(50) |
Many entropies related to ApEn and SampEn have been created. In Figure 9, some entropies related to ApEn and SampEn are represented, which are described in the following sections.
2.9.1. Quadratic Sample Entropy, Coefficient of Sample Entropy and Intrinsic Mode Entropy
The SampEn has a strong dependency on the size of the tolerance r. Normally, smaller r values lead to higher and less confident entropy estimates because of falling numbers of matches of length m and, to an even greater extent, matches of length [16].
In 2005, Lake [87] introduced the concept of quadratic sample entropy (QSE) (Lake called it quadratic differential entropy rate) to solve the aforementioned problem. QSE normalizes the value of r and allows any r for any time-series and the results to be compared with any other estimate. The QSE is defined as follows:
(51) |
where A is the number of vector pairs of length having , with and B is the total number of template matches of length m also with , as in SampEn calculation.
In 2010, derived from QSE it was introduced the coefficient of sample entropy (COSEn) [16]. This measure was first devised and applied to the detection of atrial fibrillation through the heart rate. CosEn is computed similarly as QSE:
(52) |
where is the mean value of the time-series.
In 2007, Amoud et al. [115] introduced the intrinsic mode entropy (IME). The IME is essentially the SampEn computed on the cumulative sums of the IMF [12] obtained by the EMD.
2.9.2. Dispersion Entropy and Fluctuation-Based Dispersion Entropy
The SampEn is not fast enough, especially for long signals, and PE, as a broadly used irregularity indicator, considers only the order of the amplitude values and hence some information regarding the amplitudes may be discarded. In 2016, to solve these problems, Rostaghi and Azami [18] proposed the dispersion entropy (DispEn) applied to a univariate signal whose algorithm is as follows:
First, are mapped to c classes, labeled from 1 to c. The classified signal is . To do so, there are a number of linear and nonlinear mapping techniques. For more details see [116].
Each embedding vector with m embedding dimension and time delay is created according to with . Each time-series is mapped to a dispersion pattern , where , , ..., . The number of possible dispersion patterns that can be assigned to each time-series is equal to , since the signal has m members and each member can be one of the integers from 1 to c [18].
- For each potential dispersion patterns , their relative frequency is obtained as follows:
where # represents their cardinality.(53) - Finally, the DispEn value is calculated, based on the SE definition of entropy, as follows:
(54)
In 2018, Azami and Escudero [116] proposed the fluctuation-based dispersion entropy (FDispEn) as a measure to deal with time-series fluctuations. FDispEn considers the differences between adjacent elements of dispersion patterns. According to the authors, this forms vectors with length , which each of their elements changes from to , soon, becoming potential fluctuation-based dispersion patterns. The only difference between DispEn and FDispEn algorithms is the potential patterns used in these two approaches [116].
2.9.3. Fuzzy Entropy
The uncertainty resulting from randomness is best described by probability theory, while the aspects of uncertainty resulting from imprecision are best described by the fuzzy sets introduced by Zadeh [117], in 1965. In 1972, De Luca and Termini [118] used the concept of fuzzy sets and introduced a measure of fuzzy entropy that corresponds to Shannon’s probabilistic measure of entropy. Over the years, other concepts of fuzzy entropy have been proposed [119]. In 2007, Chen et al. [120] introduced fuzzy entropy (FuzzyEn), a measure of time-series regularity, for the characterization of surface electromyography signals. In this case, FuzzyEn is the negative natural logarithm of the probability that two similar vectors for m points remain similar for the next points. This measure of FuzzyEn is similar to ApEn and SampEn, replaces the 0-1 judgment of Heaviside function associated with ApEn and SampEn by a fuzzy relationship function [121], the family of exponential functions , to get a fuzzy measurement of the two vectors similarity based on their shapes. This measure also comprises the removal of the local baseline which may allow for minimizing the effect of non-stationarity in the time-series. Besides possessing the good properties of SampEn superior to ApEn, FuzzyEn also succeeds in giving an entropy definition for the case of small parameters. The method can also be applied to noisy physiological signals with relatively short databases [50]. Consider a time-series with embedding dimension m to calculate FuzzyEn form vector sequence as follows:
(55) |
with . In Equation (55), represents m consecutive x values, starting with the ith point and generalized by removing a baseline:
(56) |
For computing FuzzyEn, consider a certain vector , define the distance function between and as the maximum absolute difference of the corresponding scalar components:
(57) |
with , .
Given n and r, calculate the similarity degree between and through a fuzzy function:
(58) |
where the fuzzy function is the exponential function
(59) |
For each vector , averaging all the similarity degree of its neighboring vectors , we get:
(60) |
Determine the function as:
(61) |
Similarly, form the vector sequence and get the function :
(62) |
(63) |
Finally, we can define the parameter of the sequence as the negative natural logarithm of the deviation of from :
(64) |
which, for finite databases, can be estimated from the statistic:
(65) |
There are three parameters that must be fixed for each calculation of FuzzyEn. The first parameter m, as in ApEn and SampEn, is the length of sequences to be compared. The other two parameters, r and n, determine the width and the gradient of the boundary of the exponential function respectively.
2.9.4. Modified Sample Entropy
The measure SampEn may have some problems in the validity and accuracy because the similarity definition of vectors is based on the Heaviside function, of which the boundary is discontinuous and hard. The sigmoid function is a smoothed and continuous version of the Heaviside function.
In 2008, a modified sample entropy (mSampEn), based on the nonlinear Sigmoid function, was proposed to overcome the limitations of SampEn [17]. The mSampEn is similar to FuzzyEn, the only differences is that instead of Equation (59), mSampEn uses the fuzzy membership function:
(66) |
2.9.5. Fuzzy Measure Entropy
In 2011, based on FuzzyEn definition, the fuzzy measure entropy (FuzzyMEn) was proposed by Liu and Zhao [122]. FuzzyMEn combines local and global similarity in a time-series and allows a has discrimination for time-series with different inherent complexities. It is defined as:
(67) |
where and are obtained by Equation (65) considering in the Equation (55) the local vector sequence and global vector sequence :
(68) |
(69) |
The vector represents m consecutive values but removing the local baseline , which is defined in Equation (56). The vector also represents m consecutive values but removing the global mean value which is defined as:
(70) |
2.9.6. Kernel Entropies
In 2005, Xu et al. proposed another modification of ApEn, the approximate entropy with Gaussian kernel [19]. It exploits the fact that the Gaussian kernel function can be used to give greater weight to nearby points by replacing the Heaviside function, in Equation (48), by . The first kernel proposed is defined as:
(71) |
and the kernel-based entropy (KbEn) [123] is given as:
(72) |
where was defined in Equation (49).
Therefore, if and is the Heaviside function then the resulting entropy value is the ApEn. The same procedure of changing the distance measure can be applied to define the sample entropy with Gaussian kernel [61].
In 2015, Mekyska et al. [124] proposed 6 other kernels based in approximate and sample entropies: exponential kernel entropy (EKE):
(73) |
Laplacian kernel entropy (LKE):
(74) |
circular kernel entropy (CKE):
(75) |
for , and zero otherwise; spherical kernel entropy (SKE):
(76) |
for , and zero otherwise; Cauchy kernel entropy (CauchyKE):
(77) |
for , and zero otherwise; triangular kernel entropy (TKE):
(78) |
for , zero otherwise.
Zaylaa et al. [123], called Gaussian entropy, exponential entropy, circular entropy, spherical entropy, Cauchy entropy and triangular entropy to kernels entropies based in ApEn.
2.10. Multiscale Entropy
The multiscale entropy approach (MSE) [60,125] was inspired by Zhang’s proposal [126], and considers the information of a system’s dynamics on different time scales. Multiscale entropy algorithms are composed of two main steps. The first one is the construction of the time-series scales: using the original signal, a scale, s, is created from the original time-series, through a coarse-graining procedure, i.e., replacing s non-overlapping points by their average. In the second step, the entropy (sample entropy, permutation entropy, fuzzy entropy, dispersion entropy, and others) is computed for the original signal and for the coarse-grained time-series to evaluate the irregularity for each scale. For this reason, methods of multiple scale entropy (such as entropy of entropy [127], composite multiscale entropy [128], refined multiscale entropy [129], modified multiscale entropy [130], generalized multiscale entropy [131], multivariate multiScale entropy [132], and others) were not explored in this paper.
3. The Entropy Universe Discussion
The Entropy Universe is presented in Figure 10.
Entropy is the uncertainty of a single random variable. We can define conditional entropy , which is the entropy of a random variable conditional on another random variable’s knowledge. The reduction in uncertainty due to the knowledge of another random variable is called mutual information.
The notion of Kolmogorov entropy as a particular case of conditional entropy includes various entropy measures, and estimates proposed to quantify the complexity of a time-series aimed at the degree of predictability of the underlying process. Measures such as approximate entropy, sample entropy, fuzzy entropy, and permutation entropy are prevalent for estimating Kolmogorov entropy in several fields [65].
Other measures of complexity are also related to entropy measures such as the Kolmogorov complexity. The original ideas of Kolmogorov complexity were put forth independently and almost simultaneously by Kolmogorov [133], Solomonoff [134], and Chaitin [135]. Teixeira et al. [136] studied relationships between Kolmogorov complexity and entropy measures. Consequently, in Figure 10 a constellation stands out that relates these entropies.
The concept of entropy has a complicated history. It has been the subject of diverse reconstructions and interpretations, making it difficult to understand, implement, and interpret. The concept of entropy can be extended to continuous distributions. However, as we mentioned, this extension can be very problematic.
Another problem we encountered was the use of the same name for different measures of entropy. The quadratic entropy was defined by Lake [87] and by Rao [89], but despite having the same name, the entropies are very different. The same is valid with graph entropy, which was defined differently by Mowshowitz [102] and by János Körne [106]. Different fuzzy entropies have been proposed over time, as mentioned in Section 2.9.3, but always using the same name. On the other hand, we find different names for the same entropy. It is common in the literature to find thermodynamic entropy, differential entropy, metric entropy, Rényi entropy, geometric entropy, and quantum entropy with different names (see Figure 10). Moreover, the same name is used for a family of entropies and a particular case. For example, the term Rényi entropy is used for all Rényi entropy families and to the particular case .
It is also common to have slightly different definitions of some entropies. The Tsallis entropy formula in some papers appears with k, but in other papers, the authors consider , which is not mentioned in the text.
Depending on the application area, different log bases are used for the same entropy. In this paper, in the Shannon entropy formula, we use the natural logarithm, and in the Rényi entropy formula, we use base 2 logarithm. However, it is common to find these entropies defined with other logarithmic bases in the scientific community, which needs to be considered when relating entropies.
In this paper, we study only single variables, but many other entropies have not been addressed. Nonetheless, some entropies were created to measure the information between two or more time-series. The relative entropy (also called Kullback–Leibler divergence) is a measure of how one probability distribution is different from a second, reference probability distribution. The cross-entropy is an index for determining the divergence between two sets or distributions. Many other like, conditional entropy, mutual information, information storage [65], relative entropy, will not be covered in this paper. While belonging to the same universe, they should be considered in a different galaxy.
Furthermore, in the literature, we find connections between the various entropies and other dispersion and uncertainty measures. Ronald Fisher defined a quantity called Fisher information as a strict limit on the ability to estimate the parameters that define a system [137,138]. Several researchers have shown that there is a connection between Fisher’s information and Shannon’s differential entropy [46,139,140,141]. The relationship between the variance and entropies was also explored in many papers such as [141,142,143]. Shannon defined what he called the derived quantity of entropy power (also called entropy rate power) to be the power in a Gaussian white noise limited to the same band as the original ensemble and having the same entropy. The entropy power defined by Shannon is the minimum variance that can be associated with the Gaussian differential entropy [144]. However, the entropy power can be calculated associated with any given entropy. The in-depth study of the links between The Entropy Universe and other measures of dispersion and uncertainty remains as future work. As well as the exploration of the possible log-concavity of these measures along with the heat flow, for example, as recently established for Fisher information [145].
Like our universe, the universe of entropies can be considered infinite because it is continuously expanding.
4. Entropy Impact in the Scientific Community
The discussion and the reflection on the definitions of entropy and how entropies are related are fundamental. However, it is also essential to understand each entropy’s impact in the scientific community and in which areas each entropy applies. The main goals of this section are two-fold. The first is to analyze the number of citations for each entropy in recent years. The second is to understand the application areas of each entropy discussed in the scientific community.
For the impact analysis, we have used the Web of Science (WoS) and Scopus databases. Those databases provide access to the analysis of citations of scientific works from all scientific branches and have become an important tool in bibliometric analysis[146]. On both websites, researchers and the scientific community can access databases, analysis, and insights. Several researchers have studied the advantages and limitations of using the WoS or Scopus databases [147,148,149,150,151].
The methodology used was the following in both databases: we searched for the paper’s title in which entropy was proposed and collected the year publication, the number of citations total and by year, for the last 16 years. In WoS, we selected in Research Areas the ten areas with a higher record count, and in Scopus, we selected the ten Documents by subject area with a larger record count for each entropy.
4.1. Number of Citations
In Table 1, we list the paper in which each entropy was proposed. We also present the number of citations of the paper in Scopus and WoS since the paper’s publication.
Table 1.
Name of Entropy | Reference | Year | Scopus | Web of Science |
---|---|---|---|---|
Boltzmann entropy | [25] | 1900 | 5 | - |
Gibbs entropy | [26] | 1902 | 1343 | - |
Hartley entropy | [29] | 1928 | 902 | - |
Von Newmann entropy | [30] | 1932 | 1887 | - |
Shannon/differential entropies | [33] | 1948 | 34,751 | 32,040 |
Boltzmann-Gibbs-Shannon | [42] | 1955 | 8 | 7 |
Topological information content | [100] | 1955 | 204 | - |
Maximum entropy | [83] | 1957 | 6661 | 6283 |
Kolmogorov entropy | [55] | 1958 | 693 | 662 |
Rényi entropy | [79] | 1961 | 3149 | - |
Topological entropy | [66] | 1965 | 728 | 682 |
Havrda–Charvát structural -entropy | [90] | 1967 | 744 | - |
Graph entropy | [102] | 1968 | 207 | 195 |
Fuzzy entropy | [118] | 1972 | 1395 | - |
Minimum entropy | [84] | 1975 | 22 | 17 |
Geometric entropy | [74] | 1988 | 71 | - |
Tsallis entropy | [91] | 1988 | 5745 | 5467 |
Approximate entropy | [112] | 1991 | 3562 | 3323 |
Spectral entropy | [49] | 1992 | 915 | 26 |
Tone-entropy | [51] | 1997 | 85 | 76 |
Sample entropy | [15] | 2000 | 3770 | 3172 |
Wavelet entropy | [52] | 2001 | 582 | 465 |
Permutation/sorting entropies | [13] | 2002 | 1900 | 1708 |
Smooth Rényi entropy | [86] | 2004 | 112 | 67 |
Kernel-based entropy | [19] | 2005 | 15 | 13 |
Quadratic sample entropy | [87] | 2005 | 65 | 68 |
Empirical mode decomposition energy entropy | [53] | 2006 | 391 | 359 |
Intrinsic mode dispersion entropy | [115] | 2007 | 59 | 55 |
Tsallis permutation entropy | [92] | 2008 | 35 | 37 |
Modified sample entropy | [17] | 2008 | 58 | 51 |
Coefficient of sample entropy | [16] | 2011 | 159 | 136 |
entropy | [54] | 2011 | 13 | 10 |
Fuzzy entropy | [122] | 2011 | 23 | 18 |
Rényi permutation entropy | [14] | 2013 | 28 | 26 |
Horizontal visibility graph entropy | [109] | 2014 | 22 | - |
Rank-based entropy | [94] | 2014 | 6 | 6 |
Kernels entropies | [124] | 2015 | 46 | 39 |
Dispersion entropy | [18] | 2016 | 98 | 84 |
Buble entropy | [95] | 2017 | 25 | 21 |
Fluctuation-based dispersion entropy | [116] | 2018 | 16 | 10 |
Legend: -paper not found in database. |
In general, as expected, the paper that firstly defined each entropy of the universe has more citations in Scopus than in WoS. Based on the number of citations in the two databases, the five most popular entropies in the scientific community are the Shannon/differential, maximum, Tsallis, SampEn, and ApEn entropies. Considering the papers published until 2000, only the papers that proposed the Boltzmann, the Boltzmann-Gibbs-Shannon, the minimum, the geometric, and the tone-entropy entropies have less than one hundred citations in each database. The papers that proposed the empirical mode decomposition energy entropy and the coefficient of sample entropy are the most cited ones among the entropies introduced in the last sixteen years (see Table 1). Of the entropies proposed in the last five years, the dispersion entropy paper was the most cited one, followed by the one introducing the kernels entropies.
Currently, the WoS platform only covers the registration and analysis of papers after 1900. Therefore, we have not found all the documents needed for the analysis to be developed in this section. Ten papers were not found in the WoS, as shown in Table 1. In particular, entropies’ papers widely cited in the scientific community were not found, such as Gibbs, Quantum, Rényi, and Fuzzy entropies. Currently, Scopus has more than 76.8 million main records: 51.3 million records after 1995 with references and 25.3 million records before 1996, the oldest from 1788 [152]. We found the relevant information regarding all papers that proposed each entropy in Scopus.
Next, we analyzed the impact of each entropy on the scientific community in recent years. Figure 11 and Figure 12 show the number of citations of the paper proposing each entropy in the last 16 years (from 2004 to 2019), respectively, in the Scopus and WoS databases. Note that the colors used in the Figure 11 and Figure 12 are in accordance with the colors in Figure 1 and we did not use the 2020 information once we considered that those values were incomplete at the research time. The range of citations of the papers that introduced each entropy is extensive; therefore, we chose to use a logarithmic scale in Figure 11 and Figure 12. For entropies whose paper that introduced it has been cited more than three times in the past 16 years, we have drawn the linear regression of the number of citations as a function of the year. In the legends of Figure 11 and Figure 12, we present the slope of the regression line, b, and the respective p-value. For the slope was considered significant. Shannon/Differential paper is the most cited in the last sixteen years, as shown in Figures and Table 1.
In recent years, Shannon/differential, sample, maximum, Tsallis, permutation/sorting, and approximate were the entropies most cited, as we can see in both Figures. However, Rényi entropy joins that group in Figure 12. Fluctuation-based dispersion entropy was proposed in 2018, so we only have citations in 2019 that correspond to the respective values in Table 1.
In the last sixteen years, there were several years in which Boltzmann-Gibbs-Shannon, graph, minimum, spectral, kernel-based, Tsallis permutation entropy, entropy, and fuzzy measure entropies had not been cited, according to WoS information (Figure 11). While, Boltzmann, Boltzmann-Gibbs-Shannon, minimum, geometric, entropy, kernel-based, Tsallis permutation, and fuzzy measure entropies papers have not been cited in all years as reported in Scopus (Figure 12). In particular, the paper that introduced Boltzmann-Gibbs-Shannon entropy, with the least number of citations in the last sixteen years, had only one quote in 2012 on the WoS platform, and in Scopus, it had three citations (2005, 2010, and 2012).
Figure 11 shows that the number of citations of the paper proposing sample entropy has increased significantly and is the fastest: the number of citations per year went from 13 in 2004 to 409 in 2019. The permutation/sorting paper had an increasing number of citations as well; the number of citations per year increased significantly from 3 in 2004 to 300 in 2019. From this information, we infer that these entropies have increased their importance in the scientific community. Based on the WoS database, the two entropies proposed in the last eight years, whose paper has strictly increased the number of citations, are dispersion and bubble entropies, respectively, from 2 in 2016 to 32 in 2019 and 1 in 2017 to 11 in 2019. However, from the ones proposed in the last eight years, the two most cited are dispersion and kernels entropies’ papers. Regarding the paper that introduced the wavelet entropy, and according to Figure 11, the year with the most citations was 2015. Since then, the number of citations per year has been decreasing.
The results from Scopus, displayed in Figure 12, show that among the entropy articles where the number of citations increased significantly are the Shannon/differential (from 754 in 2004 to 2875 in 2019) and the sample (from 16 in 2004 to 525 in 2019). More recently, the number of citations of the papers that introduced dispersion and bubble entropies continuously increased from 1 in 2016 to 42 in 2019 and from 3 in 2017 to 13 in 2019, respectively. In 2014, there were 12 entropy papers more cited than the sample entropy paper, but in 2019 the sample entropy paper was the second most cited. This progress implies that the impact of SampEn on the scientific community has increased significantly. On the other hand, the spectral entropy paper lost positions in 2004, it was the 6th most cited, and moved to position 16 in 2019 (see Figure 12). The results obtained on the WoS and Scopus platform on the impact of entropies on scientific communities complement and reinforce each other.
4.2. Areas of Application
To understand each entropy’s areas of application in the scientific community, we consider the ten areas that most cited the paper that introduced the entropy, according to the largest number of records obtained in Web of Science and the Scopus databases. The results are displayed in Figure 13 and Figure 14, respectively. The entropies are ordered in the chronological order of the work that proposed them, meaning the first line corresponds to the oldest entropy paper. There are ten color tones in the figures representing the range from 0 (blue) to 10 (red), where 0 is assigned to the research area that least mentioned the paper that introduced the entropy. In contrast, 10 is assigned to the research area that most cited the paper.
Note that there are papers in which the entropies were proposed that do not have ten application areas in the databases used. For example, the Rényi permutation entropy has only nine areas in the WoS, and the ranked-based entropy has only six areas of application in Scopus.
Figure 13 and Figure 14 show that, over the years, the set of areas that apply entropies is increasingly numerous. However, we did not obtain the same results for the two figures.
Based on the ten Research Areas with the largest number of records obtained on the WoS, we obtained 45 Research Areas (Figure 13).
Initially, the research areas that most cited the entropy papers were Physics, Mathematics, and Computer Science. However, according to data from the WoS, there is a time interval between 1991 (approximate entropy) and 2007 (intrinsic model entropy) in which the emerging entropies have more citations in the area of Engineering. Of the thirty works covered by the WoS database, sixteen were most cited in the Engineering area, nine in the Physics area, five in the Computer Science area, and two in the Mathematics area. The second area of research in which the paper that introduced tone-entropy was most cited is Psychology. The second area in which the paper that proposes Tsallis permutation entropy was most cited is Geology. According to the WoS database, the papers that introduced entropies were little cited in medical journals.
Figure 14, based on the top 10 Documents by subject area of Scopus, covers all papers that proposed each entropy described in this paper. However, this top 10 contains about half of the Research Areas provided on the WoS (25 Documents by area subject). Also, we observe that the set of research areas that most cite the works is broader. Of the forty works covered by the Scopus database, the twelve most cited papers are in the area of Computer Science, eight in the area of Physics and Astronomy, seven in the area of Engineering, seven in the area of Mathematics, three in the area of Medicine, one in the area of Chemistry and two papers with the same number of citations in different areas. Over the years, the papers that introduced entropies have been most cited in medical journals by the Scopus database.
According to the Scopus platform, areas such as Business, Management and Accounting, and Agricultural and Biological Sciences are part of the ten main areas that cite the papers that proposed the universe’s entropies. When we did the data collection on the WoS, the paper that introduced minimum entropy was mentioned by papers from only four areas, but it is complete in the top 10 of Scopus. On the WoS, the papers that proposed approximate entropy, tone-entropy, and coefficient of sample entropy are the most cited in the Engineering area. In contrast, in the Scopus database, they are most cited in the medical field.
There are several differences between the results obtained in the application areas of the two platforms. However, we also find similarities. Shannon/differential, maximum, topological, graph, minimum, tone-entropy, wavelet, EMDEnergyEn, intrinsic mode, Rényi permutation, kernels, dispersion and fluctuation-based dispersion entropies’ introductory papers show the same application areas with more citations in Scopus and WoS database.
We believe that the differences found may be due to the number of citations in the papers and the list of areas for each platform being different. Therefore, it is important to show the results obtained based on the WoS and Scopus since, when the results are the same, it strengthens, while when the results are different, the two analyzes complement each other.
5. Conclusions
In this paper, we introduce The Entropy Universe. We built this universe by describing in-depth the relationship between the most applied entropies in time-series for different scientific fields, establishing bases for researchers to properly choose the variant of entropy most suitable for their data and intended objectives. We have described the obstacles surrounding the concept of entropy in scientific research. We aim that this work help researchers choosing the best entropy for their time-series analysis. Among the problems, we discussed and reflected on the extension of the concept of entropy from a discrete variable to a continuous variable. Also, we point out that different entropies have the same name and the same entropies have different names.
The papers that proposed entropies have been increasing in the number of citations, and the Shannon/differential, Tsallis, sample, permutation, and approximate had been the most cited entropies. Permutation/sorting entropy were the ones that most increased the impact on scientific works, in the last sixteen years. Of the entropies proposed in the past five years, kernels and dispersion entropies are the ones that have the greatest impact on scientific research. Based on the ten areas, with the largest number of records of the paper that introduced each new entropy, obtained from WoS and Scopus, the areas that most applied the entropies are Computer Science, Physics, Mathematics, and Engineering. However, there are differences between the results achieved by the two databases. According to the WoS database, the papers that introduced the entropies were rarely cited in medical journals. However, we did not obtain the same result from the Scopus database.
The Entropy Universe is an ongoing work since the number of entropies is continually expanding.
Abbreviations
The following abbreviations are used in this manuscript:
ApEn | approximate entropy |
BEn | bubble entropy |
CauchyKE | Cauchy kernel entropy |
CKE | circular kernel entropy |
CosEn | coefficient of sample entropy |
DE | differential entropy |
DispEn | dispersion entropy |
EKE | exponential kernel entropy |
EMD | empirical mode decomposition |
EMDEnergyEn | empirical mode decomposition energy entropy |
FDispEn | fluctuation-based dispersion entropy |
FuzzyEn | fuzzy entropy |
FuzzyMEn | fuzzy measure entropy |
i.i.d. | independent and identically distributed |
IME | intrinsic mode entropy |
IMF | intrinsic mode functions |
InMDEn | intrinsic mode dispersion entropy |
KbEn | kernel-based entropy |
LKE | Laplacian kernel entropy |
mSampEn | modified sample entropy |
NTPE | normalized Tsallis permutation entropy |
PE | permutation entropy |
QSE | quadratic sample entropy |
RbE | rank-based entropy |
RE | Rényi entropy |
RPE | Rényi permutation entropy |
SampEn | sample entropy |
SE | shannon entropy |
SKE | spherical kernel entropy |
SortEn | sorting entropy |
SpEn | spectral entropy |
TE | Tsallis entropy |
T-E | tone-entropy |
TKE | triangular kernel entropy |
TopEn | topological entropy |
WaEn | wavelet entropy |
WoS | web of science |
Author Contributions
Conceptualization, M.R., L.A., A.S. and A.T.; methodology and investigation, M.R., L.C., T.H., A.T., and A.S.; writing, editing and review of manuscript, M.R., L.C., T.H., A.T. and A.S.; review and supervision, L.A. and C.C.-S. All authors have read and agreed to the published version of the manuscript.
Funding
M.R. acknowledges Fundação para a Ciência e a Tecnologia (FCT) under scholarship SFRH/BD/138302/2018. A.S. acknowledges funds of Laboratório de Sistemas Informáticos de Grande Escala (LASIGE) Research Unit, Ref. UIDB/00408/2020, funds of Instituto de Telecomunicações (IT) Research Unit, Ref. UIDB/EEA/50008/2020, granted by FCT/MCTES, and the FCT projects Confident PTDC/EEI-CTP/4503/2014, QuantumMining POCI-01-0145-FEDER-031826, and Predict PTDC/CCI-CIF/29877/2017 supported by the European Regional Development Fund (FEDER), through the Competitiveness and Internationalization Operational Programme (COMPETE 2020), and by the Regional Operational Program of Lisboa. This article was supported by National Funds through FCT—Fundação para a Ciência e a Tecnologia, I.P., within CINTESIS, R&D Unit (reference UIDB42552020).
Conflicts of Interest
The authors declare no conflict of interest.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Flores Camacho F., Ulloa Lugo N., Covarrubias Martínez H. The concept of entropy, from its origins to teachers. Rev. Mex. Física. 2015;61:69–80. [Google Scholar]
- 2.Harris H.H. Review of Entropy and the Second Law: Interpretation and Misss-Interpretationsss. J.Chem. Educ. 2014;91:310–311. doi: 10.1021/ed500035f. [DOI] [Google Scholar]
- 3.Shaw D., Davis C.H. Entropy and information: A multidisciplinary overview. J. Am. Soc. Inf. Sci. 1983;34:67–74. doi: 10.1002/asi.4630340110. [DOI] [Google Scholar]
- 4.Kostic M.M. The elusive nature of entropy and its physical meaning. Entropy. 2014;16:953–967. doi: 10.3390/e16020953. [DOI] [Google Scholar]
- 5.Popovic M. Researchers in an entropy wonderland: A review of the entropy concept. arXiv. 20171711.07326 [Google Scholar]
- 6.Batten D.F. Spatial Analysis of Interacting Economies. Springer; Berlin/Heidelberg, Germany: 1983. A review of entropy and information theory; pp. 15–52. [Google Scholar]
- 7.Amigó J.M., Balogh S.G., Hernández S. A brief review of generalized entropies. Entropy. 2018;20:813. doi: 10.3390/e20110813. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Tempesta P. Beyond the Shannon–Khinchin formulation: The composability axiom and the universal-group entropy. Ann. Phys. 2016;365:180–197. doi: 10.1016/j.aop.2015.08.013. [DOI] [Google Scholar]
- 9.Namdari A., Li Z. A review of entropy measures for uncertainty quantification of stochastic processes. Adv. Mech. Eng. 2019;11 doi: 10.1177/1687814019857350. [DOI] [Google Scholar]
- 10.Rong L., Shang P. Topological entropy and geometric entropy and their application to the horizontal visibility graph for financial time series. Nonlinear Dyn. 2018;92:41–58. doi: 10.1007/s11071-018-4120-6. [DOI] [Google Scholar]
- 11.Blanco S., Figliola A., Quiroga R.Q., Rosso O., Serrano E. Time-frequency analysis of electroencephalogram series. III. Wavelet packets and information cost function. Phys. Rev. E. 1998;57:932. doi: 10.1103/PhysRevE.57.932. [DOI] [Google Scholar]
- 12.Huang N.E., Shen Z., Long S.R., Wu M.C., Shih H.H., Zheng Q., Yen N.C., Tung C.C., Liu H.H. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. London. Ser. A Math. Phys. Eng. Sci. 1998;454:903–995. doi: 10.1098/rspa.1998.0193. [DOI] [Google Scholar]
- 13.Bandt C., Pompe B. Permutation entropy: A natural complexity measure for time series. Phys. Rev. Lett. 2002;88:174102. doi: 10.1103/PhysRevLett.88.174102. [DOI] [PubMed] [Google Scholar]
- 14.Zhao X., Shang P., Huang J. Permutation complexity and dependence measures of time series. EPL Europhys. Lett. 2013;102:40005. doi: 10.1209/0295-5075/102/40005. [DOI] [Google Scholar]
- 15.Richman J.S., Moorman J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000;278:H2039–H2049. doi: 10.1152/ajpheart.2000.278.6.H2039. [DOI] [PubMed] [Google Scholar]
- 16.Lake D.E., Moorman J.R. Accurate estimation of entropy in very short physiological time series: The problem of atrial fibrillation detection in implanted ventricular devices. Am. J. Physiol. Heart Circ. Physiol. 2011;300:H319–H325. doi: 10.1152/ajpheart.00561.2010. [DOI] [PubMed] [Google Scholar]
- 17.Xie H.B., He W.X., Liu H. Measuring time series regularity using nonlinear similarity-based sample entropy. Phys. Lett. A. 2008;372:7140–7146. doi: 10.1016/j.physleta.2008.10.049. [DOI] [Google Scholar]
- 18.Rostaghi M., Azami H. Dispersion entropy: A measure for time-series analysis. IEEE Signal Process. Lett. 2016;23:610–614. doi: 10.1109/LSP.2016.2542881. [DOI] [Google Scholar]
- 19.Xu L.S., Wang K.Q., Wang L. Gaussian kernel approximate entropy algorithm for analyzing irregularity of time-series; Proceedings of the 2005 international conference on machine learning and cybernetics; Guangzhou, China. 15–21 August 2005; pp. 5605–5608. [Google Scholar]
- 20.Martin J.S., Smith N.A., Francis C.D. Removing the entropy from the definition of entropy: Clarifying the relationship between evolution, entropy, and the second law of thermodynamics. Evol. Educ. Outreach. 2013;6:30. doi: 10.1186/1936-6434-6-30. [DOI] [Google Scholar]
- 21.Chakrabarti C., De K. Boltzmann-Gibbs entropy: Axiomatic characterization and application. Int. J. Math. Math. Sci. 2000;23:243–251. doi: 10.1155/S0161171200000375. [DOI] [Google Scholar]
- 22.Haubold H., Mathai A., Saxena R. Boltzmann-Gibbs entropy versus Tsallis entropy: Recent contributions to resolving the argument of Einstein concerning “Neither Herr Boltzmann nor Herr Planck has given a definition of W”? Astrophys. Space Sci. 2004;290:241–245. doi: 10.1023/B:ASTR.0000032616.18776.4b. [DOI] [Google Scholar]
- 23.Cariolaro G. Quantum Communications. Springer; Berlin/Heidelberg, Germany: 2015. Classical and Quantum Information Theory; pp. 573–637. [Google Scholar]
- 24.Lindley D., O’Connell J. Boltzmann’s atom: The great debate that launched a revolution in physics. Am. J. Phys. 2001;69:1020. doi: 10.1119/1.1383602. [DOI] [Google Scholar]
- 25.Planck M. On the theory of the energy distribution law of the normal spectrum. Verh. Deut. Phys. Ges. 1900;2:237–245. [Google Scholar]
- 26.Gibbs J.W. Elementary Principles in Statistical Mechanics: Developed with Especial Reference to the Rational Foundation of Thermodynamics. C. Scribner’s Sons; Farmington Hills, MI, USA: 1902. [Google Scholar]
- 27.Rondoni L., Cohen E. Gibbs entropy and irreversible thermodynamics. Nonlinearity. 2000;13:1905. doi: 10.1088/0951-7715/13/6/303. [DOI] [Google Scholar]
- 28.Goldstein S., Lebowitz J.L., Tumulka R., Zanghi N. Gibbs and Boltzmann entropy in classical and quantum mechanics. arXiv. 20191903.11870 [Google Scholar]
- 29.Hartley R.V. Transmission of information 1. Bell Syst. Tech. J. 1928;7:535–563. doi: 10.1002/j.1538-7305.1928.tb01236.x. [DOI] [Google Scholar]
- 30.Von Neumann J. Mathematische Grundlagen der Quantenmechanik. Springer; Berlin/Heidelberg, Germany: 1932. [Google Scholar]
- 31.Legeza Ö., Sólyom J. Optimizing the density-matrix renormalization group method using quantum information entropy. Phys. Rev. B. 2003;68:195116. doi: 10.1103/PhysRevB.68.195116. [DOI] [Google Scholar]
- 32.Coles P.J., Berta M., Tomamichel M., Wehner S. Entropic uncertainty relations and their applications. Rev. Mod. Phys. 2017;89:015002. doi: 10.1103/RevModPhys.89.015002. [DOI] [Google Scholar]
- 33.Shannon C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948;27:379–423. doi: 10.1002/j.1538-7305.1948.tb01338.x. [DOI] [Google Scholar]
- 34.Shannon C.E., Weaver W. The Mathematical Theory of Communication. University of Illinois Press; Urbana, IL, USA: 1949. pp. 1–117. [Google Scholar]
- 35.Weaver W. Recent contributions to the mathematical theory of communication. ETC Rev. Gen. Semant. 1953;10:261–281. [Google Scholar]
- 36.Rioul O., LTCI T.P. This is it: A primer on Shannon’s entropy and information. L’Information, Semin. Poincare. 2018;23:43–77. [Google Scholar]
- 37.Kline R.R. The Cybernetics Moment: Or Why We Call Our Age the Information Age. JHU Press; Baltimore, MA, USA: 2015. [Google Scholar]
- 38.Shannon C.E., Weaver W. The Mathematical Theory of Communication. University of Illinois Press; Urbana, IL, USA: 1963. Number pt. 11 in Illini books. [Google Scholar]
- 39.Smith J.D. Some observations on the concepts of information-theoretic entropy and randomness. Entropy. 2001;3:1–11. doi: 10.3390/e3010001. [DOI] [Google Scholar]
- 40.Ochs W. Basic properties of the generalized Boltzmann-Gibbs-Shannon entropy. Rep. Math. Phys. 1976;9:135–155. doi: 10.1016/0034-4877(76)90050-1. [DOI] [Google Scholar]
- 41.Plastino A., Plastino A., Vucetich H. A quantitative test of Gibbs’ statistical mechanics. Phys. Lett. A. 1995;207:42–46. doi: 10.1016/0375-9601(95)00640-O. [DOI] [Google Scholar]
- 42.Stratonovich R. The entropy of systems with a random number of particles. Sov. Phys. JETP-USSR. 1955;1:254–261. [Google Scholar]
- 43.Khinchin A.Y. Mathematical Foundations of Information Theory. Courier Corporation; Dover, NY, USA: 2013. [Google Scholar]
- 44.Cover T.M., Thomas J.A. Elem. Inf. Theory. John Wiley & Sons; Hoboken, NJ, USA: 2012. [Google Scholar]
- 45.Aczél J., Daróczy Z. On Measures of Information and Their Characterizations. Academic Press; New York, NY, USA: 1975. p. 168. [Google Scholar]
- 46.Kullback S. Information Theory and Statistics. Courier Corporation; Dover, New York, USA: 1997. [Google Scholar]
- 47.Chakrabarti C., Chakrabarty I. Shannon entropy: Axiomatic characterization and application. Int. J. Math. Math. Sci. 2005;2005:2847–2854. doi: 10.1155/IJMMS.2005.2847. [DOI] [Google Scholar]
- 48.Marsh C. Introduction to Continuous Entropy. Department of Computer Science, Princeton University; Princeton, NJ, USA: 2013. [Google Scholar]
- 49.Kapur J.N., Kesavan H.K. Entropy and Energy Dissipation in Water Resources. Springer; Berlin/Heidelberg, Germany: 1992. Entropy optimization principles and their applications; pp. 3–20. [Google Scholar]
- 50.Borowska M. Entropy-based algorithms in the analysis of biomedical signals. Stud. Logic Gramm. Rhetor. 2015;43:21–32. doi: 10.1515/slgr-2015-0039. [DOI] [Google Scholar]
- 51.Oida E., Moritani T., Yamori Y. Tone-entropy analysis on cardiac recovery after dynamic exercise. J. Appl. Physiol. 1997;82:1794–1801. doi: 10.1152/jappl.1997.82.6.1794. [DOI] [PubMed] [Google Scholar]
- 52.Rosso O.A., Blanco S., Yordanova J., Kolev V., Figliola A., Schürmann M., Başar E. Wavelet entropy: A new tool for analysis of short duration brain electrical signals. J. Neurosci. Methods. 2001;105:65–75. doi: 10.1016/S0165-0270(00)00356-3. [DOI] [PubMed] [Google Scholar]
- 53.Yu Y., Junsheng C. A roller bearing fault diagnosis method based on EMD energy entropy and ANN. J. Sound Vib. 2006;294:269–277. doi: 10.1016/j.jsv.2005.11.002. [DOI] [Google Scholar]
- 54.Chen B., Zhu Y., Hu J., Prı J.C. Δ-Entropy: Definition, properties and applications in system identification with quantized data. Inf. Sci. 2011;181:1384–1402. doi: 10.1016/j.ins.2010.11.037. [DOI] [Google Scholar]
- 55.Kolmogorov A.N. A New Metric Invariant of Transient Dynamical Systems and Automorphisms in Lebesgue Spaces. Volume 119. Russian Academy of Sciences; Moscow, Russai: 1958. pp. 861–864. Doklady Akademii Nauk. [Google Scholar]
- 56.Wong K.S., Salleh Z. A note on the notions of topological entropy. Earthline J. Math. Sci. 2018:1–16. doi: 10.34198/ejms.1119.0116. [DOI] [Google Scholar]
- 57.Sinai I. On the concept of entropy for a dynamic system. Dokl. Akad. Nauk. SSSR. 1959;124:768–771. [Google Scholar]
- 58.Farmer J.D. Information dimension and the probabilistic structure of chaos. Z. Naturforschung A. 1982;37:1304–1326. doi: 10.1515/zna-1982-1117. [DOI] [Google Scholar]
- 59.Frigg R. In what sense is the Kolmogorov-Sinai entropy a measure for chaotic behaviour?—bridging the gap between dynamical systems theory and communication theory. Br. J. Philos. Sci. 2004;55:411–434. doi: 10.1093/bjps/55.3.411. [DOI] [Google Scholar]
- 60.Costa M., Goldberger A.L., Peng C.K. Multiscale entropy analysis of biological signals. Phys. Rev. E. 2005;71:021906. doi: 10.1103/PhysRevE.71.021906. [DOI] [PubMed] [Google Scholar]
- 61.Orozco-Arroyave J.R., Arias-Londono J.D., Vargas-Bonilla J.F., Nöth E. International Conference on Nonlinear Speech Processing. Springer; Berlin/Heidelberg, Germany: 2013. Analysis of speech from people with Parkinson’s disease through nonlinear dynamics; pp. 112–119. [Google Scholar]
- 62.Zmeskal O., Dzik P., Vesely M. Entropy of fractal systems. Comput. Math. Appl. 2013;66:135–146. doi: 10.1016/j.camwa.2013.01.017. [DOI] [Google Scholar]
- 63.Henriques T., Gonçalves H., Antunes L., Matias M., Bernardes J., Costa-Santos C. Entropy and compression: Two measures of complexity. J. Eval. Clin. Pract. 2013;19:1101–1106. doi: 10.1111/jep.12068. [DOI] [PubMed] [Google Scholar]
- 64.Eckmann J.P., Ruelle D. The Theory of Chaotic Attractors. Springer; Berlin/Heidelberg, Germany: 1985. Ergodic theory of chaos and strange attractors; pp. 273–312. [Google Scholar]
- 65.Xiong W., Faes L., Ivanov P.C. Entropy measures, entropy estimators, and their performance in quantifying complex dynamics: Effects of artifacts, nonstationarity, and long-range correlations. Phys. Rev. E. 2017;95:062114. doi: 10.1103/PhysRevE.95.062114. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Adler R.L., Konheim A.G., McAndrew M.H. Topological entropy. Trans. Am. Math. Soc. 1965;114:309–319. doi: 10.1090/S0002-9947-1965-0175106-9. [DOI] [Google Scholar]
- 67.Feng D.J., Huang W. Variational principles for topological entropies of subsets. J. Funct. Anal. 2012;263:2228–2254. doi: 10.1016/j.jfa.2012.07.010. [DOI] [Google Scholar]
- 68.Nilsson J. On the entropy of a family of random substitutions. Monatshefte Math. 2012;168:563–577. doi: 10.1007/s00605-012-0401-1. [DOI] [Google Scholar]
- 69.Bowen R. Entropy for group endomorphisms and homogeneous spaces. Trans. Am. Math. Soc. 1971;153:401–414. doi: 10.1090/S0002-9947-1971-0274707-X. [DOI] [Google Scholar]
- 70.Cánovas J., Rodríguez J. Topological entropy of maps on the real line. Topol. Appl. 2005;153:735–746. doi: 10.1016/j.topol.2005.01.006. [DOI] [Google Scholar]
- 71.Bowen R. Topological entropy for noncompact sets. Trans. Am. Math. Soc. 1973;184:125–136. doi: 10.1090/S0002-9947-1973-0338317-X. [DOI] [Google Scholar]
- 72.Handel M., Kitchens B., Rudolph D.J. Metrics and entropy for non-compact spaces. Isr. J. Math. 1995;91:253–271. doi: 10.1007/BF02761650. [DOI] [Google Scholar]
- 73.Addabbo R., Blackmore D. A dynamical systems-based hierarchy for Shannon, metric and topological entropy. Entropy. 2019;21:938. doi: 10.3390/e21100938. [DOI] [Google Scholar]
- 74.Ghys E., Langevin R., Walczak P. Entropie géométrique des feuilletages. Acta Math. 1988;160:105–142. doi: 10.1007/BF02392274. [DOI] [Google Scholar]
- 75.Hurder S. Entropy and Dynamics of C1 Foliations. University of Illinois; Chicago, IL, USA: 2020. [Google Scholar]
- 76.Biś A. Entropy of distributions. Topol. Appl. 2005;152:2–10. doi: 10.1016/j.topol.2004.08.011. [DOI] [Google Scholar]
- 77.Hurder S. Lectures on foliation dynamics: Barcelona 2010. arXiv. 20111104.4852 [Google Scholar]
- 78.Lacasa L., Luque B., Ballesteros F., Luque J., Nuno J.C. From time series to complex networks: The visibility graph. Proc. Natl. Acad. Sci. USA. 2008;105:4972–4975. doi: 10.1073/pnas.0709247105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Rényi A. Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics. The Regents of the University of California; California, CA, USA: 1961. On measures of entropy and information. [Google Scholar]
- 80.Bosyk G., Portesi M., Plastino A. Collision entropy and optimal uncertainty. Phys. Rev. A. 2012;85:012108. doi: 10.1103/PhysRevA.85.012108. [DOI] [Google Scholar]
- 81.Easwaramoorthy D., Uthayakumar R. Improved generalized fractal dimensions in the discrimination between healthy and epileptic EEG signals. J. Comput. Sci. 2011;2:31–38. doi: 10.1016/j.jocs.2011.01.001. [DOI] [Google Scholar]
- 82.Müller M.P., Pastena M. A generalization of majorization that characterizes Shannon entropy. IEEE Trans. Inf. Theory. 2016;62:1711–1720. doi: 10.1109/TIT.2016.2528285. [DOI] [Google Scholar]
- 83.Jaynes E.T. Information theory and statistical mechanics. Phys. Rev. 1957;106:620. doi: 10.1103/PhysRev.106.620. [DOI] [Google Scholar]
- 84.Posner E. Random coding strategies for minimum entropy. IEEE Trans. Inf. Theory. 1975;21:388–391. doi: 10.1109/TIT.1975.1055416. [DOI] [Google Scholar]
- 85.Chevalier C., Fouque P.A., Pointcheval D., Zimmer S. Annual International Conference on the Theory and Applications of Cryptographic Techniques. Springer; Berlin/Heidelberg, Germany: 2009. Optimal randomness extraction from a Diffie-Hellman element; pp. 572–589. [Google Scholar]
- 86.Renner R., Wolf S. Smooth Rényi entropy and applications; Proceedings of the International Symposium on Information Theory; Chicago, IL, USA. 27 June–2 July 2004; p. 233. [Google Scholar]
- 87.Lake D.E. Renyi entropy measures of heart rate Gaussianity. IEEE Trans. Biomed. Eng. 2005;53:21–27. doi: 10.1109/TBME.2005.859782. [DOI] [PubMed] [Google Scholar]
- 88.Botta-Dukát Z. Rao’s quadratic entropy as a measure of functional diversity based on multiple traits. J. Veg. Sci. 2005;16:533–540. doi: 10.1111/j.1654-1103.2005.tb02393.x. [DOI] [Google Scholar]
- 89.Rao C.R. Diversity and dissimilarity coefficients: A unified approach. Theor. Popul. Biol. 1982;21:24–43. doi: 10.1016/0040-5809(82)90004-1. [DOI] [Google Scholar]
- 90.Havrda J., Charvát F. Quantification method of classification processes. Concept of structural a-entropy. Kybernetika. 1967;3:30–35. [Google Scholar]
- 91.Tsallis C. Possible generalization of Boltzmann-Gibbs statistics. J. Stat. Phys. 1988;52:479–487. doi: 10.1007/BF01016429. [DOI] [Google Scholar]
- 92.Zunino L., Pérez D., Kowalski A., Martín M., Garavaglia M., Plastino A., Rosso O. Fractional Brownian motion, fractional Gaussian noise, and Tsallis permutation entropy. Phys. A Stat. Mech. Its Appl. 2008;387:6057–6068. doi: 10.1016/j.physa.2008.07.004. [DOI] [Google Scholar]
- 93.Bandt C. Ordinal time series analysis. Ecol. Model. 2005;182:229–238. doi: 10.1016/j.ecolmodel.2004.04.003. [DOI] [Google Scholar]
- 94.Citi L., Guffanti G., Mainardi L. Rank-based multi-scale entropy analysis of heart rate variability; Proceedings of the Computing in Cardiology 2014; Cambridge, MA, USA. 7–10 September 2014; pp. 597–600. [Google Scholar]
- 95.Manis G., Aktaruzzaman M., Sassi R. Bubble entropy: An entropy almost free of parameters. IEEE Trans. Biomed. Eng. 2017;64:2711–2718. doi: 10.1109/TBME.2017.2664105. [DOI] [PubMed] [Google Scholar]
- 96.Friend E.H. Sorting on electronic computer systems. J. ACM (JACM) 1956;3:134–168. doi: 10.1145/320831.320833. [DOI] [Google Scholar]
- 97.Astrachan O. Bubble Sort: An Archaeological Algorithmic Analysis. Volume 35. ACM; New York, NY, USA: 2003. pp. 1–5. ACM SIGCSE Bulletin. [Google Scholar]
- 98.Bodini M., Rivolta M.W., Manis G., Sassi R. Analytical Formulation of Bubble Entropy for Autoregressive Processes; Proceedings of the 2020 11th Conference of the European Study Group on Cardiovascular Oscillations (ESGCO); Pisa, Italy. 15 July 2020; pp. 1–2. [Google Scholar]
- 99.Dehmer M., Mowshowitz A. A history of graph entropy measures. Inf. Sci. 2011;181:57–78. doi: 10.1016/j.ins.2010.08.041. [DOI] [Google Scholar]
- 100.Rashevsky N. Life, information theory, and topology. Bull. Math. Biophys. 1955;17:229–235. doi: 10.1007/BF02477860. [DOI] [Google Scholar]
- 101.Trucco E. A note on the information content of graphs. Bull. Math. Biophys. 1956;18:129–135. doi: 10.1007/BF02477836. [DOI] [Google Scholar]
- 102.Mowshowitz A. Entropy and the complexity of graphs: I. An index of the relative complexity of a graph. Bull. Math. Biophys. 1968;30:175–204. doi: 10.1007/BF02476948. [DOI] [PubMed] [Google Scholar]
- 103.Mowshowitz A. Entropy and the complexity of graphs: II. The information content of digraphs and infinite graphs. Bull. Math. Biophys. 1968;30:225–240. doi: 10.1007/BF02476692. [DOI] [PubMed] [Google Scholar]
- 104.Mowshowitz A. Entropy and the complexity of graphs: III. Graphs with prescribed information content. Bull. Math. Biophys. 1968;30:387–414. doi: 10.1007/BF02476603. [DOI] [PubMed] [Google Scholar]
- 105.Mowshowitz A. Entropy and the complexity of graphs: IV. Entropy measures and graphical structure. Bull. Math. Biophys. 1968;30:533–546. doi: 10.1007/BF02476673. [DOI] [PubMed] [Google Scholar]
- 106.Körner J. Coding of an information source having ambiguous alphabet and the entropy of graphs; Proceedings of the 6th Prague Conference on Information Theory; Prague, Czech Republic. 18–23 August 1973; pp. 411–425. [Google Scholar]
- 107.Csiszár I., Körner J., Lovász L., Marton K., Simonyi G. Entropy splitting for antiblocking corners and perfect graphs. Combinatorica. 1990;10:27–40. doi: 10.1007/BF02122693. [DOI] [Google Scholar]
- 108.Simonyi G. Graph entropy: A survey. Comb. Optim. 1995;20:399–441. [Google Scholar]
- 109.Zhu G., Li Y., Wen P.P., Wang S. Analysis of alcoholic EEG signals based on horizontal visibility graph entropy. Brain Informatics. 2014;1:19–25. doi: 10.1007/s40708-014-0003-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 110.Luque B., Lacasa L., Ballesteros F., Luque J. Horizontal visibility graphs: Exact results for random time series. Phys. Rev. E. 2009;80:046103. doi: 10.1103/PhysRevE.80.046103. [DOI] [PubMed] [Google Scholar]
- 111.Zhu G., Li Y., Wen P.P. Epileptic seizure detection in EEGs signals using a fast weighted horizontal visibility algorithm. Comput. Methods Programs Biomed. 2014;115:64–75. doi: 10.1016/j.cmpb.2014.04.001. [DOI] [PubMed] [Google Scholar]
- 112.Pincus S.M. Approximate entropy as a measure of system complexity. Proc. Natl. Acad. Sci. USA. 1991;88:2297–2301. doi: 10.1073/pnas.88.6.2297. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 113.Liang Z., Wang Y., Sun X., Li D., Voss L.J., Sleigh J.W., Hagihira S., Li X. EEG entropy measures in anesthesia. Front. Comput. Neurosci. 2015;9:16. doi: 10.3389/fncom.2015.00016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 114.Pincus S.M., Gladstone I.M., Ehrenkranz R.A. A regularity statistic for medical data analysis. J. Clin. Monit. 1991;7:335–345. doi: 10.1007/BF01619355. [DOI] [PubMed] [Google Scholar]
- 115.Amoud H., Snoussi H., Hewson D., Doussot M., Duchene J. Intrinsic mode entropy for nonlinear discriminant analysis. IEEE Signal Process. Lett. 2007;14:297–300. doi: 10.1109/LSP.2006.888089. [DOI] [Google Scholar]
- 116.Azami H., Escudero J. Amplitude-and fluctuation-based dispersion entropy. Entropy. 2018;20:210. doi: 10.3390/e20030210. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 117.Zadeh L.A. Fuzzy sets. Inf. Control. 1965;8:338–353. doi: 10.1016/S0019-9958(65)90241-X. [DOI] [Google Scholar]
- 118.De Luca A., Termini S. A definition of a nonprobabilistic entropy in the setting of fuzzy sets theory. Inf. Control. 1972;20:301–312. doi: 10.1016/S0019-9958(72)90199-4. [DOI] [Google Scholar]
- 119.Parkash C. Fuzzy and Non Fuzzy Measures of Information and Their Applications to Queueing Theory. Guru Nanak Dev University; Punjab, India: 2014. [Google Scholar]
- 120.Chen W., Wang Z., Xie H., Yu W. Characterization of surface EMG signal based on fuzzy entropy. IEEE Trans. Neural Syst. Rehabil. Eng. 2007;15:266–272. doi: 10.1109/TNSRE.2007.897025. [DOI] [PubMed] [Google Scholar]
- 121.Yeniyayla Y. Ph.D. Thesis. Dokuz Eylul University; Fen Bilimleri Enstitüsü, Izmir, Turkey: 2011. Fuzzy Entropy and Its Application. [Google Scholar]
- 122.Liu C., Zhao L. Using fuzzy measure entropy to improve the stability of traditional entropy measures; Proceedings of the 2011 Computing in Cardiology; 18–21 September 2011; Hangzhou, China. pp. 681–684. [Google Scholar]
- 123.Zaylaa A., Saleh S., Karameh F., Nahas Z., Bouakaz A. Cascade of nonlinear entropy and statistics to discriminate fetal heart rates; Proceedings of the 2016 3rd International Conference on Advances in Computational Tools for Engineering Applications (ACTEA); Beirut, Lebanon. 13–15 July 2016; pp. 152–157. [Google Scholar]
- 124.Mekyska J., Janousova E., Gomez-Vilda P., Smekal Z., Rektorova I., Eliasova I., Kostalova M., Mrackova M., Alonso-Hernandez J.B., Faundez-Zanuy M., et al. Robust and complex approach of pathological speech signal analysis. Neurocomputing. 2015;167:94–111. doi: 10.1016/j.neucom.2015.02.085. [DOI] [Google Scholar]
- 125.Costa M., Goldberger A.L., Peng C.K. Multiscale entropy analysis of complex physiologic time series. Phys. Rev. Lett. 2002;89:068102. doi: 10.1103/PhysRevLett.89.068102. [DOI] [PubMed] [Google Scholar]
- 126.Zhang Y.C. Complexity and 1/f noise. A phase space approach. J. Phys. I. 1991;1:971–977. doi: 10.1051/jp1:1991180. [DOI] [Google Scholar]
- 127.Hsu C.F., Wei S.Y., Huang H.P., Hsu L., Chi S., Peng C.K. Entropy of entropy: Measurement of dynamical complexity for biological systems. Entropy. 2017;19:550. doi: 10.3390/e19100550. [DOI] [Google Scholar]
- 128.Wu S.D., Wu C.W., Lin S.G., Wang C.C., Lee K.Y. Time series analysis using composite multiscale entropy. Entropy. 2013;15:1069–1084. doi: 10.3390/e15031069. [DOI] [Google Scholar]
- 129.Valencia J.F., Porta A., Vallverdu M., Claria F., Baranowski R., Orlowska-Baranowska E., Caminal P. Refined multiscale entropy: Application to 24-h holter recordings of heart period variability in healthy and aortic stenosis subjects. IEEE Trans. Biomed. Eng. 2009;56:2202–2213. doi: 10.1109/TBME.2009.2021986. [DOI] [PubMed] [Google Scholar]
- 130.Wu S.D., Wu C.W., Lee K.Y., Lin S.G. Modified multiscale entropy for short-term time series analysis. Phys. A Stat. Mech. Appl. 2013;392:5865–5873. doi: 10.1016/j.physa.2013.07.075. [DOI] [Google Scholar]
- 131.Costa M.D., Goldberger A.L. Generalized multiscale entropy analysis: Application to quantifying the complex volatility of human heartbeat time series. Entropy. 2015;17:1197–1203. doi: 10.3390/e17031197. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 132.Ahmed M.U., Mandic D.P. Multivariate multiscale entropy: A tool for complexity analysis of multichannel data. Phys. Rev. E. 2011;84:061918. doi: 10.1103/PhysRevE.84.061918. [DOI] [PubMed] [Google Scholar]
- 133.Kolmogorov A.N. Three approaches to the quantitative definition ofinformation. Probl. Inf. Transm. 1965;1:1–7. [Google Scholar]
- 134.Solomonoff R.J. A formal theory of inductive inference. Part I. Inf. Control. 1964;7:1–22. doi: 10.1016/S0019-9958(64)90223-2. [DOI] [Google Scholar]
- 135.Chaitin G.J. On the length of programs for computing finite binary sequences. J. ACM (JACM) 1966;13:547–569. doi: 10.1145/321356.321363. [DOI] [Google Scholar]
- 136.Teixeira A., Matos A., Souto A., Antunes L. Entropy measures vs. Kolmogorov complexity. Entropy. 2011;13:595–611. doi: 10.3390/e13030595. [DOI] [Google Scholar]
- 137.Zegers P. Fisher information properties. Entropy. 2015;17:4918–4939. doi: 10.3390/e17074918. [DOI] [Google Scholar]
- 138.Fisher R.A. Theory of statistical estimation. Mathematical Proceedings of the Cambridge Philosophical Society. Volume 22. Cambridge University Press; Cambridge, UK: 1925. pp. 700–725. [Google Scholar]
- 139.Blahut R.E. Principles and Practice of Information Theory. Addison-Wesley Longman Publishing Co., Inc.; Cambridge, MA, USA: 1987. [Google Scholar]
- 140.Stam A.J. Some inequalities satisfied by the quantities of information of Fisher and Shannon. Inf. Control. 1959;2:101–112. doi: 10.1016/S0019-9958(59)90348-1. [DOI] [Google Scholar]
- 141.Borzadaran G.M. Proceedings of the AIP Conference Proceedings. Volume 568. American Institute of Physics; Melville, NY, USA: 2001. Relationship between entropies, variance and Fisher information; pp. 139–144. [Google Scholar]
- 142.Mukher jee D., Ratnaparkhi M.V. On the functional relationship between entropy and variance with related applications. Commun. Stat. Theory Methods. 1986;15:291–311. [Google Scholar]
- 143.Toomaj A., Di Crescenzo A. Connections between weighted generalized cumulative residual entropy and variance. Mathematics. 2020;8:1072. doi: 10.3390/math8071072. [DOI] [Google Scholar]
- 144.Gibson J. Entropy power, autoregressive models, and mutual information. Entropy. 2018;20:750. doi: 10.3390/e20100750. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 145.Ledoux M., Nair C., Wang Y.N. Log-Convexity of Fisher Information along Heat Flow. University of Toulouse; Toulouse, France: 2021. [Google Scholar]
- 146.Vieira E., Gomes J. A comparison of Scopus and Web of Science for a typical university. Scientometrics. 2009;81:587–600. doi: 10.1007/s11192-009-2178-0. [DOI] [Google Scholar]
- 147.Liu W., Tang L., Hu G. Funding information in Web of Science: An updated overview. arXiv. 2020 doi: 10.1007/s11192-020-03362-3.2001.04697 [DOI] [Google Scholar]
- 148.Franceschini F., Maisano D., Mastrogiacomo L. Empirical analysis and classification of database errors in Scopus and Web of Science. J. Inf. 2016;10:933–953. doi: 10.1016/j.joi.2016.07.003. [DOI] [Google Scholar]
- 149.Meho L.I., Yang K. Impact of data sources on citation counts and rankings of LIS faculty: Web of Science versus Scopus and Google Scholar. J. Am. Soc. Inf. Sci. Technol. 2007;58:2105–2125. doi: 10.1002/asi.20677. [DOI] [Google Scholar]
- 150.Mugnaini R., Strehl L. Revista Eletrônica de Biblioteconomia e ciência da Informação. Encontros Bibli; Florianopolis, Brazil: 2008. Recuperação e impacto da produção científica na era Google: Uma análise comparativa entre o Google Acadêmico e a Web of Science; pp. 92–105. n. esp. [Google Scholar]
- 151.Falagas M.E., Pitsouni E.I., Malietzis G.A., Pappas G. Comparison of PubMed, Scopus, Web of Science, and Google scholar: Strengths and weaknesses. FASEB J. 2008;22:338–342. doi: 10.1096/fj.07-9492LSF. [DOI] [PubMed] [Google Scholar]
- 152.Scopus Content Selection and Advisory Board (CSAB) [(accessed on 18 June 2020)]; Available online: https://www.elsevier.com/solutions/scopus/how-scopus-works/content.