Skip to main content
Proceedings of the National Academy of Sciences of the United States of America logoLink to Proceedings of the National Academy of Sciences of the United States of America
. 2014 Apr 29;111(19):6905–6910. doi: 10.1073/pnas.1406071111

How multiplicity determines entropy and the derivation of the maximum entropy principle for complex systems

Rudolf Hanel a, Stefan Thurner a,b,c, Murray Gell-Mann b,1
PMCID: PMC4024900  PMID: 24782541

Significance

The maximum entropy principle (MEP) states that for many statistical systems the entropy that is associated with an observed distribution function is a maximum, given that prior information is taken into account appropriately. Usually systems where the MEP applies are simple systems, such as gases and independent processes. The MEP has found thousands of practical applications. Whether a MEP holds for complex systems, where elements interact strongly and have memory and path dependence, remained unclear over the past half century. Here we prove that a MEP indeed exists for complex systems and derive the generalized entropy. We find that it belongs to the class of the recently proposed (c,d)-entropies. The practical use of the formalism is shown for a path-dependent random walk.

Keywords: thermodynamics, out-of-equilibrium process, driven systems, random walk

Abstract

The maximum entropy principle (MEP) is a method for obtaining the most likely distribution functions of observables from statistical systems by maximizing entropy under constraints. The MEP has found hundreds of applications in ergodic and Markovian systems in statistical mechanics, information theory, and statistics. For several decades there has been an ongoing controversy over whether the notion of the maximum entropy principle can be extended in a meaningful way to nonextensive, nonergodic, and complex statistical systems and processes. In this paper we start by reviewing how Boltzmann–Gibbs–Shannon entropy is related to multiplicities of independent random processes. We then show how the relaxation of independence naturally leads to the most general entropies that are compatible with the first three Shannon–Khinchin axioms, the Inline graphic-entropies. We demonstrate that the MEP is a perfectly consistent concept for nonergodic and complex statistical systems if their relative entropy can be factored into a generalized multiplicity and a constraint term. The problem of finding such a factorization reduces to finding an appropriate representation of relative entropy in a linear basis. In a particular example we show that path-dependent random processes with memory naturally require specific generalized entropies. The example is to our knowledge the first exact derivation of a generalized entropy from the microscopic properties of a path-dependent random process.


Many statistical systems can be characterized by a macrostate for which many microconfigurations exist that are compatible with it. The number of configurations associated with the macrostate is called the phase-space volume or multiplicity, M. Boltzmann entropy is the logarithm of the multiplicity,

graphic file with name pnas.1406071111eq1.jpg

and has the same properties as the thermodynamic (Clausius) entropy for systems such as the ideal gas (1). We set Inline graphic. Boltzmann entropy scales with the degrees of freedom f of the system. For example, for N noninteracting point particles in three dimensions, Inline graphic. Systems where Inline graphic scales with system size are called extensive. The entropy per degree of freedom Inline graphic is a system-specific constant. Many complex systems are nonextensive, meaning that if two initially insulated systems A and B, with multiplicities Inline graphic and Inline graphic, respectively, are brought into contact, the multiplicity of the combined system is Inline graphic. For such systems, which are typically strongly interacting, non-Markovian, or nonergodic, Inline graphic and the effective degrees of freedom Inline graphic do no longer scale as N. Given the appropriate scaling for Inline graphic, the entropy Inline graphic is a finite and nonzero constant in the thermodynamic limit, Inline graphic.

A crucial observation in statistical mechanics is that the distribution of all macrostate variables gets sharply peaked and narrow as system size N increases. The reason behind this is that the multiplicities for particular macrostates grow much faster with N than those for other states. In the limit Inline graphic the probability of measuring a macrostate becomes a Dirac delta, which implies that one can replace the expectation value of a macrovariable by its most likely value. This is equivalent to maximizing the entropy in Eq. 1 with respect to the macrostate. By maximizing entropy one identifies the “typical” microconfigurations compatible with the macrostate. This typical region of phase space dominates all other possibilities and therefore characterizes the system. Probability distributions associated with these typical microconfigurations can be obtained in a constructive way by the maximum entropy principle (MEP), which is closely related to the question of finding the most likely distribution functions (histograms) for a given system.

We demonstrate the MEP in the example of coin tossing. Consider a sequence of N independent outcomes of coin tosses, Inline graphic, where Inline graphic is either head or tail. The sequence x contains Inline graphic heads and Inline graphic tails. The probability of finding a sequence with exactly Inline graphic heads and Inline graphic tails is

graphic file with name pnas.1406071111eq2.jpg

where Inline graphic is the binomial factor. We use the shorthand notation Inline graphic for the histogram of Inline graphic heads and Inline graphic tails and Inline graphic for the marginal probabilities for throwing head or tail. For the relative frequencies Inline graphic we write Inline graphic. We also refer to θ as the “biases” of the system. The probability of observing a particular sequence x with histogram k is given by Inline graphic. It is invariant under permutations of the sequence x because the coin tosses are independent. All possible sequences x with the same histogram k have identical probabilities. Inline graphic is the respective multiplicity, representing the number of possibilities to throw exactly Inline graphic heads and Inline graphic tails. As a consequence Eq. 2 becomes the probability of finding the distribution function p of relative frequencies for a given N. The MEP is used to find the most likely p. We denote the most likely histogram by Inline graphic and the most likely relative frequencies by Inline graphic.

We now identify the two components that are necessary for the MEP to hold. The first is that Inline graphic in Eq. 2 factorizes into a multiplicity Inline graphic that depends on k only and a factor Inline graphic that depends on k and the biases θ. The second necessary component is that the multiplicity is related to an entropy expression. By using Stirling’s formula, the multiplicity of Eq. 2 can be trivially rewritten for large N,

graphic file with name pnas.1406071111eq3.jpg

where an entropy functional of Shannon type (2) appears,

graphic file with name pnas.1406071111eq4.jpg

The same arguments hold for multinomial processes with sequences x of N independent trials, where each trial Inline graphic takes one of W possible outcomes (3). In that case the probability for finding a given histogram k is

graphic file with name pnas.1406071111eq5.jpg

Inline graphic is the multinomial factor and Inline graphic. Asymptotically Inline graphic holds. Extremizing Eq. 5 for fixed N with respect to k yields the most likely histogram, Inline graphic. Taking logarithms on both sides of Eq. 5 gives

graphic file with name pnas.1406071111eq6.jpg

Obviously, extremizing Eq. 6 leads to the same histogram Inline graphic. The term Inline graphic in Eq. 6 is sometimes called relative entropy or Kullback–Leibler divergence (4). We identify the first term on the right-hand side of Eq. 6 with Shannon entropy Inline graphic, and the second term is the so-called cross-entropy Inline graphic. Eq. 6 states that the cross-entropy is equal to entropy plus the relative entropy. The constraints of the MEP are related to the cross-entropy. For example, let the marginal probabilities Inline graphic be given by the so-called Boltzmann factor, Inline graphic, for the “energy levels” Inline graphic, where β is the inverse temperature and α the normalization constant. Inserting the Boltzmann factor into the cross-entropy, Eq. 6 becomes

graphic file with name pnas.1406071111eq7.jpg

which is the MEP in its usual form, where Shannon entropy gets maximized under linear constraints. α and β are the Lagrangian multipliers for the normalization and the “energy” constraint Inline graphic, respectively. Note that in Eq. 6 we used Inline graphic to scale Inline graphic. Any other nonlinear Inline graphic would yield nonsensical results in the limit of Inline graphic, either 0 or ∞. Comparing Inline graphic with Eq. 1 shows that indeed, up to a constant multiplicative factor, Inline graphic. This means that the Boltzmann entropy per degree of freedom of a (uncorrelated) multinomial process is given by a Shannon-type entropy functional. Many systems that are nonergodic, are strongly correlated, or have long memory will not be of multinomial type, implying that Inline graphic is not invariant under permutations of a sequence x. For this situation it is not a priori evident that a factorization of Inline graphic into a θ-independent multiplicity and a θ-dependent term, as in Eq. 5, is possible. Under which conditions such a factorization is both feasible and meaningful is discussed in the next section.

Results

When Does a MEP Exist?

The Shannon–Khinchin (SK) axioms (2, 5) state requirements that must be fulfilled by any entropy. [Shannon–Khinchin axioms: SK1, entropy is a continuous function of the probabilities Inline graphic only and should not explicitly depend on any other parameters; SK2, entropy is maximal for the equidistribution Inline graphic; SK3, adding a state Inline graphic to a system with Inline graphic does not change the entropy of the system; and SK4, entropy of a system composed of two subsystems, A and B, is Inline graphic.] For ergodic systems all four axioms hold. For nonergodic ones the composition axiom (SK4) is explicitly violated, and only the first three (SK1–SK3) hold. If all four axioms hold, the entropy is uniquely determined to be Shannon’s; if only the first three axioms hold, the entropy is given by the Inline graphic-entropy (6, 7). The SK axioms were formulated in the context of information theory but are also sensible for many physical and complex systems.

The first Shannon–Khinchin axiom (SK1) states that entropy depends on the probabilities Inline graphic only. Multiplicity depends on the histogram Inline graphic only and must not depend on other parameters. Up to an N-dependent scaling factor the entropy is the logarithm of multiplicity. The scaling factor Inline graphic removes this remaining N dependence from entropy, so that SK1 is asymptotically fulfilled. In fact, SK1 ensures that the factorization Inline graphic into a θ-independent characteristic multiplicity Inline graphic and a θ-dependent characteristic probability Inline graphic is not arbitrary.

For systems that are not of multinomial nature, we proceed as before: To obtain the most likely distribution function we try to find Inline graphic that maximizes Inline graphic for a given N. We denote the generalized relative entropy by

graphic file with name pnas.1406071111eq8.jpg

Note that whenever an equation relates terms containing k with terms containing p, we always assume Inline graphic. The maximal distribution Inline graphic therefore minimizes Inline graphic and is obtained by solving

graphic file with name pnas.1406071111eq9.jpg

for all Inline graphic. α is the Lagrange multiplier for normalization of p.

The histogram Inline graphic can be seen as a vector in a W-dimensional space. Let Inline graphic be a W-dimensional vector whose ith component is 1, and all of the others are 0. With this notation the derivative in Eq. 9 can be expressed asymptotically as

graphic file with name pnas.1406071111eq10.jpg

where we write Inline graphic for the log term. We interpret Inline graphic as the ith component of a vector Inline graphic. Let Inline graphic be the ith component of the jth basis vector for any given k; then Inline graphic has uniquely determined coordinates Inline graphic,

graphic file with name pnas.1406071111eq11.jpg

Inline graphic has coordinates Inline graphic in any basis Inline graphic. However, as can be easily verified, not all bases are compatible with SK1–SK3 (see condition i in the Theorem below). The problem of factorizing Inline graphic therefore reduces to the problem of finding an appropriate basis. For reasons that become clear below, we choose the following Ansatz for the basis

graphic file with name pnas.1406071111eq12.jpg

where the functions Inline graphic are so-called deformed multinomial factors, and Inline graphic are some appropriately chosen constants. Inline graphic is a factor depending on a continuous, monotonic, and increasing function T, with Inline graphic, and Inline graphic. Inline graphic Inline graphic are positive, monotonic increasing functions on the natural numbers. The freedom of choosing Inline graphic, u, and T in this basis provides a well-defined framework that allows us to derive the conditions for the existence of a MEP. Deformed multinomials are based on deformed factorials that are well known in the mathematical literature (813) and are defined as

graphic file with name pnas.1406071111eq13.jpg

For a specific choice of u, deformed multinomials are then defined in a general form as

graphic file with name pnas.1406071111eq14.jpg

where Inline graphic is the largest integer less than x. With the basis of Eq. 12 we can write

graphic file with name pnas.1406071111eq15.jpg

Note that this can be done for any process that produces sequences Inline graphic, where Inline graphic takes one of W values. We can now formulate the following:

Theorem.

Consider the class of processes Inline graphic, with Inline graphic, parameterized by the biases θ and the number of elements N. The process produces histograms k with probability Inline graphic. Let N be large and Inline graphic be the histogram that maximizes Inline graphic. Assume that a basis of the form given in Eq. 12 can be found, for which (i) Inline graphic, for all Inline graphic, and (ii) for fixed values of N and θ, the coordinate Inline graphic of Inline graphic in this basis, as defined in Eq. 11, becomes a nonzero constant at Inline graphic. [Condition ii means that the first derivatives of Inline graphic vanish at Inline graphic under the condition Inline graphic, N being constant.] Under these conditions Inline graphic factorizes, Inline graphic, with

graphic file with name pnas.1406071111eq16.jpg

Moreover, there exists a MEP with generalized entropy Inline graphic, for some scaling function Inline graphic. The factors Inline graphic in Eq. 16 represent the constraint terms in the MEP. The solution of the MEP is given by Inline graphic.

The physical meaning of the Theorem is that the existence of a MEP can be seen as a geometric property of a given process. This reduces the problem to one of finding an appropriate basis that does not violate axioms SK1–SK3 and that is also convenient. The former is guaranteed by the Theorem, and the latter is achieved by using the particular choice of the basis in Eq. 12.

Condition ii of the Theorem guarantees the existence of primitive integrals Inline graphic and Inline graphic. If condition i is violated, the first basis vector Inline graphic of Eq. 12 introduces a functional in p that will in general violate the second Shannon–Khinchin axiom SK2. Conditions i and ii together determine Inline graphic up to a multiplicative constant Inline graphic, which can be absorbed in a normalization constant. Inline graphic may be difficult to construct in practice. However, for solving the MEP it is not necessary to know Inline graphic explicitly; it is sufficient to know the derivatives of the logarithm for the maximization. These derivatives are obtained simply by taking the logarithm of Eq. 16. For systems that are compatible with the conditions of the Theorem, in analogy to Eq. 6, a corresponding MEP for the general case of nonmultinomial processes reads

graphic file with name pnas.1406071111eq17.jpg

Inline graphic has to be chosen such that for large N the generalized relative entropy Inline graphic neither becomes 0 nor diverges for large N. Inline graphic is the generalized entropy, and Inline graphic is the generalized cross-entropy. In complete analogy to the multinomial case, the generalized cross-entropy equals generalized entropy plus generalized relative entropy. Note that in general the generalized cross-entropy Inline graphic will not be linear in Inline graphic. In ref. 14 it was shown that the first three Shannon–Khinchin axioms allow only two options for the constraint terms. They can be either linear or of the so-called “escort” type (15), where constraints are given by specific nonlinear functions in Inline graphic (14). No other options are allowed. For the escort case we have shown in refs. 14 and 16 that a duality exists such that the generalized entropy S, in combination with the escort constraint, can be transformed into the dual generalized entropy Inline graphic with a linear constraint. In other words, the nonlinearity in the constraint can literally be subtracted from the cross-entropy and added to the entropy. Compare with the notion of the “corrector” discussed in ref. 17.

The Generalized Entropy.

We can now compute the generalized entropy from Eq. 17,

graphic file with name pnas.1406071111eq18.jpg

where Inline graphic is the derivative with respect to z. Further, we replace the sum over r by an integral that is correct for large N. The resulting generalized entropy is clearly of trace form. In refs. 14, 18, and 19 it was shown that the most general form of trace form entropy that is compatible with the first three Shannon–Khinchin axioms is

graphic file with name pnas.1406071111eq19.jpg

where Inline graphic is a so-called generalized logarithm, which is an increasing function with Inline graphic, Inline graphic; compare refs. 14 and 16. Comparison of the last line of Eq. 18 with Eq. 19 yields the generalized logarithm

graphic file with name pnas.1406071111eq20.jpg

with Inline graphic and b constants. By taking derivatives of Eq. 20, first with respect to z and then with respect to N, one solves the equation by separation of variables with a separation constant ν. Setting Inline graphic, we get

graphic file with name pnas.1406071111eq21.jpg

By choosing T and ν appropriately one can find examples for all entropies that are allowed by the first three SK axioms, which are the Inline graphic-entropies (6, 7). Inline graphic-entropies include most trace form entropies that were suggested in the past decades as special cases. The expressions Inline graphic and Inline graphic from Eq. 21 can be used in Eqs. 9 and 15 to finally obtain the most likely distribution from the minimum relative entropy,

graphic file with name pnas.1406071111eq22.jpg

which must be solved self-consistently. Inline graphic is the inverse function of T. In the case that only the first two basis vectors are relevant (the generalized entropy and one single constraint term), we get distributions of the form

graphic file with name pnas.1406071111eq23.jpg

with Inline graphic, Inline graphic. In a polynomial basis, specified by Inline graphic, the equally spaced “energy levels” are given by Inline graphic. Note that Inline graphic, and Inline graphic depends on bias terms.

For a specific example let us specify Inline graphic and Inline graphic. Eqs. 21 and 19 yield

graphic file with name pnas.1406071111eq24.jpg

which is the so-called Tsallis entropy (20). Inline graphic for this choice of T. Any other choice of T leads to Inline graphic-entropies. Assuming that the basis has two relevant components and using the same Inline graphic as above, the derivative of the constraint term in the example is obtained from Eq. 16,

graphic file with name pnas.1406071111eq25.jpg

This constraint term is obviously nonlinear in Inline graphic and is therefore of escort type. Here the expression Inline graphic plays the role of equidistant energy levels. The example shows explicitly that finding the most likely distribution function Inline graphic by maximization of Inline graphic (minimization of relative entropy) is equivalent to maximizing the generalized entropy of Eq. 24 under a nonlinear constraint term, Inline graphic. In ref. 14 it was shown that a duality exists that allows us to obtain exactly the same result for Inline graphic, when maximizing the dual entropy of Eq. 24, given by Inline graphic, under the linear constraint, Inline graphic.

Example: MEP for Path-Dependent Random Processes.

We now show that path-dependent stochastic processes exist that are out of equilibrium and whose time-dependent distribution functions can be predicted by the MEP, using the appropriate, system-specific generalized entropy. We consider processes that produce sequences x that increase in length at every step. At a given time the sequence is Inline graphic. At the next time step a new element Inline graphic will be added. All elements take one of W different values, Inline graphic. The system is path dependent, meaning that for a sequence x of length N the probability Inline graphic for producing Inline graphic depends on the histogram k and the biases θ only. For such processes the probability to find a given histogram, Inline graphic can be defined recursively by

graphic file with name pnas.1406071111eq26.jpg

For a particular example let the process have the transition probability

graphic file with name pnas.1406071111eq27.jpg

where Inline graphic is a normalization constant, and Inline graphic. Let us further fix Inline graphic. Note that fixing the biases θ in multinomial systems means that as N gets large one obtains Inline graphic, for all i. Obviously Inline graphic approaches a steady state and N becomes an irrelevant degree of freedom in the sense that changing N will not change Inline graphic. Fixing all Inline graphic asymptotically determines Inline graphic completely and leaves no room for any further constraint. For path-dependent processes the situation can be very different. For example, the relative frequencies Inline graphic of the process defined in Eq. 27 never reach a steady state as N gets larger. [One can show that for such systems the inverse temperature Inline graphic approximately grows (sub)logarithmically with N.] Here, fixing θ for all i still allows Inline graphic to evolve with growing N, such that 1 df remains that can be fixed by an additional constraint. [Additional constraints may become necessary for intermediate ranges of N, where some coordinates Inline graphic that need to vanish asymptotically (in the appropriately chosen basis) are not yet sufficiently small.] The process defined in Eq. 27 is a path-dependent, W-dimensional random walk that gets more and more persistent as the sequence gets longer. This means that in the beginning of the process all states are equiprobable Inline graphic. With every realization of state i in the process, all states Inline graphic become more probable in a self-similar way, and a monotonic distribution function of frequencies emerges as N grows. The process appears to “cool” as it unfolds. Adequate basis vectors Inline graphic can be obtained with deformed multinomials Inline graphic based on Inline graphic, Inline graphic, and a polynomial basis for Inline graphic. For this u, in Fig. 1A (solid lines), we show normalized deformed binomials for Inline graphic and Inline graphic. Dashed lines represent the usual binomial. Clearly, generalized multiplicities become more peaked and narrow as N increases, which is a prerequisite for the MEP to hold. In Fig. 1B the variance of deformed binomials is seen to diminish as a function of sequence length N for various values of ν. The dashed line shows the variance for the usual binomial. Distribution functions Inline graphic obtained for numerical simulations of sequences with W states are shown in Fig. 1C for sequence lengths Inline graphic, 5,000, and 10,000 (solid lines). Averages are taken over normalized histograms from 150 independent sequences that were generated with Inline graphic, and Inline graphic Inline graphic. The distributions follow exactly the theoretical result from Eq. 23, confirming that a basis with two relevant components (one for the entropy one for a single constraint fixing N) is sufficient for the given process with Inline graphic. Dashed lines are the functions suggested by the theory, Inline graphic with Inline graphic, where β is obtained from a fit to the empirical distribution. β determines Inline graphic. α is a normalization constant. Although the power exponent Inline graphic does not change with N, the “inverse temperature” β increases with N (Fig. 1C, Inset), which shows that the process becomes more persistent as it evolves—it “ages.” Because Inline graphic, the observed distribution p can also be obtained by maximizing the generalized entropy S (Eq. 24) under a nonlinear constraint or, equivalently, by maximizing its dual, Inline graphic with a linear constraint, as discussed above. For other parameter values a basis with more than two components might become necessary. Note that the nonlinear (escort) constraints can be understood as a simple consequence of the fact that the relative frequencies p have to be normalized for all N. In particular, the escort constraints arise from Inline graphic and Eq. 23, which states that Inline graphic does not change its functional shape as θ or N is varied.

Fig. 1.

Fig. 1.

Numerical results for the path-dependent random process determined by the deformed factorial Inline graphic with Inline graphic. (A) Normalized generalized binomial factors Inline graphic (solid lines). Distributions get narrower as N increases, which is necessary for the MEP to hold. Dashed lines show the usual binomial factor (Inline graphic and Inline graphic). (B) Variance Inline graphic of the normalized generalized binomial factors (solid lines), as a function of sequence length N, for various values ν and Inline graphic. The dashed line is the variance of the usual binomial multiplicity. (C) Probability distributions for the Inline graphic states i from numerical realizations of processes following Eq. 27, with Inline graphic and Inline graphic Inline graphic for various lengths N (solid lines). Distributions follow the theoretical result from Eq. 23. Dashed lines are Inline graphic with Inline graphic. α and β are obtained from fits to the distributions and clearly dependent on N (Inset). They can be used to determine Inline graphic.

Discussion

We have shown that for generalized multinomial processes, where the order of the appearance of events influences the statistics of the outcome (path dependence), it is possible to constructively derive an expression for their multiplicity. We are able to show that a MEP exists for a much wider class of processes and not only for independent multinomial processes. We can explicitly determine the corresponding entropic form from the transition probabilities of a system. We show that the logarithm of the obtained generalized multiplicity is one-to-one related to the concept of Boltzmann entropy. The expressions for the obtained generalized entropies are no longer of Shannon type, Inline graphic, but assume generalized forms that are known from the entropies of superstatistics (21, 22) and that are compatible with the first three Shannon–Khinchin axioms and violate the fourth (6, 7, 14). Further, we find that generalized entropies are of trace form and are based on known generalized logarithms (14, 16, 18, 23). Our findings enable us to start from a given class of correlated stochastic processes and derive their unique entropy that is needed when using the maximum entropy principle. We are able to determine the time-dependent distribution functions of specific processes, either through minimization of the relative entropy or through maximization of the generalized entropy under nonlinear constraints. A previously discovered duality allows us to obtain the same result by maximization of the dual generalized entropy under linear constraints. Systems for which the new technology applies include out-of-equilibrium, path-dependent processes and possibly even aging systems. In an explicit example of a path-dependent random walk we show how the corresponding generalized entropy is derived. We implement a numerical realization of the process to show that the corresponding maximum entropy principle perfectly predicts the correct distribution functions as the system ages in the sense that it becomes more persistent as it evolves. Systems of this kind often never reach equilibrium as Inline graphic.

Supplementary Material

Acknowledgments

R.H. and S.T. thank the Santa Fe Institute for hospitality. M.G.-M. acknowledges the generous support of Insight Venture Partners and the Bryan J. and June B. Zwan Foundation.

Footnotes

The authors declare no conflict of interest.

References

  • 1.Kittel C. Elementary Statistical Physics. New York: Wiley; 1958. [Google Scholar]
  • 2.Shannon CE. A mathematical theory of communication. Bell Syst Tech J. 1948;27:379–423, 623–656. [Google Scholar]
  • 3.Jaynes ET. Probability Theory: The Logic of Science. Cambridge, UK: Cambridge Univ Press; 2003. pp. 351–355. [Google Scholar]
  • 4.Kullback S, Leibler RA. On information and sufficiency. Ann Math Stat. 1951;22(1):79–86. [Google Scholar]
  • 5.Khinchin AI. Mathematical Foundations of Information Theory. New York: Dover; 1957. [Google Scholar]
  • 6.Hanel R, Thurner S. A comprehensive classification of complex statistical systems and an ab-initio derivation of their entropy and distribution functions. Europhys Lett. 2011;93:20006. [Google Scholar]
  • 7.Hanel R, Thurner S. When do generalized entropies apply? How phase space volume determines entropy. Europhys Lett. 2011;96:50003. [Google Scholar]
  • 8.Bhargava M. The factorial function and generalizations. Am Math Mon. 2000;107:783–799. [Google Scholar]
  • 9.Jackson FH. On q-definite integrals. Q J Pure Appl Math. 1910;41(9):193–203. [Google Scholar]
  • 10.Carlitz L. A class of polynomials. Trans Am Math Soc. 1938;43:167–182. [Google Scholar]
  • 11.Polya G. Über ganzwertige polynome in algebraischen zahlkörpern [On integer-valued polynomials in algebraic number fields] J Reine Angew Math. 1919;149:97–116. German. [Google Scholar]
  • 12.Ostrowski A. Über ganzwertige polynome in algebraischen zahlkörpern [On integer-valued polynomials in algebraic number fields] J. Reine Angew. Math. 1919;149:117–124. German. [Google Scholar]
  • 13.Gunji H, McQuillan DL. On a class of ideals in an algebraic numberfield. J Number Theory. 1970;2:207–222. [Google Scholar]
  • 14.Hanel R, Thurner S, Gell-Mann M. Generalized entropies and the transformation group of superstatistics. Proc Natl Acad Sci USA. 2011;108:6390–6394. [Google Scholar]
  • 15.Beck C, Schlögl F. Thermodynamics of Chaotic Systems. Cambridge, UK: Cambridge Univ Press; 1995. [Google Scholar]
  • 16.Hanel R, Thurner S, Gell-Mann M. Generalized entropies and logarithms and their duality relations. Proc Natl Acad Sci USA. 2012;109(47):19151–19154. doi: 10.1073/pnas.1216885109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Topsoe F. 2007 Exponential families and maxent calculations for entropy measures of statistical physics. Complexity, Metastability and Nonextensivity: An International Conference. AIP Conf Proc 965:104–113. [Google Scholar]
  • 18.Hanel R, Thurner S. Generalized Boltzmann factors and the maximum entropy principle: Entropies for complex systems. Physica A. 2007;380:109–114. [Google Scholar]
  • 19.Thurner S, Hanel R. 2007 Entropies for complex systems: Generalized-generalized entropies. Complexity, Metastability, and Nonextensivity: An International Conference. AIP Conf Proc 965:68–75. [Google Scholar]
  • 20.Tsallis C. A possible generalization of Boltzmann-Gibbs statistics. J Stat Phys. 1988;52:479–487. [Google Scholar]
  • 21.Beck C, Cohen EDG. Superstatistics. Physica A. 2003;322:267–275. [Google Scholar]
  • 22.Beck C, Cohen EGD, Swinney HL. From time series to superstatistics. Phys Rev E Stat Nonlin Soft Matter Phys. 2005;72(5 Pt 2):056133. doi: 10.1103/PhysRevE.72.056133. [DOI] [PubMed] [Google Scholar]
  • 23.Naudts J. Deformed exponentials and logarithms in generalized thermostatistics. Physica A. 2002;316:323–334. [Google Scholar]

Articles from Proceedings of the National Academy of Sciences of the United States of America are provided here courtesy of National Academy of Sciences

RESOURCES