May 20, 2011
We prove that information-theoretic maximum entropy (MaxEnt) approach to canonical ensemble is mathematically equivalent to the classic approach of Boltzmann, Gibbs and Darwin-Fowler. The two approaches, however, "interpret" a same mathematical theorem differently; most notably observing mean-energy in the former and energy conservation in the latter. However, applying the same MaxEnt method to grand canonical ensemble fails; while carefully following the classic approach bas...
June 17, 2008
We define a diagonal entropy (d-entropy) for an arbitrary Hamiltonian system as $S_d=-\sum_n \rho_{nn}\ln \rho_{nn}$ with the sum taken over the basis of instantaneous energy states. In equilibrium this entropy coincides with the conventional von Neumann entropy $S_n=-{\rm Tr}\, \rho\ln\rho$. However, in contrast to $S_n$, the d-entropy is not conserved in time in closed Hamiltonian systems. If the system is initially in stationary state then in accord with the second law of ...
September 16, 2018
The entropy of Boltzmann-Gibbs, as proved by Shannon and Khinchin, is based on four axioms, where the fourth one concerns additivity. The group theoretic entropies make use of formal group theory to replace this axiom with a more general composability axiom. As has been pointed out before, generalised entropies crucially depend on the number of allowed number degrees of freedom $N$. The functional form of group entropies is restricted (though not uniquely determined) by assum...
September 11, 2012
In order to undestand how the complex concept of entropy emerged,we propose a trip towards the past reviewing the works of Clausius, Boltzmann, Gibbs and Planck. In particular, since the Gibbs's work is not very well known, we present a detailed analysis, recalling the three definitions of the entropy that Gibbs gives. May be one of the most important aspect of the entropy is to see it as a thermodynamic potential like the other thermodynamic potentials as proposed by Callen....
January 27, 2015
Thermodynamic entropy, as defined by Clausius, characterizes macroscopic observations of a system based on phenomenological quantities such as temperature and heat. In contrast, information-theoretic entropy, introduced by Shannon, is a measure of uncertainty. In this Letter, we connect these two notions of entropy, using an axiomatic framework for thermodynamics [Lieb, Yngvason, Proc. Roy. Soc.(2013)]. In particular, we obtain a direct relation between the Clausius entropy a...
August 22, 2014
The recent experimental realization of exotic matter states in isolated quantum systems and the ensuing controversy about the existence of negative absolute temperatures demand a careful analysis of the conceptual foundations underlying microcanonical thermostatistics. Here, we provide a detailed comparison of the most commonly considered microcanonical entropy definitions, focussing specifically on whether they satisfy or violate the zeroth, first and second law of thermodyn...
April 8, 2014
The expression for entropy sometimes appears mysterious - as it often is asserted without justification. This short manuscript contains a discussion of the underlying assumptions behind entropy as well as simple derivation of this ubiquitous quantity.
January 7, 2016
Very recently, the validity of the concept of negative temperature has been challenged by several authors since they consider Boltzmann's entropy (that allows negative temperatures) inconsistent from a mathematical and statistical point of view, whereas they consider Gibbs' entropy (that does not admit negative temperatures) the correct definition for microcanonical entropy. In the present paper we prove that for systems with equivalence of the statistical ensembles Boltzma...
October 31, 2003
A critical examination of some basic conceptual issues in classical statistical mechanics is attempted, with a view to understanding the origins, structure and statuts of that discipline. Due attention is given to the interplay between physical and mathematical aspects, particularly regarding the role of probability theory. The focus is on the equilibrium case, which is currently better understood, serving also as a prelude for a further discussion of non-equilibrium statisti...
March 6, 2019
We show that the generalized Boltzmann distribution is the only distribution for which the Gibbs-Shannon entropy equals the thermodynamic entropy. This result means that the thermodynamic entropy and the Gibbs-Shannon entropy are not generally equal, but rather than the equality holds only in the special case where a system is in equilibrium with a reservoir.