Entropy: The Markov Ordering Approach

Reading time: 6 minute
...

📝 Original Info

  • Title: Entropy: The Markov Ordering Approach
  • ArXiv ID: 1003.1377
  • Date: 2013-11-12
  • Authors: Researchers from original ArXiv paper

📝 Abstract

The focus of this article is on entropy and Markov processes. We study the properties of functionals which are invariant with respect to monotonic transformations and analyze two invariant "additivity" properties: (i) existence of a monotonic transformation which makes the functional additive with respect to the joining of independent systems and (ii) existence of a monotonic transformation which makes the functional additive with respect to the partitioning of the space of states. All Lyapunov functionals for Markov chains which have properties (i) and (ii) are derived. We describe the most general ordering of the distribution space, with respect to which all continuous-time Markov processes are monotonic (the {\em Markov order}). The solution differs significantly from the ordering given by the inequality of entropy growth. For inference, this approach results in a convex compact set of conditionally "most random" distributions.

💡 Deep Analysis

Deep Dive into Entropy: The Markov Ordering Approach.

The focus of this article is on entropy and Markov processes. We study the properties of functionals which are invariant with respect to monotonic transformations and analyze two invariant “additivity” properties: (i) existence of a monotonic transformation which makes the functional additive with respect to the joining of independent systems and (ii) existence of a monotonic transformation which makes the functional additive with respect to the partitioning of the space of states. All Lyapunov functionals for Markov chains which have properties (i) and (ii) are derived. We describe the most general ordering of the distribution space, with respect to which all continuous-time Markov processes are monotonic (the {\em Markov order}). The solution differs significantly from the ordering given by the inequality of entropy growth. For inference, this approach results in a convex compact set of conditionally “most random” distributions.

📄 Full Content

arXiv:1003.1377v5 [physics.data-an] 9 Nov 2013 Entropy 2010, 12, 1145-1193; doi:10.3390/e12051145 OPEN ACCESS entropy ISSN 1099-4300 www.mdpi.com/journal/entropy Article Entropy: The Markov Ordering Approach Alexander N. Gorban 1,⋆, Pavel A. Gorban 2 and George Judge 3 1Department of Mathematics, University of Leicester, Leicester, UK 2Institute of Space and Information Technologies, Siberian Federal University, Krasnoyarsk, Russia 3Department of Resource Economics, University of California, Berkeley, CA, USA ⋆Author to whom correspondence should be addressed; E-mail: ag153@le.ac.uk. Received: 1 March 2010; in revised form: 30 April 2010 / Accepted: 4 May 2010 / Published: 7 May 2010 / Corrected Postprint 9 November 2013 Abstract: The focus of this article is on entropy and Markov processes. We study the properties of functionals which are invariant with respect to monotonic transformations and analyze two invariant “additivity” properties: (i) existence of a monotonic transformation which makes the functional additive with respect to the joining of independent systems and (ii) existence of a monotonic transformation which makes the functional additive with respect to the partitioning of the space of states. All Lyapunov functionals for Markov chains which have properties (i) and (ii) are derived. We describe the most general ordering of the distribution space, with respect to which all continuous-time Markov processes are monotonic (the Markov order). The solution differs significantly from the ordering given by the inequality of entropy growth. For inference, this approach results in a convex compact set of conditionally “most random” distributions. Keywords: Markov process; Lyapunov function; entropy functionals; attainable region; MaxEnt; inference 1. Introduction 1.1. A Bit of History: Classical Entropy Two functions, energy and entropy, rule the Universe. In 1865 R. Clausius formulated two main laws [1]: Entropy 2010, 12 1146 1. The energy of the Universe is constant. 2. The entropy of the Universe tends to a maximum. The universe is isolated. For non-isolated systems energy and entropy can enter and leave, the change in energy is equal to its income minus its outcome, and the change in entropy is equal to entropy production inside the system plus its income minus outcome. The entropy production is always positive. Entropy was born as a daughter of energy. If a body gets heat ∆Q at the temperature T then for this body dS = ∆Q/T. The total entropy is the sum of entropies of all bodies. Heat goes from hot to cold bodies, and the total change of entropy is always positive. Ten years later J.W. Gibbs [2] developed a general theory of equilibrium of complex media based on the entropy maximum: the equilibrium is the point of the conditional entropy maximum under given values of conserved quantities. The entropy maximum principle was applied to many physical and chemical problems. At the same time J.W. Gibbs mentioned that entropy maximizers under a given energy are energy minimizers under a given entropy. The classical expression R p ln p became famous in 1872 when L. Boltzmann proved his H-theorem [3]: the function H = Z f(x, v) ln f(x, v)dxdv decreases in time for isolated gas which satisfies the Boltzmann equation (here f(x, v) is the distribution density of particles in phase space, x is the position of a particle, v is velocity). The statistical entropy was born: S = −kH. This was the one-particle entropy of a many-particle system (gas). In 1902, J.W. Gibbs published a book “Elementary principles in statistical dynamics” [4]. He considered ensembles in the many-particle phase space with probability density ρ(p1, q1, . . . pn, qn), where pi, qi are the momentum and coordinate of the ith particle. For this distribution, S = −k Z ρ(p1, q1, . . . pn, qn) ln(ρ(p1, q1, . . . pn, qn))dq1 . . . dqndp1 . . . dpn (1) Gibbs introduced the canonical distribution that provides the entropy maximum for a given expectation of energy and gave rise to the entropy maximum principle (MaxEnt). The Boltzmann period of history was carefully studied [5]. The difference between the Boltzmann entropy which is defined for coarse-grained distribution and increases in time due to gas dynamics, and the Gibbs entropy, which is constant due to dynamics, was analyzed by many authors [6,7]. Recently, the idea of two functions, energy and entropy which rule the Universe was implemented as a basis of two-generator formalism of nonequilibrium thermodynamics [8,9]. In information theory, R.V.L. Hartley (1928) [10] introduced a logarithmic measure of information in electronic communication in order “to eliminate the psychological factors involved and to establish a measure of information in terms of purely physical quantities”. He defined information in a text of length n in alphabet of s symbols as H = n log s. In 1948, C.E. Shannon [11] generalized the Hartley approach and developed “a mathematical theory of communication”, that is informat

…(Full text truncated)…

📸 Image Gallery

cover.png page_2.webp page_3.webp

Reference

This content is AI-processed based on ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut