Information content

From Vero - Wikipedia
Jump to navigation Jump to search

Template:Short description Template:Cleanup

In information theory, the information content, self-information, surprisal, or Shannon information is a basic quantity derived from the probability of a particular event occurring from a random variable. It can be thought of as an alternative way of expressing probability, much like odds or log-odds, but which has particular mathematical advantages in the setting of information theory.

The Shannon information can be interpreted as quantifying the level of "surprise" of a particular outcome. As it is such a basic quantity, it also appears in several other settings, such as the length of a message needed to transmit the event given an optimal source coding of the random variable.

The Shannon information is closely related to entropy, which is the expected value of the self-information of a random variable, quantifying how surprising the random variable is "on average". This is the average amount of self-information an observer would expect to gain about a random variable when measuring it.<ref>Jones, D.S., Elementary Information Theory, Vol., Clarendon Press, Oxford pp 11–15 1979</ref>

The information content can be expressed in various units of information, of which the most common is the "bit" (more formally called the shannon), as explained below.

The term 'perplexity' has been used in language modelling to quantify the uncertainty inherent in a set of prospective events.Template:Citation needed

Definition

Claude Shannon's definition of self-information was chosen to meet several axioms:

  • An event with probability 100% is perfectly unsurprising and yields no information.
  • The less probable an event is, the more surprising it is and the more information it yields.
  • If two independent events are measured separately, the total amount of information is the sum of the self-informations of the individual events.

The detailed derivation is below, but it can be shown that there is a unique function of probability that meets these three axioms, up to a multiplicative scaling factor. Broadly, given a real number <math>b>1</math> and an event <math>x</math> with probability <math>P</math>, the information content is defined as the negative log probability:<math display="block">\mathrm{I}(x) := - \log_b{\left[\Pr{\left(x\right)}\right]} = -\log_b{\left(P\right)}. </math>The base <math>b</math> corresponds to the scaling factor above. Different choices of b correspond to different units of information: when <math>b=2</math>, the unit is the shannon (symbol Sh), often called a 'bit'; when <math>b = e</math>, the unit is the natural unit of information (symbol nat); and when <math>b = 10</math>, the unit is the hartley (symbol Hart).

Formally, given a discrete random variable <math>X</math> with probability mass function <math>p_{X}{\left(x\right)}</math>, the self-information of measuring <math>X</math> as outcome <math>x</math> is defined as:<ref name=":0">Template:Cite book</ref><math display="block">\operatorname{I}_{X}(x) := - \log{\left[p_{X}{\left(x\right)}\right]} = \log{\left(\frac{1}{p_{X}{\left(x\right)}}\right)}. </math>The use of the notation <math>I_X(x)</math> for self-information above is not universal. Since the notation <math>I(X;Y)</math> is also often used for the related quantity of mutual information, many authors use a lowercase <math>h_X(x)</math> for self-entropy instead, mirroring the use of the capital <math>H(X)</math> for the entropy.

Properties

Template:Expand section

Monotonically decreasing function of probability

For a given probability space, the measurement of rarer events are intuitively more "surprising", and yield more information content than more "common" events. Thus, self-information is a strictly decreasing monotonic function of the probability, or sometimes called an "antitonic" function.<ref name="CoverThomas">Template:Cite book</ref>

While standard probabilities are represented by real numbers in the interval <math>[0, 1]</math>, self-information values are non-negative extended real numbers in the interval <math>[0, \infty]</math>. Specifically:

  • An event with probability <math>\Pr(x) = 1</math> (a certain event) has an information content of <math>\mathrm{I}(x) = -\log_b(1) = 0</math>. Its occurrence is perfectly unsurprising and reveals no new information.
  • An event with probability <math>\Pr(x) = 0</math> (an impossible event) has an information content of <math>\mathrm{I}(x) = -\log_b(0)</math>, which is undefined but is taken to be <math>\infty</math> by convention. This reflects that observing an event believed to be impossible would be infinitely surprising.<ref name="MacKay">Template:Cite book</ref>

This monotonic relationship is fundamental to the use of information content as a measure of uncertainty. For example, learning that a one-in-a-million lottery ticket won provides far more information than learning it lost (See also Lottery mathematics.) This also establishes an intuitive connection to concepts like statistical dispersion; events that are far from the mean or typical outcome (and thus have low probability in many common distributions) have high self-information.

Relationship to log-odds

The Shannon information is closely related to the log-odds. The log-odds of an event <math>x</math>, with probability <math>p(x)</math>, is defined as the logarithm of the odds, <math>\frac{p(x)}{1-p(x)}</math>. This can be expressed as a difference of two information content values:<math display="block">{\displaystyle \begin{align} \text{log-odds}(x) &= \ \log_b\left(\frac{p(x)}{1-p(x)}\right) \\ &= \ \log_b(p(x)) - \log_b(1-p(x)) \\ &= \ \ \mathrm{I}(\lnot x) \ - \ \mathrm{I}(x), \end{align} }</math>where <math>\lnot x</math> denotes the event not <math>x</math>.

This expression can be interpreted as the amount of information gained (or surprise) from learning the event did not occur, minus the information gained from learning it did occur. This connection is particularly relevant in statistical modeling where log-odds are the core of the logit function and logistic regression.<ref name="Bishop">Template:Cite book</ref>

Additivity of independent events

The information content of two independent events is the sum of each event's information content. This property is known as additivity in mathematics. Consider two independent random variables <math>X</math> and <math>Y</math> with probability mass functions <math>p_X(x)</math> and <math>p_Y(y)</math>. The joint probability of observing the outcome <math>(x, y)</math> is given by the product of the individual probabilities due to independence:<math display="block"> p_{X, Y}(x, y) = \Pr(X=x, Y=y) = p_X(x) \ p_Y(y)</math>The information content of this joint event is:<math display="block"> {\displaystyle \begin{align} \operatorname{I}_{X,Y}(x, y) &= \ -\log_b \left[p{X,Y}(x, y)\right] \\ &= \ -\log_b \left[p_X(x) \ p_Y(y)\right] \\ &= \ -\log_b \left[p_X(x)\right] \ - \ \log_b \left[p_Y(y)\right] \ \\ &= \ \ \operatorname{I}_X(x) \ + \ \operatorname{I}_Y(y), \end{align} } </math>This additivity makes information content a more mathematically convenient measure than probability in many applications, such as in coding theory where the amount of information needed to describe a sequence of independent symbols is the sum of the information needed for each symbol.<ref name="CoverThomas" />

The corresponding property for likelihoods is that the log-likelihood of independent events is the sum of the log-likelihoods of each event. Interpreting log-likelihood as "support" or negative surprisal (the degree to which an event supports a given model: a model is supported by an event to the extent that the event is unsurprising, given the model), this states that independent events add support: the information that the two events together provide for statistical inference is the sum of their independent information.

Relationship to entropy

The Shannon entropy of the random variable <math>X</math> is defined as:<math display="block">{\displaystyle \begin{align}

\mathrm{H}(X) \ &= \ \sum_{x} {-p_{X}{\left(x \right)} \ \log{p_{X}{\left(x\right)}}} \\ &= \ \sum_{x} {p_{X}{\left(x\right)} \ \operatorname{I}_X(x)} \ \

{\overset{\underset{\mathrm{def}}{}}{=}} \ \

\operatorname{E}{\left[\operatorname{I}_X (X)\right]},

\end{align} } </math>by definition equal to the expected information content of measurement of <math>X</math>.<ref>Template:Cite book</ref>Template:Rp<ref>Template:Cite book</ref>Template:Rp

The expectation is taken over the discrete values over its support.

Sometimes, the entropy itself is called the "self-information" of the random variable, possibly because the entropy satisfies <math>\mathrm{H}(X) = \operatorname{I}(X; X)</math>, where <math>\operatorname{I}(X;X)</math> is the mutual information of <math>X</math> with itself.<ref>Thomas M. Cover, Joy A. Thomas; Elements of Information Theory; p. 20; 1991.</ref>

For continuous random variables the corresponding concept is differential entropy.

Notes

This measure has also been called surprisal, as it represents the "surprise" of seeing the outcome (a highly improbable outcome is very surprising). This term (as a log-probability measure) was introduced by Edward W. Samson in his 1951 report "Fundamental natural concepts of information theory".<ref name="samson53"> Template:Cite journal </ref><ref name="attneave">Template:Cite book</ref> An early appearance in the Physics literature is in Myron Tribus' 1961 book Thermostatics and Thermodynamics.<ref name="Bernstein1972">Template:Cite journal</ref><ref name="Tribus1961">Myron Tribus (1961) Thermodynamics and Thermostatics: An Introduction to Energy, Information and States of Matter, with Engineering Applications (D. Van Nostrand, 24 West 40 Street, New York 18, New York, U.S.A) Tribus, Myron (1961), pp. 64–66 borrow.</ref>

When the event is a random realization (of a variable) the self-information of the variable is defined as the expected value of the self-information of the realization.Template:Citation-needed

Examples

Fair coin toss

Consider the Bernoulli trial of tossing a fair coin <math>X</math>. The probabilities of the events of the coin landing as heads <math>\text{H}</math> and tails <math>\text{T}</math> (see fair coin and obverse and reverse) are one half each, <math display="inline">p_X{(\text{H})} = p_X{(\text{T})} = \tfrac{1}{2} = 0.5</math>. Upon measuring the variable as heads, the associated information gain is <math display="block">\operatorname{I}_X(\text{H})

= -\log_2 {p_X{(\text{H})}}
= -\log_2\!{\tfrac{1}{2}} = 1,</math>so the information gain of a fair coin landing as heads is 1 shannon.<ref name=":0" /> Likewise, the information gain of measuring tails <math>T</math> is<math display="block">\operatorname{I}_X(T)
= -\log_2 {p_X{(\text{T})}}
= -\log_2 {\tfrac{1}{2}} = 1 \text{ Sh}.</math>

Fair die roll

Suppose we have a fair six-sided die. The value of a die roll is a discrete uniform random variable <math>X \sim \mathrm{DU}[1, 6]</math> with probability mass function <math display="block">p_X(k) = \begin{cases} \frac{1}{6}, & k \in \{1, 2, 3, 4, 5, 6\} \\ 0, & \text{otherwise} \end{cases}</math>The probability of rolling a 4 is <math display="inline">p_X(4) = \frac{1}{6}</math>, as for any other valid roll. The information content of rolling a 4 is thus<math display="block">\operatorname{I}_{X}(4) = -\log_2{p_X{(4)}} = -\log_2{\tfrac{1}{6}} \approx 2.585\; \text{Sh}</math>of information.

Two independent, identically distributed dice

Suppose we have two independent, identically distributed random variables <math display="inline">X,\, Y \sim \mathrm{DU}[1, 6]</math> each corresponding to an independent fair 6-sided dice roll. The joint distribution of <math>X</math> and <math>Y</math> is<math display="block"> \begin{align}

p_{X, Y}\!\left(x, y\right) & {} = \Pr(X = x,\, Y = y) 
= p_X\!(x)\,p_Y\!(y) \\
& {} = \begin{cases}
 \displaystyle{1 \over 36}, \ &x, y \in [1, 6] \cap \mathbb{N} \\
 0 & \text{otherwise.} \end{cases}

\end{align}</math>

The information content of the random variate <math> (X, Y) = (2,\, 4)</math> is <math display="block"> \begin{align} \operatorname{I}_{X, Y}{(2, 4)}

&= -\log_2\!{\left[p_{X,Y}{(2, 4)}\right]}
= \log_2\!{36} = 2 \log_2\!{6} \\
& \approx 5.169925 \text{ Sh},

\end{align} </math> and can also be calculated by additivity of events <math display="block"> \begin{align} \operatorname{I}_{X, Y}{(2, 4)}

&= -\log_2\!{\left[p_{X,Y}{(2, 4)}\right]}
= -\log_2\!{\left[p_X(2)\right]} -\log_2\!{\left[p_Y(4)\right]} \\
& = 2\log_2\!{6} \\
& \approx 5.169925 \text{ Sh}.

\end{align} </math>

Information from frequency of rolls

If we receive information about the value of the dice without knowledge of which die had which value, we can formalize the approach with so-called counting variables <math display="block"> C_k := \delta_k(X) + \delta_k(Y) = \begin{cases}

0, & \neg\, (X = k \vee Y = k) \\
1, & \quad X = k\, \veebar \, Y = k \\
2, & \quad X = k\, \wedge \, Y = k

\end{cases} </math> for <math> k \in \{1, 2, 3, 4, 5, 6\}</math>, then <math display="inline"> \sum_{k=1}^{6}{C_k} = 2</math> and the counts have the multinomial distribution <math display="block"> \begin{align}

f(c_1,\ldots,c_6) & {} = \Pr(C_1 = c_1 \text{ and } \dots \text{ and } C_6 = c_6) \\
& {} = \begin{cases} { \displaystyle {1\over{18}}{1 \over c_1!\cdots c_k!}}, 
   \ & \text{when } \sum_{i=1}^6 c_i=2 \\
 0 & \text{otherwise,} \end{cases} \\
& {} = \begin{cases} {1 \over 18}, 
 \ & \text{when 2 } c_k \text{ are } 1 \\
 {1 \over 36}, \ & \text{when exactly one } c_k = 2 \\
 0, \ & \text{otherwise.}
\end{cases}

\end{align}</math>

To verify this, the 6 outcomes <math display="inline">(X, Y) \in \left\{(k, k)\right\}_{k = 1}^{6} = \left\{

(1, 1), (2, 2), (3, 3), (4, 4), (5, 5), (6, 6)

\right\}</math> correspond to the event <math>C_k = 2</math> and a total probability of Template:Sfrac. These are the only events that are faithfully preserved with identity of which dice rolled which outcome because the outcomes are the same. Without knowledge to distinguish the dice rolling the other numbers, the other <math display="inline"> \binom{6}{2} = 15</math> combinations correspond to one die rolling one number and the other die rolling a different number, each having probability Template:Sfrac. Indeed, <math display="inline"> 6 \cdot \tfrac{1}{36} + 15 \cdot \tfrac{1}{18} = 1</math>, as required.

Unsurprisingly, the information content of learning that both dice were rolled as the same particular number is more than the information content of learning that one die was one number and the other was a different number. Take for examples the events <math> A_k = \{(X, Y) = (k, k)\}</math> and <math> B_{j, k} = \{c_j = 1\} \cap \{c_k = 1\}</math> for <math> j \ne k, 1 \leq j, k \leq 6</math>. For example, <math> A_2 = \{X = 2 \text{ and } Y = 2\}</math> and <math> B_{3, 4} = \{(3, 4), (4, 3)\}</math>.

The information contents are <math display="block"> \operatorname{I}(A_2) = -\log_2\!{\tfrac{1}{36}} = 5.169925 \text{ Sh}</math> <math display="block"> \operatorname{I}\left(B_{3, 4}\right) = - \log_2 \! \tfrac{1}{18} = 4.169925 \text{ Sh}</math>

Let <math display="inline"> \text{Same} = \bigcup_{i = 1}^{6}{A_i}</math> be the event that both dice rolled the same value and <math> \text{Diff} = \overline{\text{Same}}</math> be the event that the dice differed. Then <math display="inline"> \Pr(\text{Same}) = \tfrac{1}{6}</math> and <math display="inline"> \Pr(\text{Diff}) = \tfrac{5}{6}</math>. The information contents of the events are <math display="block"> \operatorname{I}(\text{Same}) = -\log_2\!{\tfrac{1}{6}} = 2.5849625 \text{ Sh}</math> <math display="block"> \operatorname{I}(\text{Diff}) = -\log_2\!{\tfrac{5}{6}} = 0.2630344 \text{ Sh}.</math>

Information from sum of dice

The probability mass or density function (collectively probability measure) of the sum of two independent random variables is the convolution of each probability measure. In the case of independent fair 6-sided dice rolls, the random variable <math> Z = X + Y</math> has probability mass function <math display="inline"> p_Z(z) = p_X(x) * p_Y(y) = {6 - |z - 7| \over 36} </math>, where <math> *</math> represents the discrete convolution. The outcome <math> Z = 5 </math> has probability <math display="inline"> p_Z(5) = \frac{4}{36} = {1 \over 9} </math>. Therefore, the information asserted is<math display="block"> \operatorname{I}_Z(5) = -\log_2{\tfrac{1}{9}} = \log_2{9}

\approx 3.169925 \text{ Sh}. 

</math>

General discrete uniform distribution

Generalizing the Template:Section link example above, consider a general discrete uniform random variable (DURV) <math>X \sim \mathrm{DU}[a,b]; \quad a, b \in \mathbb{Z}, \ b \ge a.</math> For convenience, define <math display="inline">N := b - a + 1</math>. The probability mass function is <math display="block">p_X(k) = \begin{cases}

\frac{1}{N}, & k \in [a, b] \cap \mathbb{Z}  \\
0,  & \text{otherwise}.

\end{cases}</math>In general, the values of the DURV need not be integers, or for the purposes of information theory even uniformly spaced; they need only be equiprobable.<ref name=":0" /> The information gain of any observation <math>X = k</math> is<math display="block">\operatorname{I}_X(k) = -\log_2{\frac{1}{N}} = \log_2{N} \text{ Sh}.</math>

Special case: constant random variable

If <math>b = a</math> above, <math>X</math> degenerates to a constant random variable with probability distribution deterministically given by <math>X = b</math> and probability measure the Dirac measure <math display="inline">p_X(k) = \delta_{b}(k)</math>. The only value <math>X</math> can take is deterministically <math>b</math>, so the information content of any measurement of <math>X</math> is<math display="block">\operatorname{I}_X(b) = - \log_2{1} = 0.</math>In general, there is no information gained from measuring a known value.<ref name=":0" />

Categorical distribution

Generalizing all of the above cases, consider a categorical discrete random variable with support <math display="inline">\mathcal{S} = \bigl\{s_i\bigr\}_{i=1}^{N}</math> and probability mass function given by

<math display="block">p_X(k) = \begin{cases}

p_i, & k = s_i \in \mathcal{S}
\\ 0,  & \text{otherwise} .

\end{cases}</math>

For the purposes of information theory, the values <math>s \in \mathcal{S}</math> do not have to be numbers; they can be any mutually exclusive events on a measure space of finite measure that has been normalized to a probability measure <math>p</math>. Without loss of generality, we can assume the categorical distribution is supported on the set <math display="inline">[N] = \left\{1, 2, \dots, N \right\}</math>; the mathematical structure is isomorphic in terms of probability theory and therefore information theory as well.

The information of the outcome <math>X = x</math> is given

<math display="block">\operatorname{I}_X(x) = -\log_2{p_X(x)}.</math>

From these examples, it is possible to calculate the information of any set of independent DRVs with known distributions by additivity.

Derivation

By definition, information is transferred from an originating entity possessing the information to a receiving entity only when the receiver had not known the information a priori. If the receiving entity had previously known the content of a message with certainty before receiving the message, the amount of information of the message received is zero. Only when the advance knowledge of the content of the message by the receiver is less than 100% certain does the message actually convey information.

For example, quoting a character (the Hippy Dippy Weatherman) of comedian George Carlin:

Weather forecast for tonight: dark. ] Continued dark overnight, with widely scattered light by morning.<ref>Template:Cite web</ref>

Assuming that one does not reside near the polar regions, the amount of information conveyed in that forecast is zero because it is known, in advance of receiving the forecast, that darkness always comes with the night.

Accordingly, the amount of self-information <math>\operatorname{I}</math> contained in a message conveying an occurrence of event, <math>\omega_n</math>, depends only on the probability <math>\Pr(\omega_n)</math> of that event.<math display="block">\operatorname{I}(\omega_n) = f(\Pr(\omega_n)), </math>for some function <math>f</math> to be determined. If <math>\Pr(\omega_n) = 1</math>, then <math>\operatorname{I}(\omega_n) = 0</math>. If <math>\Pr(\omega_n) < 1</math>, then <math>\operatorname{I}(\omega_n) > 0</math>.

Further, by definition, the measure of self-information is nonnegative and additive. If an event <math>C</math> is the intersection of two independent events <math>A</math> and <math>B</math>, then the information of event <math>C</math> occurring is the sum of the amounts of information of the individual events <math>A</math> and <math>B</math>:<math display="block">\operatorname{I}(C) = \operatorname{I}(A \cap B) = \operatorname{I}(A) + \operatorname{I}(B).</math>Because of the independence of events <math>A</math> and <math>B</math>, the probability of event <math>C</math> is:<math display="block">\Pr(C) = \Pr(A \cap B) = \Pr(A) \cdot \Pr(B).</math>Relating the probabilities to the function <math>f</math>:<math display="block">f(\Pr(A) \cdot \Pr(B)) = f(\Pr(A)) + f(\Pr(B)).</math>This is a functional equation. The only continuous functions <math>f</math> with this property are the logarithm functions. Therefore, <math>f(p)</math> must be of the form:<math display="block">f(p) = K \log_b(p),</math>for some base <math>b</math> and constant <math>K</math>. Since a low-probability event must correspond to high information content, the constant <math>K</math> must be negative. We can write <math>K = -1</math> and absorb any scaling into the base <math>b</math> of the logarithm. This gives the final form:<math display="block">\operatorname{I}(\omega_n) = -\log_b(\Pr(\omega_n)) = \log_b \left(\frac{1}{\Pr(\omega_n)} \right). </math>The smaller the probability of event <math>\omega_n</math>, the larger the quantity of self-information associated with the message that the event indeed occurred. If the above logarithm is base 2, the unit of <math> I(\omega_n)</math> is shannon. This is the most common practice. When using the natural logarithm of base <math> e</math>, the unit will be the nat. For the base 10 logarithm, the unit of information is the hartley.

As a quick illustration, the information content associated with an outcome of 4 heads (or any specific outcome) in 4 consecutive tosses of a coin would be 4 shannons (probability 1/16), and the information content associated with getting a result other than the one specified would be <math>-\log_2(15/16) \approx 0.09</math> shannons. See above for detailed examples.

See also

References

Template:Reflist

Further reading

Template:Authority control