Processing math: 100%
Google
 
Web unafbapune.blogspot.com

Saturday, January 11, 2014

 

Entropy

The entropy H(x) of a random variable X is defined as
  H(X)=xp(x)logp(x)
which measures the uncertainty of X. Note H(X) depends only on the distribution of X but not on the actual values taken by X. For example, here is a cute binary entropy function for X{γ,1γ}:
  hb(γ)=γlogγ(1γ)log(1γ)
Via the use of expectation:
  Eg(X)=xp(x)g(x)
Entropy can also be expressed compactly in terms of expectation:
  H(X)=Elogp(X)=xp(x)logp(x)
Interestingly, the joint entropy:
  H(X,Y)=Elogp(X,Y)=x,yp(x,y)logp(x,y)
But the conditional entropy of Y given X:
  H(Y|X)=Elogp(Y|X)=x,yp(x,y)logp(y|x)=xyp(x)p(y|x)logp(y|x)=xp(x)[yp(y|x)logp(y|x)]=xp(x)H(Y|X=x)
Similarly,
  H(Y|X,Z)=zp(z)H(Y|X,Z=z)
where
  H(Y|X,Z=z)=x,yp(x,y|z)logp(y|x,z)
Now it's not hard to show that
  H(X,Y)=H(X)+H(Y|X)
for
  H(X,Y)=Elogp(X,Y)=Elog[p(X)p(Y|X)]=Elogp(X)Elogp(Y|X)=H(X)+H(Y|X)
In general, there is the Chain rule for Entropy:
  H(X1,X2,,Xn)=ni=1H(Xi|X1,,Xi1)
which can be easily verified via induction. The base case is known to be true above. Suppose it is true for n, then
  H(X1,X2,,Xn,Xn+1)=H(Xn+1|X1,,Xn)+H(X1,,Xn)=H(Xn+1|X1,,Xn)+ni=1H(Xi|X1,,Xi1)induction hypothesis=n+1i=1H(Xi|X1,,Xi1)

Source: Information Theory.


Comments: Post a Comment

<< Home

This page is powered by Blogger. Isn't yours?