Thursday, December 22, 2011

What is Entropy in Information Theory?

What is the concept of entropy?



A communication system transmit a long sequence of symbols from an information source that is , a communication system not deal with a single message but with all possible messages.
Thus we are more interested in the average information that a source produces rather than the information content of a single source.
Hence entropy is the average information per symbol .
Or   Average information per individual message.

It is a measure of uncertainty in a random variable.

Assumptions:
1 ) The source is stationary so that the probabilities remain constant with time.
2 ) The successive symbols are statistically independent  and come from the source at an average rate of being symbols per second.

Mathematical expression :

H(X) = E(I(x)) = -∑P(x)logP(x)  bits/symbol
Where H (X) is known as entropy of the source X.



Entropy for Binary source:





1 comment:

  1. Slot Machine Games Near Me - CBSDetroit.com
    It is in 경기도 출장샵 fact the 의왕 출장안마 world's 전라북도 출장안마 best casino slot machine game for 광주 출장마사지 American players and can be played at the Harrah's in Detroit. This game 하남 출장안마 includes 2.

    ReplyDelete

About The author

My photo
Himanshu Dureja is an engineering student and part time blogger.