Entropy
Entropy in Information Theory
The choice of base for log, the logarithm, varies for different applications.
- Base 2 gives the unit of bits (or "shannons"), while
- Base e gives "natural units" nat, and
- Base 10 gives units of "dits", "bans", or "hartleys".
An equivalent definition of entropy is the expected value of the self-information of a variable.
Entropy is a measure of the uncertainty in a variable.
Entropy is measured in bits and comes as a number between zero and 1.
Note: Entropy bits are not the same bits as used in computing terminology.
Entropy is given by the following equation, where n is the number of outcomes and P(xi) is the probability of the outcome i .
Common values for b are 2 , e , and 10 .
Because the log of a number less than one will be negative, the entire sum is
negated to return a positive value.
That is, only one bit is required to represent the two equally probable outcomes, heads and tails.
If the coin has the same face on both sides, the variable representing its outcome has 0 bits of entropy; that is, we are always certain of the outcome and the variable will never represent new information.
Entropy can also be represented as a fraction of a bit.
For example, an unfair coin has two different faces, but is weighted such that the faces are not equally likely to land in a toss. Assume that the probability that an unfair coin will land on heads is 0.8, and the probability that it will land on tails is 0.2.
The entropy of a single toss of this coin is equal to the following:
The outcome of a single toss of an unfair coin can have a fraction of one bit of entropy.
For the unfair type coin, there are two possible outcomes of the toss, but we are NOT totally uncertain since one outcome is more frequent.



Comments
Post a Comment