Definition: Two Random Variables
and
for the same Sample Space
and
Sigma Algebra
on
are called jointly distributed.
Notes:
We assume we are given , a Sample Space
,
Sigma Algebra
on
, and Probability Measure
on
.
We will simplify the notation a bit by using the notation for , for ,and for
However we need to be careful with this notation. In particular, would be the simplified notation for
See the definition immediately below.
Definitions:
The Joint Entropy
of a pair of Finite Random Variables
and
on
is
defined as:
Note that.
For each
, we have the Conditional Entropy
of
given
,
Finally The Conditional Entropy
of a pair of Finite Random Variables
and
is defined as follows:
Reversing the roles of and for each we have:
and
For Read: Having learned the value has take is the Information you get when you learn the value has taken
is the Expected Value of
Theorem:
with equality if and only
and
are
independent, that is
for all
.
Proof:
By Gibb's inequality, with playing the role of the 's for the pairs
With equality if and only if and thus
and,
since,
similarly for
Theorem (The Chain Rule):
equivalently
The Information you receive when you learn plus the Information you receive when you learn about given you know
equals the Information you receive what you learn when you learn about both and
The Information you receive when you learn about both and minus the Information you receive when you learn about
equals the Information you receive when you learn about about given you know
Proof (the second version):
and ,
,
No Noise,
and thus
We learn nothing new when we know what character was received given that we know what was transmitted.
since
and all
All Noise,
and
since the Random Variables are independent.
Exercise ( Due March 5): Compute and for a Binary Symmetric Channel and input vector