Definition: Two Random Variables
and
for the same Sample Space
and
Sigma Algebra
on
are called jointly distributed.
Notes:
We assume we are given , a Sample Space
,
Sigma Algebra
on
, and Probability Measure
on
.
We will simplify the notation a bit by using the notation
for
,
for
,and
for
However we need to be careful with this notation. In particular,
would
be the simplified notation for
See the definition immediately below.
Definitions:
The Joint Entropy
of a pair of Finite Random Variables
and
on
is
defined as:
Note that.
For each
, we have the Conditional Entropy
of
given
,
Finally The Conditional Entropy
of a pair of Finite Random Variables
and
is defined as follows:
Reversing the roles of
and
for each
we have:
and
For
Read:
Having learned the value
has
take
is the Information you get when you learn the value
has
taken
is the Expected Value of
Theorem:
with equality if and only
and
are
independent, that is
for all
.
Proof:
By Gibb's inequality, with
playing the role of the
's for the pairs
With equality if and only if
and
thus
and,
since,
similarly for
Theorem
(The Chain Rule):
equivalently
The Information you receive when you learn
plus the Information you receive when you learn about
given you know
equals the Information you receive what you learn when you learn about both
and
The Information you receive when you learn about both
and
minus the Information you receive when you learn about
equals the Information you receive when you learn about about
given you know
Proof (the second version):
and
,
,
No Noise,
and thus
We learn nothing new when we know what character was received given that we know what was transmitted.
since
and
all
All Noise,
and
since
the Random Variables are independent.
Exercise ( Due March 5): Compute
and
for a Binary Symmetric Channel and input vector