By composing any function with a random variable we obtain a new random variable . The new one is called a derived random variable.
Notation
The derived random variable may be written “”.
Expectation of derived variables
Discrete case:
(Here the sum is over all possible values of , i.e. where .)
Continuous case:
Notice: when applied to outcome :
is the output of
is the output of
The proofs of these formulas are tricky because we must relate the PDF or PMF of to that of .
Proof - Discrete case - Expectation of derived variable
Linearity of expectation
For constants and :
For any and on the same probability model:
Exercise - Linearity of expectation
Using the definition of expectation, verify both linearity formulas for the discrete case.
Be careful!
Usually .
For example, usually .
We distribute over sums but not products (unless the factors are independent).
Variance squares the scale factor
For constants and :
Thus variance ignores the offset and squares the scale factor. It is not linear!
Proof - Variance squares the scale factor
Extra - Moments
The moment of is defined as the expectation of :
Discrete case:
Continuous case:
A central moment of is a moment of the variable :
The data of all the moments collectively determines the probability distribution. This fact can be very useful! In this way moments give an analogue of a series representation, and are sometimes more useful than the PDF or CDF for encoding the distribution.
The exponential distribution is memoryless.
This means that knowledge that an event has not yet occurred does not affect the probability of its occurring in future time intervals:
This is easily checked using the PDF:
No other continuous distribution is memoryless.
This means any other (continuous) memoryless distribution agrees in probability with the exponential distribution. The reason is that the memoryless property can be rewritten as . Consider as a function of , and notice that this function converts sums into products. Only the exponential function can do this.
The geometric distribution is the discrete memoryless distribution.
and by substituting , we also know .
Then:
Extra - Inversion of decay rate factor in exponential
For constants and :
Derivation:
Let and observe that (the “tail probability”).
Now observe that:
Let . So we see that:
Since the tail event is complementary to the cumulative event, these two distributions have the same CDF, and therefore they are equal.
Extra - Geometric limit to exponential
Divide the waiting time into small intervals. Let be the probability of at least one success in the time interval for any . Assume these events are independent.
A random variable measuring the end time of the first interval containing a success would have a geometric distribution with in place of :
By taking the sum of a geometric series, one finds: