This course is a short series of lectures on Introductory Statistics. Topics
covered are listed in the Table of Contents. The notes were prepared by EwaPaszek and Marek Kimmel.
The development of this course has been supported by NSF 0203396 grant.
The mean, variance, and standard deviation
Mean and variance
Certain mathematical expectations are so important that they have special names. In this section we consider two of them: the mean and the variance.
Mean Value
If
X is a random variable with p.d.f.
of the discrete type and space
R =
, then
is the weighted average of the numbers belonging to
R , where the weights are given by the p.d.f.
.
We call
the mean of
X (or
the mean of the distribution ) and denote it by
. That is,
.
In mechanics, the weighted average of the points
in one-dimensional space is called the centroid of the system. Those without the mechanics background can think of the centroid as being the point of balance for the system in which the weights
are places upon the points
.
The example below shows that if the outcomes of
X are equally likely (i.e., each of the outcomes has the same probability), then the mean of
X is the arithmetic average of these outcomes.
Roll a fair die and let
X denote the outcome. Thus
X has the p.d.f.
Then,
which is the arithmetic average of the first six positive integers.
Got questions? Get instant answers now!
Variance
It was denoted that the mean
is the centroid of a system of weights of measure of the central location of the probability distribution of
X .
A measure of the dispersion or spread of a distribution is defined as follows:
If
and
exists,
the variance , frequently denoted by
or
, of a random variable
X of the discrete type (or variance of the distribution) is defined by
The positive square root of the variance is called
the standard deviation of
X and is denoted by
Let the p.d.f. of
X by defined by
The mean of
X is
To find the variance and standard deviation of
X we first find
Thus the variance of
X is
and the standard deviation of
X is
Got questions? Get instant answers now!
Let
X be a random variable with mean
and variance
. Of course,
, where a and b are constants, is a random variable, too. The mean of
Y is
Moreover, the variance of
Y is
Got questions? Get instant answers now!
Moments of the distribution
Let
r be a positive integer. If
exists, it is called
the
r th moment of the distribution about the origin. The expression moment has its origin in the study of mechanics.
In addition, the expectation
is called
the
r th moment of the distribution about
b . For a given positive integer r.
is called
the
r th factorial moment .
The second factorial moment is equal to the difference of the second and first moments:
There is another formula that can be used for computing the variance that uses the second factorial moment and sometimes simplifies the calculations.
First find the values of
and
. Then
since using the distributive property of
E , this becomes
Let continue with
example 4 , it can be find that
Thus
Got questions? Get instant answers now!
Recall the empirical distribution is defined by placing the weight (probability) of 1/
n on each of
n observations
. Then the mean of this empirical distribution is
The symbol
represents
the mean of the empirical distribution . It is seen that
is usually close in value to
; thus, when
is unknown,
will be used to estimate
.
Similarly,
the variance of the empirical distribution can be computed. Let
v denote this variance so that it is equal to
This last statement is true because, in general,
There is a relationship between the sample variance
and variance
v of the empirical distribution, namely
. Of course, with large
n , the difference between
and
v is very small. Usually, we use
to estimate
when
is unknown.
BERNOULLI TRIALS and BINOMIAL DISTRIBUTION