Professional Documents
Culture Documents
!
!
!
!
!
A random variable is a variable that takes on numerical values realized by the outcomes in the
sample space generated by a random experiment.
A random variable is discrete if it can take on no more than a countable number of values.
A random variable is continuous if it can take any value in an interval.
The probability distribution function, P(x), of a discrete random variable X represents the
probability that X takes the value x, as a function of x.
The probability distribution function of a discrete random variable must satisfy the following two
properties:
1. 0 P(x) 1 for any value of x.
2. The individual probabilities sum to 1.
The cumulative probability distribution of a random variable X represents the probability that
X does not exceed a value, i.e. F(x 0 ) = P(X x 0 ).
!
!
The properties of cumulative probability distributions for discrete random variables are quite similar
to that of the normal one.
The expected value of a discrete random variable X is defined as:
-
E[X ] = = xP(x).
x
!
!
Y2 =Var(a + bX ) = b 2X2
Summary results for the mean and variance of special linear functions:
- If a random variable always takes the value a, it will have mean a and variance 0.
The number of sequences with x successes in n independent trials is:
-
Y = E[a + bX ] = a + bX
C xn =
n!
x !(n x)!
If n independent trials are carried out, the distribution of the number of resulting successes, x, is
called the binomial distribution:
-
P(x) =
n!
x !(n x)!
Let X be the number of successes in n independent trials, each with probability of success P. Then X
follows a binomial distribution with mean and variance follows:
= E[X ] = nP
X2 = E[(X X )2 ] = nP(1 P)
!
!
e x
, for x = 0,1,2,
x!
- The constant is the expected number of successes per time or space unit. And it is the mean
and variance of the distribution.
The Poisson distribution can be used to approximate the binomial probabilities when n is large and
P is small (preferably such that = nP 7 ):
-
P(x) =
enP (nP)x
for x = 0,1,2,
x!
It can also be shown that when n 20 and P 0.05 , and the population mean is the same, both the
binomial and the Poisson distributions generate approximately the same probability values.
Suppose that random sample of n objects is chosen from a group of N objects, S of which are succeses. The distribution of X (number of successes) is the hypergeometric distribution:
-
P(x) =
N s
P(x) = C xsC nx
/C nN
Let X and Y be a pair of discrete random variables. Their joint probability distribution expresses the probability that simultaneously X takes the value x and Y takes the value y:
- P(x,y) = P(X = x Y = y).
In jointly distributed random variables, the probability distribution of the random variable X is its
marginal probability distribution, which is obtained by summing the joint probabilities over all
possible values: P(x) = P(x,y).
y
The conditional probability distribution of Y given that X takes the value x is defined as:
- P(y | x) = P(x,y) / P(x).
The jointly distributed random variables X and Y are independent if: P(x, y) = P(x)P(y).
The conditional mean is computed using the following: Y |X = E[Y | X ] = (y | x)P(y | x).
y
The expectation of any function g(X, Y) of these random variables is defined as follows:
-
E[g(X,Y )] = g(x,y)P(x,y).
x
Cov(X,Y ) = E[XY ] X Y .
If two random variables are statistically independent, the covariance between them is 0. However, the converse is not necessarily true.
2