Professional Documents
Culture Documents
Covariance Correlation Sampling estimators Unbiasedness Efficiency Consistency Normal distribution Chi2 distribution F distribution T distribution Hypothesis testing
Probability Distribution Mean of the distribution, Variance of a Random Variable Mutual and Exhaustive events Joint, Marginal Probability
Econometrics
Quantitative Economic Analysis also goes by the name of Econometrics. What is Econometrics? The term literally means Economic Measurement. There is, however, more to Econometrics than just tricks.
Definition Of Econometrics
Econometrics may be defined as the social science in which the tools of economic theory, mathematics and statistical inference are applied to the analysis of economic phenomena.
Applications
Numerous other areas apart from Economics now use the techniques of econometrics: politics, psychology, history, marketing and advertising, finance, among others. Hence, Financial Econometric, Economic econometric, Whichever area econometrics is used in, the role of theory is crucial.
Theory: Hypothesis
Without theory there is no way of evaluating results. Theory enables the setting up of the hypothesis which is to be tested with the data. Unless the hypothesis is precisely specified accepting it or rejecting it will not be unambiguous.
e.g. 1. Forecasting the correlation between stock indices of two countries 2. Testing the hypothesis that earnings have no effect on stock prices. 3. Modeling long run relationship between prices and exchange rates.
Role of theory
Theory precedes data and not the other way round.
e.g. For most goods, relationship between consumption and disposable income is positive
Reversing this order consistently is bad econometrics and goes by the name of data mining
Methodology Of Econometrics
1. 2. 3. 4. 5. 6. 7. 8. Statement of theory or hypothesis Specification of mathematical model of the theory Specification of econometric model of the theory Obtaining the data Estimation of the parameters of the model Testing of hypothesis Validation of model by forecasting Using the model for control or policy purposes
Before We do Econometrics
Before we actually plunge into doing econometrics, we need to take a tour of a vitally important branch of statistics known as probability theory. Quite often, the emphasis on probability meets with resistance and, possibly, understandably so. Probability theory tends to be quite technical. But we cannot should not avoid studying it. Doing econometrics with knowing probability is a bit like trying to learn a language but avoiding learning the grammar of that language.
ECON 939
PROBABILITY CONCEPTS
PROBABILITY CONCEPTS
An investigator conducts a variety of experiments, such as, tossing a coin or rolling a pair of dice or examining the impact of advertising on consumer purchase decisions. All such experiments result in outcomes: a coin may show heads or tails; the dice may show a combination of numbers from one to six; advertising may increase consumption. Increase in disposable income may increase consumption.
Random Variables
The outcome of the experiment, results in some variables which can take on a variety of values. Such variables are known as random variables.
Random Variables
Continuous Random Variable: A continuous
random variable can take any real value (not just whole numbers) in at least one interval on the real line.
Probability Distribution
Associated
f(x) = P(X=x)
Therefore,
+ f(x2)+. . .+f(xn) = 1.
RANGE 0 - 0.5 0.5 - 1.0 1.0 - 1.5 1.5 - 2.0 2.0 - 2.5 2.5 - 3.0 3.0 - 3.5 3.5 - 4.0
20 5 No.ofStu 20 0 10 5 10 0 5 0 0
0 .5
1 .5
2 X
2 .5
3 .5
0 5 .3 0 .3 0 5 .2 0 .2 0 5 .1 0 .1 0 5 .0 0
F(X)
0 .5
1 .5
2 X
2 .5
3 .5
Rules of Summation
n
Rule 1:
i= 1 xi = x1 + x2 + . . . + xn
Rule 2:
Rule 3:
Note that summation is a linear operator which means it operates term by term.
Rule 4:
1 i = (axi + byi) = ai 1 xi + ib 1 yi = =
Rule 5:
= n 1 xi = i=
x1 + x2 + . . . + xn
n
Rule 6:
Notation:
n m
Rule 7:
i i = 1 = 1 f(xi,yj) == 1 j
i=1 j=1
f(xi,yj) =j = 1
i=1
f(xi,yj)
Mean of a Distribution
An important characteristic of a random variable is its mathematical expectation or expected value or its mean. The expected value or mean of a random variable X is the average value of the random variable in an infinite number of repetitions of the experiment (repeated samples); it is denoted by E(X).
E[X] = i = 1i f(xi) x
E[X] = x1f(x1) + x2f(x2) + . . . + xnf(xn)
Where f(xi) = P(X=xi)
i=1
n
xi f(xi)
2
i=1
n
i=1
E [g(X)] =
i = 1
= 1 i
n
i = 1 g1(xi) f(xi) + i1 =
g2(xi) f(xi)
1 9 25 7.4
6 48 130 39.4
E(X+5X2)= E(X)+5E(X2)
39.4
Variance:
Var(X) = E(X ) -
2
Definition: Var(X) = 2 = E[(X )2] = (xi )2f(xi) This can also be written as: Var(X)
= E [ [X E(X) ]2] = E [X2 - 2XE(X) + (E(X))2 ] = E(X2) - 2 E(X) E(X) + E[ E(X)2] = E(X2) - 2 E(X)2 + E(X)2 = E(X2) - E(X)2
Empirical proof
random variable X 1 3 5 2.4 P(X) X2 X+5X2 (X-u)^2
1 9 25 7.4
6 48 130 39.4
E(X)
2.4
7.4
39.4
1.64
1.64
Variance (contd.)
Some properties of the variance are: If c is a constant, Var(c) = 0 If a and b are constants,
Definitions Two or more event are called mutually exclusive if at most one of them occur when the experiment is performed, that is, if no two of them have outcomes in common. A B=empty set Two or more event are called collectively exhaustive if at least one of them occurs when the experiment is performed, that is, if their union makes the sample space. A B=S P(A) + P(B) =1
Joint Probabilities
To answer probability questions involving two or more random variables, we need to know their joint
probability distribution.
If X and Y are discrete random variables (r.v.), the probability that X=x and Y=y, is the joint probability function for X and Y and denoted by fXY(x, y). Thus:
Marginal Probability
In the above example, the probability of drawing a male student is 0.56; probability of female student is 0.44. These values are given at the bottom of the second table. These values are called f(g) are probability density function of G (=Gender) alone. Likewise, we can obtain values h(e) which are probability density function of E (=Education) alone. f(g) and h(e) are called marginal probabilities or marginal distributions of G and E respectively.
Conditional Probability
We are sometimes interested in knowing the probability of a r.v. (say, Y) given that another r.v. (say, X) has already taken place. Thus, in the above example we may wish to know what is probability a student is studying science, given that the student is male. Such probability is know as conditional probability.
P( A and B) P( A | B) = P( B)
Covariance (contd.)
cov(X,Y)
= E [(X - EX)(Y-EY)] = E [XY - X EY - Y EX + EX EY] = E(XY) - EX EY - EY EX + EX EY = E(XY) - 2 EX EY + EX EY
Excel
X 1 3 Covariance =0 does not mean that the two 4 random variables are independent.
var(X+Y)
Covariance (contd.)
The sign of the covariance between two r.v.'s indicates whether their association is positive or negative.
In quadrant (1), both X and Y are greater than E(X) and E(Y). Hence, the product (XEX)(Y-EY) will be positive. In quadrant (3) both are less than their means and hence the product is again positive. In quadrants (2) and (4) the product will be negative. Since there are more points in (1) and (3) than in (2) and (4), covariance will tend to be positive.
E(Y) 3 4
E(X)
Correlation (contd.)
Even though covariance indicates whether the association between two r.v.'s is positive or negative, the value of the covariance does not indicate the strength of their association. This is because the size of covariance is dependent on the units in which the variables are measured.
4 0 3 5 3 0 2 5 2 0 1 5 1 0 5 0 0 2 0 4 0 6 0 8 0 10 0 10 2 10 4
5 0 0 1 0 2 0 3 0 4 0 5 0 6 0 7 0 1 5 1 0 2 0
The two graphs above are identical and yet their covariances are different. Covariance of graph on the left = 288.8 Covariance of graph on the right = 72.2 Why is this so? The reason is that the values of X and Y in the left graph are double those in the right graph. Thus, the value of covariance depends on the units of measurement of X and Y.
Correlation (contd.)
To overcome the problem of units of measurement in the computation of the covariance, a normalised covariance measure is used. That is, this normalised measure does not change with a change in the units of measurement.
Definition: XY
1/2
If X and Y are positively related, the correlation coefficient () is positive; If they are negatively related, is negative. If = 0, X and Y are uncorrelated. is a number that lies between -1 and +1 when X and Y are perfectly linearly correlated. For the 2 graphs given above, = 0.8854.
Random Sampling
A statistical investigation arises out of the need to solve a problem, e.g. what is the effect of price reduction on sales of Coke? In formulating the answer to the question, it is important to identify the population, which is the totality of elements about which some information is desired.
Sampling Distribution
A function of the observed values of the observed r.v.'s that does not contain any unknown parameters is called a sample statistic. The two most frequently used sample statistics are:
Estimation of Parameters
In an empirical investigation, the analyst very often knows the general form of the probability distribution of the r.v.'s. However, the specific values of the population parameters say, mean ()and variance (2) are not known. Since a complete census of the population is out of the question, the analyst uses a sample to draw inferences about the population parameters i.e. the underlying probability distribution.
Properties of Estimators
We often need to know which is a good estimator of a population parameter. Is the sample mean a good estimator of the population parameter? Or should we add the largest and smallest values in a sample and divide the sum by 2? Unless we know the properties that good estimators should have, we will not know how to choose.
Unbiasedness
Suppose an unknown parameter is and its estimator is ^. ^ is a function of observations x1, x2,xn and does not depend on any unknown parameters. Since the xs are random, ^ is also random. Since ^ is a r.v., it has a probability distribution with a certain mean, E(^). Definition: An estimator ^ is said to be an unbiased estimator of if E(^) = . If this is equality does not hold, the estimator is said to be biased and the bias is E(^) .
Unbiasedness (contd.)
Obviously, a particular value of ^ from a given trial/sample would not equal . But drawing a large number samples, computing ^ for each sample, the average of these values must equal , if the estimator is to be unbiased. Unbiasedness is, however, not enough. If we draw only one sample, we cannot be sure if the value of ^ is close to or away from it.
Efficiency
Since it is possible to consider an infinite number of unbiased estimators, unbiasedness is not enough. We know that variance of a r.v. measures its dispersion around the mean. A smaller variance implies that the values are clustered closer to the mean as compared to a sample which has a larger variance. Thus, given two unbiased estimators, we would choose the one with the lower variance since on average it is closer to the true mean .
Efficiency (contd.)
a)
Definition:
Let 1^ and 2^ be two unbiased estimators of the parameter . If Var(1^) < Var(2^), then 1^ is said to be more efficient than 2^. The ratio [Var(1^)]/[Var(2^) is called relative efficiency. Among all the unbiased estimators, the one with the smallest variance is called the minimum variance unbiased estimator.
b) c)
c)
Consistency
Sometimes an estimator may not possess the desirable properties in small sample, but when the sample size is large, many of the desirable properties may hold. In such cases we let the sample size n increase indefinitely and the associated estimator is labelled n^. The most frequently used large-sample property is consistency.
Consistency (contd.)
Consistency means that as n increases, the estimator approaches the true . Definition: An estimator n^ is said to be a consistent estimator of if limnP( n^ + ) = 1, for all > 0. This property is expressed as plim(n^) = . Note that the above definition is valid for any , however small.
Normal Distribution
We have been discussing r.v.s and their probability distributions. In many contexts we need to deal specific distributions. The most important is the normal distribution. If X is a normally distributed r.v. with mean and variance 2, written as X~N(, 2), then its probability density functions is given by:
1 (x )2 f(x) = exp < x< 2 2 2 2 where exp[ a ] denotes the exponential function ea
The normal distribution is symmetric around and has a bell shape. The area under the normal curve between - and + is 68.26%; between -2 and +2, it is 95.44%; between -3 and +3, it is 99.73%.
1 x2 f(x) = exp 2 2
< x<
Property
Normal random variable has two parameters ( , 2) If X1 and X2 are random variables from a normal distribution => Y= aX1+bX2 => Y is a normal random variable Y(a 1+ b 2, var( 1)+var( 2)+2cov( 1 2) ) Generating normal random variables with excel
=NORMINV(RAND(),mean,std)
Excel
Mean std Experimental Mean Experimental std
20 5
X
11.7773 3 22.0765 9 21.1331 12.1274 5 21.4586 9 19.6138 8 12.5582 6 17.4003 2 25.7012 6 24.0501 1 22.3272 2 19.7688 8 25.6151 4 17.4763 9 14.2706 8 18.3260 5 14.6457 15.4612 8 23.8288 8 12.7164 7 22.1382 1 19.9142 25.7595 2 17.3297 4 15.4520 2 17.5318
19.79 5.039
bin
0 5 10 15 20 25 30 35 40
Bin 0 5 10 15 20 25 30
Frequen Cumulat cy ive % 0 0.00% 1 22 104 254 229 98 0.14% 2.50% 16.41% 49.51% 86.23% 98.47%
Chi-Square Distribution
The distribution of the sum of squares of n independent standard normal r.v.s is called chi-square (2) distribution with n degrees
Consider n r.v.s Z1, Z2,Zn, all of which are (independent) standard normally distributed N(0,1). Define a new r.v. U such that:
Students t-Distribution
Suppose Z~N(0,1) and U~ 2n with Z and U independent. Define the r.v. t = Z/U/n = Zn/U. The distribution of t is the t-distribution with n d.f. The t-distribution is symmetric around the origin and has a shape similar to the normal distribution. For large n, the t-distribution is approximately as N(0,1).
F-Distribution
F-distribution is the ratio of two independent chisquares. If U~ 2m and V~ 2n are independent of each other, then F=(U/m) (V/n) is called the Fdistribution with m and n d.f. and written F~Fm,n. The F-distribution has a distribution similar to that of chi-square. If the r.v. t has the t-distribution with n d.f., then t2 has the F-distribution with 1 and n d.f.
Testing of Hypothesis
A null hypothesis (Ho) is a statement of the status quo, one of no difference or no effect. If the null hypothesis is not rejected, no changes will be made. An alternative hypothesis (H1) is one in which some difference or effect is expected. Accepting the alternative hypothesis will lead to changes in opinions or actions.
Type II error : Do not reject H0 when it is in fact false. The probability of type II error is denoted by P(II) . Unlike P(I) , which is specified by the researcher, the magnitude of P(II) depends on the actual value of the population parameter (proportion).
Men
Find the probability of randomly selecting a man and a boy. Find the probability of randomly selecting a man and someone who survived.
W 332 1360
Survived Died
Men
Find the probability of randomly selecting a man Find the probability of randomly selecting someone who survived.
Survived Died
1
332 1360
k
Men
Find the probability of randomly selecting a man or a boy. Find the probability of randomly selecting a man or someone who survived.
W 332 1360
1692
Survived Died
Total