74 views

Uploaded by Fanny Sylvia C.

- 4.Kaleab Kebede_39-47
- chapter8 (1).ppt
- 7_Hypothesis Testing z and T_PDF
- New Research Report of Sheema
- Marketing Research
- Research Activity
- Ch 11 Sample
- business research
- 03_MEEG_GLM
- 32 Introduction to Biostatistics
- Anova Practice after midterm 2.pdf
- data8 uc berkeley
- Fallsem2018-19 Mee2013 Eth Mb218 Vl2018191002622 Reference Material i Simulation III Unit
- Six Sigma BooK Part2
- researchmethodology-091107043803-phpapp02
- 14625thegoodnessoffittest-160909005600
- collocations.pdf
- 11.15 Psych 101 Lecture
- MS-8-june2015.pdf
- ty

You are on page 1of 8

Hypothesis testing is one of the most commonly used statistical techniques

ever. Most often, scientists are interested in estimating the differences be-

tween two populations and use the sample means as the statistic of interest.

For this reason, the normal distribution is the one most often used to esti-

mate the probabilities of interest. Drawing conclusions about populations

from data is termed inference. Scientists wish to draw inferences about the

populations they are interested in from the data they have collected.

Power calculations also can be important in the case that we failed to

reject the null hypothesis of no effect or no significant difference. This process

can be important in the regulatory community, where failing to reject the

null hypothesis of no effect, is unconvincing unless accompanied by a power

analysis that shows that if there were an effect, the sample size was large

enough to detect it.

This lecture will explore the basic concepts behind power analysis using

the normal assumption. Power and sample size calculations can be more

complicated when using other distributions but the basic idea is the same.

Most hypothesis testing is conducted using the sample mean as the statistic

of interest–to estimate the true population mean. Consider a sample of size n

of random variables X1 , X2 , · · · , Xn , with E{Xi } = µ and V ar{Xi } = σ 2 ∀i.

Let

Pn

Xi

x̄ = i=1 .

n

Then,

½ Pn ¾

i=1 Xi 1

E{x̄} = E = nµ = µ

n n

and

½ Pn ¾

i=1 Xi 1 2 1 2

V ar{x̄} = V ar = nσ = σ

n n2 n

1

2

If Xi ∼ N {µ, σ 2 }, then we know from earlier results that x̄ ∼ N {µ, σn }.

Additionally, even if the data do not come from a normal distribution

½√ ¾

n (x̄ − µ)

lim P ≤ x = Φ(x).

n→∞ σ

Hence, even if our data are not normal, for a large enough sample size,

we can calculate probabilities for x̄ by applying the Central Limit Theorem,

and our answers will be close enough.

2. Hypothesis Testing

Hypothesis testing is a formal statistical procedure that attempts to answer

the question “Is the observed sample consistent with a given hypothesis.”

This boils down to calculating the probability of the sample given the hy-

pothesis, P {X|H}. To set up the procedure, scientists propose what is called

a null hypothesis. The null hypothesis is usually of the form: these data

were generated by strictly random processes, with no underlying mechanism.

Always, the null hypothesis is the opposite of the hypothesis that we are

interested in. The scientist will then set up a hypothesis test to compare the

null hypothesis to the mechanistic, scientific hypothesis consistent with their

scientific theory.

Example 1 Examples of null hypotheses are:

1. no difference in the response of patients to a drug versus a placebo,

and a well some distance away,

the national average, and

limit.

A hypothesis test is usually represented as follows:

H 0 : µ = µ0

vs.

Ha : µ 6= µ0

2

The null and alternative hypotheses should be specified before the test is

conducted and before the data are observed. The investigators also need to

specify a value for P {X|H0 } at which they will reject H0 . The idea is that if

the data are quite unlikely under the null hypothesis, then we conclude that

they are inconsistent with the null, and hence accept the alternative. Notice

that the null and the alternative are mutually exclusive and exhaustive–that

is, one or the other must be true, but it’s impossible that both are.

The probability the we reject the null is denoted α and is called the size

of the test. It’s complement 1 − α is called the significance level of the test,

though sometimes you will see these terms used interchangeably.

Note that for some α = P {X|H0 } small enough, we reject H0 . Hence,

α = P {we reject H0 , when H0 is true}, also called the probability of a type I

error.

The probability of a Type II error is given by P {we fail to reject H0 when

H0 is false} = β.

Truth

Test H0 Ha

H0 OK Type II Error

Ha Type I Error OK

3

The values under the normal curve that are equal or more extreme than

our test statistic constitute the rejection region.

Let’s begin with an example. Say that regulators desire a high certainty

that emissions are below 5 parts per billion for a particular contaminant, and

the regulatory limit is 8 parts per billion. They may conduct the following

test

H0 : µ < 5ppb

versus

Ha : µ ≥ 5ppb

At what value will we reject H0 ? Say, we would like to reject the null

hypothesis at the 95% confidence level. This means we wish to fix the prob-

ability of falsely rejecting H0 (type I error) at no greater than 5%. Here,

under H0 , µ can be fixed at µ = 5 without altering the size (confidence level)

of the test. Now we need to find the rejection region, i.e. the value of x̄ at

which we can reject H0 at 95% confidence.

We need to find a c such that

Now, since the standard deviation is taken to be 3 and the sample size is 5,

we can standardize x̄ under H0 so that it has a standard normal distribution

(a mean of 0 and a standard deviation of 1), and then we can make use of

the standard normal probability charts. We have

( ) ( )

x̄ − 5 c−5 c−5

P r{x̄ > c|H0 } = P r > = Pr z> (2.2)

√3 √3 √3

5 5 5

charts, we know that

c−5

= 1.64 (2.3)

√3

5

4

3. Power Calculations

Let’s continue with our example. In the event that the managers fail to reject

H0 , that is, they conclude that there is insufficient evidence that emissions

are above 5 ppb, they and their stakeholders, may want to ask the question:

“Was there sufficient information in our sample (i.e. is the lack of evidence

due to insufficient sample size) to have detected a difference of 3 ppb? Hence,

they need to calculate the power of the test when µ = 8 ppb.

The power of a test is defined as the probability that we correctly reject

the null hypothesis, given that a particular alternative is true. Power can

also be defined as

require an estimate of the variability of the statistic used to conduct the

test, in most cases this statistic is the sample mean. The standard deviation

of the sample mean is given by the sample standard deviation divided by the

square root of the sample size.

σx

σx̄ = √ (3.1)

n

Continuing with the example, say we wish to calculate the power of the

above test, for a normal sample of size 5 and with known standard deviation

3. As with the size calculation, for the purposes of this power calculation,

under Ha , µ can be fixed at µ = 8. The test statistic is the sample mean. We

will reject the null hypothesis for some value of x̄. This value can be easily

calculated using elementary statistics since we have made the assumption

that the sample mean is normally distributed. This means that we assume

that if we were to repeat the experiment a large number of times and calcu-

lated the mean each time, that the resulting sample of means would show a

normal distribution.

We know that the rejection region was x̄ ≥ 7.2. We can calculate the

power of this test at some alternative, say µ = 8.

5

We need

( )

x̄ − 8 7.2 − 8

P r {x̄ > 7.2|µ = 8} = P r > = P r {z > −0.5963} = 0.7245

√3 √3

5 5

(3.2)

The power of this test at the specified alternative is then 0.7245. Alterna-

tively, we can say that the probability of type II error, or the probability that

we failed to reject the null when the true mean was 8 is 1 − 0.7245 = 0.2755.

We can conduct a full power analysis by plotting the power at a wide variety

of alternatives, or distances from µ0 , assuming that the standard deviation

remains constant across all concentrations.

1.0

0.8

0.6

Power

0.4

0.2

6 8 10 12 14

Even better than performing a power analysis after an experiment has been

conducted is to perform it before any data are collected. Careful experimen-

tal design can save untold hours and dollars from being wasted. As Quinn&

Keogh point out, too often a “post hoc” power calculation reveals nothing

6

more than our inability to design a decent experiment. If we have any reason-

able estimate of the variability and a scientifically justifiable and interesting

alternative, or even a range of alternatives, we can estimate before hand

whether or not the experiment is worth doing given the limitations on our

time and budget.

Say we would like to set the probability of a type I error at no greater

than 5% and of a type II error at no greater than 10%, what sample size

would we need for the test shown above? We saw that we rejected H0 at

µ ¶

σ

x̄ ≥ z1−α √ + µ0 .

n

Now consider the desired power. We need to repeat the same process as

we did above for the α level, but this time solving for c using the z value for

the corresponding power.

µ ¶

σ

x̄ ≥ zβ √ + µa .

n

Now recall that zβ = −z1−β . Setting the two expressions for x̄ equal to

one another we have

µ ¶ µ ¶

σ σ

z1−α √ + µ0 = −z1−β √ + µa

n n

with zα and zβ being the corresponding z values and solving for n in the

above equation yields

· ¸2

2 z1−α + z1−β

n=σ (4.1)

µa − µ0

So, for the example above, from a standard normal probability chart we

have zα = 1.645 and zβ = 1.282. For this test, µa = 8 and µ0 = 5, yielding

· ¸2

1.645 + 1.282

n=9 = 8.56. (4.2)

3

for this experiment.

7

Of course, often we will have no preliminary data from which to estimate

a standard deviation. In this case, we must use a conservative “best guess”

for the variance. In practice, we may also not know exactly what is a sci-

entifically meaningful alternative. However, as practitioners of science we

should be working to move our community towards more careful planning of

experiments and more careful thinking about our questions before we begin

the experiment.

5. References

1. Pagano M and K. Gauvreau, 1993. Principles of Biostatistics, Duxbury

Press, Belmont, California.

Press, Cambridge.

- 4.Kaleab Kebede_39-47Uploaded byiiste
- chapter8 (1).pptUploaded byAashir Ahmad
- 7_Hypothesis Testing z and T_PDFUploaded byMatt Joseph Cabanting
- New Research Report of SheemaUploaded byParshant Garg
- Marketing ResearchUploaded byManoj Jhamb
- Research ActivityUploaded byAnthony Chan
- Ch 11 SampleUploaded byRavneet Singh
- business researchUploaded byKaran Kumar
- 03_MEEG_GLMUploaded byrk
- 32 Introduction to BiostatisticsUploaded byAamir AnwarAli
- Anova Practice after midterm 2.pdfUploaded byNataliAmiranashvili
- data8 uc berkeleyUploaded bySteven Truong
- Fallsem2018-19 Mee2013 Eth Mb218 Vl2018191002622 Reference Material i Simulation III UnitUploaded byRaj Patel
- Six Sigma BooK Part2Uploaded byfoofool
- researchmethodology-091107043803-phpapp02Uploaded byvihangimadu
- 14625thegoodnessoffittest-160909005600Uploaded byJasMisionMXPachuca
- collocations.pdfUploaded bygodoydelarosa2244
- 11.15 Psych 101 LectureUploaded byKasem Ahmed
- MS-8-june2015.pdfUploaded bydebaditya_hit326634
- tyUploaded byHideous Pika
- final project 2015Uploaded byapi-302636468
- A STUDY OF SECONDARY SCHOOL TEACHERS ENTREPRENEURIAL MOTIVATION IN RELATION TO THEIR ENTREPRENEURIAL COMPETENCIES.Uploaded byIJAR Journal
- Ashaolu_Resource Use 1Uploaded byismaildeji
- MOS_Group15 - Letter&Replies.docxUploaded bySudhanshu N Ranjan
- Attrition ReportUploaded byNinoy Patel
- Chap11 Chi SquareUploaded byRandy Vd
- RM_5___6_Group_no_10_Uploaded byPramod Khanvilkar
- T- ESPE-ENG-032917Uploaded bytruehellknight
- Divya Therasal j - 098001602010Uploaded byDeepak Satheesan
- Bio Statistics Pg CD mUploaded bycrneetesh

- Model- vs. design-based sampling and variance estimationUploaded byFanny Sylvia C.
- Chapter 11Uploaded byFanny Sylvia C.
- Non%26ParaBootUploaded byFanny Sylvia C.
- SampleSizeCalcRevisitedUploaded byFanny Sylvia C.
- Chapter 21Uploaded byFanny Sylvia C.
- Charles TaylorUploaded byFanny Sylvia C.
- Chapter 8Uploaded byFanny Sylvia C.
- ReviewChaps1-2Uploaded byFanny Sylvia C.
- Chapter 9Uploaded byFanny Sylvia C.
- Chapter 13Uploaded byFanny Sylvia C.
- Chapter 20Uploaded byFanny Sylvia C.
- ReviewChaps3-4Uploaded byFanny Sylvia C.
- Chapter 10Uploaded byFanny Sylvia C.
- Chapter 14Uploaded byFanny Sylvia C.
- Chapter 12Uploaded byFanny Sylvia C.
- Chapter 6Uploaded byFanny Sylvia C.
- The not so Short Introduction to LaTeXUploaded byoetiker
- Chapter5p2LectureUploaded byFanny Sylvia C.
- Data Modeling: General Linear Model &Statistical InferenceUploaded byFanny Sylvia C.
- GRM: Generalized Regression Model for Clustering Linear SequencesUploaded byFanny Sylvia C.
- Bio Math 94 CLUSTERING POPULATIONS BY MIXED LINEAR MODELSUploaded byFanny Sylvia C.
- Chapter 5Uploaded byFanny Sylvia C.
- Chapter 7Uploaded byFanny Sylvia C.
- Chapter 7Uploaded byFanny Sylvia C.
- Clustering in the Linear ModelUploaded byFanny Sylvia C.
- Good Article on Standard Error vs Standard DeviationUploaded byAshok Kumar Bharathidasan
- An Ova PowerUploaded byFanny Sylvia C.
- Intro BootstrapUploaded byMichalaki Xrisoula
- R Matrix TutorUploaded byFanny Sylvia C.

- Child Development Refers to the Biological and Psychological and Emotional Changes That Occur in Human Beings Between Birth and the End of AdolescenceUploaded byKitur Caroline
- Sample SizeUploaded bysaravan1891
- StatsUploaded bykkveluri
- A Live Project Report OnUploaded bysunil
- hoeUploaded bytikusbiru
- Using Rapid Appraisal MethodsUploaded byapi-3731030
- Caso Met II Esthely Blanco-Noe GuzmanUploaded byNoé B. Guzmán
- Module 3 a -Research DesignUploaded bysafssim
- BiasUploaded byr4aden
- Chapter2.Methods.qrUploaded byMiyYen
- Introduction to Qualitative AnalysisUploaded bySwami Gurunand
- RESEARCH - 1st SEM (Second Quarter)Uploaded byho la
- Assignment - MB0050 - Research Methodology - Set 2Uploaded byBhupinder Singh
- 5 Hypothesis Testing.pptUploaded byJosé Luis González
- 6 methods of data collection.pdfUploaded byKristine Agustin
- volvoUploaded byparinay202
- Bias RecallUploaded byAndre Chundawan
- Evaluation Beginners GuideUploaded byusamaknight
- PSY202 Lecture 11 Slides - Proposal Templates Sampling and Analyses in Mixed-Methods ResearchUploaded byEugene Joseph
- Preparing_a_Research_Proposal - Prof. Chandra aUploaded bymbmrameez
- Thesis Proposal RaniUploaded byRaniena Chokyuhyun
- PGP29370 SPSS AssignmentUploaded byDebsoumo Das
- Bakeman - meta-analizaUploaded byAna16A
- one sample t Test for MeanUploaded byshadowann
- The Research Prospectus is a Brief Overview of the Key Components of a Research StudyUploaded byLucciana Bolaños
- SampleUploaded byMaysoun Atoum
- Rangkuman Makul ResearchUploaded byWahyu Asikin
- Review-on-Fundamentals-of-Statistics_STUDENTS.pptxUploaded byKen Soriano
- What is a Systematic ReviewUploaded byjohnjkjk
- Sample Size and Power CalculationUploaded byRendy Adhitya Pratama