You are on page 1of 297

1

346EN150
1 – 24

ANNAMALAI UNIVERSITY
DIRECTORATE OF DISTANCE EDUCATION

Master of Business Administration


First Year

F
RESEARCH METHODOLOGY AND STATISTICS

LESSONS: 1 – 24

Copyright Reserved
(For Private Circulation Only)
2

MASTER OF BUSINESS ADMINISTRATION


First Year

RESEARCH METHODOLOGY AND STATISTICS

Editorial Board
Members
Dr. K. Rajendran
Dean
Faculty of Arts
Annamalai University
Annamalainagar.

Dr. A. Rajamohan
Dr. C. Samudhrarajakumar
Professor and Head
Professor and Head Management Wing, D.D.E.
Department of Business Administration Annamalai University
Annamalai University Annamalainagar.
Annamalainagar.

Internals
Dr. K.Anandanatarajan Dr. K.Tamizh Jyothi
Assistant Professor Assistant Professor
Management Wing, D.D.E Management Wing, D.D.E
Annamalai University Annamalai University
Annamalainagar. Annamalainagar.
Externals
Dr. P.Vikkraman Dr. C. Vethirajan
Director i/c Professor of corporate Secreteriship
Department of Management Studies Alagappa University
Anna University Karaikudi
Coimbatore
Lesson Writers
Units: I – III Units: IV – VI
Dr. S. Arul Kumar Dr. N. Kathirvel
Assistant Professor Assistant Professor
Management Wing, D.D.E Department of Commerce
Annamalai University
Goverenment Arts College
Annamalainagar.
Udumalpet
3

MASTER OF BUSINESS ADMINISTRATION


First Year
RESEARCH METHODOLOGY AND STATISTICS

SYLLABUS

Objective
To equip the students with the basic understanding of the research
methodology and provide an insight into the application of modern analytical tools
and statistical techniques for the purpose of management decision making.
Unit–I
Research – Meaning – Types – Nature and scope of research – Review of
Literature - Problem formulation – Statement of research Objective – Value and cost
of information – Research Questions – Research Gap - Decision theory –Research
process – Research designs – Exploratory – Descriptive – Experimental Research.
Unit–II
Methods of data collection – Observational and Survey methods – Field work
plan – Administration of surveys – Training the field investigators – Sampling
methods – Sample size – Source of Data – Primary – Secondary - Preparation of
Interview Schedule - Questionnaire Design - Attitude measurement techniques –
Scaling Techniques.
Unit–III
Testing of Hypothesis - Hypothesis testing: Chi-square test - one sample and
two samples tests for means and proportions of large samples (z-test), one sample
and two sample tests for means of small samples (t-test), Standard Deviations – F-
test (ANOVA) – Correlation – Regression.
Unit–IV
Non-Parametric Methods - Sign test for paired data - Rank sum test - Mann –
Whitney U test and Kruskal Wallis test - One sample run test - Chi-square tests for
independence of attributes and goodness of fit.
Unit V
Multivariate Analysis - Factor Analysis - MANOVA - Discriminant Analysis,
Structural Equation Modeling - Conjoint analysis – Multidimensional Scaling –
Cluster Analysis (Theoretical Applications Only). Role of Statistics Software in
Business Decision.
Unit–VI
Research Report – Structure of Research Report – Title of a Report – Table of
Contents – Synopsis – Introduction and Body of a Report – Tables and Diagrams -
Results and Discussions – Recommendations, Conclusion and Scope for Further
Research –References – Appendices.
4

References Books
1. Gupta S: Research Methodology and Statistical Techniques, New Delhi, Deep
& Deep Publication (P) Ltd., 2002.
2. Tripathi P.C.: A Textbook of Research Methodology, New Delhi, Sultan Chand
& Sons, 2002.
3. R. Cauvery, U.K. Sudhanayak, R.Meenakshi, Research Methodology,
S.Chand, Chennai, 2012.
4. Dr.D.R. Krishnaswamy and Dr.M.Ranganathan, Methodology of Research in
Social Sciences, Second Edition, Himalaya Publishing House, 2014.
5. Dr.B.G. Satyaprasad and Dr.O.R. Krishnaswamy, Business Research
Methods, Fifth Edition, 2010.
6. J.K. Sachdeva, Business Research Methodology, Second Edition, Himalaya
Publishing House, 2013.
Journals and Magazines
1. Journal of Management Research
2. International Journal of Social Science Research
3. Journal of Educational and Social Science Research
4. International Journal of Research in Humanities and Social Sciences
5. Advances in Social Science Research Journal.
Web Resources
1. www.uk.sagepub.com
2. wps.pearsoned.co.uk/ema_uk_he_saunders_resmethbus_6
3. global.oup.com/uk/orc/busecon/business/brymanbrm3e
4. highered.mcgraw-hill.com/sites/0073373702/student_view0
5. www.wiley.com/WileyCDA/.../productCd-111994225X.html
5

MASTER OF BUSINESS ADMINISTRATION


First Year
RESEARCH METHODOLOGY AND STATISTICS

CONTENT
Lesson Title Page No.
No.
1 Basics of Research 1
2 Review of Literature 13
3 Research Process 29
4 Research Design 37
5 Sources of Data 43
6 Measurement and Scaling Techniques 66
7 The Nature of Field Work 80
8 Sampling 88
9 Hypotheis Testing 104
10 T-Test and Z-Test 137
11 F-Test (ANOVA) 151
12 Measure of Relationship – Correlation and 160
Regression
13 Non-Parametric Test 181
14 Sign Test 186
15 CHI Square Test 198
16 Mann-Whitney Test 208
17 Multivariate Statistical Analysis 218
18 Kruskal-Wallis Test 231
19 Structural Equation Modeling 239
20 Cluster Analysis 253
21 Research Report 263
22 Types of Report 269
23 Mechanics of writing Report 274
24 Precautions for writing Report 282
6
1

LESSON – 1

BASICS OF RESEARCH
1.1 INTRODUCTION
Research in simple terms, refers to a search for knowledge. It is also known as
a scientific and systematic search for information on particular topic or issue. It is
also known as the art of scientific investigation. Several social scientists have
defined research in different ways
In the Encyclopedia of Social Sciences, D. Slesinger and M. Stephension (1930)
defined research as "the manipulation of things, concept or symbols for the purpose
of generalizing to extend, correct or verify knowledge, whether that knowledge
aids in construction of theory or in practice of an art".
According to Redman and Mory (1923), defined research is a "systematized
effort to gain new knowledge". It is an academic activity and therefore the term
should be used in a technical sense. According to Clifford Woody (Ko thari 1988)
research comprises "defining and redefining problems, formulating hypothesis or
suggested solutions; collecting, organizing and evaluating data; making deductions
and reaching conclusions; and finally, carefully testing the conclusions to
determine whether they fit the formulating hypothesis".
Thus, research is an original addition to the available knowledge, which
contributes to its further advancement. It is an attempt to pursue truth through the
methods of study, observation, comparison and e xperiment. In sum, research is the
search for knowledge, using objective and systematic methods to find solution to a
problem.
1.2 OBJECTIVES
After studying this lesson, you will be able to understand:
 Meaning of research
 Scope and Objectives of research
 Criteria for good research
 Types of research
1.3 CONTENTS
1.3.1 Objectives of Research
1.3.2 Motivation in Research
1.3.3 Significance of Research
1.3.4 Scope of Research
1.3.5 Criteria of Good Research
1.3.6 Types of Research
1.3.1 Objectives of Research
The objective of research is to discover answers to questions by applying
scientific procedures. In the other words, the main aim of research is to find out
2

truth which is hidden and has not yet been discovered. Although every research
study has its own specific objectives, research objectives may be broadly grouped as
follows:
 To gain familiarity with or new insights into a phenomenon (i.e.,
formulative research studies);
 To accurately portray the characteristics of a particular individual, group,
or a situation (i.e., descriptive research studies);
 To analyse the frequency with which something occurs (i.e., diagnostic
research studies); and
 To examine a hypothesis of a causal relationship between two variables
(i.e., hypothesis-testing research studies).
1.3.2 Motivation in Research
What makes people to undertake research? This is a question of fundamental
importance. The possible motives for doing research may be either one or more of
the following:
 Desire to get a research degree along with its conseque ntial benefits;
 Desire to face the challenge in solving the unsolved problems, i.e., concern
over practical problems initiates research;
 Desire to get intellectual joy of doing some creative work;
 Desire to be of service to society;
 Desire to get respectability.
However, this is not an exhaustive list of factors motivating people to
undertake research studies. Many more factors such as directives of government,
employment conditions, curiosity about new things, desire to understand causal
relationships, social thinking and awakening, and the like may as well motivate (or
at times compel) people to perform research operations.
1.3.3 Significance of Research
"All progress is born of inquiry. Doubt is often better than overconfidence, for
it leads to inquiry, and inquiry leads to invention" is a famous Hudson Maxim in
context of which the significance of research can well be understood. Increased
amounts of research make progress possible. Research inculcates scientific and
inductive thinking and it promotes the development of logical habits of thinking and
organization.
The role of research in several fields of applied economics, whether related to
business or to the economy as a whole, has greatly increased in modern times. The
increasingly complex nature of business and government has focused attention on
the use of research in solving operational problems. Research, as an aid to
economic policy, has gained added importance, both for government and business.
Research provides the basis for nearly all government policies in our economic
system. For instance, government's budgets rest in part on an analysis of the needs
and desires of the people and on the availability of revenues to meet these needs.
3

The cost of needs has to be equated to probable revenues and this is a field where
research is most needed. Through research we can devise alternative policies and
can as well examine the consequences of each of these alternatives.
Decision-making may not be a part of research, but research certainly
facilitates the decisions of the policy maker. Government has also to chalk out
programmes for dealing with all facets of the country's existence and most of these
will be related directly or indirectly to economic conditions. The plight of cultivators,
the problems of big and small business and industry, working conditions, trade
union activities, the problems of distribution, even the size and nature of defense
services are matters requiring research. Thus, research is considered necessary
with regard to the allocation of nation's resources. Another area in government,
where research is necessary, is collecting information on the economic and social
structure of the nation. Such information indicates what is happening in the
economy and what changes are taking place. Colle cting such statistical information
is by no means a routine task, but it involves a variety of research problems. These
day nearly all governments maintain large staff of research technicians or experts
to carry on this work. Thus, in the context of govern ment, research as a tool to
economic policy has three distinct phases of operation, viz., (i) investigation of
economic structure through continual compilation of facts; (ii) diagnosis of events
that are taking place and the analysis of the forces underlyi ng them; and (iii) the
prognosis, i.e., the prediction of future developments.
Research has its special significance in solving various operational and
planning problems of business and industry. Operations research and market
research, along with motivational research, are considered crucial and their results
assist, in more than one way, in taking business decisions. Market research is the
investigation of the structure and development of a market for the purpose of
formulating efficient policies for purc hasing, production and sales. Operations
research refers to the application of mathematical, logical and analytical techniques
to the solution of business problems of cost minimisation or of profit maximisation
or what can be termed as optimisation problems. Motivational research of
determining why people behave as they do is mainly concerned with market
characteristics. In other words, it is concerned with the determination of
motivations underlying the consumer (market) behaviour. All these are of great help
to people in business and industry who are responsible for taking business
decisions. Research with regard to demand and market factors has great utility in
business. Given knowledge of future demand, it is generally not difficult for a firm,
or for an industry to adjust its supply schedule within the limits of its projected
capacity. Market analysis has become an integral tool of business policy these days.
Business budgeting, which ultimately results in a projected profit and loss account,
is based mainly an on sales estimate which in turn depends on business research.
Once sales forecasting is done, efficient production and investment programmes
can be set up around which are grouped the purchasing and financing plans.
Research, thus, replaces intuitive business decisions by more logical and scientific
decisions.
4

Research is equally important for social scientists in studying social


relationships and in seeking answers to various social problems. It provides the
intellectual satisfaction of knowing a few things just for the sake of knowledge and
also has practical utility for the social scientist to know for the sake of being able to
do something better or in a more efficient manner. Research in social sciences is
concerned both with knowledge for its own sake and with knowledge for what it can
contribute to practical concerns. "This double emphasis is perhaps especially
appropriate in the case of social science. On the one hand, its responsibility as a
science is to develop a body of principles that make possible the understanding and
prediction of the whole range of human interactions. On the other hand, because of
its social orientation, it is increasingly being looked to for practical guidance in
solving immediate problems of human relations."
In addition to what has been stated above, the significance of research can
also be understood keeping in view the following points:
a) To those students who are to write a master's or Ph.D. thesis, research
may mean a careerism or a way to attain a high position in the social
structure;
b) To professionals in research methodology, research may mean a source of
livelihood;
c) To philosophers and thinkers, research may mean the outlet for new ideas
and insights;
d) To literary men and women, research may mean the development of new
styles and creative work;
e) To analysts and intellectuals, research may mean the generalisations of
new theories.
Thus, research is the fountain of knowledge for the sake of knowledge and an
important source of providing guidelines for solving differen t business,
governmental and social problems. It is a sort of formal training which enables one
to understand the new developments in one's field in a better way.
1.5 Scope of research
Very little research in the field is pure in nature. That is, very few people do
research in research methodology. Most research is applied research, which has
wide application in many disciplines. Every profession uses research methods in
varying amounts in many areas. They use the methods and procedures developed
by research methodologists in order to increase understanding in their own
profession and to advance the professional knowledge base. It is through the
application of research methodology that they strengthen and advance their own
profession. Examine your own field. You will find that its professional practice
follows procedures and practices tested and developed by others over a long period
of time. It is in this testing process that you need research skills, the developments
5

of which fall in the category of pure re search. As a matter of fact, the validity of
your findings entirely depends upon the soundness of the research methods and
procedures adopted by you.
Within any profession, where you directly or indirectly provide a service, such
as health (nursing, occupational therapy, physiotherapy, community health, health
promotion and public health), education, psychology or social work, the application
of research can be viewed from four different perspectives:
 the service provider;
 the service administrator, manager and/or planner;
 the service consumer; and
 the professional.
1.6 Criteria of Good Research
Whatever may be the types of research works and studies, one thing that is
important is that they all meet on the common ground of scientific method
employed by them. One expects scientific research to satisfy the following criteria:
 The purpose of the research should be clearly defined and common
concepts be used.
 The research procedure used should be described in sufficient detail to
permit another researcher to repeat the research for further advancement,
keeping the continuity of what has already been attained.
 The procedural design of the research should be carefully planned to yield
results that are as objective as possible.
 The researcher should report with complete frankness, flaws in procedural
design and estimate their effects upon the findings.
 The analysis of data should be sufficiently adequate to reveal its
significance and the methods of analysis used should be appropriate. The
validity and reliability of the data should be checked carefully.
 Conclusions should be confined to those justified by the data of the
research and limited to those for which the data provide an adequate
basis.
 Greater confidence in research is warranted if the researcher is
experienced, has a good reputation in research and is a person of
integrity.
6
7

In other words, we can state the qualities of a good research as under:


1. d research is systematic: It means that research is structured with specified
steps to be taken in a specified sequence in accordance with the well defined set of
rules. Systematic characteristic of the research does not rule out creative thinking
but it certainly does reject the use of guessing and intuition in arriving at
conclusions.
2. Good research is logical: This implies that research is guided by the rules of
logical reasoning and the logical process of induction and deduction are of great
value in carrying out research. Induction is the process of reasoning from a part to
the whole whereas deduction is the process of reasoning from some premise to a
conclusion which follows from that very premise. In fact, logical reasoning makes
research more meaningful in the context of decision making.
3. Good research is empirical: It implies that research is related basically to one
or more aspects of a real situation and deals with concrete data that provides a
basis for external validity to research results.
4. Good research is replicable: This characteristic allows research results to be
verified by replicating the study and thereby building a sound basis for decisions.
1.7 Types of research
Types of research can be looked at from three different perspectives:
 Applications of the findings of the research study;
 Objectives of the study;
 Mode of enquiry used in conducting the study.
The classification of the types of a study on the basis of these perspectives is
not mutually exclusive: that is, a research study classified from the viewpoint of
'application' can also be classified from the perspectives of 'objectives' and 'enquiry
mode' employed. For example, a research project may be classified as pure or
applied research (from the perspective of application), as descriptive, co relational,
explanatory or exploratory (from the perspective of objectives) and as qu alitative or
quantitative (from the perspective of the enquiry mode employed).
Types of research: application perspective
If you examine a research endeavor from the perspective of its application,
there are two broad categories: pure research and applied research. In the social
sciences Pure research involves developing and testing theories and hypotheses
that are intellectually challenging to the researcher but may or may not have
practical application at the present time or in the future. Thus such work often
involves the testing of hypotheses containing very abstract and specialised
concepts.
Pure research is also concerned with the development, examination, verification
and refinement of research methods, procedures, techniques and tools that form
the body of research methodology. Examples of pure research include developing a
8

sampling technique that can be applied to a particular situation; developing a


methodology to assess the validity of a procedure; developing an instrument, say, to
measure the stress level in people; and finding the best way of measuring people's
attitudes. The knowledge produced through pure research is sought in order to add
to the existing body of knowledge of research methods.
Most of the research in the social sciences is appli ed. In other words, the
research techniques, procedures and methods that form the body of research
methodology are applied to the collection of information about various aspects of a
situation, issue, problem or phenomenon so that the information gathered can be
used in other ways - such as for policy formulation, administration and the
enhancement of understanding of a phenomenon.
Types of research: objectives perspective
If you examine a research study from the perspective of its objectives, broadly
a research endeavor can be classified as descriptive, co relational, explanatory or
exploratory.
A research study classified as a descriptive study attempts to describe
systematically a situation, problem, phenomenon, service or programme, or
provides information about, say, the living conditions of a community, or describes
attitudes towards an issue. For example, it may attempt to describe the types of
service provided by an organization, the administrative structure of an organization,
the living conditions of Aboriginal people in the outback, the needs of a community,
what it means to go through a divorce, how a child feels living in a house with
domestic violence, or the attitudes of employees towards management. The main
purpose of such studies is to descri be what is prevalent with respect to the
issue/problem under study.
The main emphasis in a co relational study is to discover or establish the
existence of a relationship/association/interdependence between two or more
aspects of a situation. What is the i mpact of an advertising campaign on the sale of
a product? What is the relationship between stressful living and the incidence of
heart attack? What is the relationship between fertility and mortality? What is the
relationship between technology and unempl oyment? What is the effect of a health
service on the control of a disease, or the home environment on educational
achievement? These studies examine whether there is a relationship between two or
more aspects of a situation or phenomenon and, therefore, are called co relational
studies.
Explanatory research attempts to clarify why and how there is a relationship
between two aspects of a situation or phenomenon. This type of research attempts
to explain, for example, why stressful living results in heart attacks; why a decline
in mortality is followed by a fertility decline; or how the home environment affects
children's level of academic achievement.
The fourth type of research, from the viewpoint of the objectives of a study, is
called exploratory research. This is when a study is undertaken with the objective
9

either to explore an area where little is known or to investigate the possibilities of


undertaking a particular research study. When a study is carried out to determine
its feasibility it is also called a feasibility study or a pilot study. It is usually carried
out when a researcher wants to explore areas about which she/he has little or no
knowledge. A small-scale study is undertaken to decide if it is worth carrying out a
detailed investigation. On the basis of the assessment made during the exploratory
study, a full study may eventuate. Exploratory studies are also conducted to
develop, refine and/or test measurement tools and procedures. Although,
theoretically, a research study can be classified i n one of the above objectives-
perspective categories, in practice, most studies are a combination of the first three;
that is, they contain elements of descriptive, correlational and explanatory
research. In this book the guidelines suggested for writing a research report
encourage you to integrate these aspects.
The third perspective in our typology of research concerns the process you
adopt to find answers to your research questions. Broadly, there are two
approaches to enquiry:
 The structured approach;
 The unstructured approach.
In the structured approach everything that forms the research process -
objectives, design, sample, and the questions that you plan to ask of respondents -
is predetermined. The unstructured approach, by contrast, allows flexibili ty in all
these aspects of the process. The structured approach is more appropriate to
determine the extent of a problem, issue or phenomenon, whereas the unstructured
approach is predominantly used to explore its nature, in other words, variation/
diversity per se in a phenomenon, issue, problem or attitude towards an issue. For
example, if you want to research the different perspectives of an issue, the
problems experienced by people living in a community or the different views people
hold towards an issue, then these are better explored using unstructured enquiries.
On the other hand, to find out how many people have a particular perspective, how
many people have a particular problem, or how many people hold a particular view,
you need to have a structured approach to enquiry. Before undertaking a
structured enquiry, in the author's opinion, an unstructured enquiry must be
undertaken to ascertain the diversity in a phenomenon which can then be
quantified through the structured enquiry. Both approaches have their place in
research. Both have their strengths and weaknesses. Therefore, you should not
'lock' yourself solely into a structured or unstructured approach.
Types of research: mode of enquiry perspective
The structured approach to enquiry is usually cl assified as quantitative
research and unstructured as qualitative research. The choice between quantitative
and qualitative approaches (or structured or unstructured) should depend upon:
 Aim of your enquiry - exploration, confirmation or quantification.
10

 Use of the findings - policy formulation or process understanding.


 The distinction between quantitative and qualitative research, in addition to
the structured/unstructured process of enquiry, is also dependent upon
some other considerations.
A study is classified as qualitative if the purpose of the study is primarily to
describe a situation, phenomenon, problem or event; if the information is gathered
through the use of variables measured on nominal or ordinal scales (qualitative
measurement scales); and if the analysis is done to establish the variation in the
situation, phenomenon or problem without quantifying it. The description of an
observed situation, the historical enumeration of events, an account of the different
opinions people have about an issue, and a description of the living conditions of a
community are examples of qualitative research.
On the other hand, the study is classified as quantitative if you want to
quantify the variation in a phenomenon, situation, problem or issue; if information
is gathered using predominantly quantitative variables; and if the analysis is geared
to ascertain the magnitude of the variation. Examples of quantitative aspects of a
research study are: How many people have a particular problem? How many people
hold a particular attitude?
The use of statistics is not an integral part of a quantitative study. The main
function of statistics is to act as a test to confirm or contradict the conclusions that
you have drawn on the basis of your understanding of analysed data. Statistics,
among other things, help you to quantify the magnitude of an association or
relationship, provide an indication of the confidence you can place in your findings
and help you to isolate the effect of different variables.
It is strongly recommended that you do not 'lock yourself' into becoming either
solely a quantitative or solely a qualitative researcher. It is true that there are
disciplines that lend themselves predominantly either to qualitative or to
quantitative research. For example, such disciplines as anthropology, history and
sociology are more inclined towards qualitative research, whereas psychology,
epidemiology, education, economics, public health and marketing are more inclined
towards quantitative research. However, this does not mean that an economist or a
psychologist never uses the qualitative approach, or that an anthropologist never
uses quantitative information. There is increasing recognition by most disciplines in
the social sciences that both types of research are important for a good research
study. The research problem itself should determine whether the study is carried
out using quantitative or qualitative methodologies.
As both qualitative and quantitative approaches have their strengths and
weaknesses, and advantages and disadvantages, 'neither one is markedly superior
to the other in all respects' (Ackroyd & Hughes 1992: 30). The measurement and
analysis of the variables about which information is obtained in a research study
are dependent upon the purpose of the study. In many studies you need to combine
both qualitative and quantitative approaches. For example, suppose you want to
find out the types of service available to victims of domestic violence in a city and
11

the extent of their utilisation. Types of service are the qualitative aspect of the study
as finding out about them entails description of the services. The extent of
utilisation of the services is the quantitative aspect as it involves estimating the
number of people who use the services and calculating other indicators that reflect
the extent of utilisation.
1.4 REVISION POINTS
1. Research is a "systematized effort to gain new knowledge
2. The objective of research is to discover answers to questions by applying
scientific procedures
3. Research provides the basis for nearly all government policies in our
economic system
4. Every profession uses research methods in varying amounts in many areas
5. Pure research involves developing and testing theories and hypotheses that
are intellectually challenging to the researcher but may or may not have
practical application at the present time or in the future.
6. A descriptive study attempts to describe systematically a situation, problem,
phenomenon, service or programme, or provides information about, say, the
living conditions of a community, or describes attitudes towards an issue
1.5 INTEXT QUESTIONS
1. What are the objectives of research?
2. State the significance of research.
3. What is the importance of knowing how to do research?
1.6 SUMMARY
This chapter outlined the Importance of research. Difference between basic
and applied research have been dealt in detail. This chapter has given the Meaning,
scope, types, structure of the Research.
1.7 TERMINAL EXERCISES
1. Research in simple terms, refers to a ---------------- for knowledge
2. The objective of research is to discover answers to questions by applying ----
------------- procedures
3. Research has its special significance in solving various operational and
planning problems of ----------------
1.8 SUPPLEMENTARY MATERIALS
1. Cherryholmes, C. H. (1992). Notes on pragmatism and scientific realism.
Educational Researcher,l 4, August-September, 13-1 7.
2. http://www.usg.edu/galileo/skills/unit01/infoage01_03.phtml
3. http://library.queensu.ca/research/subjects
12

1.9 ASSIGNMENTS
1. “Creative management, whether in public administration or private industry,
depends on methods of inquiry that maintain objectivity, clarity, accuracy
and consistency”. Discuss this statement and examine the significance of
research”.
2. “Research is much concerned with proper fact finding, analysis and
evaluation.” Do you agree with this statement? Give reasons in support of
your answer.
1.10 SUGGESTED READINGS/ REFERENCE BOOKS
1. Hollway, W. and Jefferson, T. (2000) Doing Qualitative Research Differently:
Free Association, Narrative and the Interview Method, London: Sage.
2. Schwandt, T. (1997) Qualitative Inquiry: A Dictionary of Terms, Thousand
Oaks, CA: Sage.
3. Kothari, C.R, 1985, Research Methodology- Methods and Techniques, New
Delhi, Wiley Eastern Limited.
4. Kumar, Ranjit, 2005, Research Methodology-A Step-by-Step Guide for
Beginners, (2nd.ed.), Singapore, Pearson Education.
1.11 LEARNING ACTIVITIES
1. Using your local and national newspapers compile a list of career
opportunities in the area of marketing research.
1.12 KEYWORDS
Research: It is one of the ways of finding answers to your professional and
practice questions. However, it is characterised by the use of tested procedures and
methods and an unbiased and objective attitude in the process of exploration
Systematic research: It means that research is structured with specified steps
to be taken in a specified sequence in accordance with the well defined set of rules.
Systematic characteristic of the research does not rule out creative thinking but it
certainly does reject the use of guessing and intuition in arriving at conclusions.
Pure research: In the social sciences Pure research involves developing and
testing theories and hypotheses that are intellectually challenging to the researcher
but may or may not have practical application at the present time or in the future.
Thus such work often involves the testing of hypotheses containing very abstract
and specialised concepts.

13

LESSON – 2

REVIEW OF LITERATURE
2.1 INTRODUCTION
One of the essential preliminary tasks when you undertake a research study is
to go through the existing literature in order to acquaint yourself with the available
body of knowledge in your area of i nterest. Reviewing the literature can be time
consuming, daunting and frustrating, but it is also rewarding. The literature review
is an integral part of the research process and makes a valuable contribution to
almost every operational step. It has value even before the first step; that is, when
you are merely thinking about a research question that you may want to find
answers to through your research journey. In the initial stages of research it helps
you to establish the theoretical roots of your study , clarify your ideas and develop
your research methodology. Later in the process, the literature review serves to
enhance and consolidate your own knowledge base and helps you to integrate your
findings with the existing body of knowledge. Since an importan t responsibility in
research is to compare your findings with those of others, it is here that the
literature review plays an extremely important role. During the write -up of your
report it helps you to integrate your findings with existing knowledge - that is, to
either support or contradict earlier research. The higher the academic level of your
research, the more important a thorough integration of your findings with existing
literature becomes. In summary, a literature review has the following functions:
 It provides a theoretical background to your study.
 It helps you establish the links between what you are proposing to
examine and what has already been studied.
 It enables you to show how your findings have contributed to the existing
body of knowledge in your profession.
 It helps you to integrate your research findings into the existing body of
knowledge.
In relation to your own study, the literature review can help in four ways. It
can:
 Bring clarity and focus to your research problem;
 Improve your research methodology;
 Broaden your knowledge base in your research area; and
 Contextualize your findings.
2.2 OBJECTIVES
 After studying this lesson, you will be able to understand:
 Review of literature
 Research problem
 Research objectives
14

 Research question
 Research gap
2.3 CONTENTS
2.3.1 How to review the literature?
2.3.2 The Research problem
2.3.3 Objectives
2.3.4 Research questions
2.3.5 Research gap
Bringing clarity and focus to your research problem
The literature review involves a paradox. On the one hand, you cannot
effectively undertake a literature search without some idea of the problem you wish
to investigate. On the other hand, the literature review can play an extremely
important role in shaping your research problem because the process of review ing
the literature helps you to understand the subject area better and thus helps you to
conceptualise your research problem clearly and precisely and makes it more
relevant and pertinent to your field of enquiry. When reviewing the literature you
learn what aspects of your subject area have been examined by others, what they
have found out about these aspects, what gaps they have identified and what
suggestions they have made for further research. All these will help you gain a
greater insight into your own research questions and provide you with clarity and
focus which are central to a relevant and valid study. In addition, it will help you to
focus your study on areas where there are gaps in the existing body of knowledge,
thereby enhancing its relevance.
Improving your research methodology
Going through the literature acquaints you with the methodologies that have
been used by others to find answers to research questions similar to the one you
are investigating. A literature review tells you if others have used procedures and
methods similar to the ones that you are proposing, which procedures and methods
have worked well for them and what problems they have faced with them. By
becoming aware of any problems and pitfalls, you will be better positioned to select
a methodology that is capable of providing valid answers to your research question.
This will increase your confidence in the methodology you plan to use and will
equip you to defend its use.
Broadening your knowledge base in your research area
The most important function of the literature review is to ensure you read
widely around the subject area in which you intend to conduct your research study.
It is important that you know what other researchers have found in regard to the
same or similar questions, what theories have been put forward and what gaps
exist in the relevant body of knowledge. When you undertake a research project for
a higher degree (e.g. an MA or a PhD) you are expected to be an expert in your area
of research. A thorough literature review helps you to fulfil this expectation.
15

Another important reason for doing a literature review is that it helps you to
understand how the findings of your study fit into the existing body of knowledge.
Enabling you to contextualise your findings
Obtaining answers to your research questions is comparatively easy: the
difficult part is examining how your findings fit into the existing body of knowledge.
How do answers to your research questions compare with what others have found?
What contribution have you been able to make to the existing body of knowledge?
How are your findings different from those of others? Undertaking a literature
review will enable you to compare your findings with those of others and answer
these questions. It is important to place your findings in the context of what is
already known in your field of enquiry
2.3.1 How to review the literature?
If you do not have a specific research problem, you should review the literature
in your broad area of interest with the aim of gradually narrowing it down to what
you want to find out about. After that the literature review should be focused
around your research problem. There is a danger in reviewing the literature without
having a reasonably specific idea of what you want to study. It can condition your
thinking about your study and the methodology you might use, resulting in a less
innovative choice of research problem and methodology than otherwise would have
been the case. Hence, you should try broadly to conceptualize your research
problem before undertaking your major literature review. There are four steps
involved in conducting a literature review:
Searching for the existing literature in your area of study.
Reviewing the selected literature.
Developing a theoretical framework.
Developing a conceptual framework.
The skills required for these tasks are different. Developing theoretical and
conceptual frameworks is more difficult than the other tasks.
1. Searching for the existing literature
To search effectively for the literature in your field of enquiry, it is imperative
that you have at least some idea of the broad subject area and of the problem you
wish to investigate, in order to set parameters for your search. Next, compile a
bibliography for this broad area. There are three sou rces that you can use to
prepare a bibliography:
books; journals; the Internet.
2. Reviewing the selected literature
Now that you have identified several books and articles as useful, the next step
is to start reading them critically to pull together theme s and issues that are of
relevance to your study. Unless you have a theoretical framework of themes in
mind to start with, use separate sheets of paper for each theme or issue you
16

identify as you go through selected books and articles. While going through the
literature you should carefully and critically examine it with respect to the following
aspects:
 Note whether the knowledge relevant to your theoretical framework has
been confirmed beyond doubt.
 Note the theories put forward, the criticisms of these and their basis, the
methodologies adopted (study design, sample size and its characteristics,
measurement procedures, etc.) and the criticisms of them.
 Examine to what extent the findings can be generalized to other situations.
 Notice where there are significant differences of opinion among researchers
and give your opinion about the validity of these differences.
 Ascertain the areas in which little or nothing is known - the gaps that exist
in the body of knowledge.
3. Developing a theoretical framework
Examining the literature can be a never-ending task, but as you have limited
time it is important to set parameters by reviewing the literature in relation to some
main themes pertinent to your research topic. As you start reading the literature,
you will soon discover that the problem you wish to investigate has its roots in a
number of theories that have been developed from different perspectives. The
information obtained from different books and journals now needs to be sorted
under the main themes and theories, highlighting agreements and disagreements
among the authors and identifying the unanswered questions or gaps. You will also
realise that the literature deals with a number of aspects that have a direct or
indirect bearing on your research topic. Use these aspects as a basis for developing
your theoretical framework. Your review of the literature should sort out the
information, as mentioned earlier, within this framework. Unless you review the
literature in relation to this framework, you will not be able to develop a focus in
your literature search: that is, your theoretical framework provides you with a guide
as you read. This brings us to the paradox mentioned previously: until you go
through the literature you cannot develop a theoretical framework, and until you
have developed a theoretical framework you cannot effectively review the literature.
The solution is to read some of the literature and then attempt to develop a
framework, even a loose one, within which you can organize the rest of the
literature you read. As you read more about the area, you are likely to change the
framework. However, without it, you will get bogged down in a great deal of
unnecessary reading and note-taking that may not be relevant to your study.
4. Developing a conceptual framework
The conceptual framework is the basis of your research problem. It stems from
the theoretical framework and usually focuses on the section(s) which become the
basis of your study. Whereas the theoretical framework consists of the theories or
issues in which your study is embedded, the conceptual framework describes the
aspects you selected from the theoretical framework to become the basis of your
17

enquiry. For instance, the theoretical framework includes all the theories that have
been put forward to explain the relationship between fertility and mortality.
However, out of these, you may be planning to test only one, say the fear of non -
survival. Similarly the conceptual framework is focused on indicators to measure
the success or failure of the strategies to enhance community responsiveness.
Hence the conceptual framework grows out of the theoretical framework and relates
to the specific research problem.
2.3.2 The Research Problem
Broadly speaking, any question that you want answered and any assumption
or assertion that you want to challenge or investigate can become a research
problem or a research topic for your study. However, it is important to remember
that not all questions can be transformed into research problems and some may
prove to be extremely difficult to study. According to Powers, Meenaghan and
Twoomey (1985: 38), 'Potential research questions may occur to us on a regular
basis, but the process of formulating them in a meaningful way is not at all an easy
task.' As a newcomer it might seem easy to formulate a problem but it requires
considerable knowledge of both the subject area and research methodology. Once
you examine a question more closely you will soon realize the complexity of
formulating an idea into a problem which is re searchable. 'First identifying and
then specifying a research problem might seem like research tasks that ought to be
easy and quickly accomplished. However, such is often not the case' (Yegidis &
Weinback 1991: 35). It is essential for the problem you formulate to be able to
withstand scrutiny in terms of the procedures required to be undertaken. Hence
you should spend considerable time in thinking it through.
The importance of formulating a research problem
The formulation of a research problem is the first and most important step of
the research process. It is like the identification of a destination before undertaking
a journey. In the absence of a destination, it is impossible to identify the shortest -
or indeed any - route. Similarly, in the absence of a clear research problem, a clear
and economical plan is impossible. To use another analogy, a research problem is
like the foundation of a building. The type and design of the building are dependent
upon the foundation. If the foundation is well designe d and strong you can expect
the building to be also. The research problem serves as the foundation of a research
study: if it is well formulated, you can expect a good study to follow. According to
Kerlinger: If one wants to solve a problem, one must generally know what the
problem is. It can be said that a large part of the problem lies in knowing what one
is trying to do. You must have a clear idea with regard to what it is that you want to
find out about and not what you think you must find.
A research problem may take a number of forms, from the very simple to the
very complex. The way you formulate a problem determines almost every step that
follows: the type of study design that can be used; the type of sampling strategy
that can be employed; the research instrument that can be used or developed; and
the type of analysis that can be undertaken. Suppose your broad area of interest is
18

depression. Further suppose you want to conduct a research study regarding


services available to patients with depression living in a community. If your focus is
to find out the types of service available to patients with depression, the study will
dominantly be descriptive and qualitative in nature. These types of studies fall in
the category of qualitative research and are carried out using qualitative research
methodologies. On the other hand, if you want to find out the extent of use of these
services, that is the number of people using them, it will dominantly use
quantitative methodologies even though it is descriptive i n nature describing the
number of people using a service. If your focus is to determine the extent of use in
relation to the personal attributes of the patients, the study will be classified as
correlational (and quantitative). The methodology used will be different than the one
used in the case of a descriptive study. Similarly, if your aim is to find out the
effectiveness of these services, the study will again be classified as correlational and
the study design used, methods of collecting data and its an alysis will be a part of
the quantitative methodology. Hence, it is important for you to understand that the
way you formulate a research problem determines all the subsequent steps that
you have to follow during your research journey.
The formulation of a problem is like the 'input' to a study, and the 'output' -
the quality of the contents of the research report and the validity of the associations
or causation established - is entirely dependent upon it. Hence the famous saying
about computers, 'garbage in, garbage out', is equally applicable to a research
problem. Initially, you may become more confused but this is normal and a sign of
progression. Remember: confusion is often but a first step towards clarity. Take time
over formulating your problem, for the clearer you are about your research
problem/question, the easier it will be for you later on. Remember, this is the most
crucial step.
Technique Involved In Defining A Problem
Let us start with the question: What does one mean when he/she wants to
define a research problem? The answer may be that one wants to state the problem
along with the bounds within which it is to be studied. In other words, defining a
problem involves the task of laying down boundaries within which a researcher
shall study the problem with a pre-determined objective in view.
How to define a research problem is undoubtedly a herculean task. However, it
is a task that must be tackled intelligently to avoid the perplexity encountered in a
research operation. The usual approach is that the researcher should himself pose
a question (or in case someone else wants the researcher to carry on research, the
concerned individual, organisation or an authority should pose the question to the
researcher) and set-up techniques and procedures for throwing light on the
question concerned for formulating or defining the research problem. But such an
approach generally does not produce definitive results because the question
phrased in such a fashion is usually in broad general terms and as such may not
be in a form suitable for testing.
19

Defining a research problem properly and clearly is a crucial part of a research


study and must in no case be accomplished hurriedly. However, in practice this a
frequently overlooked which causes a lot of problems later on. Hence, the research
problem should be defined in a systematic manner, giving due weightage to all
relating points. The technique for the purpose involves the undertaking of the
following steps generally one after the other: (i) statement of the pro blem in a
general way; (ii) understanding the nature of the problem; (iii) surveying the
available literature (iv) developing the ideas through discussions; and (v) rephrasing
the research problem into a working proposition. A brief description of all these
points will be helpful.
1. Statement of the problem in a general way
First of all the problem should be stated in a broad general way, keeping in
view either some practical concern or some scientific or intellectual interest. For
this purpose, the researcher must immerse himself thoroughly in the subject
matter concerning which he wishes to pose a problem. In case of social research, it
is considered advisable to do some field observation and as such the researcher
may undertake some sort of preliminary survey or what is often called pilot survey.
Then the researcher can himself state the problem or he can seek the guidance of
the guide or the subject expert in accomplishing this task. Often, the guide puts
forth the problem in general terms, and it is the n up to the researcher to narrow it
down and phrase the problem in operational terms. In case there is some directive
from an organizational authority, the problem then can be stated accordingly. The
problem stated in a broad general way may contain various ambiguities which must
be resolved by cool thinking and rethinking over the problem. At the same time the
feasibility of a particular solution has to be considered and the same should be kept
in view while stating the problem.
2. Understanding the nature of the problem
The next step in defining the problem is to understand its origin and nature
clearly. The best way of understanding the problem is to discuss it with those who
first raised it in order to find out how the problem originally came about and w ith
what objectives in view. If the researcher has stated the problem himself, he should
consider once again all those points that induced him to make a general statement
concerning the problem. For a better understanding of the nature of the problem
involved, he can enter into discussion with those who have a good knowledge of the
problem concerned or similar other problems. The researcher should also keep in
view the environment within which the problem is to be studied and understood.
3. Surveying the av ailable literature
All available literature concerning the problem at hand must necessarily be
surveyed and examined before a definition of the research problem is given. This
means that the researcher must be well -conversant with relevant theories in the
field, reports and records as also all other relevant literature. He must devote
sufficient time in reviewing of research already undertaken on related problems.
This is done to find out what data and other materials, if any, are available for
20

operational purposes. "Knowing what data are available often serves to narrow the
problem itself as well as the technique that might be used." 2 . This would also help a
researcher to know if there are certain gaps in the theories, or whether the existing
theories applicable to the problem under study are inconsistent with each other, or
whether the findings of the different studies do not follow a pattern consistent with
the theoretical expectations and so on. All this will enable a researcher to take new
strides in the field for furtherance of knowledge i.e., he can move up starting from
the existing premise. Studies on related problems are useful for indicating the type
of difficulties that may be encountered in the present study as also the possible
analytical shortcomings. At times such studies may also suggest useful and even
new lines of approach to the present problem.
4. Developing the ideas through discussions:
Discussion concerning a problem often produces useful information. Various
new ideas can be developed through such an exercise. Hence, a researcher must
discuss his problem with his colleagues and others who have enough experience in
the same area or in working on similar problems. This is quite often known as an
experience survey. People with rich experience are in a position to enlighten the
researcher on different aspects of his proposed study and their advice and
comments are usually in valuable to the researcher. They help him sharpen his
focus of attention on specific aspects within the field. Discussi ons with such
persons should not only be confined to the formulation of the specific problem at
hand, but should also be concerned with the general approach to the given
problem, techniques that might be used, possible solutions, etc.
1. Rephrasing the research problem
Finally, the researcher must sit to rephrase the research problem into a
working proposition. Once the nature of the problem has been clearly understood,
the environment (within which the problem has got to be studied) has been defined,
discussions over the problem have taken place and the available literature has been
surveyed and examined, rephrasing the problem into analytical or operational
terms is not a difficult task. Through rephrasing, the researcher puts the research
problem in as specific terms as possible so that it may become operationally viable
and may help in the development of working hypotheses.*
2.3.3 Objectives
Having decided what to study, and knowing why s/he wants to study it, the
investigator can now formulate his study objectives. Objectives should be closely
related to the statement of the problem. For example, if the problem identified is
low utilization of health stations in a rural district, the general objective of the
study could be to assess the reasons for this l ow utilization. If we break down this
general objective into smaller and logically connected parts, then we get specific
objectives
General objectives: aim of the study in general terms
21

Specific objectives: measurable statements on the specific questions to be


answered. Unlike the general objectives, the specific objectives are more specific
and are related to the research problem situation. They indicate the variable to be
examined and measured.
How should we state our objectives? We have to make sure that our objectives:
 Cover the different aspects of the problem and its contributing factors in a
coherent way and in a logical sequence
 Are clearly expressed in measurable terms
 Are realistic considering local conditions
 Meet the purpose of the study
 Use action verbs that are specific enough to be measured
Examples of action verbs are:
- to determine
- to compare
- to verify
- to calculate
- to describe
- to find out
- to establish
Avoid the use of vague non-action verbs such as;
- to appreciate
- to understand
- to study
- to believe
Research objectives can be stated as:
Questions - the objectives of this study are to answer the following questions
 Positive sentence - the objectives of this study are to find out, to establish,
to determine
 Hypothesis - the objective of this study is to verify the following hypothesis
(examples are given below) Based on the type of the study problem, it
might be possible to develop explanations for the problem that can be
tested. If so, we can formulate hypotheses in addition to the other study
objectives.
A hypothesis is a prediction of a relationship between one or more variables
and the problem under study. That is, It specifies the relationship among variables.
These variables are to be statistically tested at a later stage. In order to measure the
22

relationship among variables to be studied the dependent and independent


variables need to be identified.
A few examples are given below:
1. The health of children living in rural villagization projects is better than
those living in traditional rural communities.
2. To examine whether there is any significant difference between district “A”
and district “B” with respect to their malaria prevalence rates
One of the most important problems usually observed among students is the
tendency of stating too many study objectives which are not appropriately
addressed (or sometimes will be forgotten) in the sections that follow. It should be
noted that it is on the bases of these specific objectives that the methods, results
and discussion sections will be presented. For example, sample size calculations for
each stated objective and identifying (selecting) the most appropriate sample size
that will answer the required research questions is not covered in the development
of most research proposals. This is also true during the write up of the completed
research work. It is not uncommon to come across a situation in which some of the
specific objectives are not addressed in the results section at all. It is therefore
advisable to limit the number of specific objectives. In most practical situations, the
number of specific objectives should not exceed three.
2.3.3 Research Questions
A research question guides and centers your research. It should be clear and
focused, as well as synthesize multiple sources to present your unique argument.
Even if your instructor has given you a specific assignment, the research question
should ideally be something that you are interested in or care about. Be careful to
avoid the “all-about” paper and questions that can be answered in a few factual
statements.
Examples: 1. For instance, the following question is too broad and does not
define the segments of the analysis: Why did the chicken cross the road?
(The question does not address which chicken or which road.)
2. Similarly, the following question could be answered by a hypothetical
Internet search: How many chickens crossed Broad Street in Durham, NC, on
February 6, 2014?
(Ostensibly, this question could be answered in one sentence and does not
leave room for analysis. It could, however, become data for a larger argument.)
3. A more precise question might be the following: What are some of the
environmental factors that occurred in Durham, NC between January and February
2014 that would cause chickens to cross Broad Street?
(This question can lead to the author taking a stand on which factors are
significant, and allows the writer to argue to what degree the results are beneficial
or detrimental.)
23

How Do You Formulate A Good Research Question?


Choose a general topic of interest, and conduct preliminary research on this
topic in current periodicals and journals to see what research has already been
done. This will help determine what kinds of questions the topic generates.
Once you have conducted preliminary research, consider: Who is the
audience? Is it an academic essay, or will it be read by a more general public? Once
you have conducted preliminary research, start asking openended “How?” “What?”
and Why?” questions. Then evaluate possible respon ses to those questions.
Examples
Say, for instance, you want to focus on social networking sites. After reading
current research, you want to examine to what degree social networking sites are
harmful. The Writing Center at George Mason University provide s the following
examples and explanations:
Possible Question: Why are social networking sites harmful? An evaluation of
this question reveals that the question is unclear: it does not specify which social
networking sites or state what harm is being cause d. Moreover, this question takes
as a given that this “harm” exists. A clearer question would be the following:
Revised Question: How are online users experiencing or addressing privacy
issues on such social networking sites as Face book and Twitter? This version not
only specifies the sites (Face book and Twitter), but also the type of harm (privacy
issues) and who is harmed (online users).
While a good research question allows the writer to take an arguable position,
it does not leave room for ambiguity.
Research Question in the Sciences and Social Sciences
While all research questions need to take a stand, there are additional
requirements for research questions in the sciences and social sciences. That is,
they need to have repeatable data. Unreliable data in the original research does not
allow for a strong or arguable research question.
In addition, you need to consider what kind of problem you want to address. Is
your research trying to accomplish one of these four goals?
1. Define or measure a specific fact or gather facts about a specific
phenomenon.
2. Match facts and theory.
3. Evaluate and compare two theories, models, or hypotheses.
4. Prove that a certain method is more effective than other methods. Moreover,
the research question should address what the variables of the experiment
are, their relationship, and state something about the testing of those
relationships. The Psychology department at California State University,
Fresno, provides the following examples and explanations:
24

Examples
Possible research question: Are females smarter than males? This question
delineates the variables to be measured: gender and intelligence. Yet, it is unclear
how they will be evaluated: What method will be used to define and measure
intelligence?
Revised question: Do females age 18-35 score higher than adult males age 18-
35 on the WAIS-III? (The WAIS-III is a standardized intelligence test.) This research
question produces data that can be replicated. From there, the author can devise a
question that takes a stand.
In essence, the research question that guides the sciences and social sciences
should do the following three things:
1. Post a problem.
2. Shape the problem into a testable hypothesis.
3. Report the results of the tested hypothesis.
There are two types of data that can help shape research questions in the
sciences and social sciences: quantitative and qualitative data. While quantitative
data focuses on the numerical measurement and analysis between variables,
qualitative data examines the social processe s that give rise to the relationships,
interactions, and constraints of the inquiry
2.3.4 Research Gap
Once you have found a promising research area of interest, you must identify
research questions that are not only unresolved but whose exploration can
meaningfully contribute to existing theory and/or practice. In this section, you will
learn to identify such promising research questions by avoiding common beginner's
mistakes, understanding how to identify a "gap" in the scholarly literature, and
making sure it is worthwhile to try to fill the gap.
Don’t decide on a research method or focus on the data to be analyzed before
finding the research problem. Because many graduate students are relatively
inexperienced in designing and conducting original research, they can forget that
research questions must be formulated before determining the most appropriate
research method or focusing on the type of data to be collected. For example, the
researchers in Scenario 1 and Scenario 2 make the mistake of not determin ing the
research questions first. The researcher in Scenario 3, on the other hand, wisely
bases her research design on the nature of the research questions.
Scenario 1: Jane would like to conduct a content analysis for her master's
thesis project. She decides to investigate how the characteristics of online gaming
walkthroughs affect online community response. Since she is doing a content
analysis, Jane decides to collect data on text features such as reference to self, use
of jargon, use of sentence fragments, etc.
25

Why constrain the analysis to word- or phrase-level data? Jane should


consider using discourse analysis, a method in which the object of analysis is the
whole piece of writing.
Scenario 2: John is interested in software tools that track users' l ine of vision
as they navigate digital interfaces. Since such tools track patterns of users' visual
focus, he decides to find out how users' interactions with social networking sites
differ based upon differences in these patterns.
Users' visual focus patterns provide only a limited view of their interaction
with a system. John should think about using click-through data or individual user
account activity data.
Scenario 3: Kristy is interested in how information architecture decisions on
educational websites affect self-directed learning behavior. Since she is interested
in comparing the detailed behavior of many users across 6-10 different websites,
Kristy decides to collect the lesson log data and survey data from users who agree
to participate in the study.
Based on her research question, Kristy determined which data to collect. She
then based her research method (survey, log transcript analysis) on the desired
data and other practical considerations.
What is a gap in the literature?
A gap in the literature is a research question relevant to a given domain that
has not been answered adequately or at all in existing peer-reviewed scholarship. A
gap in the literature may emerge if:
1. The question has not been addressed in a given domain, although it may
have been answered in a similar or related area.
Example: Existing theories on collaborative learning suggest that peers utilize
one another's knowledge and skills as a means to gradually move toward
independent problem solving. However, when online distance learning first
emerged, it was not yet known whether or how email and bulletin board posts could
similarly facilitate collaborative learning, necessitating the research question, "How
does collaborative learning occur through asynchronous textual communicatio n?"
2. The question has never been asked before, but it now merits exploration
due to changes in accepted theory, data collection technology, or culture.
Example: A study uses new eye tracking software to determine how spatial
language descriptions direct people’s judgments and parsing of a visual scene. Eye
tracking allows researchers to observe exactly how attention is allocated in visual
scenes over time, affording a way to test an existing attention model.
3. The question has been asked and tested in peer-reviewed research, but the
methods were either of questionable validity or had necessitated limited
applicability of results. Alternatively, a replication study could be run to
verify a published study's results if appropriate.
26

Example: A published study tested whether adapting website content per


established guidelines to be intellectually accessible to persons with intellectual
disabilities significantly affect its usability for this group. The study was criticized
for not testing word-level characteristics (i.e. word choice, syllable count). Thus, a
new study may ask, "How do variations in grammar construct use affect website
content comprehension among intellectually disabled users?"
Why does it matter?
Even if a legitimate gap in the literature exists, it does not necessarily mean
that the research question(s) merits pursuit. To justify to yourself and others the
investment of time and energy into designing and conducting research, you must
ensure the research will likely have valuable practical and/or theo retical
implications.
4. Practical implications: Answering the research question could improve
existing practice and/or inform professional decision -making.
Example: A study investigating the effect of the incorporation of a "threat
condition" in computer-mediated communication on users' long-term memorization
of content could potentially apply its findings to risk communication design (if
threat conditions were found to enhance long-term memory in CMC).
5. Theoretical implications: Answering the research question could revise,
build upon, or create theory informing research design and practice.
Example: A study investigating how the use of verbal expressed thanks plays a
role in establishing and maintaining relationships could contribute to existing
theory of non-informational communication use in humans.
2.4 REVISION POINTS
1. The literature review is an integral part of the research process and makes
a valuable contribution to almost every operational step
2. Literature review helps you establish the links between what you are
proposing to examine and what has already been studied
3. Potential research questions may occur to us on a regular basis, but the
process of formulating them in a meaningful way is not at all an easy task.
4. First of all the problem should be stated in a broad general way, keeping in
view either some practical concern or some scientific or intellectual interest
5. Objectives should be closely related to the statement of the problem.
6. A research question guides and centers your research
7. Once you have found a promising research area of interest, you must
identify research questions that are not only unresolved but whose
exploration can meaningfully contribute to existing theory and/or practice.
2.5 INTEXT QUESTIONS
1. Defining a research problem properly is a prerequisite for any study. Why?
2. While you define a research problem what do you try to do?
27

3. What precautions should be taken while formulating a problem?


4. What are the uses of literature review?
5. What is research objectives explain briefly?
6. What are the ways available for finding research gap?
2.6 SUMMARY
This chapter has given the Review of literature, Research problem, Research
objectives, Research question, and Research gap. These five areas are known as key
for the research. Before finalizing the researc h design these activities can be
finalized.
2.7 TERMINAL EXERCISES
1. A good topic should be small enough for a .......................... investigation.
2. A ......................... should always avoid selecting the first problem that he
encounters.
3. The research problem undertaken for study must be .......................
selected.
2.8 SUPPLEMENTARY MATERIALS
1. http://uq.edu.au/student-services/pdf/learning/lit-reviews-for-rx-students-
v7.pdf
2. https://www.tru.ca/__shared/assets/Literature_Review_Template30564.pdf
3. Boote, D.N. & Beile, P. (2005). Scholars before researchers: On the central ity
of the dissertation literature review in research preparation. Educational
Researcher 34/6, 3-15.
2.9 ASSIGNMENTS
1. It is often said that there is not a proper link between some of the activities
under way in the world of academics and in most business in our country. Account
for this state of affairs and give suggestions for improvement.
2. “Knowing what data are available often serves to narrow down the problem
itself as well as the technique that might be used.” Explain the underlying idea in
this statement in the context of defining a research problem.
2.10 SUGGESTED READINGS/ REFERENCE BOOKS
1. De Vaus, D.A. (1991) Surveys in Social Research, 3rd edition, London: UCL
Press.
2. Gash, S. (1999) Effective Literature Searching for Research, 2nd edition,
Aldershot: Gower.
3. Hart, C. (2001) Doing a Literature Search, London: Sage.
4. Spence, G. (2001) A Simple Guide to Internet Research, Harlow: Prentice
Hall
28

2.11 LEARNING ACTIVITIES


1. Describe principles and procedures of review of literature and precautions in
consulting library material
2.12 KEYWORDS
Research objectives are specific statements of goals that you set out to be
achieved at the end of your research journey.
Research problem: Any issue, problem or question that becomes the basis of
your enquiry is called a research problem. It is what you want to find out about
during your research endeavor.
Research Questions: Questions that you would like to find answers to through
your research, like ‘What does it mean to have a child with ADHD in a family ?’ or
‘What is the impact of immigration on family roles?’ Research questions become the
basis of research objectives. The main difference between research questions and
research objectives is the way they are worded. Research questions take the form of
questions whereas research objectives are statements of achievements expressed
using action oriented words.

29

LESSON – 3

RESEARCH PROCESS
3.1 INTRODUCTION
The research process is the step-by-step procedure of developing one's
research. However, one can seldom progress in a step-by-step fashion as such. The
research process involves identifying, locating, assessing, analyzing, and then
developing and expressing your ideas. All of these activities will be based on
primary and secondary sources from which recommendations or plans are
formulated. The details steps in research process are explained below.
3.2 OBJECTIVES
After studying this lesson, you will be able to understand:
 Research process
 Steps in research process
3.3 CONTENTS
3.3.1 Steps in research process
3.3.1 Steps in research process
Research process consists of a series of steps or actions required for effectively
conducting research. The following are the steps that provide useful procedural
guidelines regarding the conduct of research:
1. Formulating the research problem;
2. Extensive literature survey;
3. Developing hypothesis;
4. Preparing the research design;
5. Determining sample design;
6. Collecting data;
7. Execution of the project;
8. Analysis of data;
9. Hypothesis testing;
10. Generalization and interpretation, and
11. Preparation of the report or presentation of the results.
1. Formulating the research problem
The first and foremost stage in the research process is to select and properly
define the research problem. A researcher should first identify a problem and
formulate it, so as to make it amenable or susceptible to research. In general, a
research problem refers to an unanswered question that a researcher might
encounter in the context of either a theoretical or practical situation, which he/she
would like to answer or find a solution to. Identification of a research problem is the
pre-condition to conducting research. A research problem is said to be the one
30

which requires a researcher to find the best available solution to the given problem.
That is, the researcher needs to find out the best course of action through which
the research objective may be achieved optimally in the context of a given situation.
Several factors may contribute to making the problem complicated. For example,
the environment may alter, thus affecting the efficiencies of the alternative courses
of action taken or the quality of the outcomes. The number of alternative courses of
action might be very large and the individual not involved in making the decision
may be affected by the change in environment and may react to it favorably or
unfavorably. Other similar factors are also likely to cause such changes in the
context of research, all of which may be considered from the point of view of a
research problem.
2. Extensive literature survey
Once the problem is formulated, a brief summary of it should be written down.
It is compulsory for a research worker writing a thesis for a Ph.D. degree to write a
synopsis of the topic and submit it to the necessary Committee or the Research
Board for approval. At this juncture the researcher should undertake extensive
literature survey connected with the problem. For this purpose, the abstracting and
indexing journals and published or unpublished bibliographies are the first place to
go to. Academic journals, conference proceedings, government reports, books etc.,
must be tapped depending on the nature of the problem. In this process, it should
be remembered that one source will lead to another. The earlier studies, if any,
which are similar to the study in hand should be carefully studied. A good library
will be a great help to the researcher at this stage.
3. Development of working hypotheses
After extensive literature survey, researcher should state in clear terms the
working hypothesis or hypotheses. Working hypothesis is tentative assumption
made in order to draw out and test its logical or empiric al consequences. As such
the manner in which research hypotheses are developed is particularly important
since they provide the focal point for research. They also affect the manner in which
tests must be conducted in the analysis of data and indirectly the quality of data
which is required for the analysis. In most types of research, the development of
working hypothesis plays an important role. Hypothesis should be very specific and
limited to the piece of research in hand because it has to be tested. The role of the
hypothesis is to guide the researcher by delimiting the area of research and to keep
him on the right track. It sharpens his thinking and focuses attention on the more
important facets of the problem. It also indicates the type of data require d and the
type of methods of data analysis to be used.
How does one go about developing working hypotheses? The answer by using
the following approach:
a) Discussions with colleagues and experts about the problem, its origin and
the objectives in seeking a solution;
31

b) Examination of data and records, if available, concerning the problem for


possible trends, peculiarities and other clues;
c) Review of similar studies in the area or of the studies on similar problems;
and
d) Exploratory personal investigation which involves original field interviews
on a limited scale with interested parties and individuals with a view to
secure greater insight into the practical aspects of the problem.
Thus, working hypotheses arise as a result of a-priori thinking about the
subject, examination of the available data and material including related studies
and the counsel of experts and interested parties. Working hypotheses are more
useful when stated in precise and clearly defined terms. It may as well be
remembered that occasionally we may encounter a problem where we do not need
working hypotheses, specially in the case of exploratory or formulative researches
which do not aim at testing the hypothesis. But as a general rule, specification of
working hypotheses in another basic step of the research process in most research
problems.
4. Preparing the research design
The research problem having been formulated in clear cut terms, the
researcher will be required to prepare a research design, i.e., he will have to state
the conceptual structure within which research would be conducted. The
preparation of such a design facilitates research to be as efficient as possible
yielding maximal information. In other words, the function of research design is to
provide for the collection of relevant e vidence with minimal expenditure of effort,
time and money. But how all these can be achieved depends mainly on the research
purpose. Research purposes may be grouped into four categories, viz.,
i) Exploration,
ii) Description,
iii) Diagnosis, and
iv) Experimentation.
A flexible research design which provides opportunity for considering many
different aspects of a problem is considered appropriate if the purpose of the
research study is that of exploration. But when the purpose happens to be an
accurate description of a situation or of an association between variables, the
suitable design will be one that minimizes bias and maximizes the reliability of the
data collected and analyzed.
There are several research designs, such as, experimental and non -
experimental hypothesis testing. Experimental designs can be either informal
designs (such as before-and-after without control, after-only with control, before-
and-after with control) or formal designs (such as completely randomized design,
randomized block design, Latin square design, simple and complex factorial
designs), out of which the researcher must select one for his own project.
32

The preparation of the research design, appropriate for a particular research


problem, involves usually the consideration of the following:
a) The means of obtaining the information;
b) The availability and skills of the researcher and his staff (if any);
c) Explanation of the way in which selected means of obtaining information
will be organized and the reasoning leading to the selection;
d) The time available for research; and
e) The cost factor relating to research, i.e., the finance available for the
purpose.
5. Determining sample design
All the items under consideration in any field of inquiry constitute a 'universe'
or 'population'. A complete enumeration of all the items in the 'population' is known
as a census inquiry. It can be presumed that in such an inquiry when all the items
are covered no element of chance is left and highest accuracy is obtained. But in
practice this may not be true. Even the sli ghtest element of bias in such an inquiry
will get larger and larger as the number of observations increases. Moreover, there
is no way of checking the element of bias or its extent except through a resurvey or
use of sample checks. Besides, this type of i nquiry involves a great deal of time,
money and energy. Not only this, census inquiry is not possible in practice under
many circumstances. For instance, blood testing is done only on sample basis.
Hence, quite often we select only a few items from the uni verse for our study
purposes. The items so selected constitute what is technically called a sample.
The researcher must decide the way of selecting a sample or what is popularly
known as the sample design. In other words, a sample design is a definite plan
determined before any data are actually collected for obtaining a sample from a
given population. Thus, the plan to select 12 of a city's 200 drugstores in a certain
way constitutes a sample design. Samples can be either probability samples or non-
probability samples. With probability samples each element has a known
probability of being included in the sample but the non -probability samples do not
allow the researcher to determine this probability. Probability samples are those
based on simple random sampling, systematic sampling, stratified sampling,
cluster/area sampling whereas non-probability samples are those based on
convenience sampling, judgment sampling and quota sampling techniques.
6. Collecting the data
In dealing with any real life problem it is often found that data at hand are
inadequate, and hence, it becomes necessary to collect data that are appropriate.
There are several ways of collecting the appropriate data which differ considerably
in context of money costs, time and other resources at the disposal of the
researcher.
Primary data can be collected either through experiment or through survey. If
the researcher conducts an experiment, he observes some quantitative measurements,
33

or the data, with the help of which he examines the truth con tained in his
hypothesis. But in the case of a survey, data can be collected by any one or more of
the following ways:
By observation, through personal interview, through telephone interview,
mailing of questionnaire, through schedules:
The researcher should select one of these methods of collecting the data taking
into consideration the nature of investigation, objective and scope of the inquiry,
financial resources, available time and the desired degree of accuracy. Though he
should pay attention to all these factors but much depends upon the ability and
experience of the researcher. In this context Dr A.L. Bowley very aptly remarks that
in collection of statistical data commonsense is the chief requisite and experience
the chief teacher.
7. Execution of the project
Execution of the project is a very important step in the research process. If the
execution of the project proceeds on correct lines, the data to be collected would be
adequate and dependable. The researcher should see that the project is exec uted in
a systematic manner and in time. If the survey is to be conducted by means of
structured questionnaires, data can be readily machine -processed. In such a
situation, questions as well as the possible answers may be coded. If the data are to
be collected through interviewers, arrangements should be made for proper
selection and training of the interviewers. The training may be given with the help
of instruction manuals which explain clearly the job of the interviewers at each
step. Occasional field checks should be made to ensure that the interviewers are
doing their assigned job sincerely and efficiently. A careful watch should be kept for
unanticipated factors in order to keep the survey as much realistic as possible.
This, in other words, means that steps should be taken to ensure that the survey is
under statistical control so that the collected information is in accordance with the
pre-defined standard of accuracy. If some of the respondents do not cooperate,
some suitable methods should be designe d to tackle this problem. One method of
dealing with the non-response problem is to make a list of the non -respondents and
take a small sub-sample of them, and then with the help of experts vigorous efforts
can be made for securing response.
8. Analysis of data
After the data have been collected, the researcher turns to the task of
analyzing them. The analysis of data requires a number of closely related
operations such as establishment of categories, the application of these categories
to raw data through coding, tabulation and then drawing statistical inferences. The
unwieldy data should necessarily be condensed into a few manageable groups and
tables for further analysis. Thus, researcher should classify the raw data into some
purposeful and usable categories. Coding operation is usually done at this stage
through which the categories of data are transformed into symbols that may be
tabulated and counted. Editing is the procedure that improves the quality of the
34

data for coding. With coding the stage is re ady for tabulation. Tabulation is a part
of the technical procedure wherein the classified data are put in the form of tables.
The mechanical devices can be made use of at this juncture. A great deal of data,
specially in large inquiries, is tabulated by c omputers. Computers not only save
time but also make it possible to study large number of variables affecting a
problem simultaneously.
Analysis work after tabulation is generally based on the computation of
various percentages, coefficients, etc., by appl ying various well defined statistical
formulae. In the process of analysis, relationships or differences supporting or
conflicting with original or new hypotheses should be subjected to tests of
significance to determine with what validity data can be said to indicate any
conclusion(s). For instance, if there are two samples of weekly wages, each sample
being drawn from factories in different parts of the same city, giving two different
mean values, then our problem may be whether the two mean values are
significantly different or the difference is just a matter of chance. Through the use
of statistical tests we can establish whether such a difference is a real one or is the
result of random fluctuations. If the difference happens to be real, the inference w ill
be that the two samples come from different universes and if the difference is due to
chance, the conclusion would be that the two samples belong to the same universe.
Similarly, the technique of analysis of variance can help us in analysing whether
three or more varieties of seeds grown on certain fields yield significantly different
results or not. In brief, the researcher can analyse the collected data with the help
of various statistical measures.
9. Hypothesis testing
After analyzing the data as stated above, the researcher is in a position to test
the hypotheses, if any, he had formulated earlier. Do the facts support the
hypotheses or they happen to be contrary? This is the usual question which should
be answered while testing hypotheses. Various tests, such as Chi square test, t-
test, F-test, have been developed by statisticians for the purpose. The hypotheses
may be tested through the use of one or more of such tests, depending upon the
nature and object of research inquiry. Hypothesis-testing will result in either
accepting the hypothesis or in rejecting it. If the researcher had no hypotheses to
start with, generalizations established on the basis of data may be stated as
hypotheses to be tested by subsequent researches in times to come.
10. Generalizations and interpretation
If a hypothesis is tested and upheld several times, it may be possible for the
researcher to arrive at generalization, i.e., to build a theory. As a matter of fact, the
real value of research lies in its ability to arrive at certain generalizations. If the
researcher had no hypothesis to start with, he might seek to explain his findings on
the basis of some theory. It is known as interpretation. The process of
interpretation may quite often trigger off new questions which i n turn may lead to
further researches.
35

11. Preparation of the report or the thesis


Finally, the researcher has to prepare the report of what has been done by
him. Writing of report must be done with great care keeping in view the following:
1. The layout of the report should be as follows: (i) the preliminary pages;
(ii) the main text, and (iii) the end matter.
In its preliminary pages the report should carry title and date followed by
acknowledgements and foreword. Then there should be a table of conten ts followed
by a list of tables and list of graphs and charts, if any, given in the report.
The main text of the report should have the following parts:
a) Introduction: It should contain a clear statement of the objective of the
research and an explanation of the methodology adopted in accomplishing the
research. The scope of the study along with various limitations should as well be
stated in this part.
b) Summary of findings: After introduction there would appear a statement of
findings and recommendations in non-technical language. If the findings are
extensive, they should be summarized.
c) Main report: The main body of the report should be presented in logical
sequence and broken-down into readily identifiable sections.
d) Conclusion: Towards the end of the main text, researcher should again put
down the results of his research clearly and precisely. In fact, it is the final
summing up.
At the end of the report, appendices should be enlisted in respect of all
technical data. Bibliography, i.e., list of books, journals, reports, etc., consulted,
should also be given in the end. Index should also be given specially in a published
research report.
3.4 REVISION POINTS
1. Research process consists of a series of steps or actions required for
effectively conducting research
3.5 INTEXT QUESTIONS
1. Briefly describe the different steps involved in a research process
3.6 SUMMARY
Research involves several interrelated steps. When we talk of particular step in
research, our focus is mainly on it. At every stage of the research process, there is
likely to be some degree of error. The researcher should exercise sufficient care to
minimize the errors. This chapter has outlined the stages in business research
prawn.
3.7 TERMINAL EXERCISES
1. The analysis of data requires a number of ---------- operations
2. The hypotheses may be tested through the use of one or more of such tests,
depending upon ------------ research inquiry
36

3.8 SUPPLEMENTARY MATERIALS


1. http://www.library.illinois.edu/learn/research/researchprocess .html

2. http://as.wiley.com/Wile yCDA/WileyTitle/productCd-1118682092.html

3. http://scholarcommons.usf.edu/cgi/viewcontent.cgi?article=1002&c ontext
=oa_textbooks
3.9 ASSIGNMENTS
1. “Empirical research in India in particular creates so many problems for the
researchers”. State the problems that are usually faced by such researchers
3.10 SUGGESTED READINGS/ REFERENCE BOOKS
1. Balnaves, M. and Caputi, P. (2001) Introduction to Quantitative Research
Methods: An Investigative Approach, London: Sage.
2. Bernard, H.R. (2000) Social Research Methods: Qualitative and Quantitative
Approaches, Thousand Oaks, CA: Sage.
3. Kothari, C.R.,1985, Research Methodology- Methods and Techniques, New
Delhi, Wiley Eastern Limited.
Kumar, Ranjit, 2005, Research Methodology-A Step-by-Step Guide for
Beginners,(2nd.ed.),Singapore, Pearson Education.
3.11 LEARNING ACTIVITIES
1. Plan an descriptive research project and describe the various steps which
are followed in completing the project
3.12 KEYWORDS
Bias is a deliberate attempt either to conceal or highlight something that you
found in your research or to use deliberately a procedure or method that you know
is not appropriate but will provide information that you are looking for because you
have a vested interest in it.
Control group: The group in an experimental study which is not exposed to
the experimental intervention is called a control group. The sole purpose of the
control group is to measure the impact of extraneous and chance variables on the
dependent variable.
Editing consists of scrutinising the completed research instruments to identify
and minimise, as far as possible, errors, incompleteness, misclassification and gaps
in the information obtained from respondents.
Pie chart: The pie chart is another way of representing data graphically. As
there are 360 degrees in a circle, the full circle can be used to represent 100 per
cent or the total population. The circle or pie is divided into sections in accordance
with the magnitude of each subcategory comprising the total population. Hence
each slice of the pie is in proportion to the size of each subcategory of a frequency
distribution.

37

LESSON – 4

RESEARCH DESIGN
4.1 INTRODUCTION
The most important problem after defining the research problem is preparing
the design of the research project, which is popularly known as the 'research
design'. A research design helps to decide upon issues like what, when, where, how
much, by what means, etc., with regard to an enquiry or a research study. "A
research design is the arrangement of conditions for collection and analysis of data
in a manner that aims to combine relevance to the research purpose with economy
in procedure. In fact, the research design is the conceptual structures within which
research is conducted; it constitutes the blueprint for the collection, measurement
and analysis of data" (Selltiz, et.al. 1962). Thus, research design provi des an outline
of what the researcher is going to do in terms of framing the hypothesis, its
operational implications, and the final data analysis. Specifically, the research
design highlights decisions which include:
i) The nature of the study
ii) The purpose of the study
iii) The location where the study would be conducted
iv) The nature of data required
v) From where the required data can be collected
vi) What time period the study would cover
vii) The type of sample design that would be used
viii) The techniques of data collection that would be used
ix) The methods of data analysis that would be adopted
x) The manner in which the report would be prepared
In view of the stated research design decisions, the overall research design
may be divided into the following (Kothari 1988)
a) The sampling design that deals with the method of selecting items to be
observed for the selected study;
b) The observational design that relates to the conditions under which the
observations are to be made;
c) The statistical design that concerns with the question of how many items
are to be observed, and how the information and data gathered are to be
analysed; and
d) The operational design that deals with the techniques by which the
procedures specified in the sampling, statistical and observational designs
can be carried out.
38

4.2 OBJECTIVES
After studying this lesson, you will be able to understand:
 Research design
 Features of research design
 Characteristics of good research design
 Types of research design
4.3 CONTENTS
4.3.1 Features of research design
4.3.2 Characteristics of research design
4.3.3 Importance of research design
4.3.4 Types
4.3.1 Features of research design
The important features of research design may be outlined as follows:
i) It constitutes a plan that identifies the types and sources of information
required for the research problem;
ii) It constitutes a strategy that specifies the methods of data collection and
analysis which would be adopted; and
iii) It also specifies the time period of research and monetary budget involved
in conducting the study, which comprise the two major constraints of
undertaking any research.
4.3.2 Characteristics of a good research design
A good research design often possesses the qualities such as being flexible,
suitable, efficient, economical, and so on. Generally, a research design which
minimizes bias and maximizes the reliability of the data collected and analyzed is
considered a good design (Kothari 1988).
A research design which involves the smallest experimental error is said to be
the best design for investigation. Further, a researc h design that yields maximum
information and provides an opportunity of viewing the various dimensions of a
research problem is considered to be the most appropriate and efficient design.
Thus, the question of a good design relates to the purpose or objective and nature
of the research problem studied. While a research design may be good, it may not
be equally suitable to all studies. In other words, it may be lacking in one aspect or
the other in the case of some other research problems. Therefore, no sing le research
design can be applied to all types of research problems.
A research design suitable for a specific research problem would usually
involve the following considerations:
i) The methods of gathering the information;
ii) The skills and availability of the researcher and his/her staff, if any;
iii) The objectives of the research problem being studied;
39

iv) The nature of the research problem being studied; and


v) The available monetary fund’s and time duration for the research work.
4.3.3 Importance of research design
The need for a research design arises out of the fact that it facilitates the
smooth conduct of the various stages of research. It contributes to making research
as efficient as possible, thus yielding the maximum information with minimum
effort, time and expenditure. A research design helps to plan in advance of the
methods to be employed for collecting the relevant data and the techniques to be
adopted for their analysis, so as to pursue the objectives of the research in the best
possible manner, given the available staff, time and money. Hence, the research
design should be prepared with utmost care, so as to avoid any error that may
disturb the entire project. Thus, research design plays a crucial role in attaining the
reliability of the results obtained, which forms the strong foundation of the entire
process of the research work.
Despite its significance, the purpose of a well -planned design is not realized at
times. This is because it is not given the importance that this problem deserves. As
a consequence, many researchers are not able to achieve the purpose for which the
research designs are formulated, due to which they end up arriving at misleading
conclusions. Therefore, faulty designing of the research project tends to render the
research exercise meaningless. This makes it imperative that an efficient and
suitable research design must be planned before commencing the process of
research. The research design helps the researcher to organize his/her ideas in a
proper form, which would facilitate him/her to identify the inadequacies and faults
in them. The research design may also be discussed with other experts for their
comments and critical evaluation, without which it would be difficult for any critic
to provide a comprehensive review and comment on the proposed study.
4.3.4 Types of research design
1. Exploratory research design
The exploratory research design is known as formulative research design. The
main objective of using such a research design is for formulating a research
problem for an in-depth or more precise investigation, or for developing a working
hypothesis from an operational aspect. The major purpose of such studies is the
discovery of ideas and insights. Therefore, such a research design suitable for such
a study should be flexible enough to provide opportunity for considering different
dimensions of the problem under study. The in -built flexibility in research design is
required as the initial research problem would be transformed into a more precise
one in the exploratory study, which in turn may necessitate changes in the
research procedure for collecting relevant data. Usually, the following three
methods are considered in the context of a research design for such studies. They
are (a) a survey of related literature; (b) experi ence survey; and (c) analysis of'
insight-stimulating' instances.
40

2. Descriptive and diagnostic research design


A descriptive research design is concerned with describing the characteristics
of a particular individual, or a group. Meanwhile, a diagnostic research design
determines the frequency with which a variable occurs or its relationship with
another variable. In other words, the study analyzing whether a certain variable is
associated with another comprises a diagnostic research study. On the other hand,
a study that is concerned with specific predictions or with the narration of facts
and characteristics relating to an individual, group or situation, are instances of
descriptive research studies. Generally, most of the social research design falls
under this category. As a research design, both the descriptive and diagnostic
studies share common requirements, and hence they may group together. However,
the procedure to be used must be planned carefully, and so the research design
should also be planned carefully. The research design must also make appropriate
provision for protection against bias and thus maximize reliability, with due regard
to the completion of the research study in as economical manner as possible. The
research design in such studies should be rigid and not flexible. Besides, it must
also focus attention on the following: (a) formulation of the objectives of the study,
 Proper Designing of the methods of data collection ,
 Sample selection,
 Data collection,
 Processing and analysis of the collected data, and
 Reporting the findings.
2. Experimental research design
Experimental studies are those that are intended to test cause -effect
relationships (hypotheses) in a tightly controlled setting by separating the cause
from the effect in time, administering the cause to one group of subjects (the
"treatment group") but not to another group ("control group"), and observing how
the mean effects vary between subjects in these two groups. For instance, if we
design a laboratory experiment to test the efficacy of a new drug in treating a
certain ailment, we can get a random sample of people afflicted with that ailment,
randomly assign them to one of two groups (treatment and control groups),
administer the drug to subjects in the treatment group, bu t only give a placebo
(e.g., a sugar pill with no medicinal value). More complex designs may include
multiple treatment groups, such as low versus high dosage of the drug, multiple
treatments, such as combining drug administration with dietary intervention s. In a
true experimental design, subjects must be randomly assigned between each
group. If random assignment is not followed, then the design becomes quasi-
experimental. Experiments can be conducted in an artificial or laboratory setting
such as at a university (laboratory experiments) or in field settings such as in an
organization where the phenomenon of interest is actually occurring (field
experiments). Laboratory experiments allow the researcher to isolate the variables
of interest and control for extraneous variables, which may not be possible in field
41

experiments. Hence, inferences drawn from laboratory experiments tend to be


stronger in internal validity, but those from field experiments tend to be stronger in
external validity. Experimental data is analyzed using quantitative statistical
techniques. The primary strength of the experimental design is its strong internal
validity due to its ability to isolate, control, and intensively examine a small number
of variables, while its primary weakness is limited external generalizability since
real life is often more complex (i.e., involve more extraneous variables) than
contrived lab settings. Furthermore, if the research does not identify ex ante
relevant extraneous variables and control for such variabl es, such lack of controls
may hurt internal validity and may lead to spurious correlations.
4.4 REVISION POINTS
1. The most important problem after defining the research problem is
preparing the design of the research project, which is popularly known as
the 'research design
2. Research design provides an outline of what the researcher is going to do in
terms of framing the hypothesis, its operational implications, and the final
data analysis.
3. The need for a research design arises out of the fact that it facilitates the
smooth conduct of the various stages of research.
4. Research design plays a crucial role in attaining the reliability of the results
obtained, which forms the strong foundation of the entire process of the
research work
4.5 INTEXT QUESTIONS
1. Outline the features of research design.
2. Discuss the features of a good research design.
3. Describe the different types of research design.
4. Explain the significance of research design.
4.6 SUMMARY
This chapter focuses on research design ad its types. A few definitio ns of
research designs and it is pointed out that there are three broad categories of
research design – Exploratory, descriptive and casual. This chapter also discusses
Features of research design and Characteristics of good research design
4.7 TERMINAL EXERCISES
1. ........................... research is used to seek insights into general nature of
the problem.
2. Research design helps to plan in advance the methods and techniques to be
used for collecting and ………………..data.
3. The major emphasis in exploratory research is on converting ......................,
vague problem statements into ............... and .............................. sub-problem
statements.
4. Exploratory research is ........................ and very .......................
42

4.8 SUPPLEM ENTARY MATERIALS


1. Denzin, Norman, The Research Act, Chicago: Aldine, 1973
2. Fiebleman, J.K., Scientific Method, Netherlands: Martinus Nijhoff, The
Hague, 1972
3. Giles, G.B., Marketing, 2nd ed., London: Macdonald & Evans Ltd., 1974.
4. https://www.nyu.edu/classes/bkg/methods/005847ch1.pdf
4.9 ASSIGNMENTS
1. What is research design? Discuss the basis of stratification to be employed
in sampling public opinion on inflation.
2. “Research design in exploratory studies must be flexible but in descriptive
studies, it must minimize bias and maximize reliability.” Discuss.
4.10 SUGGESTED READINGS/ REFERENCE BOOKS
1. Denscombe, M. (1998) The Good Research Guide: for small -scale social
research projects, Buckingham: Open University Press.
2. Eichler, M. (1988) Non-sexist Research Methods, Boston: Allen & Unwin.
3. Fowler, F. (2001) Survey Research Methods, 3rd edition, Thousand Oaks,
CA: Sage.
4.11 LEARNING ACTIVITIES
1. Visit the Web Pages of any three marketing research firms and identify what
type of research design has been implemented recen tly by these firms?
4.8 KEYWORDS
Research design: A research design is a procedural plan that is adopted by
the researcher to answer questions validly, objectively, accurately and
economically. A research design therefore answers questions that would determine
the path you are proposing to take for your research journey. Through a research
design you decide for yourself and communicate to others your decisions regarding
what study design you propose to use, how you are going to collect information
from your respondents, how you are going to select your respondents, how the
information you are going to collect is to be analysed and how you are going to
communicate your findings.

43

LESSON – 5

SOURCES OF DATA
5.1 INTRODUCTION – MEANING, TYPES AND IMPORTANCE
“Data are facts, figures and other relevant materials past and present serving
as bases for study and analysis”.
Meaning of Data
The search for answers to research questions calls collection of Data. “Data are
facts, figures and other relevant materials, past and present, serving as bases for
study and analysis”.
Types of Data
The Data needed for social science Research may be broadly classified into:
a) Data pertaining to human beings
b) Data relating to Organisations
c) Data pertaining to territorial area.
A) Personal Data (relating to Human beings) are of two types.
1. Demographic and socio-economic characteristics of individuals. Like name,
sex, race, social class, relation, education, occupation, income etc.
2. Behavioural Variables: Attitudes, opinion knowledge, practice, intensions
etc.
B) Organistion Data:- Consist of data relating to an organizations, origin
ownership, function, performance etc.
C) Territorial Data: - are related to geo-physical characteristic, population,
infrastructure etc of divisions like villages, cities, taluks, distinct, state etc.
Importance of Data
The data serve as the bases or raw materials for analysis without Data no
specific inferences can be drawn on our study. The reliability of data determines the
quality of research. Data form the basis of testing hypothesis data provides the
facts and figures for constructing measuring scale. The scientific process of
research can be carried out only through accurate data.
5.2 OBJECTIVES
After studying this lesson, you will be able to understand:
 Sources of data
 Types of data
 Various data collection methods
5.3 CONTENTS
5.3.1 Sources of Data
5.3.2 Data Collection Methods
44

5.3.1 Sources of Data


The sources of Data may be classified into a) primary sources b) Secondary
sources.
a) Primary Sources
Primary sources are original sources from which the researcher directly
collects data that have not been previously collected. Primary Data are first –hand
information collected through various methods such as observation, interview etc.
Some examples of sources of primary data are individuals, focus groups, pan-
els of respondents specifically set up by the researcher and from whom opinions
may be sought on specific issues from time to time, or some unobtrusive sources
such as a trash can. The Internet could also serve as a primary data source when
questionnaires are administered over it.
b) Secondary Sources
These are sources containing data which have been collected and compiled for
another purpose. The secondary sources consist of readily available and compiled
statistical statements and reports. Secondary sources consist of not only published
but also unpublished records. They consist of Data over which a researche r has no
original control.
Some examples of secondary sources are, company records or archives,
government publications, industry analyses offered by the media, web sites, the
Internet, and so on. In some cases, the environment or particular settings and
events may themselves be sources of data, as for example, studying the layout of a
plant.
Primary Sources of Data
Focus Groups
Focus groups consist typically of 8 to 10 members with a moderator leading
the discussions for about 2 hours on a particular topic, concept, or product. Mem-
bers are generally chosen on the basis of their expertise in the topic on which
information is sought. For example, computer specialists may be selected to form a
focus group to discuss matters related to computers and computin g, and women
with children may compose the focus group to identify how organizations can help
working mothers.
The focus sessions are aimed at obtaining respondents' impressions, interpre -
tations, and opinions, as the members talk about the event, concept, product, or
service. The moderator plays a vital role in steering the discussions in a manner
that would draw out the information sought, and keeping the members on track.
Focus group discussions on a specific topic at a particular location and at a
specified time provide the opportunity for a flexible, free -flowing format for the
members. The unstructured and spontaneous responses are expected to reflect the
genuine opinions, ideas, and feelings of the members about the topic under
discussion. Focus groups are relatively inexpensive and can provide fairly
dependable data within a short time frame.
45

Panels
Panels, like focus groups, are another source of primary information for
research purposes. Whereas focus groups meet for a one -time group session,
panels (of members) meet more than once. In cases where the effects of certain
interventions or changes are to be studied over a period of time, panel studies are
very useful. Individuals are randomly chosen to serve as panel members for a
research study. For instance, if the effects of a proposed advertisement for a certain
brand of coffee are to be assessed quickly, the panel members can be exposed to
the advertisement and their intentions of purchasing that brand assessed. This can
be taken as the response that could be expected of consumers if, in fact, they had
been exposed to the advertisement. A few months later, the product manager might
think of introducing a change in the flavor of the same product and explore its
effects on this panel. Thus, a continuing set of "experts" serves as the sample base
or the sounding board for assessing the effects of change. Such expert members
compose the panel, and research that uses them is called a panel study.
Unobtrusive Measures
Trace measures, or unobtrusive measures as they are also called, originate
from a primary source that does not involve people. One example is the wear and
tear of journals in a university library, which offers a good indication of their
popularity, frequency of use, or both. The number of di fferent brands of soft drink
cans found in trash bags also provides a measure of their consumption levels.
Signatures on checks exposed to ultraviolet rays could indicate the extent of forgery
and frauds; actuarial records are good sources for collecting d ata on the births,
marriages, and deaths in a community; company records disclose a lot of personal
information about employees, the level of company efficiency, and other data as
well. Thus these unobtrusive sources of data and their use are also importan t in
research.
Secondary Sources
Secondary data are indispensable for most organizational research. Secondary
data refer to information gathered by someone other than the researcher
conducting the current study. Such data can be internal or external to the
organization and accessed through the Internet or perusal of recorded or published
information.
Secondary data can be used, among other things, for forecasting sales by con -
structing models based on past sales figures, and through extrapolation. There are
several sources of secondary data, including books and periodicals, government
publications of economic indicators, census data, Statistical Abstracts, data bases,
the media, annual reports of companies, etc. Case studies, and other archival
records—sources of secondary data—provide a lot of information for research and
problem solving. Such data are, as we have seen, mostly qualitative in nature. Also
included in secondary sources are schedules maintained for or by key personnel in
organizations, the desk calendar of executives, and speeches delivered by them.
Much of such internal data, though, could be proprietary and not accessible to all.
46

Financial databases readily available for research are also secondary data
sources. The Compustat Database contains information on thousands of companies
organized by industry, and information on global companies is also avail able
through Compustat.
The advantage of seeking secondary data sources is savings in time and costs
of acquiring information. However, secondary data as the sole source of information
has the drawback of becoming obsolete, and not meeting the specific needs of the
particular situation or setting. Hence, it is important to refer to sources that offer
current and up-to-date information.
5.3.2 Data collection methods
Having examined the various sources of data, let us now look into the data
collection methods.
The important data collection methods are:
1. Observation
2. Interviewing
3. Questionnaire
4. Schedules
1. Observational method
Whereas interviews and questionnaires elicit responses from the subjects, it is
possible to gather data without asking questions of respondents. People can be
observed in their natural work environment or in the lab setting, and their activities
and behaviors or other items of interest can be noted and recorded.
Apart from the activities performed by the individuals under study, their move -
ments, work habits, the statements made and meetings conducted by them, their
facial expressions of joy, anger, and other emotions, and body language can be
observed. Other environmental factors such as layout, work-flow patterns, the
closeness of the seating arrangement, and the like, can also be noted. Children can
be observed as to their interests and attention span with various stimuli, such as
their involvement with different toys. Such observation would help toy manufac -
turers, child educators, day-care administrators, and others deeply involved in or
responsible for children's development, to design and model ideas based on c hil-
dren's interests, which are more easily observed than traced in any other manner.
The researcher can play one of two roles while gathering field observational data—
that of a nonparticipant-observer or participant-observer.
Advantages and Disadvantages of Observational Studies
There are some specific advantages and disadvantages to gathering data
through observation as listed below.
Advantages of Observational Studies
The following are among the advantages of observational studies.
47

1. The data obtained through observation of events as they normally occur are
generally more reliable and free from respondent bias.
2. In observational studies, it is easier to note the effects of environmental
influences on specific outcomes. For example, the weather (hot, c old, rainy), the day
of the week (midweek as opposed to Monday or Friday), and such other factors that
might have a bearing on, for example, the sales of a product, traffic patterns,
absenteeism, and the like, can be noted and meaningful patterns might emerge
from this type of data.
3. It is easier to observe certain groups of individuals—for example, very young
children and extremely busy executives—from whom it may be otherwise difficult to
obtain information.
Drawbacks of Observational Studies
The following drawbacks of observational studies have also to be noted.
1. It is necessary for the observer to be physically present (unless a camera or
another mechanical system can capture the events of interest), often for prolonged
periods of time.
2. This method of collecting data is not only slow, but also tedious and
expensive.
3. Because of the long periods for which subjects are observed, observer
fatigue could easily set in, which might bias the recorded data.
4. Though moods, feelings, and attitudes can be guessed by observing facial
expressions and other nonverbal behaviors, the cognitive thought processes of
individuals cannot be captured.
5. Observers have to be trained in what and how to observe, and ways to avoid
observer bias.
2. Interview
One method of collecting data is to interview respondents to obtain
information on the issues of interest. Interviews could be unstructured or
structured, and conducted either face to face or by telephone or online. The
unstructured and structured interviews are discussed first. Some important factors
to be borne in mind while interviewing will then be detailed; the advan tages and
disadvantages of face-to-face interviewing and telephone interviews enumerated
thereafter; and finally, computer-assisted interviews described.
Unstructured and Structured Interviews
Unstructured Interviews
Unstructured interviews are so labeled because the interviewer does not enter
the interview setting with a planned sequence of questions to be asked of the
respondent. The objective of the unstructured interview is to bring some pre -
liminary issues to the surface so that the researcher can determine what variables
need further in-depth investigation. In the discussion of the "Broad Problem Area,"
we saw several situations where the man ager might entertain a vague idea of
48

certain changes taking place in the situation without knowing what exactly they
are. Such situations call for unstructured interviews with the people concerned. In
order to understand the situation in its totality, the researcher will interview
employees at several levels. In the initial stages, only broad, open -ended questions
would be asked, and the replies to them would inform the researcher of the
perceptions of the individuals. The type and nature of the questions asked of the
individuals might vary according to the job level and type of work done by them.
For instance, top and middle-level managers might be asked more direct questions
about their perceptions of the problem and the situation. Employees at lower leve ls
may have to be approached differently.
Clerical and other employees at lower hierarchical levels may be asked broad,
open-ended questions about their jobs and the work environment during the
unstructured interviews. Supervisors may be asked broad questi ons relating to
their department, the employees under their supervision, and the organization.
Structured Interviews
Structured interviews are those conducted when it is known at the outset what
information is needed. The interviewer has a list of predetermined questions to be
asked of the respondents either personally, through the telephone, or through the
medium of a PC. The questions are likely to focus on factors that had surfaced
during the unstructured interviews and are considered relevant to the problem. As
the respondents express their views, the researcher would note them down. The
same questions will be asked of everybody in the same manner. Sometimes,
however, based on the exigencies of the situation, the experienced researcher might
take a lead from a respondent's answer and ask other relevant questions not on the
interview protocol. Through this process, new factors might be identified, resulting
in a deeper understanding. However, to be able to recognize a probable response,
the interviewer must comprehend the purpose and goal of each question. This is
particularly important when a team of trained interviewers conducts the survey.
Face-to-Face and Telephone Interviews
Interviews can be conducted either face to face or over the telephone. They
could also be computer-assisted. Although most unstructured interviews in orga-
nizational research are conducted face to face, structured interviews could be either
face to face or through the medium of the telephone, depending on the level of
complexity of the issues involved, the likely duration of the interview, the
convenience of both parties, and the geographical area covered by the survey.
Telephone interviews are best suited when information from a large number of
respondents spread over a wide geographic area is to be obtained quickly, and the
likely duration of each interview is, say, 10 minutes or less. Many market surveys,
for instance, are conducted through structured telephone interviews. In addition,
computer-assisted telephone interviews (CATI) are also possible, and easy to
manage.
49

Face-to-face interviews and telephone interviews have other advantages and


disadvantages. These will now be briefly discussed.
Face-to-Face Interviews
Advantages: The main advantage of face-to-face or direct interviews is that
the researcher can adapt the questions as necessary, clarify doubts, and ensure
that the responses are properly understood, by repeating or rephrasing the
questions. The researcher can also pick up nonverbal cues from the respondent.
Any discomfort, stress, or problems that the respondent experiences can be
detected through frowns, nervous tapping, and other body language unconsciously
exhibited by her. This would be impossible to detect in a telephone interview.
Disadvantages: The main disadvantages of face-to-face interviews are the
geographical limitations they may impose on the surveys and the vast resources
needed if such surveys need to be done nationally or internationally. The costs of
training interviewers to minimize interviewer biases (e.g., differences in questioning
methods, interpretation of responses) are also high. Another drawback is that
respondents might feel uneasy about the anonymity of their responses when they
interact face to face with the interviewer.
Telephone Interviews
Advantages: The main advantage of telephone interviewing, from the
researcher's point of view, is that a number of different people can be reached (if
need be, across the country or even internationally) in a relatively short period of
time. From the respondents' standpoint it would eliminate any discomfort that
some of them might feel in facing the interviewer. It is also possible that most of
them would feel less uncomfortable disclosing personal information over the phone
than face to face.
Disadvantages: A main disadvantage of telephone interviewing is that the
respondent could unilaterally terminate the interview without warning or expla -
nation, by hanging up the phone. Caller ID might further aggravate the situation.
This is understandable, given the numerous telemarketing calls people are bom-
barded with on a daily basis. To minimize this type of a nonresponsive problem, it
would be advisable to call the interviewee ahead of time to request participation in
the survey, giving an approximate idea of how long the interview would last, and
setting up a mutually convenient time. Interviewees usually tend to appreciate this
courtesy and are more likely to cooperate. It is a good policy not to prolong the
interview beyond the time originally stated. A s mentioned earlier, another
disadvantage of the telephone interview is that the researcher will not be able to see
the respondent to read the nonverbal communication.
Interviewing is a useful data collection method, especially during the
exploratory stages of research. Where a large number of interviews are conducted
with a number of different interviewers, it is important to train the interviewers with
care in order to minimize interviewer biases manifested in such ways as voice
50

inflections, differences in wordings, and interpretation. Good training decreases


interviewer biases.
3. Questionnaire
The questionnaire may be regarded as a form of interview on paper. Procedure
may be regarded as a form of interview on paper. Procedure for the construction of
a questionnaire follows a pattern similar to that of the interview schedule. However,
because the questionnaire is impersonal it is all the more important to take care
over its construction. Since there is no interviewer to explain ambiguities or to
check misunderstandings, the questionnaire must be especially clear in its
working. The variety of possible answers to each question must be anticipated more
fully than for an interview. The questionnaire is probably the most used and most
abused of the data gathering devices.
Definition
A questionnaire is a form which is prepared and distributed for the purpose of
securing responses. Generally these questions are factual and designed for securing
information about certain conditions or practices, of which recipient is presumed to
have knowledge. Goode and Hatt have defined questionnaire as a device for
securing answers to questions by using a form which the respondent fills himself.
According to Barr, Davis and Johnson "A questionnaire is a systematic
compilation of questions that are submitted to a sampling of population from which
information is desired."
As the term generally used in educational researches, "the questionnaire
consists of a sense of questions or statements to which individuals are asked to
respond the questions frequently asked for facts or the opinions, attitudes or
preferences of the respondents."
Goode and Hatt have given an illustrated definition of questionnaire.
"In general words questionnaire refers to a device for securing answers to
questions by using a form which the respondent fills in himself"
In questionnaire forms are used and the respondents fill in themselves, thus
questionnaire place heavy reliance on the validity of the verbal reports. This
instrument is widely used by researchers for a number of reasons.
Significance of Questionnaire
Beginners are more commonly tempted to this tool, because they imagine that
planning and using a questionnaire is easier than the use of other tools. It is also
considered to be the most flexible of tools and possesses a unique advantage over
others in collecting both qualitative and quantitative information . Critics speak of it
as the lazy man's way of gaining information, because it is comparatively easy to
plan and administer a questionnaire. "As a matter of fact, preparation of a good
questionnaire takes a great deal of time, ingenuity and hard work."
Common Faults
Questionnaire prepared by novices suffer from such errors as:
51

i) Too lengthy they contain a large number of questions requiring lengthy


answers.
ii) Vague: Items are imperfectly worded and improperly arranged.
iii) The proforma itself is poorly conceived and badly organized.
iv) The subjects touched by the items of the questionnaire are trivial
importance.
We should also beware that this device suffers from certain inherent
drawbacks. These according to T.K. Kelly are:
i) It is not possible to justify the selection of the sample of respondents with
demonstrable fairness.
ii) Benefits of the study seldom acquire to the respondents.
iii) It requires a large amount of investment on the part of the investigator.
iv) The respondent may not possess the expertise required for giving the
answer.
v) Educational administrators are allergic towards questionnaire since they
receive a large number of them, day in and day out. The arrival of a new
questionnaire irritates them as they believe that it constitutes an
encroachment upon their busy schedule of work.
vi) It is also wrong to imagine that questionnaire technique is a quick and
easy method of investigation. On the contrary, the responses are rather
slow and disappointing because of their incompleteness, indefiniteness and
hostile attitude of the recipients, who feel bored by the quantum of the
questionnaires that they receive.
vii) Their casual responses furnish a very flimsily basis of valid ge neralization.

Design of the Questionnaire


In order to gain acceptance for his questionnaire, the researcher should design
an appealing format. Many unattractive questionnaires end up in a wastebasket
rather in the hands of the sender. To improve the attractiveness of the instrument,
choose a title that is clear, concise, and descriptive of the research project, and use
well typed or printed questions that are properly spaced and easy to read. It is
generally advisable to group questions of a similar nature together.
Type of Questionnaire Items
The two commonly used types of questionnaire items are the unrestricted, or
open form items, and the restricted, or close form items. Each type has its
advantages and disadvantages, so the researcher must decide which is more likely
to yield the data needed in a particular research project.
The open formed item is also referred to as the "Open end", "Short-answer", or
"Free-response" item because offer the question there is a space provided in which
52

the respondent is asked to write his answer. This type of item permits explanation,
but responses can be difficult to summarize and tabulate. The responses also may
be too brief, or the respondent may have omitted important information.
The open form item is of most value when unrestricted, in depth responses are
needed. An example of this type of questionnaire item follows:
In your opinion, what might your high school have done to prepare you more
adequately for college?
The closed form item sometimes is referred to as the "restricted" or
"structured" type. It consists of a question or a statement to which a person
responds by selecting one or more choices, such as "Yes" or "No". In one variation of
this type the respondent may be asked to underline award from the two or more
alternatives. Another variation requires the ranking of choices.
The close form item facilities the tabulation and analysis of data. It also
improves the reliability and consistency of the data. One limitation of this kind of
item is that the respondent does not have the opportunity to explain why he has
given certain responses, and this may be important in some kinds of research
studies. The close form item also limits the scope and depth of responses, so its use
in measuring attitudes, feelings and certain aspects of behavior may be limited. It is
possible also that the answers from which the respondents must choose are not the
proper ones for all of the individuals who are given the questionnaire. An example
of a closed form questionnaire item follows: If group tests are used in your school,
by whom are they administered? (a) Administrators (b) Counselors,
(c) Psychologists, (d) Psychometricians,
(e) Teachers, and (e) Others
Elements of Questioning Method
In using the questioning method there are six ele ments to be considered:
i) The content of the questioning; which includes both why the researcher is
asking the questions and why the respondent believes they are being asked,
since these are not necessarily the same;
ii) Content of the questioning;
iii) The question which can be defined as the verbal stimulus to which the
respondent will be exposed;
iv) The provision for answering which refers to the response the respondent will
be.
v) The way in which the response will be recorded, which refers to variation like
writing or speaking or recording;
vi) The nature of the researcher respondent interaction, which considers
whether researcher respondent interaction , which considers whether
researcher and respondent actually meets face to face or whether they
communicate impersonally.
53

Preparing and Administering the Questionnaire


Get all the help that you can in planning and constructing your questionnaire.
Study other questionnaire and submit your items for criticism to other member of
your class or your faulty, especially to those who have had experience in
questionnaire construction.
Try out your questionnaire on a few friends and questionnaires when you do
this personally, you may find that a number of your items are ambiguous.
Choose respondents carefully. It is an important that qu estionnaire be sent
only to those who possess the desired information or those who are likely to be
sufficiently interested to respond conscientiously and objectively. A preliminary
card, asking whether or not the individual would be willing to participate in the
proposed study, is recommended by research authorities. In a study on
questionnaire returns see (1) discovered that a greater proportions of returns was
obtained when the original request was sent to the administrative head of an
organization, rather than directly to the person who had the desired information. It
is possible then when a superior officer turns over a questionnaire to a staff
member to fill out, there is implied some feeling of obligation.
If schedules or questionnaires are planned for use in a public school, asking
for the responses of teachers or pupils, it is essential that approval of the project be
secured from the principal, who may, then, wish to secure approval from the
superintendents of schools.
If the desired information is delicate or intimate in nature, consider the
possibility of proving for anonymous responses. The anonymous instrument is most
likely to produce objective responses. There are occasions, however, for purposes of
classification, when the identity of the respondent is necessary. If a signature is
needed it is essential to convince the respondent that his responses will be held in
strict confidence, and that his answer will in no way jeopardize the status and
security of his position.
Try to get the aid of sponsorship. Recipients are more likely to answer if a
person, organization, or institution for prestige has endorsed the project.
Be sure to include a courteous, carefully constructed cover letter to explain the
purpose of the study. The letter should promise some sort of Inducement to the
respondent for compliance with the request.
The cover letter should assure the respondent that delicate information will be
held in strict confidence. The explanation sponsorship might well be mentioned, of
course, a stamped addressed return envelope should be included. To omit this
courtesy would be practically to guarantee that many of the questionnaires would
go into the waste paper basket. It has been suggested that two copies of the
questionnaires be sent, one to be returned when completed and the other for
respondent's own life.
54

Recipients are often slow to return completed questionnaire to increase the


number of returns, a vigorous follow up procedure may be necessary. A courteous
postcard reminding the recipient that the completed questionnaire has not been
received will bring in some additional responses. This reminder will be effective with
those who have just put in the follow up process may involve a personal letter or
reminder. In extreme cases a telegram, phone call, or personal visit may bring
additional responses.
It is difficult to estimate, in the abstract, what percentage of questionnaire
responses is to be considered adequate or satisfactory. The importance of the
project, the quality of questionnaire, the c ase used in selecting recipients, the time
of year, and many other factors may be significant in determining the proportion of
responses. Needless to say, the smaller the percentage of responses, the smaller the
degree of confidence one may place in the adequacy of the data collected. However,
objectivity of reporting requires that the proportion of responses received should
always be included in the research report.
Choosing the Questions
Stage in a question is to define the problem precisely. It is desirable also to
define at an early stage the population to whom the question is to be directed and
to decide the nature of the sample to be drawn, as this may influence the drafting
of the question. The definition of the problem should set out one by one the aspects
of the problem to be explored and stage, II, follows on from this with construction of
questions or items to deal with each aspect in turn. The question should be
compiled with definite hypotheses or theories in mind. Each question should
contribute clear information on a specific aspect to be explored or else the question
should be discarded ruthlessly.
If the investigation is on a small scale and the investigator himself will classify
all the responses, it may not be necessary to put all questions into a multiple
choice pattern. Multiple choice can be tiresome, especially if the alternatives offered
for choice do not express adequately the response a subject wishes to express
opinion questions if they are included must not be too general or they wil l be
unclassifiable and should be restricted to fairly straightforward topics where
answers can easily be coded subsequently.
It may be necessary to choose questions and to specify possible answers in
such a way as to facilitate coding. The art of drafting a question consists of being
able to do this without forcing your respondents into an uneasy choice a difficulty
which they can too easily resolve by giving up altogether.
The questionnaire should start with simple factual questions, so that the
person completing it gets off to a good start. Complex or awkward topics should
come towards the end. An open ended- general question at the end will allow
expression of points which the responder thinks important, though they are not
covered by the questionnaire.
55

A questionnaire need not to be restricted to questions attitude scales ratings


and check lists may be included, providing they are brief and straight forward and
the instructions are kept simple. Similarity and brevity are cardinal virtues .
The distinction between a leading question and a neutral question is
sometimes difficult to decide. For example, which of following is a neutral wording?
1. Are you in favour of abolishing corporal punishment?
2. Should corporal punishment be abolished?
3. Should corporal punishment be retained etc.?
The questionnaire is not a scientific instrument. It is a cheap, easy and rapid
method of obtaining information and non - information one never knows which he
points the weaknesses of both questionnaire and interview-the tested-validity of the
responses. Questionnaires show what people say, not what do or are. Any form of
questionnaire implies a certain frame of reference and thereby influences the
answer given.
Try out of the Questionnaire
It is very helpful to obtain criticism of qualified persons about the
questionnaire before the final form is got printed and mailed out. The investigator
may request some of his friends to look over the items critically with the idea of
discovering possible misinterpretations. A few copies may then be got typed and
sent or administered personally to a random sample of the respondents. Their
responses must be carefully examined to see if they suggest any modifications of
the questionnaire before it is used on a large sample or for study purpose . This
presetting of all the measuring tools is essential to ascertain the workability of the
tool. Different persons will differently interpret the same lines. A preliminary form of
the questionnaire used in this way will require revision and improvement by
eliminating some of the items. It may also suggest addition of some more items. The
result of the try out should also be analyzed in a preliminary way to determine
whether they lead to certain conclusions for the significant purpose of the study.
The presetting of a questionnaire indicates the following things:
i) Relative effectiveness and costs of alternative questions instruments and
procedures.
ii) Acceptability and intelligibility of the questions from the respondent’s point
of view.
iii) Completeness of questions for correct coding and interpretation.
iv) Response rates and estimates of mean and variance.
Improving Questionnaire Items
The questionnaire maker must depend on words alone. It is apparent that he
cannot be too careful in phrasing questions to insure the ir clarity of purpose. While
there are no certain ways of producing full proof questions, there are principles that
might be employed to make items more precise. A few are suggested with the hope
56

that students constructing questionnaires will become critic al of their first efforts
and strive to make each question as clear as possible.
1. Define or qualify terms that could easily be a misinterpreted. "What is the
value of your horse? The meaning of the term value is not clear. These
values may differ considerably. It is essential to frame specific questions
such as, "what is the present market value of your horse?"
2. Be careful in using descriptive adjectives and adverbs that have no agreed -
upon-meaning. This fault is frequently found in rating scales as well as
questionnaire.
3. Beware of double negatives (1) Federal aid should not granted for those
states in which education' is not equal regardless of race, creed, or colour.
4. Be careful of inadequate alternatives (1) Married? Yes/No.
5. Avoid the double barreled question. Example: Do you believe that gifted
students should be placed in separate groups for instructional purposes
and assigned to special schools?
6. Underline the word if you wish to indicate special emphasis. Example
7. Should all schools offer modern foreign language?
8. When asking for ratings or comparisons a point of reference is necessary.
Example: How would you rate this student teacher's class room teaching?
Superior Average-Below Average .
Phrase questions so that they are appropriate for all respondents. Desi gn
questions that will give a complete response. Such questions which as answer "Yes
or no" would not reveal much information about the reading habits of the
respondents.
There must be a provision for the systematic quantification of response. One
type of question that asks respondents to check a number of items from a list is
difficult to summarize especially all respondents do not check the same number.
The students should bear in mind that these suggestions are used in constructing
questionnaire items.
Scope of Questionnaire
In descriptive studies where the sources are varied and widely scattered, the
questionnaire is a major instrument for gathering data. It is very handy in cases,
where one cannot conveniently see personally all the people from whom the
responses are required. This technique finds favour in determination of present
status in certain aspects of education-current practice in schools, financial aspects
service conditions of teachers, etc. It can be used over extensive range of territory -
national and international. "As research techniques are becoming more and more
refined day after day, it is hoped that this wayward child of the science of education
will soon curb its unruly disposition and also mend its unseemly ways".
57

Characteristics of a Good Questionnaire


The following are the characteristics of a good questionnaire
1. The covering letter of the questionnaire is drafted in a befriending tone and
indicates its importance to the respondents.
2. The questionnaire contains directions which are cle ar and complete.
Important items are clearly defined and each question deals with a single
idea defined in unambiguous terms.
3. It is reasonable short, through comprehensive enough to secure all relevant
information.
4. It does not seek information which may be obtainable from other sources
such as school records and University results.
5. It is attractive in appearance, neatly arranged, clearly duplicated and free
from typographical errors.
6. It avoids annoying or embracing questions, which arouse hostility in the
respondent.
7. Items are arranged in categories which ensure easy and accurate
responses.
8. Questions do not contain leading suggestions for the respondents and are
objective in nature.
9. They are arranged in good order. Simple and general questions should
precede the specific and complex ones. Questions that create favourable
atmosphere should precede those that are personal and touch delicate
points.
10. They are so worded, that it is easy to tabulate and interpret the responses.
It is always advisable to base them upon a preconceived tabulation sheet.
Suggestions for Construction of a Good Questionnaire
1. The first thing to consider is the psychology of the respondents. His
willingness, honesty and ability to answer questions are of utmost
importance. He is usually a stranger, who is neither interested in the
investigator nor in his project. He may be a very busy person, and his
administrative responsibility may be making heavy demands on his time.
The investigator must put himself in the respondents position to see how
his attention, sympathy and cooperation can be secured. He should also
try to minimize the demands on his time.
2. One can make the responses simple by providing a variety of possible
answer and requiring the respondents to put a check mark against the
correct one.
3. One should not undertake a questionnaire study, unless the problem is
really important from the point of view of others. It should appear to be
worth investigation to the respondents. It should always be accompanied
58

by a brief introductory letter which will explain the purpose of investigation


and show how the respondent's cooperation will benefit him or a noble
cause. It should also contain suitable instructions for filling in it.
4. Efforts should be made eliminated irrelevant items, which may not be
helpful in interpretation of results. Weeding out of unnecessary and trivial
questions for making the questionnaire as brief as possible is very
important. Experience shows that in this way the number of questions can
in certain cases be reduced to one half.
5. In framing the questionnaire the situation in which the average respondent
works should be taken into average respondent works should be taken into
consideration. It is frequently noticed that questionnaires are prepared
with reference to institutions of smaller size or with reference to
institutions which are financially or otherwise more happily placed and
these are institutions of smaller size where those conditions do not prevail.
Thus the responses are irrelevant and useless.
6. The importance of each question i.e. what is actually required must be
clear to the maker as well as the recipient. All the items must be free from
technical and ambiguous terms so that the recipients can give their
responses without consulting others. If any technical terms have to be
used they must be explained in the body of the questionnaire.
7. As far as possible the response should be of such a nature as can be easily
summarized in some form. However, they need not necessarily be
quantitative or of yes or no character, or che ck mark type. These are
preferable; because they are easier to summarize.
8. Questions eliciting opinion should be avoided unless opinion is worth
getting. Questions permitting subjective responses should be minimized.
9. Where the responses are too large to be counted, it is always advisable to
resort to coding symbols, which can be punched and weighed for
summarizing purposes.
10. The group to whom a questionnaire is administered must be judiciously
selected. One must carefully avoid asking people questions, they d o not
know or about which they have strongly coloured convictions. Sometimes
the respondents may possess the information required, but he may not be
free or willing to divulge it. Asking people working in denominational
institutions on the eve of vacation whether they would return to work after
vacation may not beget truthful replies. In such cases asking respondents
not to append their signatures would be helpful in getting frank responses.
11. If there is any doubt that corrects information about certain items will not
be forthcoming from all the respondents questions about such items
should not be asked if the investigator wants to avoid landing himself in
difficulties at the interpretation stage.
59

12. A good questionnaire will arouse the curiosity as regards resu lts and will
stimulate them to make supplementary inquiries and promise them a copy
of the published results.
13. It should be so organized as to avoid overlapping of items in different
sections.
14. Evidently respondents will not like to incur expenditure on postage for
returning the questionnaire. The investigator should therefore see that it
accompanied by self-addressed stamped envelope, in case; it is to be
returned by post.
Limitations of the Questionnaire Technique
In this technique, a research worker has to depend on several hundred
persons from whom response is expected and it is not an easy job to get active and
willing cooperation of all the respondents. One may be very diligent and sincere
about his work but one cannot be sure that the responses would be forthcoming.
Some of the respondents may hold back their replies because they are skeptical
about the value of research, others may not respond for want of time, or because
they do not feel interested in the problem in hand or because they have not been
sufficiently motivated by the introductory letter. The research worker may therefore
remain in a state of expectancy and his work may be delayed.
1. Even when he gets back a sufficiently large number of questionnaires, he
may find that all of them have not been completely filled in Omissions in
some cases may be such as will materially vitiate the results and
significantly affect the interpretations.
2. If the questionnaire is sent to different areas or to people of different
categories and one set of respondents deliberately with holds its responses,
the inquiry would be affected adversely. Suppose there is a judicial inquiry
on discrimination against a minority and section of the minority concerned
refuses to cooperate as a protest, evidently the results of the i nquiry will
not present a true picture.
3. Sometimes the respondents fill in their responses very indifferently,
without bothering about their correctness and sometimes they deliberately
give wrong information.
4. Interview Schedule
Scheduling is defined as the translation of the developed plan a time table,
showing the calendar date for the start and completion of work. The scheduled start
and end of each activity or work package as well as the total project is emphasized.
The schedule helps to determine the operating is emphasized. The schedule helps
to determine the operating budget for the project and permits us to allocate
resources to' the activities.
It is within the scheduling process that we become concerned with competition
for one or more of the resources that may exist in an organization. In establishing a
60

schedule for the project, we are concerned not only with the time to do a job but
also with the exact data that the office involved may be able to do our work, having
considered all the other projects in the organization that will utilize this same
resources. The planned schedule, which is generated as an output of the
scheduling process, enable the project director to judge event progress and forecast
a data of completion. The scheduling' process also enables the project director to
inform various organizational units of the schedules that they must maintain for
their particular task.
Some Scheduling Constraints
On the surface, the process of scheduling appears relatively simple. Several
constraints, however, make the process somewhat difficult and often lead to the
development of less than an ideal or optimum schedule. Some constraints are given
in the following list:
1. The availability of particular resources during specific calendar periods.
2. The general sequence of the work in the project plan.
3. Consideration of resource requirements of other present or future projects.
4. Different or conflicting demands on the same resource.
5. A desire to avoid peak load for particular skills.
6. The available local capacity to do a particular task.
7. Limitations and requirements imposed by funding agencies.
8. Desire to minimize overtime and idle time.
9. Necessary integration with other plans or projects using the same
resources.
10. The manager's judgement of a reasonable time for pe rforming activities of
an uncertain nature.
11. Technical constraints such as uncertainties which may require extra time.
12. Local personnel policies concerning work practices (vacations, sick leave,
etc.)
13. National, states, and local laws governing work practices .
14. Difficulties inherent in scheduling far in advance.
15. The varying number of working days in a month and their translation into
calendar dates.
16. Many of these constraints relate to project scheduling in the business-
military complex, but have some application to the educational situation.
The educational situation, in contrast, has some unique constraints which
can affect schedules. Notable among these is the nine to ten month period
in which educational personnel operate.
61

Characteristics of a Good Schedule


Unfortunately, an Ideal or optimum schedule rarely can be prepared in a
practical situation. Instead, the project manager must generate a "reasonable'
schedule. In doing so, he must form the criterion of reasonableness. Some possible
criteria are: (1) to complete the project in a minimum amount of time: (2) To
complete the project with a minimum amount of cost: (3) to maximize performance
in the project. A further criterion might be to "level" the utilization of resources over
a period of time. A schedule that is developed for only one of these criteria probably
will not meet the needs of the criterion.
The most typical criterion is the one associated with the least cost , i.e.
schedules are prepared to minimize the costs that are associated with the resources
used in the project.
Scheduling Steps
Designing a schedule begins with the individual activities and work package.
Giving consideration to the resource availability, a scheduled elapsed time may be
shorter/longer that the estimated time for the activity. A schedule end date can be
determined for the terminal event, if a schedule start date has been provided.
Forward and backward time calculations can be used to establish the earliest
schedule completion dates and the latest schedule completion dates for e ach event,
as we did in the time estimating phase .
If, upon completion of this step, the total time exceeds the time available or the
total costs are exceeded, readjustments will have to be made, using the procedure
of paralleling activities, eliminating tasks, redefining work scope, and adjusting
resources. The adjustment continues until a schedule is devised that meets the
criterion established by the manager.
In developing the schedule, the manager should realize the slackness existing
on the pathways, in the work flow. If the manager is aware of how much and where
slackness activities on non-critical paths can be moved until resources are
available; their duration can be extended in order to reduce resource utilization
during a given period of time, or e ven possibly split into segments for different
Scheduled periods. Splitting into segments can only be done when it is possible to
complete the task in segments. The activities on the critical path are not changed
unless it is absolutely necessary .
Rescheduling
Once the project is initiated and operations are begun, the project will need to
be rescheduled from time to time. The project manager should not assume that the
initial schedule will be maintained throughout the life of the project. This condition
could be desirable but schedules can be disrupted for many reasons. A change in
the major or supporting objective of a project may lead to a rescheduling of the
project internally.
Another factor that may cause rescheduling might be change in the work flow
necessary to achieve an objective. The schedule may have to be changed because of
62

certain slippages which have occured because of unusual delays in completing an


activity, or because of increase in slack time due to the early completion of an
activity. Another cause of rescheduling may be change in funding. The funding may
go higher or lower. In either case, a different schedule may have to be generated .
Problems in Scheduling
There are four different problems associated with the scheduling process.
Several general problems appear to give the most concern to project directors these
four problems have been the object of research and study. Each of them will now be
discussed briefly.
2. The Fixed Duration Time of Problem
This scheduling problem arises when there is a constrain upon the total
project duration time. That is there is a fixed time by which the project must be
completed. This situation usually arises and others have pointed out, when the
project manager has produced sufficient resources to carry out the project but
wishes to carry them at a constant rate, making the most effective use of them. This
concept is often referred to as "manpower leveling".
3. The Fixed Resources Problem
This scheduling problem arises when a project managed has a pool of
resources which cannot be exceeded. His objective is to schedule the activities in a
way that minimizes any possible increase in the total project time. He should realize
that, because of the limitations on resources, the project will have to be extended
by some small amount, but this increase must be kept to a minimum.
The solution of this problem lies in the following steps; Activities on non -
critical pathways throughout the project are delayed until the manpower needed for
them is available. Activities on the various slack pathways or non-critical paths are
also lengthened. Instead of taking one week to do the job, perhaps two weeks are
scheduled with the resources in effect for half of that activity. If the resource
requirement is still excessive after these two procedures have been employed the
critical path should be examined to determine the greatest reduction in resource
requirements per unit of time increase in the project duration.
4. The Time/Cost Trade off concept Problem
Scheduling can become a major problem if either time or resources are
constrained to certain limits. Some of the specific scheduling models do not take
into account limited resources. Therefore, they do not as useful as scheduling
techniques. If there are no constraints on either resources or time, the problem
becomes at time/cost trade off. The concept has limited value in education research
and development situations. Its greatest application has been made in the
construction industry and other trades where there is enough data accumulated
about the time and cost associated with a particular job.
4. Multi-Project Scheduling Problem
We have been discussing scheduling as it relates to a single project directed by
one project manager. In many organizations, several projects go on simultaneously
63

under different project managers. These projects draw upon several common
resources of the organization. Consequently conflicts arise about the utilization of
the resources. The employment of the previous procedures makes the work flow for
a particular department more even and insures that the work of the projects will be
accomplished according to the schedule.
Limitations of a Schedule
The following are the main disadvantages of a schedule:
1. It is very time consuming and costly instrument in admini stering to the
subject personally.
2. Sometimes some subjects have several time queries about the schedule
and difficult to explain and satisfy them.
3. Some subjects are to be conducted to get-data.
4. Some of the subjects e.g. principals administrators are not easily
approachable and get appointment for administering tool.
5. Sometimes subjects are more alert or intelligent than the researcher.
Researcher has the difficulty to administer the tools.
6. On a large sample of subjects this tool can not be used effectively and
easily.
5.4 REVISION POINTS
1. Data are facts, figures and other relevant materials, past and present,
serving as bases for study and analysis
2. Primary sources are original sources from which the researcher directly
collects data that have not been previously collected.
3. The secondary sources consist of readily available and compiled statistical
statements and reports
4. The focus sessions are aimed at obtaining respondents' impressions,
interpretations, and opinions, as the members talk about the event,
concept, product, or service
5. Panels, like focus groups, are another source of primary information for
research purposes
6. Secondary data are indispensable for most organizational research
5.5 INTEXT QUESTIONS
1. Explain primary and secondary data and distinguish between them.
2. Explain the different methods of collecting primary data.
3. Explain direct personal, indirect oral interview, information received
4. through agencies etc.)
5. Explain the merits and demerits of different methods of collecting primary
data.
64

6. Explain the different sources of secondary data and the precautions in


using secondary data.
7. What is editing of secondary data? Why is it required?
8. What are the different types of editing of secondary data?
5.6 SUMMARY
This chapter focuses on primary and secondary data. To start with meaning of
data, importance have been discussed. Since the use of primary and secondary
data may be hazardous at times, the marketing researcher must satisfy himself
regarding their suitability before use. At the end, the lesson conc ludes the various
data collection methods in detail.
5.7 TERMINAL EXERCISES
1. ………………..and …………………are two broad approaches available for
primary data collection.
2. The major difference between the observation and questioning approaches is
that in the ………………….process
3. A causal observation could enlighten the researcher to …………….the
problem.
4. Survey is widely used in ……………………social science research
5.8 SUPPLEMENTARY MATERIALS
1. https://www.researchgate.net/publication/228772638_Press_releases_ann
ual_reports_and_newspaper_articles-
using_alternative_data_sources_for_studies_on_business_network_dynamics
2. Magazines for Libraries. Cheryl LaGuardia, editor. 22nd ed. New York:
Bowker, 2014.
5.9 ASSIGNMENTS
1. “It is never safe to take published statistics at their face value without
knowing their meaning and limitations.” Elucidate this statement by enumerating
and explaining the various points which you would consider before using any
published data. Illustrate your answer by examples wherever possible.
2. Examine the merits and limitations of the observation method in c ollecting
material. Illustrate your answer with suitable examples.
3. “Experimental method of research is not suitable in management field.”
Discuss, what are the problems in the introduction of this research design in
business organisation?
5.10 SUGGESTED READINGS/ REFERENCE BOOKS
1. Fowler, F. (2001) Survey Research Methods, 3rd edition, Thousand Oaks,
CA: Sage.
2. Marsh, C. (1982) The Survey Method: The Contribution of Surveys to
Sociological Explanation, London: George
3. Allen and Unwin. Sapsford, R. (1999) Survey Research, London: Sage
65

4. Kothari, C.R.,1985, Research Methodology- Methods and Techniques, New


Delhi, Wiley Eastern Limited.
5. Kumar, Ranjit, 2005, Research Methodology-A Step-by-Step Guide for
Beginners,(2nd.ed.),Singapore, Pearson Education.
5.11 LEARNING ACTIVITIES
1. Select an industry of your choice. Using secondary sources, obtain
industry sales and sales of major firms in that firms for the past year and
also estimate market share of each firm.
5.12 KEYWORDS
Questionnaire: A questionnaire is a written list of questions, the answers to
which are recorded by respondents. In a questionnaire respondents read the
questions, interpret what is expected and then write down the answers. The only
difference between an interview schedule and a questionnaire is that in the former
it is the interviewer who asks the questions (and, if necessary, explains them) and
records the respondent’s replies on an interview schedule, while in the latter replies
are recorded by the respondents themselves.
Primary data: Information collected for the specific purpose of a study either
by the researcher or by someone else is called primary data.
Primary sources: Sources that provide primary data such as interviews,
observations, and questionnaires are called primary sources.
Participant observation: is when you, as a researcher, participate in the
activities of the group being observed in the same manner as its members, with or
without their knowing that they are being observed. Participant observation is
principally used in qualitati ve research and is usually done by developing a close
interaction with members of a group or ‘living’ in with the situation which is being
studied.











66

LESSON – 6

MEASUREMENT AND SCALING TECHNIQUES


6.1 INTRODUCTION
In our daily life we are said to measure when we use some yardstick to
determine weight, height, or some other feature of a physical object. We also
measure when we judge how well we like a song, a painting or the personalities of
our friends. We, thus, measure physical objects as well as abstract concepts.
Measurement is a relatively complex and demanding task, specially so when it
concerns qualitative or abstract phenomena. By measurement we mean the process
of assigning numbers to objects or observations, the level of measurement being a
function of the rules under which the numbers are assigned.
It is easy to assign numbers in respect of properties of some objects, but it is
relatively difficult in respect of others. For instance, measuring such things as
social conformity, intelligence, or marital adjustment is much less obvious and
requires much closer attention than measuring physical weight, biological age or a
person's financial assets. In other words, properties like weight, height, etc., can be
measured directly with some standard unit of measurement, but it is not that easy
to measure properties like motivation to succeed, ability to stand stress and the
like. We can expect high accuracy in measuring the length of pipe with a yard stick,
but if the concept is abstract and the measurement tools are not standardized, we
are less confident about the accuracy of the results of measurement.
Technically speaking, measurement is a process of mapping aspects of a domain
onto other aspects of a range according to some rule of correspondence. In
measuring, we devise some form of scale in the range (in terms of set theory, range
may refer to some set) and then transform or map the properties of objects from the
domain (in terms of set theory, domain may refer to some other set) onto this scale.
For example, in case we are to find the male to female attendance ratio while
conducting a study of persons who attend some show, then we may tabu late those
who come to the show according to sex. In terms of set theory, this process is one of
mapping the observed physical properties of those coming to the show (the domain)
on to a sex classification (the range). The rule of correspondence is: If the object in
the domain appears to be male, assign to "0" and if female assign to "1". Similarly, we
can record a person's marital status as 1, 2, 3 or 4, depending on whether the person
is single, married, widowed or divorced. We can as well record "Yes or No" answers to
a question as "0" and "1" (or as 1 and 2 or perhaps as 59 and 60). In this artificial or
nominal way, categorical data (qualitative or descriptive) can be made into
numerical data and if we thus code the various categories, we refer to the numbers
we record as nominal data. Nominal data are numerical in name only, because they
do not share any of the properties of the numbers we deal in ordinary arithmetic. For
instance if we record marital status as 1, 2, 3, or 4 as stated above, we cannot write >
2 or 3 < 4 and we cannot write 3 - 1 = 4 - 2, 1 + 3 = 4 or 4 -s- 2 = 2.
67

In those situations when we cannot do anything except set up inequalities, we


refer to the data as ordinal data. For instance, if one mineral can scratch another, it
receives a higher hardness number and on Mohs' scale the numbers from 1 to 10
are assigned respectively to talc, gypsum, calcite, fluorite, apatite, feldspar, quartz,
topaz, sapphire and diamond. With these numbers we can write > 2 or 6 < 9 as
apatite is harder than gypsum and feldspar is softer than sapphire, but we cannot
write for example 10 - 9 = 5 - 4, because the difference in hardness between
diamond and sapphire is actually much greater than that between apatite and
fluorite. It would also be meaningless to say that topaz is twice as hard as fluorite
simply because their respective hardness numbers on Mohs' scale are 8 and 4. The
greater than symbol (i.e., >) in connection with ordinal data may be used to
designate "happier than" "preferred to" and so on.
When in addition to setting up inequalities we can also form differences, we
refer to the data as interval data. Suppose we are given the following temperature
readings (in degrees Fahrenheit): 58°, 63°, 70°, 95°, 110°, 126° and 135°. In this
case, we can write 100° > 70° or 95° < 135° which simply means that 110° is
warmer than 70° and that 95° is cooler than 135°. We can also write for example
95° - 70° = 135° - 110°, since equal temperature differences are equal in the sense
that the same amount of heat is required to raise the temperature of an object from
70° to 95° or from 110° to 135°. On the other hand, it would not mean much if we
said that 126° is twice as hot as 63°, even though 126° + 63° = 2. To show the
reason, we have only to change to the centigrade scale, where the first temperature
becomes 5/9 (126 - 32) = 52°, the second temperature becomes 5/9 (63 -32) = 17°
and the first figure is now more than three times the second. This difficulty arises
from the fact that Fahrenheit and Centigrade scales both have artificial origins
(zeros) i.e., the number 0 of neither scale is indicative of the absence of whatever
quantity we are trying to measure.
When in addition to setting up inequalities and forming differences we can also
form quotients (i.e., when we can perform all the customary operations of
mathematics), we refer to such data as ratio data. In this sense, ratio data includes
all the usual measurement (or determinations) of length, height, money amounts,
weight, volume, area, pressures etc.
The above stated distinction between nominal, ordinal, interval and ratio data
is important for the nature of a set of data may suggest the use of particular
statistical techniques*. A researcher has to be quite alert about this aspect while
measuring properties of objects or of abstract concepts.
6.2 OBJECTIVES
After studying this lesson, you will be able to:
 Measurement
 Basic measurement scales
 Measurement accuracy
68

 Scaling and its types


6.3 CONTENTS
6.3.1 Basic Measurement scales
6.3.2 Measurement accuracy
6.3.3 Scaling
6.3.1 Basic measurement scales
Nominal scale
A nominal scale is a figurative labeling scheme in which the numbers serve
only as labels or tags for identifying and classifying objects. For example, the
numbers assigned to the respondents in a study constitute a nominal scale, thus a
female respondent may be assigned a number 1 and a male respondent 2. When a
nominal scale is used for the purpose of identification, there is a strict one -to-one
correspondence between the numbers and the objects. Each number is assigned to
only one object, and each object has only one number assigned to it. Common
examples include student registration numbers at their college or university and
numbers assigned to football players or jockeys in a horse race. In marketing
research, nominal scales are used for identifying respondents, brands, attributes,
banks and other objects. When used for classification purposes, the nominally
scaled numbers serve as labels for classes or categories. For example, you might
classify the control group as group 1 and the experimental group as group 2. The
classes are mutually exclusive and collectively exhaustive. The objects in each class
are viewed as equivalent with respect to the characteristic represented by the
nominal number. All objects in the same class have the same number, and no two
classes have the same number. The numbers in a nominal scale do not reflect the
amount of the characteristic possessed by the objects. For example, a high number
on a football player’s shirt does n ot imply that the footballer is a better player than
one with a low number or vice versa. The same applies to numbers assigned to
classes. The only permissible operation on the numbers in a nominal scale is
counting. Only a limited number of statistics, al l of which are based on frequency
counts, are permissible. These include percentages, mode, chi -square and binomial
tests. It is not meaningful to compute an average student registration number, the
average gender of respondents in a survey, or the number assigned to an average
bank.
Ordinal scale
An ordinal scale is a ranking scale in which numbers are assigned to objects to
indicate the relative extent to which the objects possess some characteristic. An
ordinal scale allows you to determine whether an object has more or less of a
characteristic than some other object, but not how much more or less. Thus, an
ordinal scale indicates relative position, not the magnitude of the differences
between the objects. The object ranked first has more of the characte ristic as
compared with the object ranked second, but whether the object ranked second is a
close second or a poor second is not known. Common examples of ordinal scales
include quality rankings, rankings of teams in a tournament and occupational
69

status. In marketing research, ordinal scales are used to measure relative attitudes,
opinions, perceptions and preferences. Measurements of this type include ‘greater
than’ or ‘less than’ judgments from the respondents. In an ordinal scale, as in a
nominal scale, equivalent objects receive the same rank. Any series of numbers can
be assigned that preserves the ordered relationships between the objects’ In other
words; any monotonic positive (order preserving) transformation of the scale is
permissible, since the differences in numbers are void of any meaning other than
order (see the following example). For these reasons, in addition to the counting
operation allowable for nominal scale data, ordinal scales permit the use of
statistics based on percentiles. It is me aningful to calculate percentile, quartile,
median, rank-order correlation or other summary statistics from ordinal data.
Interval scale
In an interval scale, numerically equal distances on the scale represent equal
values in the characteristic being measured. An interval scale contains all the
information of an ordinal scale, but it also allows you to compare the differences
between objects. The difference between any two scale values is identical to the
difference between any other two adjacent values of an interval scale. There is a
constant or equal interval between scale values. The difference between 1 and 2 is
the same as the difference between 2 and 3, which is the same as the difference
between 5 and 6. A common example in everyday life is a temperature scale. In
marketing research, attitudinal data obtained from rating scales are often treated
as interval data. In an interval scale, the location of the zero point is not fixed.
Both the zero point and the units of measurement are arbitrary. Hence, any positive
linear transformation of the form y = a + bx will preserve the properties of the scale.
Here, x is the original scale value, y is the transformed scale value, b is a positive
constant, and a is any constant. Therefore, two interval scales that rate objects A,
B, C and D as 1, 2, 3 and 4 or as 22, 24, 26 and 28 are equivalent. Note that the
latter scale can be derived from the former by using a = 20 and b = 2 in the
transforming equation. Because the zero point is not fixed, it is not meaningful to
take ratios of scale values. As can be seen, the ratio of D to B values changes from
2:1 to 7:6 when the scale is transformed. Yet, ratios of differences between scale
values are permissible. In this process, the constants a and b in the transforming
equation drop out in the computations. The ratio of the difference between D and B
values to the difference between C and B values is 2:1 in both the scales. Statistical
techniques that may be used on interval scale data include all those that can be
applied to nominal and ordinal data in addition to the arithmetic mean, standard
deviation (Chapter 18), product moment correlations, and other statistics
commonly used in marketing research. Certain specialized statistics such as
geometric mean, harmonic mean and coefficient of variation, however, are not
meaningful on interval scale data.
Ratio scale
A ratio scale possesses all the properties of the nominal, ordinal and interval
scales, and, in addition, an absolute zero point. Thus, in ratio scales we can
70

identify or classify objects, rank the objects, and compare intervals or differences. It
is also meaningful to compute ratios of scale values. Not only is the difference
between 2 and 5 the same as the difference between 14 and 17, but also 14 is
seven times as large as 2 in an absolute sense. Common examples of ratio scales
include height, weight, age and money. In marketing, sales, costs, market share
and number of customers are variables measured on a ratio scale. Ratio scales
allow only proportionate transformations of the form y = bx, where b is a positive
constant. One cannot add an arbitrary constant, as in the case of an interval scale.
An example of this transformation is provided by the conversion of yards to feet (b =
3). The comparisons between the objects are identical whether made in yards or
feet. All statistical techniques can be applied to ratio data. These include specialized
statistics such as geometric mean, harmonic mean and coefficient of variation.
6.3.2 Measurement accuracy
A measurement is a number that reflects some characteristic of an object. A
measurement is not the true value of the characteristic of interest but rather an
observation of it. A variety of factors can cause measurement error, which results in
the measurement or observed score being different from the true score of the
characteristic being measured. The true score model provides a framework for
understanding the accuracy of measurement.
According to this model, XO = XT + XS + XR
where
XO = The Observed Score Or Measurement
XT = The True Score Of The Characteristic
XS = Systematic Error
XR = Random Error
Note that the total measurement error includes the systematic error, XS , and
the random error, XR . Systematic error affects the measurement in a constant
way. It represents stable factors that affect the observed score in the same way each
time the measurement is made, such as mechanical factors. Random error, on the
other hand, is not constant. It represents transient factors that affect the observed
score in different ways each time the measurement is made, such as short-term
transient personal factors or situational factors . The distinction between
systematic and random error is crucial to our understanding of reliability and
validity.
Reliability It refers to the extent to which a scale produces consistent results
if repeated measurements are made. Systematic sources of error do not have an
adverse impact on reliability, because they affect the measurement in a constant
way and do not lead to inconsistency. In contrast, random error produces
inconsistency, leading to lower reliability. Reliability can be defined as the extent to
which measures are free from random error, XR . If XR = 0, the measure is perfectly
reliable. Reliability is assessed by determining the proportion of systematic
71

variation in a scale. This is done by determining the association between scores


obtained from different administrations of the scale. If the association is high, the
scale yields consistent results and is therefore reliable. Approaches for assessing
reliability include the test–re-test, alternative forms, and internal consistency
methods. In test–re-test reliability, respondents are administered identical sets of
scale items at two different times, under as nearly equivalent cond itions as
possible. The time interval between tests or administrations is typically two to four
weeks. The degree of similarity between the two measurements is determined by
computing a correlation coefficient. The higher the correlation coefficient, the
greater the reliability. Several problems are associated with the test–re-test
approach to determining reliability. First, it is sensitive to the time interval between
testing. Other things being equal, the longer the time interval, the lower the
reliability. Second, the initial measurement may alter the characteristic being
measured. For example, measuring respondents’ attitude towards low-alcohol beer
may cause them to become more health conscious and to develop a more positive
attitude towards low-alcohol beer. Third, it may be impossible to make repeated
measurements (for example, the research topic may be the respondent’s initial
reaction to a new product). Fourth, the first measurement may have a carryover
effect to the second or subsequent measurements. Respondents may attempt to
remember answers they gave the first time. Fifth, the characteristic being measured
may change between measurements. For example, favorable information about an
object between measurements may make a respondent’s attitude more positive.
Finally, the test–re-test reliability coefficient can be inflated by the correlation of
each item with itself. These correlations tend to be higher than correlations between
different scale items across administrations. Hence, it is possible to have high test–
re-test correlations because of the high correlations between the same scale items
measured at different times even though the correlations between different scale
items are quite low. Because of these problems, a test–re-test approach is best
applied in conjunction with other approaches, such as alternative -forms reliability.
In alternative-forms reliability, two equivalent forms of the scale are constructed.
The same respondents are measured at two different times, usually two to four
weeks apart (e.g. by initially using Likert scaled items and then using Stapel scaled
items). The scores from the administrations of the alternative scale forms are
correlated to assess reliability.
The two forms should be equivalent with respect to content, i.e . each scale
item should attempt to measure the same items. The main problems with this
approach are that it is difficult, time-consuming and expensive to construct an
equivalent form of the scale. In a strict sense, it is required that the alternative sets
of scale items should have the same means, variances and intercorrelations. Even if
these conditions are satisfied, the two forms may not be equivalent in content.
Thus, a low correlation may reflect either an unreliable scale or non -equivalent
forms. Internal consistency reliability is used to assess the reliability of a
summated scale where several items are summed to form a total score. In a scale of
72

this type, each item measures some aspect of the construct measured by the entire
scale, and the items should be consistent in what they indicate about the
construct. This measure of reliability focuses on the internal consistency of the set
of items forming the scale. The simplest measure of internal consistency is split-
half reliability. The items on the scale are divided into two halves and the resulting
half scores are correlated. High correlations between the variables indicate high
internal consistency. The scale items can be split into halves based on odd - and
even-numbered items or randomly. The problem is that the results will depend on
how the scale items are split. A popular approach to overcoming this problem is to
use the coefficient alpha. The coefficient alpha, or Cronbach’s alpha, is the average
of all possible split-half coefficients resulting from different ways of splitting the
scale items. This coefficient varies from 0 to 1, and a value of 0.6 or less generally
indicates unsatisfactory internal consistency reliability. An important property of
coefficient alpha is that its value tends to i ncrease with an increase in the number
of scale items. Therefore, coefficient alpha may be artificially, and inappropriately,
inflated by including several redundant scale items.
Another coefficient that can be employed in conjunction with coefficient alpha
is coefficient beta. Coefficient beta assists in determining whether the averaging
process used in calculating coefficient alpha is masking any inconsistent items.
Some multi-item scales include several sets of items designed to measure different
aspects of a multidimensional construct. For example, bank image is a
multidimensional construct that includes country of origin, range of products,
quality of products, service of bank personnel, credit terms, investment rates,
convenience of location, and physical layout of branches. Hence, a scale designed to
measure bank image would contain items measuring each of these dimensions.
Because these dimensions are somewhat independent, a measure of internal
consistency computed across dimensions would be inappropriate. If several items
are used to measure each dimension, however, internal consistency reliability can
be computed for each dimension.
Validity
The validity of a scale may be considered as the extent to which differences in
observed scale scores reflect true differences among objects on the characteristic
being measured, rather than systematic or random error. Perfect validity requires
that there be no measurement error (XO = XT , XR = 0, XS = 0). Researchers may
assess content validity, criterion vali dity or construct validity.
Content validity, sometimes called face validity, is a subjective but systematic
evaluation of how well the content of a scale represents the measurement task at
hand. The researcher or someone else examines whether the scale items adequately
cover the entire domain of the construct being measured. Thus, a scale designed to
measure bank image would be considered inadequate if it omitted any of the major
dimensions (range of products, quality of products, service of bank personnel , etc.).
Given its subjective nature, content validity alone is not a sufficient measure of the
73

validity of a scale, yet it aids in a common -sense interpretation of the scale scores.
A more formal evaluation can be obtained by examining criterion validity
Criterion validity reflects whether a scale performs as expected in relation to
other selected variables (criterion variables) as meaningful criteria. If, for example, a
scale is designed to measure loyalty in customers, criterion validity might be
determined by comparing the results generated by this scale with results generated
by observing the extent of repeat purchasing. Based on the time period involved,
criterion validity can take two forms, concurrent validity and predictive validity.
Concurrent validity is assessed when the data on the scale being evaluated
(e.g. loyalty scale) and the criterion variables (e.g. repeat purchasing) are collected
at the same time. The scale being developed and the alternative means of
encapsulating the criterion variables would be administered simultaneously and
the results compared.
Predictive validity is concerned with how well a scale can forecast a future
criterion. To assess predictive validity, the researcher collects data on the scale at
one point in time and data on the criterion variables at a future time. For example,
attitudes towards how loyal customers feel to a particular brand could be used to
predict future repeat purchases of that brand. The predicted and actual purchases
are compared to assess the predictive validity of the attitudinal scale. Construct
validity addresses the question of what construct or characteristic the scale is, in
fact, measuring. When assessing construct validity, the researcher attempts to
answer theoretical questions about why the scale works and what deductions can
be made concerning the underlying theory. Thus, construct validity requires a
sound theory of the nature of the construct being measured and how it relates to
other constructs.
Construct validity is the most sophisti cated and difficult type of validity to
establish. Construct validity includes convergent, discriminant and nomological
validity. Convergent validity is the extent to which the scale correlates positively
with other measurements of the same construct. It i s not necessary that all these
measurements be obtained by using conventional scaling techniques. Discriminant
validity is the extent to which a measure does not correlate with other constructs
from which it is supposed to differ. It involves demonstrating a lack of correlation
among differing constructs. Nomological validity is the extent to which the scale
correlates in theoretically predicted ways with measures of different but related
constructs. A theoretical model is formulated that leads to further deductions, tests
and inferences. An example of construct validity can be evaluated in the following
example. A researcher seeks to provide evidence of construct validity in a multi -
item scale, designed to measure the concept of ‘self-image’. These findings would be
sought:■ High correlations with other scales designed to measure self-concepts and
with reported classifications by friends (convergent validity) ■ Low correlations with
unrelated constructs of brand loyalty and variety-seeking (discriminant validity) ■
74

Brands that are congruent with the individual’s self-concept are more preferred, as
postulated by the theory (nomological validity) ■ A high level of reliability.
Note that a high level of reliability was included as evidence of construct
validity in this example.
6.3.3 Scaling
In research we quite often face measurement problem (since we want a valid
measurement but may not obtain it), specially when the concepts to be measured
are complex and abstract and we do not possess the standardized measurement
tools. Alternatively, we can say that while measuring attitudes and opinions, we
face the problem of their valid measurement. Similar problem may be faced by a
researcher, of course in a lesser degree, while measuring physical or institutional
concepts. As such we should study some procedures which may enable us to
measure abstract concepts more accurately. This brings us to the study of scaling
techniques.
6.3.4 Meaning of Scaling
Scaling describes the procedures of assigning numbers to various degre es of
opinion, attitude and other concepts. This can be done in two ways viz., (i) making
a judgment about some characteristic of an individual and then placing him
directly on a scale that has been defined in terms of that characteristic and (ii)
constructing questionnaires in such a way that the score of individual's responses
assigns him a place on a scale. It may be stated here that a scale is a continuum,
consisting of the highest point (in terms of some characteristic e.g., preference,
favorableness, etc.) and the lowest point along with several intermediate points
between these two extreme points. These scale -point positions are so related to
each other that when the first point happens to be the highest point, the second
point indicates a higher degree in terms of a given characteristic as compared to the
third point and the third point indicates a higher degree as compared to the fourth
and so on. Numbers for measuring the distinctions of degree in the
attitudes/opinions are, thus, assigned to indivi duals corresponding to their scale-
positions. All this is better understood when we talk about scaling technique(s).
Hence the term 'scaling' is applied to the procedures for attempting to determine
quantitative measures of subjective abstract concepts. Sc aling has been defined as
a "procedure for the assignment of numbers (or other symbols) to a property of
objects in order to impart some of the characteristics of numbers to the properties
in question.
6.3.5 M easurement of attitudes in quantitative and qualitative research
There are a number of differences in the way attitudes are measured in
quantitative and qualitative research. In quantitative research you are able to
explore measure, determine the intensity and combine attitudes to different aspects
of an issue to arrive at one indicator that is reflective of the overall attitude. In
qualitative research, you can only explore the spread of attitudes and establish the
types of attitudes prevalent. In quantitative research you can ascertain the types of
attitudes people have in a community, how many people have a particular attitude
75

and what the intensity is of those attitudes. A number of techniques have been
developed to measure attitudes and their intensity in quantitative research, but
such techniques are lacking in qualitative research. This is mainly because in
qualitative research you do not make an attempt to measure or quantify. The
concept of attitudinal scales, therefore, is only prevalent in quantitative research.
6.3.6 Attitudinal scales in quantitative research
In quantitative research there are three scales which have been developed to
'measure' attitudes. Each of these scales is based upon different assumptions and
follows different procedures in their construction. As a beginner in research
methods it is important for you to understand these procedures and the
assumptions behind them so that you can make appropriate and accurate
interpretation of the findings. As you will see, it is not very easy to construct an
attitudinal scale. Out of the three scales, the Likert scale is the easiest to construct
and therefore is used far more.
6.4.7 Functions of attitudinal scales
If you want to find out the attitude of respondents towards an issue, you can
ask either a closed or an open-ended question. For example, let us say that you
want to ascertain the attitude of students in a class towards their lecturer and that
you have asked them to respond to the following question: 'What is your attitude
towards your lecturer?' If your question is open ended, it invites each respondent to
describe the attitude that s/he holds towards the lecturer. If you have framed a
closed question, with categories such as 'extremely positive', 'positive', 'uncertain',
'negative' and 'extremely negative', this guides the respondents to select a category
that best describes their attitude. This type of questioning, whether framed
descriptively or in a categorical form, elicits an overall attitude towards the lecturer.
While ascertaining the overall attitude may be sufficient in some situations, in
many others, where the purpose of attitudinal questioning is to develop strategies
for improving a service or intervention, or to formulate policy, eliciting attitudes on
various aspects of the issue under study is required.
But as you know, every issue, including that of the attitude of students
towards their lecturers, has many aspects. For example, the attitude of the
members of a community towards the provision of a particular service comprises
their attitude towards the need for the service, its manner of delivery, its location,
the physical facilities provided to users, the behaviour of the staff, the competence
of the staff, the effectiveness and efficiency of the service, and so on. Similarly,
other examples - such as the attitude of employees towards the management of
their organisation, the attitude of employees towards occupational redeployment
and redundancy, the attitude of nurses towards death and dying, the attitude of
consumers towards a particular product, the attitude of students towards a
lecturer, or the attitude of staff towards the strategic plan for their organisation -
can be broken down in the same manner.
76

Respondents usually have different attitudes towards different aspects. Only


when you ascertain the attitude of respondents to an issue by formulating a
question for each aspect, using either open -ended or closed questions, do you find
out their attitude towards each aspect. The main limitation of this method is that it
is difficult to draw any conclusion about the overall attitude of a respondent from
the responses. Take the earlier example, where you want to find out the attitude of
students towards a lecturer. There are different aspects of teaching: the contents of
lectures; the organisation of material; the lecturer's ability to communicate
material; the presentation and style; knowledge of the subject; responsiveness;
punctuality; and so on. Students may rate the lecturer differently on different
aspects. That is, the lecturer might be considered extremely compe tent and
knowledgeable in his/her subject but may not be considered a good communicator
by a majority of students. Further, students may differ markedly in their opinion
regarding any one aspect of a lecturer's teaching. Some might consider the lecturer
to be a good communicator and others might not. The main problem is: how do we
find out the 'overall' attitude of the students towards the lecturer? In other words,
how do we combine the responses to different aspects of any issue to come up with
one indicator that is reflective of an overall attitude? Attitudinal scales play an
important role in overcoming this problem.
Attitudinal scales measure the intensity of respondents' attitudes towards the
various aspects of a situation or issue and provide technique s to combine the
attitudes towards different aspects into one overall indicator. This reduces the risk
of an expression of opinion by respondents being influenced by their opinion on
only one or two aspects of that situation or issue.
6.4.8 Difficulties in developing an attitudinal scale
In developing an attitudinal scale there are three problems:
 Which aspects of a situation or issue should be included when seeking to
measure an attitude? For instance, in the example cited above, what
aspects of teaching should be included in a scale to find out the attitude of
students towards their lecturer?
 What procedure should be adopted for combining the different aspects to
obtain an overall picture?
 How can one ensure that a scale really is measuring what it is supposed to
measure?
The first problem is extremely important as it largely determines the third
problem: the extent to which the statements on different aspects are reflective of
the main issue largely determines the validity of the scale. You can solve the third
problem by ensuring that your statements on the various aspects have a logical link
with the main issue under study - the greater the link, the higher the validity. The
different types of attitudinal scale (Likert, Thurstone and Guttman) provide an
answer to the second problem. They guide you as to the procedure for combining
77

the attitudes towards various aspects of an issue, though the degree of difficulty in
following the procedure for these scales varies from scale to scale.
6.4.6 Types of attitudinal scale
Likert’s Scale
Likert, is extremely popular for measuring attitudes, because, the method is
simple to administer. With the Likert scale, the respondents indicate their own
attitudes by checking how strongly they agree or disagree with carefully w orded
statements that range from very positive to very negative towards the attitudinal
object. Respondents generally choose from five alternatives (say strongly agree,
agree, neither agree nor disagree, disagree, strongly disagree). A Likert scale may
include a number of items or statements. Disadvantage of Likert Scale is that it
takes longer time to complete than other itemised rating scales because
respondents have to read each statement. Despite the above disadvantages, this
scale has several advantages. It is easy to construct, administer and use.
Semantic Differential Scale
This is a seven point rating scale with end points associated with bipolar labels
(such as good and bad, complex and simple) that have semantic meaning. It can be
used to find whether a respondent has a positive or negative attitude towards an
object. It has been widely used in comparing brands, products and company
images. It has also been used to develop advertising and promotion strategies and
in a new product development study.
Thurstone scale
Thurstone's method of pair comparisons can be considered a prototype of a
normal distribution-based method for scaling-dominance matrices. Even though
the theory behind this method is quite complex the algorithm itself is
straightforward. A Thurstone scale has a number of statements to which the
respondent is asked to agree or disagree.
There are three types of scale that Thurstone described:
 Equal-appearing intervals method
 Successive intervals method
 Paired comparisons method

Multi-Dimensional Scaling
Multidimensional scaling (MDS) is a means of visualizing the level of similarity
of individual cases of a dataset. It refers to a set of related ordination techniques
used in information visualization, in particular to display the information contained
in a distance matrix.
6.4 REVISION POINTS
1. Measurement is a relatively complex and demanding task, specially so
when it concerns qualitative or abstract phenomena
2. Measurement is a process of mapping aspects of a domain onto other
aspects of a range according to some rule of correspondence
78

3. Scaling describes the procedures of assigning numbers to various degrees


of opinion, attitude and other concepts
4. In quantitative research you are able to explore measure, determine the
intensity and combine attitudes to different aspects of an issue to arrive at
one indicator that is reflective of the overall attitude. In qualitative
research, you can only explore the spread of attitudes and establish the
types of attitudes prevalent.
5. If you want to find out the attitude of respondents towards an issue, you
can ask either a closed or an open-ended question
6.5 INTEXT QUESTIONS
1. Discuss the relative merits and demerits of:
(a) Rating vs. Ranking scales.
(b) Summated vs. Cumulative scales.
(c) Scalogram analysis vs. Factor analysis
2. What is the meaning of measurement in research?
3. What difference does it make whether we measure in terms of a nominal,
ordinal, interval or ratio scale? Explain giving examples.
4. Describe the different methods of scal e construction, pointing out the
merits and demerits of each.
6.6 SUMMARY
This chapter has dealt with the concept of measurement and scaling and its
components. This approach involves self reporting wherein a person is asked
directly how he feels about an object. The subsequent discussion is centered on
some attitude scales. As so many scales are available, the researcher faces major
question which scale should be used? The chapter deals with it briefly.
6.7 TERMINAL EXERCISES
1. ....................... scale may tell us "How far the objects are apart with respect
to an attribute?"
2. Ratio scale is a special kind of internal scale that has a meaningful
.................................
3. Scale construction techniques are used for measuring the ………….of a
group.
4. The comparative technique is used to determine the scale values of
……………items by performing comparisons among the items.
6.8 SUPPLEMENTARY MATERIALS
1. Likert, R. (1932). A Technique for the Measurement of Attitudes. Archives of
Psychology, 140, 1–55.
2. McLeod, S. A. (2008). Likert Scale. Retrieved from www.simplypsychology.org/
likert-scale.html
79

3. Thurstone, L. L., (1927), "A Law of Comparative Judgment", Psychological


Review 34, pp. 273-86.
6.9 ASSIGNMENTS
1. “Scaling describes the procedures by which numbers are assigned to various
degrees of opinion, attitude and other concepts.” Discuss. Also point out the bases
for scale classification.
2. Narrate the procedure for developing a scalogram and illustrate the same by
an example.
3. Workout Guttman’s coefficient of reproducibility from the following
information:
Number of cases (N) = 30
Number of items (n) = 6
Number of errors (e) = 10
Interpret the meaning of coefficient you work out in this example.
6.10 SUGGESTED READINGS/ REFERENCE BOOKS
1. Barnes, J.A. (1979) Who Should Know What? Social Science, Privacy and
Ethics, Harmondsworth: Penguin.
2. Bulmer, M. (1982) Social Research Ethics, London: Mac - Millan.
6.11 LEARNING ACTIVITIES
1. Design a likert scales to measure the usefulness of Ford Motor Company’s
website. Visit the company official site and rate it on the scales that you
have developed
6.12 KEYWORDS
Validity: The concept of validity can be applied to every aspect of the research
process. In its simplest form, validity refers to the appropriateness of each step in
finding out what you set out to. However, the concept of validity is more associated
with measurement procedures. In terms of the measurement procedure, validity is
the ability of an instrument to measure what it is designed to measure.
Thurstone scale: The Thurstone scale is one of the scales designed to measure
attitudes in the social sciences. Attitude through this scale is measured by means
of a set of statements, the ‘attitudinal value’of which has been determined by a
group of judges. A respondent’s agreement with the statement assigns a score
equivalent to the attitudinal value’ of the statement. The total score of all
statements is the attitudinal score for a respondent.
Thematic writing: A style of writing which is written around main themes
Scale: This is a method of measurement and/or classification of respondents
on the basis of their responses to questions you ask of them in a study. A scale
could be continuous or categorical. It helps you to classify a study population in
subgroups or as a spread that is reflective on the scale.


80

LESSON – 7

THE NATURE OF FIELD WORK


7.1 INTRODUCTION
A personal interviewer administering a questionnaire door to door, a telephone
interviewer calling from a central location, an observer counting pedestrians in a
shopping mall, and others involved in the collection of data and the supervision of
that process are all Field Workers in the field. The activities of the field workers may
vary in nature. This lesson helps us to understand the interview methods in data
collection process of the research and field work management.
7.2 OBJECTIVES
After studying this lesson, you will be able to understand:
 The importance of training for new interviewers
 The principle tactics of asking questions
7.3 CONTENTS
7.3.1 Who conducts the field work?
7.3.2 Interviewing
7.3.3 Training for interviewers
7.3.4 Fieldwork management

7.3.1 Who conducts the field work?


Data collection is rarely carried out by the person who designs the research
project. However, the data-collecting stage is crucial, because the research project
is no better than the data collected in the field. Therefore, it is important that the
research administrator select capable people who may be entrusted to collect the
data. An irony of business research is that highly educated and trained individuals
design the research, but the people who collect the data typically have little
research training or experience. Knowing the vital importance of data collected in
the field, research administrators must concentrate on carefully selecting field
workers.
7.3.2 Interviewing
Interviewing process is establishing rapport with the respondent Interviewer
bias may enter in if the fieldworker’s clothing or physical appearance is unattractive
or unusual. Suppose that a male interviewer, wearing a dirty T -shirt, interviews
subjects in an upper-income neighborhood. Respondents may consider the
interviewer slovenly and be less cooperative than they would be with a person
dressed more appropriately.
Interviewers and other fieldworkers are generally paid an hourly rate or a per-
interview fee. Often interviewers are part-time workers—housewives, graduate
students, secondary school teachers—from diverse backgrounds. Primary and
secondary school teachers are an excellent source for temporary interviewers
during the summer, especially when they conduct interviews outside the school
81

districts where they teach. Teachers’ educational backgrounds and experiences


with the public make them excellent candidates for fieldwork.
7.3.3 Training for Interviewers
The objective of training is to ensure that the data collection instrument is
administered uniformly by all field investigators. The goal of training sessions is to
ensure that each respondent is provided with common information. If the data are
collected in a uniform manner from all respondents, the training session will have
been a success. After personnel are recruited and selected, they must be trained.
Ex: A woman who has just sent her youngest child off to first grade is hired by
an interviewing firm. She has decided to become a working mother by becoming a
professional interviewer. The training that she will receive after being selected by a
company may vary from virtually no training to a 3-day program if she is selected
by one of the larger survey research agencies. Almost always there will be a briefing
session on the particular project. Typically, the recruits will record answers on a
practice questionnaire during a simulated training interview.
More extensive training programs are likely to cover the following topics
1. How to make initial contact with the respondent and secure the interview
2. How to ask survey questions
3. How to probe
4. How to record responses
5. How to terminate the interview
1. Making Initial Contact and Securing the Interview
Interviewers are trained to make appropriate opening remarks that will
convince the person that his or her cooperation is important.
For example
Good afternoon, my name is _____ and I’m from a national survey research
company. We are conducting a survey concerning ____. I would like to get a few of
your ideas.
Much fieldwork is conducted by research suppliers who specialize in data
collection, When a second party is employed, the job of the study designed by the
parent firm is not only to hire a research supplier but also to establish supervisory
controls over the field service.
In some cases a third party is employed. For example, a firm may contact a
survey research firm, which in turn subcontracts the fieldwork to a field se vice.
Under these circumstances it is still desirable to know the problems that might
occur in the field and the managerial practices that can minimize them.
2. Asking the Questions
The purpose of the interview is, of course, to have the interviewer ask
questions and record the respondent’s answers- Training in the art of stating
82

questions can be extremely beneficial, because interviewer bias can be a source of


considerable error in survey research.
There are five major principles for asking questions
1. Ask the questions exactly as they are worded in the questionnaire.
2. Read each question very slowly.
3. Ask the questions in the order in which they are presented in the
questionnaire.
4. Ask every question specified in the questionnaire.
5. U Repeat questions those are misunderstood or misinterpreted.
Although interviewers are generally trained in these procedures, when working
in the field many interviewers do not follow them exactly. Inexperienced
interviewers may not understand the importance of strict adherence to the
instructions. Even professional interviewers take shortcuts when the task becomes
monotonous. Interviewers may shorten questions or rephrase unconsciously when
they rely on their memory of the question rather than reading the question as it is
worded. Even the slightest change in wording can distort the meaning of the
question and cause some bias to enter into a study. By reading the question, the
interviewer may be reminded to concentrate on avoiding slight variations in tone of
voice on particular words phases in the question.
1. Probing
General training of interviewers should include instructions on how to probe
when respondents give no answer, incomplete answers, or answers that require
clarification. Probing may be needed for two types of situations. First, it is
necessary when the respondent must be motivated to enlarge on, clarify or explain
his or her answer. It is the interviewer’s job to probe for complete, unambiguous
answers. The interviewer must encourage the respondent to clarify or expand on
answers by providing a stimulus that will not suggest the interviewer’s own ideas or
attitudes. The ability to probe with neutral stimuli is the mark of an experienced
interviewer. Second, probing may be necessary in situations in which the
respondent begins to ramble or lose track of the question. In such cases the
respondent must be led to focus on the specific content of the interview and to
avoid irrelevant and unnecessary information.
The interviewer has several possible probing tactics to choose from, depending
on the situation
Repetition of the question: The respondent who remains completely silent may
not have understood the question or may not have decided how to answer it. Mere
repetition may encourage the respondent to answer in such cases. For example, if
the question is “What is there that you do not like about your supervisor?” and the
respondent does not answer, the interviewer may probe: “Just to check, is there
anything you do not like about your supervisor?”
83

An expectant pause: If the interviewer believes the respondent has more to say,
the “silent probe:’ accompanied by an expectant look, may motivate the respondent
to gather his or her thoughts and give a complete response Of course, the
interviewer must be sensitive to the respondent so that the silent probe does not
become an embarrassed silence.
Repetition of the Respondent’s Reply: Sometimes the interviewer may repeat
the verbatim of the respondent. This may help the respondent to expand the
answer.
2. Recording the Responses
The analyst who fails to instruct fieldworkers in the techniques of recording
answers for one study rarely forgets to do so in the second study. Although the
concept of recording an answer seems extremely simple, mistakes c an be made in
the recording phase of the research. All fieldworkers should use the same
mechanics of recording.
Example: It may appear insignificant to the interviewer whether she uses a pen
or pencil, but to the editor who must erase and rewrite illegibl e words, using a
pencil is extremely important.
The rules for recording responses to closed questionnaires vary with the
specific questionnaire. The general rule, however, is to place a check in the box that
correctly reflects the respondent’s answer. All too often interviewers don’t bother
recording the answer to a filter question because they believe that the subsequent
answer will make the answer to the filter question obvious. However, editors and
coders do not know how the respondent actually answered a question.
The general instruction for recording answers to open -ended-response
questions is to record the answer verbatim, a task that is difficult for most people.
Inexperienced interviewers should be given the opportunity to practice
The Interviewer’s Manual of the Survey Research Center provides the
instructions on the recording of interviews. Some of its suggestions
 Recording answers to open-ended-response questions follow:
 Record the responses during the interview.
 Use the respondent’s own words.
 Do not summarize or paraphrase the respondent’s answer.
 Include everything that pertains to the question objectives.
 Include all of your probes;
The basics of Effective Interviewing
Interviewing is a skilled occupation; not everyone can do it, and even few can
do it extremely well. A good interviewer observes the following principles:
1. Have integrity and be honest.
2. Have patience and tact.
84

3. Pay attention to accuracy and detail.


4. Exhibit the real enquiry at hand, but keep your own opinions to yourself.
5. Be a good listener.
6. Keep the inquiry and respondent’s responses confidential. Respect other’s
rights
1. Terminating the Interview
The final aspect of training deals with instructing the interviewers on how to
close the interview. Fieldworkers should not close the i nterview before pertinent
information has been secured. The interviewer whose departure hasty will not be
able to record those spontaneous comments responds sometimes offer after all
formal questions have been asked. Avoiding hasty departures is also a matter of
courtesy.
Fieldworkers should also answer to the best of their ability any quest the
respondent, concerning the nature and purpose of the study. Beat the fieldworker
may be required to reinterview the respondent at some his time, he or she should
leave the respondent with a positive feeling about having cooperated in a
worthwhile undertaking. It is extremely important thank the respondent for his or
her cooperation.
7.5 Fieldwork Management
Managers of the field operation select, train, supervise, and control
fieldworkers. Our discussion of fieldwork principles mentioned selection and
training. This section investigates the tasks of the fieldwork manager in greater
detail.
Briefing Session for Experienced Interviewers
After interviewers have been trained in fundamentals, and even when they
have become experienced, it is always necessary to inform workers about the
individual project. Both experienced and inexperienced fieldworkers must be
instructed on the background of the sponsoring organization, sampling techniques,
asking questions, callback procedures, and other matters specific to the project.
If there are special instructions—for example, about using show cards or video
equipment or about restricted interviewing times—these should also be covered
during the briefing session. Instructions for handling certain key questions are
always important. For example, the following fieldworker instructions appeared in a
survey of institutional investors who make buy-and sell decisions about stocks for
banks, pension funds, and the like.
A briefing session for experienced interviewers might go like, All interviewers
report to the central office, where the background of the firm and the general aims
of the study are briefly explained. Interviewers are not provi ded with too much
information about the p of the study, thus ensuring that they will not transmit any
preconceived notions to respondents. For example, in a survey about the banks in a
community, the interviewers would be told that the research is a bankin g study,
85

but not the name of the sponsoring bank. To train the interviewers about the
questionnaire, a field supervisor conducts an interview with another field
supervisor who acts as a respondent. The trainees observe the interviewing process,
after which they each interview and record the responses of another field
supervisor. Additional instructions are given to the trainees after the practice
interview.
Training to Avoid Procedural Errors in Sample Selection
The briefing session also covers the sampling procedure. A number of research
projects allow the interviewer to be at least partially responsible for selection of the
sample. When the fieldworker has some discretion in the selection of respondents,
the potential for selection bias exists. This is obvious in the case of quota sampling,
but less obvious in other cases.
Example: in probability sampling where every nth house is selected, the
fieldworker uses his or her discretion in identifying housing units. Avoiding
selection error may not be as simple as it sounds.
Example: in an older, exclusive neighborhood, a mansion’s coach house or
servants’ quarters may have been converted into an apartment that should be
identified as a housing unit. This type of dwelling and other unusual housing units
(apartments with alley entrances only, lake cottages, rooming houses) may be
overlooked, giving rise to selection error. Errors may also occur in the selection of
random-digit dialing samples Considerable effort should be expended in training
and supervisory control to minimize these errors.
The activities involved in collecting data in the field may be performed by the
organization needing information, by research suppliers, or by third -party field
service organizations. Proper execution of fieldwork is essential for producing
research results without substantial error.
Proper control of fieldwork begins with interviewer selection. Fieldworkers
should generally be healthy, outgoing, and well groomed. New fieldworkers must be
trained in opening the interview, aski ng the questions, probing for additional
information, recording the responses, and terminating the interview. Experienced
fieldworkers are briefed for each new project so that they are familiar with its
specific requirements. A particular concern of the briefing session is reminding
fieldworkers to adhere closely to the prescribed sampling procedures.
Careful supervision of fieldworkers is also necessary. Supervisors gather and
edit questionnaires each day. They check to see that field procedures are properly
followed and that interviews are on schedule. They also check to be sure that the
proper sampling units are used and that the proper people are responding in the
study. Finally, supervisors check for interviewer cheating and verify a portion of the
interviews by reinterviewing a certain percentage of each fieldworker’s respondents.
7.4 REVISION POINTS
1. It is important that the research administrator select capable people who
may be entrusted to collect the data
86

2. Interviewer bias may enter in if the fieldworker’s clothing or physical


appearance is unattractive or unusual
3. The objective of training is to ensure that the data collection instrument is
administered uniformly by all field investigators.
4. The goal of training sessions is to ensure that each responde nt is provided
with common information
5. The analyst who fails to instruct fieldworkers in the techniques of recording
answers for one study rarely forgets to do so in the second study.
7.5 INTEXT QUESTIONS
1. What Qualities should a field worker possess?
2. What is the proper method of asking Questions?
3. When should an interviewer Probe? Give examples of how probing should
be done?
4. How should an Interviewer terminate the interview?
5. What are the Qualities of the Interviewer make him more effective?
7.6 SUMMARY
An important aspect of the field survey is interviewing. This lesson outlined the
importance of training for new interviewers. In this chapter five major principles for
asking questions have been dealt in detail.
7.7 TERMINAL EXERCISES
1. The objective of the unstructured interview is to bring some ------- to the
surface
2. ------------ are those conducted when it is known at the outset what
information is needed.
3. Many ------------- are conducted through structured telephone interviews.
4. ------------- may be regarded as a form of interview on paper.
5. ------------------ helps to determine the operating budget for the project and
permits us to allocate resources to' the activities.
7.8 SUPPLEMENTARY MATERIALS
1. http://www.crest.fr/congres-afs/basinska.pdf
2. http://yyy.rsmas.miami.edu/IAI/Inst2000/lectures/wood_jul20/reading/qu
al_appr_1.pdf
7.9 ASSIGNMENTS
1. Discuss interview as a technique of data collection
2. ”Interviews introduce more bias than does the use of questionnaire” -
Critically examine the following
87

7.10 SUGGESTED READINGS/ REFERENCE BOOKS


1. Homan, R. (1991) The Ethics of Social Research, London: Longman.
2. Lee-Trewick, G. and Linkogle, S. (eds.) (2000) Danger in the Field: Risk and
Ethics in Social Research, London: Routledge.
7.11 LEARNING ACTIVITIES
Visit marketing research association website and examine the ethical codes
relating to data collection. Write a brief report.
7.8 KEYWORDS
Non-participant observation: When you, as a researcher, do not get involved in
the activities of the group but remain a passive observe r, watching and listening to
its activities and interactions and drawing conclusions from them, this is called
non-participant observation.
Interviewing: is one of the commonly used methods of data collection in the
social sciences. Any person-to-person interaction, either face to face or otherwise,
between two or more individuals with a specific purpose in mind is called an
interview. It involves asking questions of respondents and recording their answers.
Interviewing spans a wide spectrum in terms of its structure. On the one hand, it
could be highly structured and, on the other, extremely flexible, and in between it
could acquire any form.
Interview guide : A list of issues, topics or discussion points that you want to
cover in an in-depth interview is called an interview guide. Note that these points
are not questions. It is basically a list to remind an interviewer of the areas to be
covered in an interview.
Code book: A listing of a set of numerical values (set of rules) that you decided
to assign to answers obtained from respondents in response to each question is
called a code book.
Evaluation: It is a process that is guided by research principles for reviewing
an intervention or programme in order to make informed decisions about its
desirability and/or identifying changes to enhance its efficiency and effectiveness.

88

LESSON – 8

SAMPLING
8.1 INTRODUCTION
Sampling is the statistical process of selecting a subset (called a "sample") of a
population of interest for purposes of making observations and statistical
inferences about that population. Social science research is generally about
inferring patterns of behaviors within specific populations. We cannot study entire
populations because of feasibility and cost constraints, and hence, we must select a
representative sample from the population of interest for observati on and analysis.
It is extremely important to choose a sample that is truly representative of the
population so that the inferences derived from the sample can be generalized back
to the population of interest.
8.2 OBJECTIVES
After studying this lesson, you will be able to understand:
 Sampling
 Sampling error
 Sampling technique
 Sampling size estimation
8.3 CONTENTS
8.3.1 Sampling Issues
8.3.2 The sampling process
8.3.3 Determination of sample size
8.3.4 mImportance of sampling design and sample size
8.3.1 Sampling Issues
One of the questions researchers face when developing a research plan is who
will be chosen to be research participants. The word ‘population’ is commonly used
to define everyone of interest who could be possibly included in a research stud y.
Researchers may define a population by geographic area. In addition, they may also
define a population using such demographic data as age, gender, income, or
ethnicity. Because marketing promotion is often targeted based on psychographic
segmentation, researchers may also define a population based on interests, values
or lifestyles. Product usage can also be a means for defining a population, such as
nonusers, occasional users and frequent users. Of course these variables can also
be used in combination. The resulting
This fact might be how many consumers would buy a product, what type of
new promotion would work best, the effect of pricing on purchase behavior, or the
best store in which to distribute a product. Of course if a 100 per cent accurate
answer is needed, researchers must ask every person who is included in a
population.
89

Asking everyone in a population is called a census. Conducting a census is


possible if the number of people from whom information is needed is small and all
of the members can be reached. For example, a professor might wonder if students
in his class study each evening. A census could be used for surveying all the
members of this class about their plans for the evening. However, if even one
student is absent due to illness the census would be incomplete. Even if everyone is
present, if one student refuses to answer this question, again, the census would be
incomplete.
Obviously, there are problems with trying to conduct a census. Some people
will not be present and some may refuse to participate. A census might be
conducted despite these difficulties in reaching everyone but any errors in the
answers obtained from this research would be very expensive. For example, a
medical equipment supplier might be developing a piece of equi pment that would
be used by only a very few doctors treating a rare condition. If the cost of developing
such equipment for the marketplace is very expensive, the company may wish to
ensure that the design of the equipment is exactly right so that it will be purchased.
When only a small number of people are in the population, such as a few doctors,
conducting a census is possible.
However, most research will involve the sampling of a population rather than a
census. For example, researchers might be developing a new toothpaste flavor that
they plan to introduce in France. Asking everyone in the population will simply take
too much time and money. In addition, the company commissioning the research
does not need to know with 100 per cent accuracy that the new flavor will be
acceptable. Less accuracy will still provide the information as to whether the flavor
will be acceptable to a significant percentage of adults in France. Thus confused
with other possible responses. For example, if the question requires a y es or no
answer whether a respondent is to make a tick, cross out, circle or underline
should be clearly stated. In addition, the blanks for completion of open -ended
questions should be far enough apart to ensure there is no confusion when an
answer is indicated.
Even if participants and respondents are very careful when administering
surveys, there are still occasions for error. Once surveys are collected all the
information must be input into a computer software program. Even if the number
of surveys is so low that computer software is not used, the responses must still be
counted and tabulated. Because this work requires a low level of skill, clerical
assistants often perform the data entry task. And because of the repetitious nature
of this task, errors can result when the responses are input incorrectly. This is a
difficult type of error to eliminate, but careful hiring can help reduce errors.
Because of the possibility of data entry error the data should be doubled checked
once they are entered.
90

Data recording errors and solutions


• A researcher makes a wrong entry: training and motivation
• A respondent makes a wrong entry: clear design and instructions
• Data entry personnel make a wrong entry: careful hiring
Selection errors
Selection errors occur when the correct population has been chosen, but the
sample taken from the population is not representative of the entire population. For
example, a university may decide to ask students who are enrolled in art classes
what additional classes they should offer. A selection error occurs when the
students who are asked are still not representative of the whole student population.
For example, a young male student employed to conduct the survey might view the
task as a good opportunity to chat up lots of young women. As a result, a much
larger population of women then men will be included in the sample. If women want
different types of classes than men, this could result in the survey returning
inaccurate information due to selection error.
8.3 The Sampling Process

The sampling process

The sampling process comprises of several stage. The first stage is defining the
target population. A population can be defined as all people or items (unit of
analysis) with the characteristics that one wishes to study. The unit of analysis
may be a person, group, organization, country, object, or any other entity that you
wish to draw scientific inferences about. Sometimes the population is obvious. For
example, if a manufacturer wants to determine whether finished goods
manufactured at a production line meets certain quality requirements or must be
scrapped and reworked, then the population consists of the entire set of finished
goods manufactured at that production facility. At other times, the target
population may be a little harder to understand. If you wish to identify the primary
91

drivers of academic learning among high school students, then what is your target
population: high school students, their teachers, school principals, or parents? The
right answer in this case is high school students, because you are interested in
their performance, not the performance of their teachers, parents, or schools.
Likewise, if you wish to analyze the behavior of roulette wheels to identify biased
wheels, your population of interest is not different observatio ns from a single
roulette wheel, but different roulette wheels (i.e., their behavior over an infinite set
of wheels).
The second step in the sampling process is to choose a sampling frame. This
is an accessible section of the target population (usually a l ist with contact
information) from where a sample can be drawn. If your target population is
professional employees at work, because you cannot access all professional
employees around the world, a more realistic sampling frame will be employee lists
of one or two local companies that are willing to participate in your study. If your
target population is organizations, then the Fortune 500 list of firms or the
Standard & Poor's (S&P) list of firms registered with the New York Stock exchange
may be acceptable sampling frames.
Note that sampling frames may not entirely be representative of the population
at large, and if so, inferences derived by such a sample may not be generalizable to
the population. For instance, if your target population is organizational employees
at large (e.g., you wish to study employee self-esteem in this population) and your
sampling frame is employees at automotive companies in the American Midwest,
findings from such groups may not even be generalizable to the American workforce
at large, let alone the global workplace. This is because the American auto industry
has been under severe competitive pressures for the last 50 years and has seen
numerous episodes of reorganization and downsizing, possibly resulting in low
employee morale and self-esteem. Furthermore, the majority of the American
workforce is employed in service industries or in small businesses, and not in
automotive industry. Hence, a sample of American auto industry employees is not
particularly representative of the American workforce. Likewise, the Fortune 500
list includes the 500 largest American enterprises, which is not representative of all
American firms in general, most of which are medium and small -sized firms rather
than large firms, and is therefore, a biased sampling frame. In contrast, the S&P
list will allow you to select large, medium, and/or small companies, depending on
whether you use the S&P large-cap, mid-cap, or small-cap lists, but includes
publicly traded firms (and not private firms) and hence stil l biased. Also note that
the population from which a sample is drawn may not necessarily be the same as
the population about which we actually want information. For example, if a
researcher wants to the success rate of a new "quit smoking" program, then the
target population is the universe of smokers who had access to this program, which
may be an unknown population. Hence, the researcher may sample patients
arriving at a local medical facility for smoking cessation treatment, some of whom
92

may not have had exposure to this particular "quit smoking" program, in which
case, the sampling frame does not correspond to the population of interest.
The last step in sampling is choosing a sample from the sampling frame
using a well-defined sampling technique. Sampling techniques can be grouped into
two broad categories: probability (random) sampling and non -probability sampling.
Probability sampling is ideal if generalizability of results is important for your
study, but there may be unique circumstances where non -probability sampling can
also be justified. These techniques are discussed below.
PROBABILITY SAMPLING
Probability sampling is a technique in which every unit in the population has a
chance (non-zero probability) of being selected in the sample, and this chance can
be accurately determined. Sample statistics thus produced, such as sample mean
or standard deviation, are unbiased estimates of population parameters, as long as
the sampled units are weighted according to their probability of selection. All
probability sampling have two attributes in common: (1) every unit in the
population has a known non-zero probability of being sampled, and (2) the
sampling procedure involves random selection at some point. The different types of
probability sampling techniques include:
Simple random sampling: In this technique, all possible subsets of a population
(more accurately, of a sampling frame) are given an equal probability of being
selected. The probability of selecting any set of n units out of a total of N units in a
sampling frame is WC„. Hence, sample statistics are unbiased estimates of
population parameters, without any weighting. Simple random sampling involves
randomly selecting respondents from a sampling frame, but with large sampling
frames, usually a table of random numbers or a computerized random number
generator is used. For instance, if you wish to select 200 firms to survey from a list
of 1000 firms, if this list is entered into a spreadsheet like Excel, you can use
Excel's RAND() function to generate random numbers for each of the 1000 clients
on that list. Next, you sort the list in increasing order of their corresponding
random number, and select the first 200 clients on that sorted list. This is the
simplest of all probability sampling techniques; however, the simplicity is also the
strength of this technique. Because the sampling frame is not subdivided or
partitioned, the sample is unbiased and the inferences are most generalizable
amongst all probability sampling techniques.
Systematic sampling: In this technique, the sampling frame is ordered
according to some criteria and elements are selected at regular intervals through
that ordered list. Systematic sampling involves a random start and then proceeds
with the selection of every k th element from that point onwards, where k = N/n,
where k is the ratio of sampling frame size N and the desired sample size n, and is
formally called the sampling ratio. It is important that the starting point is not
automatically the first in the list, but is instead ran domly chosen from within the
first k elements on the list. In our previous example of selecting 200 firms from a
93

list of 1000 firms, you can sort the 1000 firms in increasing (or decreasing) order of
their size (i.e., employee count or annual revenues), randomly select one of the first
five firms on the sorted list, and then select every fifth firm on the list. This process
will ensure that there is no overrepresentation of large or small firms in your
sample, but rather that firms of all sizes are generall y uniformly represented, as it
is in your sampling frame. In other words, the sample is representative of the
population, at least on the basis of the sorting criterion.
Stratified sampling: In stratified sampling, the sampling frame is divided into
homogeneous and non-overlapping subgroups (called "strata"), and a simple
random sample is drawn within each subgroup. In the previous example of
selecting 200 firms from a list of 1000 firms, you can start by categorizing the firms
based on their size as large (more than 500 employees), medium (between 50 and
500 employees), and small (less than 50 employees). You can then randomly select
67 firms from each subgroup to make up your sample of 200 firms. However, since
there are many more small firms in a sampling frame than large firms, having an
equal number of small, medium, and large firms will make the sample less
representative of the population (i.e., biased in favor of large firms that are fewer in
number in the target population). This is called non-proportional stratified sampling
because the proportion of sample within each subgroup does not reflect the
proportions in the sampling frame (or the population of interest), and the smaller
subgroup (large-sized firms) is over-sampled. An alternative technique will be to
select subgroup samples in proportion to their size in the population. For instance,
if there are 100 large firms, 300 mid-sized firms, and 600 small firms, you can
sample 20 firms from the "large" group, 60 from the "medium" group and 120 from
the "small" group. In this case, the proportional distribution of firms in the
population is retained in the sample, and hence this technique is called proportional
stratified sampling. Note that the non-proportional approach is particularly effective
in representing small subgroups, such as large -sized firms, and is not necessarily
less representative of the population compared to the proportional approach, as
long as the findings of the non-proportional approach is weighted in accordance to
a subgroup's proportion in the overall population.
Cluster sampling: If you have a population dispersed over a wide geographic
region, it may not be feasible to conduct a simple random sampling of the entire
population. In such case, it may be reasonable to divide the population into
"clusters" (usually along geographic boundaries), randomly sample a few clusters,
and measure all units within that cluster. For instance, if you wish to sample city
governments in the state of New York, rather than travel all over the state to
interview key city officials (as you may have to do with a simple random sample),
you can cluster these governments based on their counties, randomly select a set of
three counties, and then interview officials from every official in those counties.
However, depending on between-cluster differences, the variability of sample
estimates in a cluster sample will generally be higher than that of a simple random
94

sample, and hence the results are less generalizable to the population than those
obtained from simple random samples.
Matched-pairs sampling: Sometimes, researchers may want to compare two
subgroups within one population based on a specific criterion. For instance, why
are some firms consistently more profitable than other firms? To conduct such a
study, you would have to categorize a sampling frame of firms into "high profitable"
firms and "low profitable firms" based on gross margins, earnings per share, or
some other measure of profitability. You would then select a simple random sample
of firms in one subgroup, and match each firm in this group with a firm in the
second subgroup, based on its size, industry segment, and/or other matching
criteria. Now, you have two matched samples of high-profitability and low-
profitability firms that you can study in greater detail. Such matched-pairs
sampling technique is often an ideal way of understanding bipolar differences
between different subgroups within a given population.
Multi-stage sampling: The probability sampling techniques described
previously are all examples of single-stage sampling techniques. Depending on your
sampling needs, you may combine these single -stage techniques to conduct multi-
stage sampling. For instance, you can stratify a list of businesses based on firm
size, and then conduct systematic sampling within each stratum. This is a two-
stage combination of stratified and systematic sampling. Likewise, you can start
with a cluster of school districts in the state of New York, and within each cluster,
select a simple random sample of school s; within each school, select a simple
random sample of grade levels; and within each grade level, select a simple random
sample of students for study. In this case, you have a four-stage sampling process
consisting of cluster and simple random sampling.
NON-PROBABILITY SAMPLING
Nonprobability sampling is a sampling technique in which some units of the
population have zero chance of selection or where the probability of selection
cannot be accurately determined. Typically, units are selected based on certai n
non-random criteria, such as quota or convenience. Because selection is non -
random, nonprobability sampling does not allow the estimation of sampling errors,
and may be subjected to a sampling bias. Therefore, information from a sample
cannot be generalized back to the population. Types of non -probability sampling
techniques include:
Convenience sampling: Also called accidental or opportunity sampling, this is
a technique in which a sample is drawn from that part of the population that is
close to hand, readily available, or convenient. For instance, if you stand outside a
shopping center and hand out questionnaire surveys to people or interview them as
they walk in, the sample of respondents you will obtain will be a convenience
sample. This is a non-probability sample because you are systematically excluding
all people who shop at other shopping centers. The opinions that you would get
from your chosen sample may reflect the unique characteristics of this shopping
95

center such as the nature of its stores (e.g., high end-stores will attract a more
affluent demographic), the demographic profile of its patrons, or its location (e.g., a
shopping center close to a university will attract primarily university students with
unique purchase habits), and therefore may not be representative of the opinions of
the shopper population at large. Hence, the scientific generalizability of such
observations will be very limited. Other examples of convenience sampling are
sampling students registered in a certain class or sampling patients arriving at a
certain medical clinic. This type of sampling is most useful for pilot testing, where
the goal is instrument testing or measurement validation rather than obtaining
generalizable inferences.
Quota sampling: In this technique, the population is segmented into mutually-
exclusive subgroups (just as in stratified sampling), and then a non -random set of
observations is chosen from each subgroup to meet a predefined quota. In
proportional quota sampling, the proportion of respondents in each subgroup
should match that of the population. For instance, if the American population
consists of 70% Caucasians, 15% Hispanic -Americans, and 13% African-
Americans, and you wish to understand their voting preferences in an sample of 98
people, you can stand outside a shopping center and ask people their voting
preferences. But you will have to stop asking Hispanic -looking people when you
have 15 responses from that subgroup (or African -Americans when you have 13
responses) even as you continue sampling other ethnic groups, so that the ethnic
composition of your sample matches that of the general American population. Non-
proportional quota sampling is less restrictive in that you don't have to achieve a
proportional representation, but perhaps meet a minimum size in each subgroup.
In this case, you may decide to have 50 respondents from each of the three ethnic
subgroups (Caucasians, Hispanic-Americans, and African-Americans), and stop
when your quota for each subgroup is reached. Neither type of quota sampling will
be representative of the American population, since depending on whether your
study was conducted in a shopping center in New York or Kansas, your results may
be entirely different. The non-proportional technique is even less representative of
the population but may be useful in that it allows capturing the opinions of small
and underrepresented groups through oversampling.
Expert sampling: This is a technique where respondents are chosen in a non -
random manner based on their expertise on the phenomenon being studied. For
instance, in order to understand the impacts of a new governmental policy such as
the Sarbanes-Oxley Act, you can sample an group of corporate accountants who
are familiar with this act. The advantage of this approach is that since experts tend
to be more familiar with the subject matter than non experts, opinions from a
sample of experts are more credible than a sample that includes both experts and
non-experts, although the findings are still not generalizable to the overall
population at large.
Snowball sampling: In snowball sampling, you start by identifying a few
respondents that match the criteria for inclusion in your study, and then ask them
96

to recommend others they know who also meet your selection criteria. For instance,
if you wish to survey computer network administrators and you know of only one or
two such people, you can start with them and ask them to recommend others who
also do network administration. Although this method hardly leads to
representative samples, it may sometimes be the only way to reach hard-to-reach
populations or when no sampling frame is available.
8.4 Determination of Sample Size
Determination of appropriate sample size is crucial part of any Business
Research. The decision on proper sample size tremendously requires the use of
statistical theory. When a business research report is been evaluated, the
evaluation start with the question of How big the Sample size?
Having discussed various sampling designs it is important to focus the
attention on Sample Size. Suppose we select a sample size of 30 from the
population of 3000 through a simple random sampling procedure. Will we able to
generalize the findings to the population with confidence? So in this case what is
the sample size that would be required to carry out the research?
It is the known fact that larger the sample size, the more accurate the research
is. In fact this is the fact based on the statistics. According to this fact, increasing
the sample size decreases the width of the confi dence interval at a given confidence
level. When the Standard deviation of the population is unknown, a confidence
interval is calculated by using the formula
Confidence Interval µ = X ± KSX
S
Where as x = S / √n
In sum, choosing the appropriate sampling plan is one of the important
research design decisions the researcher has to make. The choice of a specific
design will depend broadly on the goal of research, the characteristics of the
population, and considerations of cost.
Issues of Precision and Confidence in determining Sample size
We now need to focus attention on the second aspect of the sampling design
issue—the sample size. Suppose we select 30 people from a population of 3,000
through a simple random sampling procedure. Will we be able to generalize our
findings to the population with confidence? What is the sample size that would be
required to make reasonably precise generalizations with confidence? What do
precision and confidence mean?
A reliable and valid sample should enable us to generalize the findings from
the sample to the population under investigation. No sample statistic (X, for
instance) is going to be exactly the same as the population parameter (S X), no
matter how sophisticated the probability sampling design is. Remember that the
very reason for a probability design is to increase the probability that the sample
statistics will be as close as possible to the population parameters!
97

Precision
Precision refers to how close our estimate is to the true population
characteristic. Usually, we would estimate the population parameter to fall within a
range, based on the sample estimate.
Example: From a study of a simple random sample of 50 of the total 300
employees in a workshop, we find that the average daily production rate per person
is 50 pieces of a particular product (X = 50). We might then (by doing certain
calculations, as we shall see later) be able to say that the true average daily
production of the product (X) would lie anywhe re between 40 and 60 for the
population of employees in the workshop. In saying this, we offer an interval
estimate, within which we expect the true population mean production to be (µ = 50
± 10). The narrower this interval, the greater the precision. For instance, if we are
able to estimate that the population mean would fall anywhere between 45 and 55
pieces of production (µ = 50 ± 5) rather than 40 and 60 (µ = 50 ± 10), then we would
have more precision. That is, we would now estimate the mean to lie within a
narrower range, which in turn means that we estimate with greater exactitude or
precision.
Precision is a function of the range of variability in the sampling distribution of
the sample mean. That is, if we take a number of different samples from a
population, and take the mean of each of these, we will usually find that they are
all different, are normally distributed, and have a dispersion associated with them.
Even if we take only one sample of 30 subjects from the population, we will still be
able to estimate the variability of the sampling distribution of the sample mean.
This variability is called the standard error, denoted by .S’. The standard error is
calculated by the following formula:
S
x = S / √n
Where S is the standard deviation of the sample, n is the sample size, and S X
indicates the standard error or the extent of precision offered by the sample.
In sum, the closer we want our sample results to reflect the population
characteristics, the greater will be the precision we would aim at. The greater the
precision required, the larger is the sample size needed, especially when the
variability in the population itself is large.
Confidence
Whereas precision denotes how close we estimate the population parameter
based on the sample statistic, confidence denotes how certain we are that our
estimates will really hold true for the population. In the previous example of
production rate, we know we are more precise when we estimate the true mean
production (µ) to fall somewhere between 45 and 55 pieces, than somewhere
between 40 and 60.
98

In essence, confidence reflects the level of certainty with which we can state
that our estimates of the population parameters, based on our sample statistics will
hold true. The level of confidence can range from 0 to 100% . A 95% confidence is
the conventionally accepted level for most business research, most commonly
expressed by denoting the significance level as p = .05. In other words, we say that
at least 95 times out of 100, our estimate will reflect the true population
characteristic,
In sum, the sample size, n, is a function of
1. The variability in the population
2. Precision or accuracy needed
3. Confidence level desired
4. Type of sampling plan used—for example, sample random sampling versus
stratified random sampling
1. thus becomes necessary for researchers to consider at least four points
while making decisions on the sample size needed to do the research
2. Much precision is really needed in estimating the population characteristics
interest—that is, what is the margin of allowable error?
3. How much confidence is really needed—that is, how much chance can we
take of making errors in estimating the population parameters?
4. To what extent is there variability in the population on the characteristic s
investigated?
5. What is the cost—benefit analysis of increasing the sample size?
Determining the Sample Size
Now that we are aware of the fact that the sample size is governed by the
extent of precision and confidence desired, how do we determine the sample size
required for our research? The procedure can be illustrated through an example.
Suppose a manager wants to be 95% confident that the expected monthly
withdrawals in a bank will be within a confidence interval of ± $500. Example of a
sample of clients indicates that the average withdrawals made by them have a
standard deviation of $3,500. What would be the sample size needed in this case?
We noted earlier that the population mean can be estimated by using the
formula:
S
µ= X ± K x
Since the confidence level needed here is 95% , the applicable K value is 1.96 (t
table). The interval estimate of ± $500 will have to encompass a dispersion of (1.96
x standard error). That is,
S
500 = 1.96 × x
99

S
500/1.96 = 255.10
x
We already know that
S
x = S / √n
255.10= 3500/ √n
n = 188
The sample size needed in the above was 188. Let us say that this bank has
the total clientele of only 185. This means we cannot sample 188 clients. We can in
this case apply the correction formula and see what sample size would be needed to
have the same level of precision and confidence given the fact that we have a total
of only 185 clients. The correction formula is as follows:
S
x = S/√n × √N – n /N-1
Where N is the total number of elements in the population, n is the sample
size to be estimated, Sx is the standard error of estimate of the mean, and S is the
standard deviation of the sample mean.
Applying the correlation formula, we fin d that
255.10 = 3500/√n × √185-n/184
n = 94
We would now sample 94 of the total 185 clients.
To understand the impact of precision and/or confidence on the sample size,
let us try changing the confidence level required in the bank withdrawal exercise
which needed a sample size of 188 for a confidence level of 95% . Let us say that the
bank manager now wants to be 99% sure that the expected monthly withdrawals
will be within the interval of ±$500. What will be the sample size now needed?
S
will now be 500/2.576 = 194.099
x
194.099 = 3500/√n
n = 325
The sample has now to be increased 1.73 times (from 188 to 325) to increase
the confidence level from 95% to 99% . It is hence a good idea to think through how
much precision and confidence one really needs, before determining the sample size
for the research project.
So far we have discussed sample size in the context of precision and
confidence with respect to one variable only. However, in research, the theoretical
framework has several variables of interest, and the question ari ses how one should
come up with a sample size when all the factors are taken into account.
100

Krejcie and Morgan (1970) greatly simplified size decision by providing a table
that ensures a good decision model. The Table provides that generalized scientific
guideline for sample size decisions. The interested Student is advised to read
Krejcie and Morgan (1970) as well as Cohen (1969) for decisions on sample size.
Table on Sample Size for a given Population Size
N S N S N S
10 10 220 140 1200 291
15 14 230 144 1300 297
20 19 240 148 1400 302
25 24 250 152 1500 306
30 28 260 155 1600 310
35 32 270 159 1700 313
40 36 280 162 1800 317
45 40 290 165 1900 320
50 44 300 169 2000 322
55 48 320 175 2200 327
60 52 340 181 2400 331
65 56 360 186 2600 335
70 59 380 191 2800 338
75 63 400 196 3000 341
80 66 420 201 3500 346
85 70 440 205 4000 351
90 73 460 210 4500 354
95 76 480 214 5000 357
100 80 500 217 6000 361
110 86 550 226 7000 364
120 92 600 234 8000 367
130 97 650 242 9000 368
140 103 700 248 10000 370
150 108 750 254 15000 375
160 113 800 260 20000 377
170 118 850 265 30000 379
180 123 900 269 40000 380
190 127 950 274 50000 381
200 132 1000 278 75000 382
210 136 1100 285 100000 384
8.5 Importance of Sampling Design and Sample Size
It is now possible to see how both sampling design and the sample size are
important to establish the representativeness of the sample for generalizability. If
the appropriate sampling design is not used, a large sample size will not, in itself,
allow the findings to be generalized to the population. Likewise, unless the sample
101

size is adequate for the desired level of precision and confidence, no sampling
design, however sophisticated, can be useful to the researcher in meeting the
objectives of the study.
Hence, sampling decisions should consider both the sampling design and the
sample size. Too large a sample size, however (say, over 500) could also become a
problem in as much as we would be prone to committing Type II errors. Hence,
neither too large nor too small sample sizes help research projects.
Roscoe (1975) proposes the following rules of thumb for determining sample
size
1. Sample sizes larger than 30 and less than 500 are appropriate for most
research.
2. Where samples are to be broken into sub samples; (male/females, juniors/
seniors, etc.), a minimum sample size of 30 for each category is necessary.
3. In multivariate research (including multiple regression analyses), the
sample size should be several times (preferably 10 times or more) as large
as the number of variables in the study.
4. For simple experimental research with tight experimental controls
(matched pairs, etc.), successful research is possible with samples as small
as 10 to 20 in size.
8.4 REVISION POINTS
1. Sampling is the statistical process of selecting a subset (called a "sample") of
a population of interest for purposes of making observations and statistical
inferences about that population.
2. The word ‘population’ is commonly used to define everyone of interest who
could be possibly included in a research study.
3. Selection errors occur when the correct population has been chosen, but the
sample taken from the population is not representative of the entire
population.
4. A population can be defined as all people or items (unit of analysis) with the
characteristics that one wishes to study.
5. A sampling frame is an accessible section of the target population (usually a
list with contact information) from where a sample can be drawn.
6. Probability sampling is a technique in which every unit in the population
has a chance (non-zero probability) of being selected in the sample, and this
chance can be accurately determined.
7. Nonprobability sampling is a sampling technique in which some units of the
population have zero chance of selection or where the probability of selection
cannot be accurately determined
102

8.5 INTEXT QUESTIONS


1. Define the concept of sampling design.
2. Describe the steps involved in sampling design.
3. Discuss the criteria for selecting a sampling procedure.
4. Distinguish between probability and non -probability sampling.
5. How is a random sample selected?
6. Explain complex random sampling designs.
8.6 SUMMARY
The lesson has given the definitions of some basic terms used in sampling.
Then advantages and disadvantages of sampling has been mentioned. This is
followed by a discussion on the steps involved in a sampling process. Basic issues
involved in the sampling design and sampling size decision were also discussed.
8.7 TERMINAL EXERCISES
1. A sample is a part of a .................... population.
2. Sampling .................... is the list of elements from which the sample is
actually drawn.
3. A sampling plan should clearly specify the .................... population.
4. The sample size depends upon the .................... available with the
company.
8.8 SUPPLEMENTARY MATERIALS
1. http://www.investopedia.com/terms/s/sampling.asp
2. https://onlinecourses.science.psu.edu/stat100/node/18
8.9 ASSIGNMENTS
1. A team of medico research experts feels confident that a new drug they have
developed will cure about 80% of the patients. How large should the sample
size be for the team to be 98% certain that the sample proportion of cure is
within plus and minus 2% of the proportion of all cases that the drug will
cure?
2. Mr. Kishore wants to determine the average time required to complete a job
with which he is concerned. As per the last studies, the population standard
deviation is 8 days. How large should the sample be so that Mr. Kishore may
be 99% confident that the sample average may remain within 2 days of the
average?
8.10 SUGGESTED READINGS/ REFERENCE BOOKS
1. Henry, G. (1990) Practical Sampling, Newbury Park, CA: Sage.
2. Huff, D. (1994) How to Lie With Statistics, NY: Norton.
3. Field, A. (2000) Discovering Statistics Using SPSS for Windows: Advanced
Techniques for Beginners, London: Sage.
4. Owen, F. and Jones, R. (1994) Statistics, 4th edition, London: Pitman.
103

5. Kothari, C.R.,1985, Research Methodology- Methods and Techniques, New


Delhi, Wiley Eastern Limited.
6. Kumar, Ranjit, 2005, Research Methodology-A Step-by-Step Guide for
Beginners,(2nd.ed.), Singapore, Pearson Education.
8.11 LEARNING ACTIVITIES
1. Visit the gallup organization website (www.gallup.com), identify some of the
survey recently completed by the gallup. What are the sample sizes in these
surveys and how they are determined?
8.8 KEYWORDS
Random sampling: A method of selecting samples so that all members of the
population have the same chance of being selected for a sample
Sample: A relatively small subset of a population, intended to represent the
population; a subset of the complete group of scores found in any particular
situation
Sampling error: The difference, due to random chance, between a sample
statistic and the population parameter it represents
Sampling with replacement: A sampling procedure in which previously selected
individuals or events are returned to the population before any additional samples
are selected
Sampling without replacement: A sampling procedure in which previously
selected individuals or events are not returned to the population before additional
samples are selected










104

LESSON – 9

HYPOTHESIS TESTING
9.1 INTRODUCTION
Statistical hypothesis is an assumption about a population parameter. This
assumption may or may not be true.
The best way to determine whether a statistical hypothesis is true is to
examine the entire population. Since this is often impractical, researchers typically
examine a random sample from the population. If the sample data are consistent
with the statistical hypothesis, the hypothesis is accepted; if not, the hypothesis is
rejected.
There are two types of statistical hypotheses.
 Null hypothesis. The null hypothesis is usually the hypothesis that sample
observations result purely from chance effects.
 Alternative hypothesis. The alternative hypothesis is the hypothesis that
sample observations are influenced by some non -random cause.
For example, suppose we wanted to determine whether a coin was fair and
balanced. A null hypothesis might be that half the flips would result in Heads and
half, in Tails. The alternative hypothesis might be that the number of Heads and
Tails would be very different. Suppose we flipped the coin 50 times, resulting in 40
Heads and 10 Tails. Given this result, we would be inclined to reject the null
hypothesis and accept the alternative hypothesis.
9.2 OBJECTIVES
After studying this lesson, you will be able to understand:
 Hypothesis testing
 Types of hypothesis
 Meaning of probability value
 Type I error and Type II error
 Chi square test
9.3 CONTENTS
9.3.1 Hypothesis test
9.3.2 The precise meaning of the probability value
9.3.3 Statistical and practical significance
9.3.4 Type I and II errors
9.3.5 One and two tailed test
9.3.6 Confidence interval and hypothesis testing
9.3.7 Defining the decision rule and region of acceptance
9.3.8 Other considerations
9.3.9 Chi square test
105

9.3.1 Hypothesis Tests


Statisticians follow a formal process to determine whether to accept or reject a
null hypothesis, based on sample data. This process, called hypothesis testing,
consists of four steps.
 Formulate the hypotheses. This involves stating the null and alternative
hypotheses. The hypotheses are stated in such a way that they are
mutually exclusive. That is, if one is true, the other must be false; and vice
versa.
 Identify the test statistic. This involves specifying the statistic (e.g., a mean
score, proportion) that will be used to assess the validity of the null
hypothesis.
 Formulate a decision rule. A decision rule is a procedure that the
researcher uses to decide whether to accept or reje ct the null hypothesis.
 Accept or reject the null hypothesis. Use the decision rule to evaluate the
test statistic. If the statistic is consistent with the null hypothesis, accept
the null hypothesis; otherwise, reject the null hypothesis.
This section provides an introduction to hypothesis testing. Basic analysis
involves some hypothesis testing. Examples of hypotheses generated in marketing
research abound
 The department store is being patronized by more than 10 percent of the
households
 The heavy and light users of a brand differ in terms of psychographic
characteristics
 One hotel has a more upscale image than its close competitor.
 Familiarity with a restaurant results in greater preference for that
restaurant.
The null hypothesis is an hypothesis about a population parameter. The
purpose of hypothesis testing is to test the viability of the null hypothesis in the
light of experimental data. Depending on the data, the null hypothesis either will or
will not be rejected as a viable possibility.
Consider a researcher interested in whether the time to respond to a tone is
affected by the consumption of alcohol. The null hypothesis is that µ 1 - µ2 = 0 where
µ1 is the mean time to respond after consuming alcohol and µ 2 is the mean time to
respond otherwise. Thus, the null hypothesis concerns the parameter µ 1 - µ2 and
the null hypothesis is that the parameter equals zero.
The null hypothesis is often the reverse of what the experimenter actually
believes; it is put forward to allow the data to contradict it. In the experiment on the
effect of alcohol, the experimenter probably expects alcohol to have a harmful effect.
If the experimental data show a sufficiently large effect of alcohol, then the null
hypothesis that alcohol has no effect can be rejected.
106

It should be stressed that researchers very frequently put forward a null


hypothesis in the hope that they can discredit it. For a second example, consider an
educational researcher who designed a new way to teach a particular concept in
science, and wanted to test experimentally whether this new method worked better
than the existing method. The researcher would design an experiment comparing
the two methods. Since the null hypothesis would be that there is no difference
between the two methods, the researcher would be hoping to reject the null
hypothesis and conclude that the method he or she developed is the better of the
two.
The symbol H0 is used to indicate the null hypothesis. For the example ju st
given, the null hypothesis would be designated by the following symbols
H0: µ1 - µ2 = 0
or by
H0: μ 1 = μ 2.
The null hypothesis is typically a hypothesis of no difference as in this example
where it is the hypothesis of no difference between population means. That is why
the word "null" in "null hypothesis" is used -- it is the hypothesis of no difference.
Despite the "null" in "null hypothesis," there are occasions when the parameter
is not hypothesized to be 0. For instance, it is possible for the null hypothesis to be
that the difference between population means is a particular value. Or, the null
hypothesis could be that the mean SAT score in some population is 600. The null
hypothesis would then be stated as: H 0: μ = 600. Although the null hypotheses
discussed so far have all involved the testing of hypotheses about one or more
population means, null hypotheses can involve any parameter.
An experiment investigating the correlation between job satisfaction and
performance on the job would test the null hypothesis that the population
correlation (ρ) is 0.
Symbolically, H0: ρ = 0. Some possible null hypotheses are given
below:
H0: μ=0
H0: μ=10
H0: μ 1 - μ 2 = 0
H0: π = .5
H0: π1 - π2 = 0
H0: μ 1 = μ 2 = μ 3
H0: ρ1- ρ2= 0
1. The first step in hypothesis testing is to specify the null hypothesis (H 0) and
the alternative hypothesis (H1 ). If the research concerns whether one method
of presenting pictorial stimuli leads to better recognition than another, the
107

null hypothesis would most likely be that there is no difference between


methods (H0: μ 1 - μ 2 = 0). The alternative hypothesis would be H 1: μ 1 ≠ μ 2. If
the research concerned the correlation between grades and SAT scores, the
null hypothesis would most likely be that there is no correlation (H 0: ρ= 0).
The alternative hypothesis would be H1: ρ ≠ 0.
2. The next step is to select a significance level. Typically the 0.05 or the 0.01
level is used.
3. The third step is to calculate a statistic analogous to the parameter specified
by the null hypothesis. If the null hypothesis were defined by the parameter
μ 1- μ 2 , then the statistic M1 - M2 would be computed.
4. The fourth step is to calculate the probability value (often called the p value).
The p value is the probability of obtaining a statistic as different or more
different from the parameter specified in the null hypothesis as the statistic
computed from the data. The calculations are made assuming that the null
hypothesis is true.
5. The probability value computed in Step 4 is compared with the significance
level chosen in Step 2. If the probability is less than or equal to the
significance level, then the null hypothesis is rejected; if the probability is
greater than the significance level then the null hypothesis is not rejected.
When the null hypothesis is rejected, the outcome is said to be "statistically
significant"; when the null hypothesis is not rejected then the outcome is
said be "not statistically significant."
6. If the outcome is statistically significant, then the null hypothesis is rejected
in favor of the alternative hypothesis. If the rejected null hypothesis were
that μ 1- μ 2 = 0, then the alternative hypothesis would be that μ 1≠ μ 2. If M1
were greater than M2 then the researcher would naturally conclude that μ 1 ≥
μ2.
7. The final step is to describe the result and the statistical conclusi on in an
understandable way. Be sure to present the descriptive statistics as well as
whether the effect was significant or not. For example, a significant
difference between a group that received a drug and a control group might
be described as follow:
Subjects in the drug group scored significantly higher (M = 23) than did
subjects in the control group (M = 17), t(18) = 2.4, p = 0.027. The statement
that "t(18) =2.4" has to do with how the probability value (p) was calculated.
A small minority of researchers might object to two aspects of this wording.
First, some believe that the significance level rather than the probability level
should be reported. The argument for reporting the probability val ue is
presented in another section. Second, since the alternative hypothesis was
stated as µ1 ≠ µ2, some might argue that it can only be concluded that the
population means differ and not that the population mean.
This argument is misguided. Intuitively, there are strong reasons for inferring
that the direction of the difference in the population is the same as the difference in
108

the sample. There is also a more formal argument. A non significant effect might be
described as follows:
Although subjects in the drug group scored higher (M = 23) than did subjects
in the control group, (M = 20), the difference between means was not significant,
t(18) = 1.4, p = 0.179. It would not have been correct to say that there was no
difference between the performance of the two groups. There was a difference. It is
just that the difference was not large enough to rule out chance as an explanation
of the difference. It would also have been incorrect to imply that there is no
difference in the population. Be sure not to accept the null hypothesis.
9.3.2 The Precise Meaning of the Probability Value
There is often confusion about the precise meaning of the probability
computed in a significance test. As stated in Step 4 of the steps in hypothesis
testing, the null hypothesis (H0) is assumed to be true. The difference between the
statistic computed in the sample and the parameter specified by H0 is computed
and the probability of obtaining a difference this large or large is calculated. This
probability value is the probability of obtaining data as extreme or more extreme
than the current data (assuming H0 is true). It is not the probability of the null
hypothesis itself. Thus, if the probability value is 0.005, this does not mean that the
probability that the null hypothesis is true is .005. It means that the probability of
obtaining data as different or more different from the null hypothesis as those
obtained in the experiment is 0.005.
The inferential step to conclude that the null hypothesis is false goes as
follows: The data (or data more extreme) are very unlikely given that the null
hypothesis is true. This means that: (1) a very unlikely event occurred or (2) the
null hypothesis is false. The inference usually made is that the null hypothesis is
false.
To illustrate that the probability is not the probability of the hypothesis,
consider a test of a person who claims to be able to predict whether a coin will come
up heads or tails. One should take a rather skeptical attitude toward this claim and
require strong evidence to believe in its validity. The null hypothesis is that the
person can predict correctly half the time (H 0: π = 0.5). In the test, a coin is flipped
20 times and the person is correct 11 times. If the person has no special ability (H 0
is true), then the probability of being correct 11 or more times out of 20 is 0.41.
Would someone who was originally skeptical now believe that there is only a 0.41
chance that the null hypothesis is true? They almost certainly would not since they
probably originally thought H0 had a very high probability of being true (perhaps as
high as 0.9999). There is no logical reason for them to decrease their belief in the
validity of the null hypothesis since the outcome was perfectly consistent with the
null hypothesis.
The proper interpretation of the test is as follows: A person made a rather
extraordinary claim and should be able to provide strong evidence in support of the
109

claim if the claim is to believed. The test provided data consistent with the null
hypothesis that the person has no special ability since a person with no special
ability would be able to predict as well or better more than 40% of the time.
Therefore, there is no compelling reason to believe the extraordinary claim.
However, the test does not prove the person cannot predict better than chance; it
simply fails to provide evidence that he or she can. The probability that the null
hypothesis is true is not determined by the statistical analysis conducted as part of
hypothesis testing. Rather, the probability computed is the probability of obtaining
data as different or more different from the null hypothesis (given that the null
hypothesis is true) as the data actually obtained.
According to one view of hypothesis testing, the significance level should be
specified before any statistical calculations are performed. Then, when the
probability (p) is computed from a sign ificance test, it is compared with the
significance level. The null hypothesis is rejected if p is at or below the significance
level; it is not rejected if p is above the significance level. The degree to which p
ends up being above or below the significance level does not matter. The null
hypothesis either is or is not rejected at the previously stated significance level.
Thus, if an experimenter originally stated that he or she was using the 0.05
significance level and p was subsequently calculated to be 0.042, then the person
would reject the null hypothesis at the 0.05 level. If p had been 0.0001 instead of
0.042 then the null hypothesis would still be rejected at the 0.05 level. The
experimenter would not have any basis to be more confident that the nu ll
hypothesis was false with a p of 0.0001 than with a p of 0.041. Similarly, if the p
had been 0.051 then the experimenter would fail to reject the null hypothesis.
He or she would have no more basis to doubt the validity of the null
hypothesis than if p had been 0.482. The conclusion would be that the null
hypothesis could not be rejected at the 0.05 level. In short, this approach is to
specify the significance level in advance and use p only to determine whether or not
the null hypothesis can be rejected at the stated significance level.
Many statisticians and researchers find this approach to hypothesis testing
not only too rigid, but basically illogical. Who in their right mind would not have
more confidence that the null hypothesis is false with a p of 0.0001 then with a p of
0.042? The less likely the obtained results (or more extreme results) under the null
hypothesis, the more confident one should be that the null hypothesis is false. The
null hypothesis should not be rejected once and for all. The possibility that it was
falsely rejected is always present, and, all else being equal, the lower the p value,
the lower this possibility.
9.3.3 Statistical and Practical Significance
It is important not to confuse the confidence with which the null hypothe sis
can be rejected with size of the effect. To make this point concrete, consider a
researcher assigned the task of determining whether the video display used by
travel agents for booking airline reservations should be in color or in black and
110

white. Market research had shown that travel agencies were primarily concerned
with the speed with which reservations can be made. Therefore, the question was
whether color displays allow travel agents to book reservations faster. Market
research had also shown that in order to justify the higher price of color displays,
they must be faster by an average of at least 10 seconds per transaction. Fifty
subjects were tested with color displays and 50 subjects were tested with black and
white displays.
Subjects were slightly faster at making reservations on a color display (M =
504.7 seconds) than on a black and white display (M = 508.2) seconds. Although
the difference is small, it was statistically significant at the .05 significance level.
Box plots of the data are shown below.

The 95% confidence interval on the difference between means is:


-7.0 ≤ μcolor - μblack & white ≤ -0.1
which means that the experimenter can be confident that the color display is
between 7.0 seconds and 0.1 seconds faster. Clearly, the differe nce is not big
enough to justify the more expensive color displays. Even the upper limit of the
95% confidence interval (seven seconds) is below the minimum needed to justify the
cost (10 seconds). Therefore, the experimenter could feel confident in his or her
recommendation that the black and white displays should be used. The fact that
the color displays were significantly faster does not mean that they were much
faster. It just means that the experimenter can reject the null hypothesis that there
is no difference between the displays.
The experimenter presented this conclusion to management but management
did not accept it. The color displays were so dazzling that despite the statistical
analysis, they could not believe that color did not improve performance by at least
10 seconds. The experimenter decided to do the experiment again, this time using
100 subjects for each type of display. The results of the second experiment were
very similar to the first. Subjects were slightly faster at making reservations on a
color display (M = 504.7 seconds) than on a black and white display (M = 508.1
seconds) . This time the difference was significant at the 0.01 level rather than the
0.05 level found in the first experiment. Despite the fact that the size of the
difference between means was no larger, the difference was "more significant" due
to the larger sample size used. If the population difference is zero, then a sample
111

difference of 3.4 or larger with a sample size of 100 is less likely than a sample
difference of 3.5 or larger with a sample size of 50.
The 95% confidence interval on the difference between means is:
-5.8 ≤ μcolor - μblack & white ≤ -0.9 and the 99% interval is:
-6.6 ≤ μcolor - μblack & white ≤ -0.1
Therefore, despite the finding of a "more significant" difference between means,
the experimenter can be even more certain that the color displays are only slightly
better than the black and white displays. The second experiment shows
conclusively that the difference is less than 10 seconds.
This example was used to illustrate the following points: (1) an effect that is
statistically significant is not necessarily large enough to be of practical significance
and (2) the smaller of two effects can be "more significant" than the larger. Be
careful how you interpret findings reported in the media. If you read that a
particular diet lowered cholesterol significantly, this does not necessarily mean that
the diet lowered cholesterol enough to be of any health value. It means that the
effect on cholesterol in the population is greater than zero.
9.3.4 Type I and II errors (1 of 2)
There are two kinds of errors that can be made in significance testing: (1) a
true null hypothesis can be incorrectly rejected and (2) a false null hypothesis can
fail to be rejected. The former error is called a Type I error and the latter error is
called a Type II error. These two types of errors are defined in the table.
True State of the Null Hypothesis
Statistical Decision
H0 True H0 False
Reject H0 Type I error Correct
Do not Reject H0 Correct Type II error

The probability of a Type I error is designated by the Greek letter alpha () and
is called the Type I error rate; the probability of a Type II error (the Type II error
rate) is designated by the Greek letter beta (ß). A Type II error is only an error in the
sense that an opportunity to reject the null hypothesis correctly was lo st. It is not
an error in the sense that an incorrect conclusion was drawn since no conclusion is
drawn when the null hypothesis is not rejected.
A Type I error, on the other hand, is an error in every sense of the word. A
conclusion is drawn that the null hypothesis is false when, in fact, it is true.
Therefore, Type I errors are generally considered more serious than Type II errors.
The probability of a Type I error (α) is called the significance level and is set by the
experimenter. There is a tradeoff between Type I and Type II errors. The more an
experimenter protects himself or herself against Type I errors by choosing a low
level, the greater the chance of a Type II error. Requiring very strong evidence to
reject the null hypothesis makes it very unlikely that a true null hypothesis will be
112

rejected. However, it increases the chance that a false null hypothesis will not be
rejected, thus lowering power. The Type I error rate is almost always set at .05 or at
.01, the latter being more conservative since it requires stronger evidence to reject
the null hypothesis at the .01 level then at the .05 level.
9.3.5 One- and Two-Tailed Tests
In the section on "Steps in hypothesis testing" the fourth step involves
calculating the probability that a statistic would differ as much or more from
parameter specified in the null hypothesis as does the statistic obtained in the
experiment. This statement implies that a difference in either direction would be
counted. That is, if the null hypothesis were:
H0: μ- μ = 0 and the value of the statistic M 1- M2 were +5, then the probability
of M1- M2 differing from zero by five or more (in either direction) would be
computed. In other words, probability value would be the probabi lity that either M1-
M2 ≥ 5 or M1- M2 ≤ -5. Assume that the figure shown below is the sampling
distribution of M 1- M2.

The figure shows that the probability of a value of +5 or more is 0.036 and that
the probability of a value of -5 or less is .036. Therefore the probability of a value
either greater than or equal to +5 or less than or equal to -5 is 0.036 + 0.036 =
0.072.
A probability computed considering differences in both directions is called a
"two-tailed" probability. The name makes sense since both tails of the sampling
distribution are considered. There are situations in which an experimenter is
concerned only with differences in one direction. For example, an experimenter may
be concerned with whether or not μ 1 - μ 2 is greater than zero. However, if μ 1 - μ 2 is
not greater than zero, the experimenter may not care whether it equals zero or is
less than zero. For instance, if a new drug treatment is developed, the main issue is
whether or not it is better than a placebo. If the treatment is not better than a
placebo, then it will not be used. It does not really matter whether or not it is worse
than the placebo.
When only one direction is of concern to an experimenter, then a "one -tailed"
test can be performed. If an experimenter were only concerned with whether or not
μ 1 - μ 2 is greater than zero, then the one-tailed test would involve calculating the
probability of obtaining a statistic as greater than the one obtained in the
experiment.
113

In the example, the one-tailed probability would be the probability of obtaining


a value of M1- M2 greater than or equal to five given that the difference between
population means is zero.

The shaded area in the figure is greater than five. The figure shows that the
one-tailed probability is 0.036.
It is easier to reject the null hypothesis with a one -tailed than with a two-tailed
test as long as the effect is in the specified direction. Therefore, one-tailed tests
have lower Type II error rates and more power than do two-tailed tests. In this
example, the one-tailed probability (0.036) is below the conventional significance
level of 0.05 whereas the two-tailed probability (0.072) is not. Probability values for
one-tailed tests are one half the value for two-tailed tests as long as the effect is in
the specified direction.
One-tailed and two-tailed tests have the same Type I error rate. One -tailed
tests are sometimes used when the experimenter predicts the direction of the effect
in advance. This use of one-tailed tests is questionable because the experimenter
can only reject the null hypothesis if the effect is in the predicted direction. If the
effect is in the other direction, then the null hypothesis cannot be rejected no
matter how strong the effect is. A skeptic might question whether the experimenter
would really fail to reject the null hypothesis if the effect were strong enough in the
wrong direction. Frequently the most intere sting aspect of an effect is that it runs
counter to expectations. Therefore, an experimenter who committed himself or
herself to ignoring effects in one direction may be forced to choose between ignoring
a potentially important finding and using the techn iques of statistical inference
dishonestly. One-tailed tests are not used frequently. Unless otherwise indicated, a
test should be assumed to be two-tailed.
9.3.6 Confidence Intervals & Hypothesis Testing
There is an extremely close relationship between c onfidence intervals and
hypothesis testing. When a 95% confidence interval is constructed, all values in the
interval are considered plausible values for the parameter being estimated. Values
outside the interval are rejected as relatively implausible. If the value of the
parameter specified by the null hypothesis is contained in the 95% interval then the
null hypothesis cannot be rejected at the 0.05 level. If the value specified by the
null hypothesis is not in the interval then the null hypothesis can be rejected at the
0.05 level. If a 99% confidence interval is constructed, then values outside the
interval are rejected at the 0.01 level.
114

Imagine a researcher wishing to test the null hypothesis that the mean time to
respond to an auditory signal is the same as the mean time to respond to a visual
signal. The null hypothesis therefore is:
μvisual - μauditory = 0.
Ten subjects were tested in the visual condition and their scores (in
milliseconds) were: 355, 421, 299, 460, 600, 580, 474, 511, 550, and 586.
Ten subjects were tested in the auditory condition and their scores were: 275,
320, 278, 360, 430, 520, 464, 311, 529, and 326.
The 95% confidence interval on the difference between means is:
9 ≤ μvisual - μauditory ≤ 196.
Therefore only values in the interval between 9 and 196 are retained as
plausible values for the difference between population means. Since zero, the value
specified by the null hypothesis, is not in the interval, the null hypothesis of no
difference between auditory and visual presentation can be rejected at the 0.05
level. The probability value for this example is 0.034. Any time the parameter
specified by a null hypothesis is not contained in the 95% confidence interval
estimating that parameter, the null hypothesis can be rejected at the 0.05 level or
less. Similarly, if the 99% interval does not contain the parameter then the null
hypothesis can be rejected at the 0.01 level. The null hypothesis is not rejected if
the parameter value specified by the null hypothesis is in the interval since the null
hypothesis would still be plausible.
However, since the null hypothesis would be only one of an infinite number of
values in the confidence interval, accepting the null hypothesis is not justified.
There are many arguments against accepting the null hypothesis when it is not
rejected. The null hypothesis is usually a hypothesis of no difference. Thus null
hypotheses such as:
μ1 - μ2 = 0
π1 - π2 = 0
in which the hypothesized value is zero are most common. When the
hypothesized value is zero then there is a simple relationship between hypothesis
testing and confidence intervals:
If the interval contains zero then the null hypothesis cannot be rejected at the
stated level of confidence. If the interval does not contain zero then the null
hypothesis can be rejected.
This is just a special case of the general rule stating that the null hypothesis
can be rejected if the interval does not contain the hypothesized value of the
parameter and cannot be rejected if the interval contains the hypothesize d value.
Since zero is contained in the interval, the null hypothesis that μ 1 - μ 2 = 0 cannot
be rejected at the 0 .05 level since zero is one of the plausible values of μ 1 - μ 2. The
115

interval contains both positive and negative numbers and therefore μ 1 may be
either larger or smaller than μ 2. None of the three possible relationships between μ 1
and μ 2:
μ 1 - μ 2 = 0,
μ 1 - μ 2 > 0, and
μ1 - μ2 < 0
can be ruled out. The data are very inconclusive. Whenever a significance test
fails to reject the null hypothesis, the direction of the effect (if there is one) is
unknown.
Now, consider the 95% confidence interval:
6 ≤ μ1 - μ2 ≤ 15.
Since zero is not in the interval, the null hypothesis that μ 1 - μ 2 = 0 can be
rejected at the 0.05 level. Moreover, since all the values in the interval are positive,
the direction of the effect can be inferred: μ 1 > μ 2.
Whenever a significance test rejects the null hypothe sis that a parameter is
zero, the confidence interval on that parameter will not contain zero. Therefore
either all the values in the interval will be positive or all the values in the interval
will be negative. In either case, the direction of the effect is known.
9.3.7 Define the Decision Rule and the Region of Acceptance
The decision rule consists of two parts: (1) a test statistic and (2) a range of
values, called the region of acceptance. The decision rule determines whether a null
hypothesis is accepted or rejected. If the test statistic falls within the region of
acceptance, the null hypothesis is accepted; otherwise, it is rejected.
We define the region of acceptance in such a way that the chance of making a
Type I error is equal to the significance level. Here is how that is done.
 Given the significance level α , find the upper limit (UL) of the range of
acceptance. There are three possibilities, depending on the form of the null
hypothesis.
 If the null hypothesis is μ < M: The upper limit of the region of
acceptance will be equal to the value for which the cumulative
probability of the sampling distribution is equal to one minus the
significance level. That is, P( x < UL ) = 1 - α .
 If the null hypothesis is μ = M: The upper limit of the region of
acceptance will be equal to the value for which the cumulative
probability of the sampling distribution is equal to one minus the
significance level divided by 2. That is, P( x < UL ) = 1 - α/2 .
 If the null hypothesis is μ > M: The upper limit of the region of
acceptance is equal to plus infinity.
116

 In a similar way, we find the lower limit (LL) of the range of acceptance.
Again, there are three possibilities, depending on the form of the null
hypothesis.
 If the null hypothesis is μ < M: The lower limit of the region of
acceptance is equal to minus infinity.
 If the null hypothesis is μ = M: The lower limit of the region of
acceptance will be equal to the value for which the cumulative
probability of the sampling distribution is equal to the significance
level divided by 2. That is, P( x < LL ) = α/2 .
 If the null hypothesis is μ > M: The lower limit of the region of
acceptance will be equal to the value for which the cumulative
probability of the sampling distribution is equal to the significance
level. That is, P( x < LL ) = α .
The region of acceptance is defined by the range between LL and UL.
Accept or Reject the Null Hypothesis
Once the region of acceptance is defined, the null hypothesis can be tested
against sample data. Simply compute the test statistic. In this case, the test
statistic is the sample mean. If the sample mean falls within the region of
acceptance, the null hypothesis is accepted; if not, it is rejected.
9.3.8 Other Considerations
When one tests a hypothesis in the real world, other issues may come into
play. Here are some suggestions that may be helpful.
 You will need to make an assumption about the sampling distribution of
the mean score. If the sample is relatively large (i.e., greater than or equal
to 30), you can assume, based on the central limit theorem, that the
sampling distribution will be roughly normal. On the other hand, if the
sample size is small (less than 30) and if the population random variable is
approximately normally distributed (i.e., has a bell -shaped curve), you can
transform the mean score into a t score. The t score will have a (duh) t
distribution.
 Assume that the mean of the sampling distribution is equal to the test
value M specified in the null hypothesis.
 In some situations, you may need to compute the standard deviation of
the sampling distribution sx. If the standard deviation of the population σ
is known, then
sx = σ * sqrt[( 1/n ) - ( 1/N )]
where n is the sample size and N is the population size. On the other hand, if
the standard deviation of the population σ is unknown, then
sx = s * sqrt[( 1/n ) - ( 1/N )]
where s is the sample standard deviation.
117

Example 1
An inventor has developed a new, energy-efficient lawn mower engine. He
claims that the engine will run continuously for 5 hours (300 minutes) on a single
gallon of regular gasoline. Suppose a random sample of 50 e ngines is tested. The
engines run for an average of 295 minutes, with a standard deviation of 20
minutes. Test the null hypothesis that the mean run time is 300 minutes against
the alternative hypothesis that the mean run time is not 300 minutes. Use a 0.0 5
level of significance.
Solution
There are four steps in conducting a hypothesis test, as described in the
previous sections. We work through those steps below:
 Formulate hypotheses. The first step is to state the null hypothesis and an
alternative hypothesis.
Null hypothesis: μ = 300 minutes

Alternative hypothesis: μ <> 300 minutes


Note that these hypotheses constitute a two-tailed test. The null hypothesis
will be rejected if the sample mean is too big or if it is too small.
 Identify the test statistic. In this example, the test statistic is the mean
run time of the 50 engines in the sample - 295 minutes.
 Define the decision rule. The decision rule consists of two parts: (1) a test
statistic and (2) a range of values, called the region of acceptance. We
already know that the test statistic is a sample mean equal to 295. All that
remains is to describe the region of acceptance; that is, to define the lower
limit and the upper limit of the region. Here is how that is done.
 Specify the sampling distribution. Since the sample size is large
(greater than or equal to 30), we assume that the sampling
distribution of the mean is normal, based on the central limit
theorem.
 Define the mean of the sampling distribution . We assume that the
mean of the sampling distribution is equal to the mean value that
appears in the null hypothesis - 300 minutes.
 Compute the standard deviation of the sampling
 distribution. Here the standard deviation of the sampling distribution
sx is:
sx = σ * sqrt[( 1/n ) - ( 1/N )]
sx = 20 * sqrt[ 1/50 ] = 2.83
where s is the sample standard deviation, n is the sample size, and N is the
population size. In this example, we assume that the population size is very large,
so that the value 1/N is about zero.
118

 Find the lower limit of the region of acceptance. Given a two-tailed


hypothesis, the lower limit (LL) will be equal to the value for which the
cumulative probability of the sampling distribution is equal to the
significance level divided by 2. That is, P( x < LL ) = α/2 = 0.05/2 = 0.025.
To find this lower limit, we use the Normal Distribution table. From table:
cumulative probability = 0.025, mean = 300, and standard deviation =
2.83. The calculation tells us that the lower limit is 294.45, given those
inputs.
 Find the upper limit of the region of acceptance. Given a two-tailed
hypothesis, the upper limit (UL) will be equal to the value for which
the cumulative probability of the sampling distribution is equal to one
minus the significance level divided by 2. That is, P( x < UL ) = 1 - α/2
= 1 - 0.025 = 0.975. To find this upper limit, we use the Normal
Distribution Table. From table: cumulative probability = 0.975, mean
= 300, and standard deviation = 2.83. The calculation tells us that
the upper limit is 305.55, given those inputs.
 Thus, we have determined that the region of acceptance is defined by
the values between 294.45 and 305.55.
 Accept or reject the null hypothesis. The sample mean in this example was
295 minutes. This value falls within the region of acceptance. Therefore,
we cannot reject the null hypothesis that a new engine runs for 300
minutes on a gallon of gasoline.
Example 2
Bon Air Elementary School has 300 students. The principal of the school
thinks that the average IQ of students at Bon Air is at least 110. To prove her point,
she administers an IQ test to 20 randomly selected students. Among the sampled
students, the average IQ is 108 with a standard deviation of 10. Based on these
results, should the principal accept or reject her original hypothesis? Assume a
significance level of 0.01.
Solution
There are four steps in conducting a hypothesis test, as described in the
previous sections. We work through those steps below:
 Formulate hypotheses. The first step is to state the null hypothesis and an
alternative hypothesis.
Null hypothesis: μ > 110
Alternative hypothesis: μ < 110
Note that these hypotheses constitute a one -tailed test. The null hypothesis
will be rejected if the sample mean is too small.
 Identify the test statistic. In this example, the test statistic is the mean IQ
score of the 20 students in the sample. Thus, the test statistic is a mean
IQ score of 108.
119

 Define the decision rule. The decision rule consists of two parts: (1) a test
statistic and (2) a range of values, called the region of acceptance. We
already know that the test statistic is a sample mean equal to 108. All that
remains is to describe the region of acceptance; that is, to define the lower
limit and the upper limit of the region. Here is how that is done.
 Specify the sampling distribution. Since the sample size is small (less
than 30), we assume that the sampling distribution of the mean
follows a t distribution.
 Define the mean of the sampling distribution . We assume that the
mean of the sampling distribution is equal to the mean value that
appears in the null hypothesis - 110.
 Find the lower limit of the region of acceptance. Given a one-tailed
hypothesis, the lower limit (LL) will be equal to the value for which
the cumulative probability of the sampling distribution is equal to the
significance level. That is, P( x < LL ) = α = 0.01. To find this lower
limit, we use the T Distribution Table. From table: cumulative
probability = 0.01, population mean = 110, standard deviation = 2.16,
and degrees of freedom = 20 - 1 = 19. The calculation tells us that the
sample mean is 104.51, given those inputs. This is the lower limit of
our region of acceptance.
 Find the upper limit of the region of acceptance. Since we have a one-
tailed hypothesis in which the null hypothesis states that the IQ is
greater than 110, any big number is consistent with the null
hypothesis. Therefore, the upper limit is plus infinity.
Thus, we have determined that the range of acceptance is defined by the
values between 104.51 and plus infinity.
 Accept or reject the null hypothesis. The sample mean in this example was
an IQ score of 108. This value falls within the region of acceptance.
Therefore, we cannot reject the null hypothesis that the average IQ score
of students at Bon Air Elementary is equal to 110.
Power of a Hypothesis Test
When we conduct a hypothesis test, we accept or reject the null hypothesis
based on sample data. Because of the random nature of sample data, our decision
can have four possible outcomes.
 We may accept the null hypothesis when it is true. Thus, the decision is
correct.
 We may reject the null hypothesis when it is true. This kind of incorrect
decision is called a Type I error.
 We may reject the null hypothesis when it is false. Thus, the decision is
correct.
120

 We may accept the null hypothesis when it is false. This kind of incorrect
decision is called a Type II error.
The probability of committing a Type I error is called the significance level and
is denoted by α. The probability of committing a Type II error is called Beta and is
denoted by β. The probability of not committing a Type II error is called the power of
the test.
How to Compute the Power of a Test
When a researcher designs a study to test a hypothesis, he/she should
compute the power of the test (i.e., the likelihood of avoiding a Type II error). Here is
how to do that.
 Define the region of acceptance. (The process for defining the region of
acceptance is described in the previous three lessons. See, for example,
Hypothesis Tests of Mean Score, Hypothesis Tests of Proportion (Large
Sample), and Hypothesis Tests of Proportion (Small Sample).)
 Specify the critical value. The critical value is an alternative to the value
specified in the null hypothesis. The difference between the critical value
and the value from the null hypothesis is called the effect size. That is, the
effect size is equal to the critical value minus the value from the null
hypothesis.
 Compute power. Assume that the true population parameter is equal to
the critical value, rather than the value specified in the null hypothesis.
Based on that assumption, compute the probability that the sample
estimate of the population parameter will fall outside the region of
acceptance. That probability is the power of the test.
Example 1: Power of the Hypothesis Test of a Mean Score
Two inventors have developed a new, energy-efficient lawn mower engine. One
inventor says that the engine will run continuously for 5 hours (300 mi nutes) on a
single gallon of regular gasoline. Suppose a random sample of 50 engines is tested.
The engines run for an average of 295 minutes, with a standard deviation of 20
minutes. The inventor tests the null hypothesis that the mean run time is 300
minutes against the alternative hypothesis that the mean run time is not 300
minutes, using a 0.05 level of significance.
The other inventor says that the new engine will run continuously for only 290
minutes on a gallon of gasoline. Find the power of the te st to reject the null
hypothesis, if the second inventor is correct.
Solution
The steps required to compute power are presented below.
 Define the region of acceptance. In a previous lesson, we showed that the
region of acceptance for this problem consists of the values between
294.45 and 305.55
121

 Specify the critical value. The null hypothesis tests the hypothesis that the
run time of the engine is 300 minutes. We are interested in determining
the probability that the hypothesis test will reject the null hypothesis, if
the true run time is actually 290 minutes. Therefore, the critical value is
290. (Another way to express the critical value is through effect size. The
effect size is equal to the critical value minus the hypothesized value.
Thus, effect size is equal to 290 - 300 or -10.)
 Compute power. The power of the test is the probability of rejecting the
null hypothesis, assuming that the true population mean is equal to the
critical value. Since the region of acceptance is 294.45 to 305.55, the null
hypothesis will be rejected when the sampled run time is less than 294.45
or greater than 305.55.
Therefore, we need to compute the probability that the sampled run time will
be less than 294.45 or greater than 305.55. To do this, we make the following
assumptions:
 The sampling distribution of the mean is normally distributed. (Because
the sample size is relative large, this assumption can be justified by the
central limit theorem.)
 The mean of the sampling distribution is the critical value, 290.
 The standard deviation of the sampling distribution is 2.83. The standard
deviation of the sampling distribution was computed in a previous lesson
Given these assumptions, we first assess the probability that the sample run
time will be less than 294.45. This is easy to do, using the Normal Calculator. We
enter the following values into the calculator: value = 294.45; mean = 290; and
standard deviation = 2.83. Given these inputs, we find that the cumulative
probability is 0.94207551. This means the probability that the sample mean will be
less than 294.45 is 0.942.
Next, we assess the probability that the sample mean is greater than 305.55.
Again, we use the Normal Calculator. We enter the following values into the
calculator: value = 305.55; mean = 290; and standard deviation = 2.83. Given these
inputs, we find that the probability that the sample mean is less than 305.55 (i.e.,
the cumulative probability) is 0.99999998. Thus, the probability that the sample
mean is greater than 305.55 is 1 - 0.99999998 or 0.00000002.
The power of the test is the sum of these probabilities: 0.94207551 +
0.00000002 = 0.94207553. This means that if the true average run time of the new
engine were 290 minutes, we would correctly reject the hypothesis that the run
time was 300 minutes 94.2 percent of the time. Hence, the probability of a Type II
error would be very small. Specifically, it would be 1 minus 0.942 or 0.058.
122

IMPORTANT STATISTICAL DEFINITIONS OF HYPOTHESIS TESTING


Alpha
The significance level of a test of hypothesis that denotes the probability of
rejecting a null hypothesis when it is actually true. In other words, it is the
probability of committing a Type I error.
Alternative Hypothesis
A hypothesis that takes a value of a population parameter different from that
used in the null hypothesis.
Beta
The probability of not rejecting a null hypothesis when it actually is false. In
other words, it is the probability of committing a Type II error.
Critical region
The set of values of the test statistic that will cause us to reject the null
hypothesis.
Critical Value
The 'first' (or 'boundary') value in the critical region.
Decision rule
If the calculated test statistic falls within the critical region, the null
hypothesis Ho is rejected. In contrast, if the calculated test statistic does not fall
within the critical region, the null hypothesis is not rejected.
F-distribution
A continuous distribution that has two parameters (df for the numerator and
df for the denominator). It is mainly used to test hypotheses concerning variances.
F-ratio
In ANOVA, it is the ratio of between -column variance to within column
variance.
Hypothesis
An unproven proposition or supposition that tentatively explains a
phenomenon.
Null hypothesis
A statement about a status quo about a population parameter, that is being
tested.
One-tailed test
A statistical hypothesis test in which the alternative hypothesis is specified
such that only one direction of the possible distribution of values is considered.
Power of hypothesis Test
The probability of rejecting the null hypothesis when it is false.
Significance level
The value4 of a that gives the probability of rejecting the null hypothesis when
it is true. This gives rise to Type I error.
123

Test criteria
Criteria consisting of (i) specifying a level of significance ex, (ii) determining a
test statistic, (iii) determining the critical region(s), and (iv) determining the critical
value(s).
Test Statistic
The value of Z or t calculated for a sample statistic such as the sample mean
or the sample proportion.
Two tail test
A statistical hypothesis test in which the alternative hypothesis is stated in
such a way that it includes both the higher and the lower values of a parameter
than the value specified in the null hypothesis.
Type I error
An error caused by rejecting a null hypothesis that is true.
Type II error
An error caused by failing to reject a null hypothesis, that is not true.
9.3.9 Chi square test
Chi square is a non-parametric test of statistical significance for bivariate
tabular analysis (also known as cross breaks). Any appropriately performed test of
statistical significance lets you know the degree of confidence you can have in
accepting or rejecting a hypothesis. Typically, the hypothesis tested with chi square
is whether or not two different samples (of people, texts, whatever) are different
enough in some characteristic or aspect of their behavior that we can generalize
from our samples that the populations from which our samples are drawn are also
different in the behavior or characteristic.
A non-parametric test, like chi square, is a rough estimate of confidence; it
accepts weaker, less accurate data as input than parametric tests (like t-tests and
analysis of variance, for example) and therefore has less status in the pantheon of
statistical tests. Nonetheless, its limitations are also its strengths; because chi
square is more 'forgiving' in the data it will accept, it can be used in a wide variety
of research contexts.
Chi square is used most frequently to test the statistical significance of results
reported in bivariate tables, and interpreting bivariate tables is integral to
interpreting the results of a chi square test, so we'll take a look at bivariate tabular
(cross break) analysis.
Bivariate Tabular Analysis
Bivariate tabular (cross break) analysis is used when you are trying to
summarize the intersections of independent and dependent variables and
understand the relationship (if any) between those variables. For example, if we
wanted to know if there is any relationship between the biological sex of American
undergraduates at a particular university and their footwear preferences, we might
select 50 males and 50 females as randomly as possible, and ask them, "On
average, do you prefer to wear sandals, sneakers, leather shoes, boots, or
124

something else?" In this example, our independent variable is biological sex. (In
experimental research, the independent variable is actively manipulated by the
researcher; for example, whether or not a rat gets a food pellet when it pulls on a
striped bar. In most sociological research, the independent variable is not actively
manipulated in this way, but controlled by sampling for, e.g., males vs. females.)
Put another way, the independent variable is the quality or characteristic that you
hypothesize helps to predict or explain some other quality or characteristic (the
dependent variable). We control the independent variable (and as much else as
possible and natural) and elicit and measure the dependent variable to test our
hypothesis that there is some relationship between them. Bivariate tabular analysis
is good for asking the following kinds of questions:
1. Is there a relationship between any two variables in the data?
2. How strong is the relationship in the data?
3. What is the direction and shape of the relationship in the data?
4. Is the relationship due to some intervening variable(s) in the data?
To see any patterns or systematic relationship between biological sex of
undergraduates at University of X and reported footwear preferences, we co uld
summarize our results in a table like this:
Table: Male and Female Undergraduate Footwear Preferences
Sandals Sneakers Leather shoes Boots Other
Male
Female

Depending upon how our 50 male and 50 female subjects responded, we could
make a definitive claim about the (reported) footwear preferences of those 100
people.
In constructing bivariate tables, typically values on the independent variable
are arrayed on vertical axis, while values on the dependent variable are arrayed on
the horizontal axis. This allows us to read 'across' from hypothetically 'causal'
values on the independent variable to their 'effects', or values on the dependent
variable. How you arrange the values on each axis should be guided "iconically" by
your research question/hypothesis. For example, if values on an independent
variable were arranged from lowest to highest value on the variable and values on
the dependent variable were arranged left to right from lowest to highest, a positive
relationship would show up as a rising left to right line. (But remember, association
does not equal causation; an observed relationship between two variables is not
necessarily causal.)
Each intersection/cell--of a value on the independent variable and a value on
the independent variable--reports the result of how many times that combination of
values was chosen/observed in the sample being analyzed. (So you can see that
crosstabs are structurally most suitable for analyzing relationships between
125

nominal and ordinal variables. Interval and ratio variables will have to first be
grouped before they can "fit" into a bivariate table.) Each cell reports, essentially,
how many subjects/observations produced that combination of independent and
dependent variable values. So, for example, the top left cell of the table above
answers the question: "How many male undergraduates at University of X prefer
sandals?" (Answer: 6 out of the 50 sampled.)
Table 1.b. Male and Female Undergraduate Footwear Preferences
Sandals Sneakers Leather shoes Boots Other
Male 6 17 13 9 5
Female 13 5 7 16 9

Reporting and interpreting cross tabs is most easily done by converting raw
frequencies (in each cell) into percentages of each cell within the values/categories
of the independent variable. For example, in the Footwear Preferences table above,
total each row, then divide each cell by its row total, and multiply that fraction by
100.
Table 1.C. Male and Female Undergraduate Footwear Preferences (Percentages)
Sandals Sneakers Leather shoes Boots Other N
Male 12 34 26 18 10 50
Female 26 10 14 32 18 50
Percentages basically standardize cell frequencies as if there were 100
subjects/observations in each category of the independent variable. This is useful
for comparing across values on the independent variable, but that usefulness
comes at the price of a generalization --from the actual number of subjects/
observations in that column in your data to a hypothetical 100 subjects/
observations. If the raw row total was 93, then percentages do little violence to the
raw scores; but if the raw total is 9, then the generalization (on no statistical basis,
i.e., with no knowledge of sample-population representativeness) is drastic. So you
should provide that total N at the end of each row/independent variable category
(for replicability and to enable the reader to assess your interpretation of the table's
meaning).
With this caveat in mind, you can compare the patterns of distribution of
subjects/observations along the dependent variable between the values of the
independent variable: e.g., compare male and female undergraduate footwear
preference. (For some data, plotting the results on a line graph can also help you
interpret the results: i.e., whether there is a positive (/), negative (\), or curvilinear
(\/, /\) relationship between the variables.) Table 1.c shows that within our
sample, roughly twice as many females preferred sandals and boots as males; and
within our sample, about three times as many men preferred sneakers as women
and twice as many men preferred leather shoes. We might also infer from the
'Other' category that female students within our sample had a broader range of
footwear preferences than did male students.
126

Generalizing from Samples to Populations


Converting raw observed values or frequencies into pe rcentages does allow us
to see more easily patterns in the data, but that is all we can see: what is in the
data. Knowing with great certainty the footwear preferences of a particular group of
100 undergraduates at University of X is of limited use to us; we usually want to
measure a sample in order to know something about the larger populations from
which our samples were drawn. On the basis of raw observed frequencies (or
percentages) of a sample's behavior or characteristics, we can make claims about
the sample itself, but we cannot generalize to make claims about the population
from which we drew our sample, unless we submit our results to a test of statistical
significance. A test of statistical significance tells us how confidently we can
generalize to a larger (unmeasured) population from a (measured) sample of that
population.
How does chi square do this? Basically, the chi square test of statistical
significance is a series of mathematical formulas which compare the actual
observed frequencies of some phenomenon (in our sample) with the frequencies we
would expect if there were no relationship at all between the two variables in the
larger (sampled) population. That is, chi square tests our actual results against the
null hypothesis and assesses whether the actual results are different enough to
overcome a certain probability that they are due to sampling error. In a sense, chi -
square is a lot like percentages; it extrapolates a population characteristic (a
parameter) from the sampling characteristic (a statistic) similarly to the way
percentage standardizes a frequency to a total column N of 100. But chi -square
works within the frequencies provided by the sample and does not inflate (or
minimize) the column and row totals.
Chi Square Requirements
As mentioned before, chi square is a nonparametric test. It does not require
the sample data to be more or less normally distributed (as parametric tests like t-
tests do), although it relies on the assumption that the variable is normally
distributed in the population from which the sample is drawn.
But chi square does have some requirements:
1. The sample must be randomly drawn from the population.
2. Data must be reported in raw frequencies (not percentages);
3. Measured variables must be independent;
4. Values/categories on independent and dependent variables must be
mutually exclusive and exhaustive;
5. Observed frequencies cannot be too small.
1. As with any test of statistical significance, your data must be from a random
sample of the population to which you wish to generalize your claims.
2. You should only use chi square when your data are in the form of raw
frequency counts of things in two or more mutually exclusive and exhaustive
127

categories. As discussed above, converting raw frequencies into percentages


standardizes cell frequencies as if there were 100 subjects/observations in each
category of the independent variable for comparability. Part of the chi square
mathematical procedure accomplishes this standardizing, so computing the chi
square of percentages would amount to standardizing an already standardized
measurement.
3. Any observation must fall into only one category or value on each variable.
In our footwear example, our data are counts of male versus female undergraduates
expressing a preference for five different categories of footwear. Each
observation/subject is counted only once, as either male or female (an exhaustive
typology of biological sex) and as preferring sandals, sneakers, leather shoes, boots,
or other kinds of footwear. For some vari ables, no 'other' category may be needed,
but often 'other' ensures that the variable has been exhaustively categorized. (For
some kinds of analysis, you may need to include an "uncodable" category.) In any
case, you must include the results for the whole sample.
4. Furthermore, you should use chi square only when observations are
independent: i.e., no category or response is dependent upon or influenced by
another. (In linguistics, often this rule is fudged a bit. For example, if we have one
dependent variable/column for linguistic feature X and another column for number
of words spoken or written (where the rows correspond to individual speakers/texts
or groups of speakers/texts which are being compared), there is clearly some
relation between the frequency of feature X in a text and the number of words in a
text, but it is a distant, not immediate dependency.)
5. Chi-square is an approximate test of the probability of getting the
frequencies you've actually observed if the null hypothesis were true. It's based on
the expectation that within any category, sample frequencies are normally
distributed about the expected population value. Since (logically) frequencies
cannot be negative, the distribution cannot be normal when expected population
values are close to zero--since the sample frequencies cannot be much below the
expected frequency while they can be much above it (an asymmetric/non -normal
distribution). So, when expected frequencies are large, there is no problem with the
assumption of normal distribution, but the smaller the expected frequencies, the
less valid are the results of the chi -square test. We'll discuss expected frequencies
in greater detail later, but for now remember that expected frequencies are derived
from observed frequencies. Therefore, if you have cells in your bivariate table which
show very low raw observed frequencies (5 or below), your expected frequencies
may also be too low for chi square to be appropriately used. In addition, because
some of the mathematical formulas used in chi square use division, no cell in your
table can have an observed raw frequency of 0.
The following minimum frequency thresholds should be obeyed:
 for a 1 X 2 or 2 X 2 table, expected frequencies in each cell should be at
least 5;
128

 for a 2 X 3 table, expected frequencies should be at least 2;


 for a 2 X 4 or 3 X 3 or larger table, if all expected frequencies but one are at
least 5 and if the one small cell is at least 1, chi -square is still a good
approximation.
In general, the greater the degrees of freedom (i.e., the more values/categories
on the independent and dependent variables), the more lenient the minimum
expected frequencies threshold. (We'll discuss degrees of freedom in a moment.)
Collapsing Value s
A brief word about collapsing values/categories on a variable is necessary.
First, although categories on a variable --especially a dependent variable--may be
collapsed, they cannot be excluded from a chi -square analysis. That is, you cannot
arbitrarily exclude some subset of your data from your analysis. Second, a decision
to collapse categories should be carefully motivated, with consideration for
preserving the integrity of the data as it was originally collected. (For example, how
could you collapse the footwear preference categories in our example and still
preserve the integrity of the original question/data? You can't, since there's no way
to know if combining, e.g., boots and leather shoes versus sandals and sneakers is
true to your subjects' typology of footwear.) As a rule, you should perform a chi
square on the data in its uncollapsed form; if the chi square value achieved is
significant, then you may collapse categories to test subsequent refinements of your
original hypothesis.
Computing Chi Square
Let's walk through the process by which a chi square value is computed, using
Table 1.b. above (renamed 1.d., below).
The first step is to determine our threshold of tolerance for error. That is, what
odds are we willing to accept that we are wrong in generalizing from the results in
our sample to the population it represents? Are we willing to stake a claim on a 50
percent chance that we're wrong? A 10 percent chance? A five percent chance? 1
percent? The answer depends largely on our research question and the
consequences of being wrong. If people's lives depend on our interpretation of our
results, we might want to take only 1 chance in 100,000 (or 1,000,000) that we're
wrong. But if the stakes are smaller, for example, whether or not two texts use the
same frequencies of some linguistic feature (assuming this is not a forensic issue in
a capital murder case!), we might accept a greater probability --1 in 100 or even 1 in
20--that our data do not represent the population we're generalizing about. The
important thing is to explicitly motivate your threshold before you perform any test
of statistical significance, to minimize any temptation for post hoc compromise of
scientific standards. For our purposes, we'll set a probability of error threshold of 1
in 20, or p < .05, for our Footwear study.)
The second step is to total all rows and columns
Table Male and Female Undergraduate Footwear Preferences: Observed
Frequencies with Row and Column Totals
129

Sandals Sneakers Leathershoes Boots Other Total


Male 6 17 13 9 5 50
Female 13 5 7 16 9 50
Total 19 22 20 25 14 100

Remember that chi square operates by comparing the actual, or observed,


frequencies in each cell in the table to the frequencies we would expect if there were
no relationship at all between the two variables in the populations from which the
sample is drawn. In other words, chi square compares what actually happened to
what hypothetically would have happened if 'all other things were equal' (basically,
the null hypothesis). If our actual results are sufficiently different from the
predicted null hypothesis results, we can reject the null hypothesis and claim that
a statistically significant relationship exists between our variables.
Chi square derives a representation of the null hypothesis--the 'all other things
being equal' scenario--in the following way. The expected frequency in each cell is
the product of that cell's row total multiplied by that cell's column total, divided by
the sum total of all observations. So, to derive the expected frequency of the "Males
who prefer Sandals" cell, we multiply the top row total (50) by the first column total
(19) and divide that product by the sum total (100): ((50 X 19)/100) = 9.5. The logic
of this is that we are deriving the expected frequency of each cell from the union of
the total frequencies of the relevant values on each variable (in this case, Male and
Sandals), as a proportion of all observed frequencies (across all values of each
variable). This calculation is performed to derive the expected frequency of each
cell, as shown in Table 1.e below (the computation for each cell is listed below Table
Table. Male and Female Undergraduate Footwear Preferences: Observed and
Expected Frequencies
Sandals Sneakers Leather shoes Boots Other Total
Male observed 6 17 13 9 5 50
Male expected 9.5 11 10 12.5 7
Female observed 13 5 7 16 9 50
Female expected 9.5 11 10 12.5 7
Total 19 22 20 25 14 100
Male/Sandals: ((19 X 50)/100) = 9.5
Male/Sneakers: ((22 X 50)/100) = 11
Male/Leather Shoes: ((20 X 50)/100) = 10
Male/Boots: ((25 X 50)/100) = 12.5
Male/Other: ((14 X 50)/100) = 7
Female/Sandals: ((19 X 50)/100) = 9.5
Female/Sneakers: ((22 X 50)/100) = 11
130

Female/Leather Shoes: ((20 X 50)/100) = 10


Female/Boots: ((25 X 50)/100) = 12.5
Female/Other: ((14 X 50)/100) = 7

(Notice that because we originally obtained a balanced male/female sample,


our male and female expected scores are the same. This usually will not be the
case.)
We now have a comparison of the observed results versus the results we would
expect if the null hypothesis were true. We can informally analyze this table,
comparing observed and expected frequencies in each cell (Males prefer sandals
less than expected), across values on the independent variable (Males prefer
sneakers more than expected, Females less than expected), or across values on the
dependent variable (Females prefer sandals and boots more than expected, but
sneakers and shoes less than expected). But so far, the extra computation doesn't
really add much more information than interpretation of the results in percentage
form. We need some way to measure how different our observed results are from
the null hypothesis. Or, to put it another way, we need some way to determine
whether we can reject the null hypothesis, and if we can, with what degree o f
confidence that we're not making a mistake in generalizing from our sample results
to the larger population.
Logically, we need to measure the size of the difference between the pair of
observed and expected frequencies in each cell. More specifically, we calculate the
difference between the observed and expected frequency in each cell, square that
difference, and then divide that product by the difference itself. The formula can be
expressed as:
((O - E)2 /E)
Squaring the difference ensures a positive number, so that we end up with an
absolute value of differences. If we didn't work with absolute values, the positive
and negative differences across the entire table would always add up to 0. (You
really understand the logic of chi square if you can figure out why this is true.)
Dividing the squared difference by the expected frequency essentially removes the
expected frequency from the equation, so that the remaining measures of
observed/expected difference are comparable across all cells.
So, for example, the difference between observed and expected frequencies for
the Male/Sandals preference is calculated as follows:
1. Observed (6) minus Expected (9.5) = Difference (-3.5)
2. Difference (-3.5) squared = 12.25
3. Difference squared (12.25) divided by Expected (9.5) = 1.289
The sum of all products of this calculation on each cell is the total chi square
value for the table.
131

The computation of chi square for each cell is listed below Table
Table Male and Female Undergraduate Footwear Preferences: Obse rved and Expected
Frequencies Plus Chi Square
Sandals Sneakers Leather shoes Boots Other Total
Male observed 6 17 13 9 5 50
Male expected 9.5 11 10 12.5 7
Female observed 13 5 7 16 9 50
Female expected 9.5 11 10 12.5 7
Total 19 22 20 25 14 100
Male/Sandals: ((6 - 9.5)2/9.5) = 1.289
Male/Sneakers: ((17 - 11)2/11) = 3.273
Male/Leather Shoes: ((13 - 10)2/10) = 0.900
Male/Boots: ((9 - 12.5)2/12.5) = 0.980
Male/Other: ((5 - 7)2 /7) = 0.571
Female/Sandals: ((13 - 9.5)2/9.5) = 1.289
Female/Sneakers: ((5 - 11)2/11) = 3.273
Female/Leather Shoes: ((7 - 10)2/10) = 0.900
Female/Boots: ((16 - 12.5)2 /12.5) = 0.980
Female/Other: ((9 - 7)2 /7) = 0.571

(Again, because of our balanced male/female sample, our row totals were the
same, so the male and female observed-expected frequency differences were
identical. This is usually not the case.)
The total chi square value for Table 1 is 14.026.
Interpreting the Chi Square Value
We now need some criterion or yardstick against which to measure the table 's
chi square value, to tell us whether or not it is significant. What we need to know is
the probability of getting a chi square value of a minimum given size even if our
variables are not related at all in the larger population from which our sample was
drawn. That is, we need to know how much larger than 0 (the absolute chi square
value of the null hypothesis) our table's chi square value must be before we can
confidently reject the null hypothesis. The probability we seek depends in part on
the degrees of freedom of the table from which our chi square value is derived.
Degrees of freedom
Mechanically, a table's degrees of freedom (df) can be expressed by the
following formula:
df = (r-1)(c-1)
132

That is, a table's degrees of freedom equals the number of rows in the table
minus one multiplied by the number of columns in the table minus one. (For 1 X 2
tables: df = k - 1, where k = number of values/categories on the variable.)
Degrees of freedom are an issue because of the way in which expected values
in each cell are computed from the row and column totals of each cell. All but one
of the expected values in a given row or column are free to vary (within the total
observed--and therefore expected) frequency of that row or column); once the free to
vary expected cells are specified, the last one is fixed by virtue of the fact that the
expected frequencies must add up to the observed row and column totals (from
which they are derived).
Another way to conceive of a table's degrees of freedom is to think of one row
and one column in the table as fixed, with the remaining cells free to vary. Consider
the following visuals (where X = fixed):

X X

(r-1)(c-1) = (3-1)(2-1) = 2 X 1 = 2

X X X

(r-1)(c-1) = (5-1)(3-1) = 4 X 2 = 8
So, for our Table 1, df = (2-1)(5-1) = 4
Sandals Sneakers Leather shoes Boots Other
Male X X X X X
Female X

In a statistics book, the sampling distribution of chi square (also know as


'critical values of chi square') is typically listed in an appendix. You read down the
133

column representing your previously chosen probability of error threshold (e.g., p <
.05) and across the row representing the degrees of freedom in your table. If your
chi square value is larger than the critical value in that cell, your data present a
statistically significant relationship between the variables in your table.
Table 1's chi square value of 14.026, with 4 degrees of freedom, handily clears
the related critical value of 9.49, so we can reject the null hypothesis and affirm the
claim that male and female undergraduates at University of X differ in their (self-
reported) footwear preferences.
Statistical significance does not help you to interpret the nature or explanation
of that relationship; that must be done by other means (in cluding bivariate tabular
analysis and qualitative analysis of the data). But a statistically significant chi
square value does denote the degree of confidence you may hold that relationship
between variables described in your results is systematic in the larger population
and not attributable to random error.
Statistical significance also does not ensure substantive significance. A large
enough sample may demonstrate a statistically significant relationship between two
variables, but that relationship may be a trivially weak one. Statistical significance
means only that the pattern of distribution and relationship between variables
which is found in the data from a sample can be confidently generalized to the
larger population from which the sample was ran domly drawn. By itself, it does not
ensure that the relationship is theoretically or practically important or even very
large.
Measures of Association
While the issue of theoretical or practical importance of a statistically
significant result cannot be quantified, the relative magnitude of a statistically
significant relationship can be measured. Chi -square allows you to make decisions
about whether there is a relationship netween two or more variables; if the null
hypothesis is rejected, we conclude that there is a statistically significant
relationship between the variables. But we frequently want a measure of the
strength of that relationship--an index of degree oof correlation, a measure of the
degree of association between the variables represented in our table (and data).
Luckily, several related measures of association can be derived from a table's chi
square value.
For tables larger than 2 X 2 (like our Table 1), a measure called 'Cramer's phi'
is derived by the following formula (where N = the total number of observations, and
k = the smaller of the number of rows or columns):
Cramer's phi = the square root of (chi-square divided by (N times (k minus 1)))
So, for our Table 1 (a 2 X 5), we would compute Cramer's phi as follows:
1. N(k - 1) = 100 (2-1) = 100
2. chi square/100 = 14.026/100 = 0.14
3. square root of (2) = 0.37
134

The product is interpreted as a Pearson r (that is, as a correlation coefficient).


(For 2 X 2 tables, a measure called 'phi' is derived by dividing the table's chi
square value by N (the total number of observations) and then taking the square
root of the product. Phi is also interpreted as a Pearson r.)
A complete account of how to interpret correlation coefficients is unnecessary
for present purposes. It will suffice to say that r2 is a measure called shared
variance. Shared variance is the portion of the total behavior (or distribution) of the
variables measured in the sample data which is accounted for by the relationship
we've already detected with our chi square. For Table 1, r2 = 0.137, so
approximately 14% of the total footwear preference story is explained/predicted by
biological sex.
Computing a measure of association like phi or Cramer's phi is rarely done in
quantitative linguistic analyses, but it is an important benchmark of just 'how
much' of the phenomenon under investigation has been explained. For example,
Table 1's Cramer's phi of 0.37 (r2 = 0.137) means that there are one or more
variables still undetected which, cumulatively, account for and predict 86% of
footwear preferences. This measure, of course, doesn't begin to address the nature
of the relation(s) between these variables, which is a crucial part of any adequate
explanation or theory.
9.4 REVISION POINTS
1. Statistical hypothesis is an assumption about a population parameter
2. The null hypothesis is usually the hypothesis that sample observations
result purely from chance effects.
3. The alternative hypothesis is the hypothesis that sample observations are
influenced by some non-random cause.
4. This probability value is the probability of obtaining data as extreme or more
extreme than the current data (assuming H0 is true).
5. The probability of a Type I error is designated by the Greek letter alpha (
and is called the Type I error rate
6. The probability of a Type II error (the Type II error rate) is designated by the
Greek letter beta (ß)
9.5 INTEXT QUESTIONS
1. Define hypothesis.
2. What are the characteristic features of a hypothesis?
3. Distinguish between null and alternative hypothesis.
4. Differentiate type i error and type ii error.
5. How is a hypothesis tested?
6. What are the uses of chi square test?
7. What is confidence interval?
135

9.6 SUMMARY
This chapter had given clear picture about the process of Hypothesis Testing
and to find out the details of Null hypothesis and alternate hypothesis. This lesson
also has given the significance and differences between Type I error and Type II
error. The detailed step-wise procedures were given to perform the hypothesis
testing. This chapter also gave details about the calculation of P value. The above
chapter has also given the framework for performing key statistical tests like chi -
square test. Chi-square is a non parametric test of statistical significance for
bivariate tabular analysis
9.7 TERMINAL EXERCISES
1. -----------------is an assumption about a population parameter.
2. ------------------ is usually the hypothesis that sample observations result
purely from chance effects.
3. ----------- is the hypothesis that sample observations are influenced by some
non-random cause.
4. The --------------- is rejected if p is at or below the significance level; it is not
rejected if p is above the significance level.
5. A true null hypothesis can be incorrectly rejected is known as ------- error
6. A false null hypothesis can fail to be rejected is known as ------- error
9.8 SUPPLEMENTARY MATERIALS
1. Cronin, M. J. (1998a, Feb. 2). Business secrets of the billion -dollar website.
Fortune, p. 142.
2. Cohen, J. (1990). Things I have learned (so far). American Psychologist,
1304–1312
3. Chicago Manual of Style, 14th ed. (1993). Chicago: University of Chicago
Press.
9.9 ASSIGNMENTS
1. What is a hypothesis? What characteristics it must possess in order to be a
good research hypothesis?
2. A manufacturer considers his production process to be working properly if
the mean length of the rods the manufactures is 8.5". The standard
deviation of the rods always runs about 0.26". Suppose a sample of 64 rods
is taken and this gives a mean length of rods equal to 8.6". What are the null
and alternative hypotheses for this problem? Can you infer at 5% level of
significance that the process is working properly?
3. The procedure of testing hypothesis requires a researcher to adopt several
steps. Describe in brief all such steps.
4. Point out the important limitations of tests of hypotheses. What precaution
the researcher must take while drawing inferences as per the results of the
said tests?
136

9.10 SUGGESTED READINGS/ REFERENCE BOOKS


1. Bryman, A. and Cramer, D. (1990) Quantitative Data Analysis for Social
Scientists, London: Routledge
2. Bryman, A. and Cramer, D. (2001) Quantitative Data Analysis with SPSS
Release 10 for Windows: A Guide for Social Scientists, (new edition) Hove:
Routledge.
9.11 LEARNING ACTIVITIES
1. You have been hired as a marketing research analyst by a major company.
Your boss is high powered statistician who does not believe chi square test.
Convince your boss with the help of merits of that test.
9.12 KEYWORDS
Hypothesis: A hypothesis is a hunch, assumption, suspicion, assertion or an
idea about a phenomenon, relationship or situation , the reality or truth of which
you do not know and you set up your study to find this truth. A researcher refers to
these assumptions, assertions, statements or hunches as hypotheses and they
become the basis of an enquiry. In most studies the hypothesis will be based either
upon previousstudies or on your own or someone else’s observations.
Hypothesis of Association: When as a researcher you have sufficient
knowledge about a situation orphenomenon and are in a position to stipulate the
extent of the relationship between two variables and formulate a hunch that reflects
the magnitude of the relationship, such a type of hypothesis formulation is known
as hypothesis of association.
Hypothesis of Difference: A hypothesis in which a researcher stipulates that
there will be a difference but does not specify its magnitude is called a hypothesis of
difference.
Type I Error: In testing a hypothesis, many reasons you may sometimes
commit a mistake and draw the wrong conclusion with respect to the validity of
your hypothesis. If you reject a null hypothesis when it is true and you should not
have rejected it, this is called a Type I error.
Type II Error: In testing a hypothesis, for many reasons you may sometimes
commit a mistake and draw the wrong conclusion in terms of the validity of your
hypothesis. If you accept a null hypothesis when it is false and you should not have
accepted it this is called a Type II error.




137

LESSON – 10

T- TEST AND Z- TEST


10.1 INTRODUCTION
Sometimes, measuring every single piece of item is just not practical. That is
why we developed and use statistical methods to solve problems. The most practical
way to do it is to measure just a sample of the population. Some methods test
hypotheses by comparison. The two of the more known statistical hypothesis test
are the T-test and the Z-test. Let us try to breakdown the two.
A T-test is a statistical hypothesis test. In such test, the test statistic follows a
Student’s T-distribution if the null hypothesis is true. The T-statistic was
introduced by W.S. Gossett under the pen name “Student”. The T-test is also
referred as the “Student T-test”. It is very likely that the T-test is most commonly
used Statistical Data Analysis procedure for hypothesis testing since it is
straightforward and easy to use. Additionally, it is flexible and adaptable to a broad
range of circumstances.
There are various T-tests and two most commonly applied tests are the one -
sample and paired-sample T-tests. One-sample T-tests are used to compare a
sample mean with the known population mean. Two-sample T-tests, the other
hand, are used to compare either independent samples or dependent samples.
T-test is best applied, at least in theory, if you have a limited sample
size (n < 30) as long as the variables are approximately normally
distributed and the variation of scores in the two groups is not reliably
different. It is also great if you do not know the populations’ standard
deviation. If the standard deviation is known, then, it would be best to
use another type of statistical test, the Z-test. The Z-test is also applied
to compare sample and population means to know if there’s a significant
difference between them. Z-tests always use normal distribution and also
ideally applied if the standard deviation is known. Z-tests are often
applied if the certain conditions are met; otherwise, other statistical tests
like T-tests are applied in substitute. Z-tests are often applied in large
samples (n > 30). When T-test is used in large samples, the t-test
becomes very similar to the Z-test. There are fluctuations that may occur
in T-tests sample variances that do not exist in Z-tests. Because of this,
there are differences in both test results. A brief introduction to these
test are explained below.
10.2 OBJECTIVES
After studying this lesson, you will be able to understand:
 One sample Z- test
138

 One sample t – test


 Two sample z- test
 Two sample t- test
10.3 Contents
10.3.1 Testing the Difference Between Two Means: Using the Z Test
10.3.2 Testing the Difference Between Two Means of Independent Samples:
Using the t Test
10.3.3 Testing the Difference Between Two Means: Dependent Samples
10.3.1 Testing the Difference Between Two Means: Using the Z Test
Suppose we are interested in determining if a certain medication relieves
patients' headaches.
We give the drug/treatment to one group and give a placebo to a control group
and compare the mean incidences of patient rel ief from the headache between the
two groups.
If the treatment group had a statistically significant improvement in headache
symptoms over the control group, then we can conclude the drug works.
So our question might be, "Is the mean incidence of headache relief different
for the two groups?"
Let µ1 = mean headache relief from treatment group and
µ2 = mean headache relief from control group.
Then our hypotheses would be:
H0:
H1 :
Alternatively, we could state the hypotheses as:
H0 :

H1 :
Assumptions for the Test to Determine the Difference between Two Means
 The samples must be independent of each other. That is, there can be no
relationship between the subjects in each sample.
 The populations from which the samples come must be (approximately)
normally distributed or the sample sizes of both groups should be at least
30.
 The standard deviations of both populations must be known.
We can compare the groups by the difference in their population means , µ1 -
µ2, where µ1 is the population mean for group 1 and µ2 is the population mean for
group 2.
We estimate µ1-µ2 with
139

The standard deviation of

When both populations are normally distributed or the samples size for each
group is at least 30, then

has a normal distribution.


Formula for the z test for Comparing Two Means from Independent Populations

The observed difference between the sample means may be due to chance, in
which case the null hypothesis will not be rejected. If the difference is statistically
significant, the null hypothesis is rejected an d the researcher can conclude the
population means are different. The common approach to finding critical values
and P-values that was used here .
Example: Dr. Cribari would like to determine if there is a statistically
significant difference between her two Math 2830 classes. To make this comparison
she will compare the results from exam 1. Class one had 35 students take the exam
with a mean of 82.6 and a population standard deviation of 1.41. Class two had 32
students take the exam with a mean of 84 and a population standard deviation of
3.63. Can Dr. Cribari conclude that there is difference in the mean test grades
between the two classes at ∞=0.05?
Step 1: State the hypotheses and identify the claim

Step 2: Find the critical value(s) from the appropriate table .As stated, the
problem is giving the population standard deviations. This means that we will be
doing a z-test.
Two sided test => critical value = ±1.96 ∞ 0.05
Step 3: Compute the test value and determine the P-value.
140

Step 4: Make the decision to reject or not reject the null hypothesis.
Since the p-value is smaller than our ∞, the null hypothesis is rejected.
[OR, Since, our test value, -2.05, falls within the rejection region, the null
hypothesis is rejected]
Step 5: Summarize the results.
That is, there is evidence to support the claim that the exam 1 grades differ
between the two sections of MATH2830.
Practice Example: A survey found that the average hotel room in New Orleans
is $88.42 and the average room rate in Phoenix is $80.61. Assume that the data
were obtained from two samples of 50 hotels each and that the (population)
standard deviations were $5.62 and $4.83, respectively. At ∞ = 0.01, can it be
concluded that the average hotel room in New Orleans costs more than in Phoenix?
Step 1: State the hypotheses and identify the claim.
Step 2: Find the critical value(s) from the appropriate table.
Step 3: Compute the test value and determine the P-value.
Step 4: Make the decision to reject or not reject the null hypothesis.
Step 5: Summarize the results
Formula for the z Confidence Interval for Difference Between Two Means
Assumptions
 The data for each group are independent random sample s.
 The data are from normally distributed populations and/or the sample
sizes of the groups are greater than 30.
 The population standard deviation is (assumed) known

Example: Two brands of cigarettes are selected, and their nicotine content is
compared. The data are shown below. Find the 95% confidence interval of the true
difference in the means.
141

At ∞=0.05., is there convincing evidence that the mean amount of nicotine


differs between the brands?
Example: For the hotel example, construct a 98% confiden ce interval of the
true difference in the means.
10.3 Testing the Difference Between Two Means of Independent Samples: Using the t
Test
Many times the conditions set forth by the z test in previous cannot be met
(e.g., the population standard deviations are not known). In these cases, a t test is
used to test the difference between means when the two samples are independent
and when the samples are taken from two normally or approximately normally
distributed populations.
Formula for the t Test for Testing the Difference Between Two Means: Independent
Samples

Warning: Your calculator will perform a 2 sample t-test (its #4 under STATS
then TESTS). However, it uses a complicated formula to determine the degrees of
freedom that will ultimately affect how the calculator deals with confidence
intervals and p-values. We will come back to this point at the end of the section.
Example: A real estate agent wishes to determine whether tax assessors and
real estate appraisers agree on the values of homes. A random sample of the two
groups appraised 10 homes. Is there a significant difference in the values of the
142

homes for each group? Let ∞ = 0.05. Assume the data are from normally distributed
populations.

Step 1: State the hypotheses and identify the claim


H0: µ1=µ2
H1: µ1≠µ2
Step 2: Find the critical value(s] from the appropriate table.
 T-test means use the t-table (Table F).
 We have 9 degrees of freedom since n1=10 and n2=10. The smallest of n1-
1 and n2-1 is 9.
 Information we need: two-tailed test, ∞ =0.05, df=9
 T critical value is ± 2.262
Step 3: Compute the test value and determine the P-value.

Step 4: Make the decision to reject or not reject the null hypothesis.
The null hypothesis is rejected. This decision can be based on:
 The fact that the test value (-4.02) is within the critical region since it is
less than -2.262 or
 The fact that the p-value (0.003) is smaller than a=0.05
143

Step 5: Summarize the results.


There is significant evidence that tax assessors and real estate appraisers
disagree on the values of homes.
Practice Example: A researcher suggests that male nurses earn more than
female nurses. A survey of 16 male nurses and 20 female nurses reports these
data. Is there enough evidence to support the claim that male nurses earn more
than female nurses? Use ∞ = 0.01. Assume the data are from normally distributed
populations.

Step 1: State the hypotheses and identify the claim.


Step 2: Find the critical value(s) from the appropriate table.
Step 3: Compute the test value and determine the P-value.
Step 4: Make the decision to reject or not reject the null hypothesis.
Step 5: Summarize the results.
Confidence Intervals for the Difference of Two Means: Small Independent Samples
Variances assumed to be unequal:
144

Warning: The way our calculator determines the degrees of freedom is not the
same as the book. So you will NOT be able to use your calculator STAT/TESTS
function to calculate your confidence interval because you will get a VERY different
confidence interval. This is due to the fact that the t-multiplier will be sufficiently
different then what the calculator will find.
Example: Let’s find the 95% Confidence Interval for the first problem.

Example: Let’s find the 99% Confidence Interval for the second problem.
10.4 Testing the Difference Between Two Means: Dependent Samples
So far we have only compared two means when the samples were independent.
Samples are considered to be dependent when the subjects are paired or matched
in some way
Examples of paired data
 Each person is measured twice where the 2 measurements measure the
same thing but under different conditions
 Similar individuals are paired prior to an experiment and each member of a
pair receives a different treatment
 Two different variables are measured for each individual and there is
interest in the amount of difference between the 2 variables
When using paired data, you are interested primarily in the "difference" and not the
data itself
When samples are dependent, a special t test for dependent means is used.
The test uses the difference in the values of the matched pairs.
Important: We cannot use the t test we had learned for a difference in
independent means.
To determine whether one set of observations tend to be larger or different
than the paired observations, we take the difference between the matched
observations and perform analysis on the differences.
Classic example would be weight loss
Weight before
Weight after
We are interested in change
An aside: this study also used a placebo group. Why?
145

Here, µ1 is the mean of the population of the first set of measurements and µ2
is the mean of the population of the second set of measurements.
Actually, you can also use µd = µ2 -µ1 long as you are consistent with your
statement of hypotheses and calculation of D.
Formulas for the t Test for Dependent Samples
146

The good news is we can find the mean of the differences, ¯D, and the
standard deviation of the differences, sD , using the LIST and STAT functions in
your TI-83/84.
1. Go to STAT -> EDIT -> Edit .
2. Enter the first set of observations under L1.
3. Enter the second set of observations under L2.
4. Highlight L3 in list, type L1 - L2 and hit enter. The set of differences should
now be calculated.
5. Go to STAT -> CALC -> 1-Var Stats, hit enter. Type L3 (after 1-Var Stats on
your screen) and hit enter. Your calculator will calculate the sample mean
and standard deviation for you .
Example: A physical education director claims by taking a special vitamin, a
weight lifter can increase his strength. Eight athletes are selected and given a test
of strength, using the standard bench press. After 2 weeks of regular training,
supplemented with the vitamin, they are tested again. Test the effectiveness of the
vitamin regimen at a = 0.05. Each value in these data represents the maximum
number of pounds the athlete can bench press. Assume the variable is
approximately normally distributed.

Before X1 After

210 219
230 236
182 179
205 204
262 270
253 250
219 222
216 216
Step 1: State the hypotheses and identify the claim.
I will base my differences on: D = strength after - strength before.
Ho: µD=0 H1: µD>0
Step 2: Find the critical value(s) from the appropriate table.
We have 8 lifters which gives 7 degrees of freedom. Our ∞=0.05.
We have a right tailed test => critical value will be positive. t critical value =
1.895
147

Step 3: Compute the test value and determine the P-value


Use your calculator to get the mean difference and standard deviation of the
differences.

Step 4: Make the decision to reject or not reject the null hypothesis
The null hypothesis is NOT rejected. We can base this decision on either of the
two facts:
 The p-value is larger than ∞ = 0.05
The test value (1.388) is smaller than the critical value (1.895). That is, our
test value is within the non-rejection region.
Step 5: Summarize the results
There is not sufficient evidence to support the education director claims by
taking a special vitamin, a weight lifter can increase his strength.
Practice Example: A sample of 10 college students in a class were asked how
many hours per week they watch TV and how many hours a week they used a
computer. Is there a difference in the mean number of hours a college student
148

spends on a computer versus watching TV at ∞ = 0.01? Assume the population of


differences is approximately normally distributed.
The data:

Student Comp TV
1 30 2
2 20 1.5
3 10 14
4 10 2
5 10 6
6 0 20
7 35 14
8 20 1
9 2 14
10 5 10

Step 1: State the hypotheses and identify the claim.


Step 2: Find the critical value(s) from the appropriate table.
Step 3: Compute the test value and determine the P-value.
Step 4: Make the decision to reject or not reject the null hypothesis.
Confidence Interval for the Mean Difference
S
D  tα/2 D where d.f.  n  1
n
Let’s find the 99% confidence interval for the mean difference of the example of
TV watching vs. Computer Usage.
10.4 REVISION POINTS
Steps for conducting t and z test
Step 1: State the hypotheses and identify the claim.
Step 2: Find the critical value(s) from the appropriate table.
Step 3: Compute the test value and determine the P-value.
Step 4: Make the decision to reject or not reject the null hypothesis.
Step 5: Summarize the results
10.5 INTEXT QUESTIONS
1. What are the uses of z test?
2. What are the uses of t test?
3. Explain the situation to use t and z test
149

10.6 SUMMARY
Basic data analysis provides valuable insights and guides the rest of the data
analysis as well as interpretation of results. The parametric and non parametric
tests are also available for testing the hypothesis related to difference. In parametric
case, the t and z test is used to determine hypothesis related to population mean.
Different forms of t and z test are suitable for testing hypothesis based on sample
and two independent samples or paired samples.
10.7 TERMINAL EXERCISES
1. The observed difference between the sample means may be due to ------
2. A----------test is used to test the difference between means when the two
samples are independent and when the samples are taken from two
normally or approximately normally distributed populations
3. Each person is measured twice where the 2 measurements measure the
same thing but under different conditions is known as ---------
4. We cannot use the ----- test we had learned for a difference in independent
means
10.8 SUPPLEMENTARY M ATERIALS
a) Identify the most appropriate statistical test
b) Identify the independent variable(s) or factor(s)
c) Identify the dependent variable
d) Identify the scale of measurement for the independent and dependent
variables.
10.9 ASSIGNMENTS
Research suggests that the antioxidants in foods such as blueberries can
reduce perhaps reverse age-related declines in cognitive functioning. To test this
phenomenon, a researcher selects a sample of 25 adults aged between 70 and 75
years old and administers a cognitive function test to each participant. The
participants then drink a blueberry supplement every day for 4 months before they
are tested again. The researcher compares the scores before treatment with the
scores after treatment to see if there is any change in cognitive function.
10.10 SUGGESTED READING/REFERENCE BOOKS
1. De Vaus, D.A. (1991) Surveys in Social Research, 3rd edition, London: UCL
Press.
2. Gash, S. (1999) Effective Literature Searching for Research, 2nd edition,
Aldershot: Gower.
3. Hart, C. (2001) Doing a Literature Search, London: Sage.
Spence, G. (2001) A Simple Guide to Internet Research, Harlow: Prentice Hall
150

10.11 LEARNING ACTIVITIES


1. Describe principles and procedures of review of literature and precautions in
consulting library material
10.12 KEYWORDS
Alternate hypothesis: The formulation of an alternate hypothesis is a
convention in scientific circles. Its main function is to specify explicitly the
relationship that will be considered as true in case the research hypothesis proves
to be wrong. In a way, an alternate hypothesis is the opposite of the research
hypothesis.
Dependent variable : When establishing causality through a study, the variable
assumed to be the cause is called an independent variable and the variables in
which it produces changes are called the dependent variables. A dependent variable
is dependent upon the independent variable and it is assumed to be because of the
changes.
Independent variable: When examining causality in a study, there are four sets
of variables that can operate. One of them is a variable that is responsible for
bringing about change. This variable which is the cause of the changes in a
phenomenon is called an independent variable. In the study of causality, the
independent variable is the cause variable which is responsible for bringing about
change in a phenomenon.






151

LESSON – 11

F- TEST (ANOVA)
11.1 INTRODUCTION
Analysis of variance (abbreviated as ANOVA) is an extremely useful technique
concerning researches in the fields of economics, biology, education, psychology,
sociology, business/industry and in researches of several other disciplines. This
technique is used when multiple sample cases are involved. As stated earlier, the
significance of the difference be tween the means of two samples can be judged
through either z-test or the t-test, but the difficulty arises when we happen to
examine the significance of the difference amongst more than two sample means at
the same time. The ANOVA technique enables us to perform this simultaneous test
and as such is considered to be an important tool of analysis in the hands of a
researcher. Using this technique, one can draw inferences about whether the
samples have been drawn from populations having the same mean.
The ANOVA technique is important in the context of all those situations where
we want to compare more than two populations such as in comparing the yield of
crop from several varieties of seeds, the gasoline mileage of four automobiles, the
smoking habits of five groups of university students and so on. In such
circumstances one generally does not want to consider all possible combinations of
two populations at a time for that would require a great number of tests before we
would be able to arrive at a decision. This would also consume lot of time and
money, and even then certain relationships may be left unidentified (particularly
the interaction effects). Therefore, one quite often utilizes the ANOVA technique and
through it investigates the differences among the means of all the populations
simultaneously.
11.2 OBJECTIVES
After studying this lesson, you will be able to understand:
 ANOVA
 Importance of ANOVA
 Assumptions
 Conditions
 ANOVA model
10.3 CONTENTS
10.3.1 What is ANOVA?
10.3.2 Importance of ANOVA
10.3.3 Assumptions
10.3.4 Conditions for ANOVA
10.3.5 ANOVA Model
152

11.3.1 What is ANOVA?


Analysis of variance (ANOVA) is used to test hypotheses about differences
between two or more means. The t-test based on the standard error of the
difference between two means can only be used to test differences between two
means. When there are more than two means, it is possible to compare each mean
with each other mean using t-tests. However, conducting multiple t-tests can lead
to severe inflation of the Type I error rate. Analysis of variance can be used to test
differences among several means for significance without increasing the Type I error
rate. This chapter covers designs with between -subject variables. The next chapter
covers designs with within-subject variables.
The statistical method for testing the null hypothesis that the means of several
populations are equal is analysis of variance. It uses a single factor, fixed –effects
model to compare the effects of one factor (brands of coffee, varieties if residential
housing, types if retail stores) on a continuous dependent variable. In a fixed effects
model, the levels of the factor are established in advance and the results are not
generalizable to other levels of treatment.
Consider a hypothetical experiment on the effect of the intensity of distracting
background noise on reading comprehension. Subjects were randomly assigned to
one of three groups. Subjects in Group 1 were given 30 minutes to read a story
without any background noise. Subjects in Group 2 read the story with moderate
background noise, and subjects in Group 3 read the story in the presence of loud
background noise.
The first question the experimenter was interested in was whether background
noise has any effect at all. That is, whether the null hypothesis: µ 1 = µ2 = µ3 is true
where µ1 is the population mean for the "no noise" condition, µ 2 is the population
mean for the "moderate noise" condition, and µ 3 is the population mean for the
"loud noise" condition. The experimental design therefore has one factor (noise
intensity) and this factor has three levels: no noise, moderate noise, and loud noise.
Analysis of variance can be used to provide a significance test of the null
hypothesis that these three population means are equal. If the test is significant,
then the null hypothesis can be rejected and it can be concluded that background
noise has an effect.
In a one-factor between- subjects ANOVA, the letter "a" is used to indicate the
number of levels of the factor (a = 3 for the noise intensity example). The number of
subjects assigned to condition 1 is designated as n 1; the number of subjects
assigned to condition 2 is designated by n 2, etc.
If the sample size is the same for all of the treatment groups, then the letter "n"
(without a subscript) is used to indicate the number of subjects in each group. The
total number of subjects across all groups is indicated by "N." If the sample sizes
are equal, then N = (a)(n); otherwise,
N = n 1 + n 2 + ... + n a.
153

Some experiments have more than one between -subjects factor. For instance,
consider a hypothetical experiment in which two age groups (8 -year olds and 12-
year olds) are asked to perform a task either with or without distracting
background noise. The two factors are age and distraction.
11.3.2 Importance of ANOVA
It is important to know about analysis of variance because it is the most
common inferential statistical procedure used in experiments. Why? Because there
are actually many versions of ANOVA, so it can be used with many different
designs: It can be applied to an experiment involving independent samples or
related samples, to an independent variable involving any number of conditions,
and to a study involving any number of independent variables. Such complex
designs are common because, first, the hypotheses of the study may require
comparing more than two conditions of an independent variable. Second,
researchers often add more conditions because, after all of the time and effort
involved in creating two conditions, little more is needed to test additional
conditions. Then we learn even more about a behavior (which is the purpose of
research). Therefore, you’ll often encounter the ANOVA when conducting your own
research or when reading about the research of others.
11.3.3 Assumptions
Analysis of variance assumes normal distributions and homogeneity of
variance. Therefore, in a one-factor ANOVA, it is assumed that each of the
populations is normally distributed with the same variance (σ²). In betwe en-
subjects analyses, it is assumed that each score is sampled randomly and
independently. Research has shown that ANOVA is "robust" to violations of its
assumptions.
This means that the probability values computed in an ANOVA are
satisfactorily accurate even if the assumptions are violated. Moreover, ANOVA tends
to be conservative when its assumptions are violated. This means that although
power is decreased, the probability of a Type I error is as low or lower than it would
be if its assumptions were met. There are exceptions to this rule. For example, a
combination of unequal sample sizes and a violation of the assumption of
homogeneity of variance can lead to an inflated Type I error rate.
11.3.4 Conditions for ANOVA
1. The sample must be randomly selected from normal populations
2. The populations should have equal variances
3. The distance from one value to its group’s mean should be independent of
the distances of other values to that mean (independence of error).
4. Minor variations from normality and equal variances are tolerable.
Nevertheless, the analyst should check the assumptions with the diagnostic
techniques.
154

Analysis of variance breaks down or partitions total variability into component


parts. Unlike the‘t’ test, which uses the sample standard deviations, ANOVA uses
squared deviations of the variance so computation of distances of the individual
data points from their own mean or from the grand mean can be summed.
In ANOVA model, each group has its own mean and values that deviate from
that mean. Similarly, all the data points from all of the groups produce an overall
grand mean. The total deviation is the sum of the squared differences between each
data point and the overall grand mean.
The total deviation of any particular data point may be partition ed into
between-groups variance and within group variance. The between group variance
represents the effect of the treatment or factor. The difference of between -groups
means imply that each group was treated differently. The treatment will appear as
deviations of the sample means from the grand mean. Even if this were not so,
there would still be some natural variability among subjects and some variability
attributable to sampling. The within-groups variance describes the deviations of the
data points within each group from the sample mean. This results from variability
among subjects and from random variation. It is often called error.
When the variability attributable to the treatment exceeds the variability
arising from error and random fluctuations, the viability of the null hypothesis
begins to diminish. And this is exactly the way the test static for analysis of
variance works.
The test statistics for ANOVA is the F ratio. It compares the variance from the
last two sources:
F= Between group variance = Mean square between
Within group variance Mean square within
Where,
Mean square between = Sum of square between
Degree of freedom between
Mean square within = Sum of square withinn
Degree of freedom within
To compute the F ratio, the sum of the squared deviations for the numerator
and denominator are divided by their respective degrees of freedom. By dividing,
computing the variances as an average or mean, thus the term mean square. The
degrees of freedom for the numerator, the mean square between groups, is one less
than the number of groups (k-1). The degree of freedom for the denominator, the
mean square within groups, is the total number of observations minus the number
of groups (n-k).
If the null hypothesis is true, there should be no difference between the
populations, and the ratio should be close to 1. If the population means are not
equal, the numerator should manifest this difference. The F ratio should be greater
155

than 1. The f distribution determines the size of ratio necessary to reject the null
hypothesis for a particular sample size and level of significance.
11.3.5 ANOVA model
To illustrate reports one way ANOVA, consider the following hypothetical
example. To find out the number one best business school in India, 20 business
magnets were randomly selected and asked to rate the top 3 B-schools. The ratings
are given below
Data

Person Rating of B-school 1 Rating of B-school 2 Rating of B-school 3


1 40 56 92
2 28 48 56
3 36 64 64
4 32 56 72
5 60 28 48
6 12 32 52
7 32 42 64
8 36 40 68
9 44 61 76
10 36 58 56
11 40 52 88
12 68 70 79
13 20 73 92
14 33 72 88
15 65 73 73
16 40 71 68
17 51 55 81
18 25 68 95
19 37 81 68
20 44 78 78
Let’s apply the one way ANOVA test on this example.
Step 1: null hypothesis
H0: A1=A2 =A3
HA : A1 A2 A3
Step 2: statistical Test
The F test is chosen because the example has independent samples, accept
the assumptions of analysis of variance and have interval data.
Step 3: significance level
156

Let  = 0.05 and degree of freedom=[numerator (k-1) = (3-1)=2], [denominator


(n-k)=(60-3)=57]= 2,57
Step 4: Calculated value
Mean squarebetween
F
Mean squarewithin
F= 5822.017/ 205.695 = 28.304 degree of freedom (2, 57)
Step 5: critical test value
From the F-distribution table with degree of freedom (2, 57),
 = 0.05 the critical value is 3.16.
Step 6: Decision
Since the calculated value is greater than the critical value (28.3>3.16), the
null hypothesis is rejected. The conclusion is there is statistically significant
difference between two or more pairs of means. The following table shows that the p
value equals 0.0001. Since the p value (0.0001) is less than the significance level
(0.05), this has the second method for rejecting the null hypothesis.
The ANOVA model summary given in the following table is the standard way of
summarizing the results of analysis of variance. It contains the sources of variation,
degrees of freedom, sum of squares, mean squares and calculated F value. The
probability of rejecting the null hypothesis is computed up to 100 percent  -- that
is, the probability value column reports the exact significance for the F ratio being
tested.
Table: One way ANOVA model summary

Degree of freedom Sum of squares Mean squares F value ‘p’ value


2 11644.033 5822.017 28.304 0.0001
57 11724.550 205.694
Total 59 23368.583

Table on means

Business school Persons Mean Standard deviation Standard error


1 20 38.95 14.01 3.13
2 20 58.90 15.09 3.37
3 20 72.90 13.90 3.11
157

Table on Scheffes multiple comparison procedure


Critical
Business school comparison Difference ‘p’ value
difference
1&2 19.95 11.40 0.0002 S
1&3 33.95 11.40 0.0001 S
2&3 14.00 11.40 0.0122 S
S = significantly different at this level. Significance level: 0.05
All data are hypothetical
Figure on One way analysis of variance plots

Figure on One way analysis of variance plots


158

11.4 REVISION POINTS


1. Analysis of variance (abbreviated as ANOVA) is an extremely useful
technique concerning researches in the fields of economics, biology,
education, psychology, sociology, business/industry and in researches of
several other disciplines.
2. The ANOVA technique enables us to perform this simultaneous test and as
such is considered to be an important tool of analysis in the hands of a
researcher. Using this technique, one can draw inferences about whether
the samples have been drawn from populations having the same mean.
3. The statistical method for testing the null hypothesis that the means of
several populations are equal is analysis of variance.
4. ANOVA uses a single factor, fixed –effects model to compare the effects of
one factor (brands of coffee, varieties if residential housing, types if retail
stores) on a continuous dependent variable.
11.5 INTEXT QUESTIONS
1. Explain the meaning of analysis of variance. Describe briefly the technique
of analysis of variance for one-way and two-way classifications.
2. State the basic assumptions of the analysis of vari ance.
3. What do you mean by the additive property of the technique of the analysis
of variance? Explain how this technique is superior in comparison to
sampling.
11.6 SUMMARY
In ANOVA the dependent variable is metric and independent variables are all
categorical or combinations of categorical and metric variables. One way ANOVA
involves a single independent categorical variable. Interest lies in testing the null
hypothesis that category means are equal in the population.
11.7 TERMINAL EXERCISES
1. ------------ is used when multiple sample cases are involved
2. ANOVA is used to test hypothesis about difference between -------- means.
11.8 SUPPLEMENTARY MATERIALS
1. Sekaran, U. (1983 Fall). Methodological and theoretical issues and
advancements in cross- cultural research. Journal of International
Business, 61–73.
2. Sekaran, U. (1986). Dual-career families: Contemporary organizational and
counseling issues. San Francisco: Jossey-Bass
11.9 ASSIGNMENTS
1. An e-commerce research company claims that 60% or more graduate
students have bought merchandise on-line. A consumer group is suspicious of the
claim and thinks that the proportion is lower than 60% . A random sample of 80
graduate students show that only 22 students have ever done so. Is there enough
159

evidence to show that the true porportion is lower than 60% ? Conduct the test at
10% Type I error rate, and use the p-value and rejection region approaches.
11.10 LEARNING ACTIVITIES
1. Which procedure is more useful in marketing research –ANOVA or Chi
Square test or T – test? Discuss as a small group
11.11 SUGGESTED READINGS/ REFERENCE BOOKS
1. Campbell, A. A., & Katona, G. (1966). The sample survey: A technique for
social science research. In L. Festinger & D. Katz (Eds.), Research methods
in the behavioral sciences. New York: Holt, Rinehart and Winston.
2. Cronbach, L. J. (1990). Essentials of psychological testing (5th ed.).
New York: Harper & Row.
3. Davies, G. R., & Yoder, D. (1937). Business statistics. New York: John Wiley.
4. Davis, D., & Cosenza, R. M. (1988). Business research for decision
making (2nd ed.). Boston: PWS-Kent Publishing.
11.12 KEYWORDS
One-tailed test: The test used to evaluate a statistical hypothesis that predicts
that scores will only increase or only decrease
One-way ANOVA: The analysis of variance performed when an experiment has
only one independent variable
Nonsignificant: Describes results that are considered likely to result from
chance sampling error when the predicted relationship does not exist; it indicates
failure to reject the null hypothesis
Probability: A mathematical statement indicating the likelihood that an event
will occur when a particular population is randomly sampled; symbolized by p
Mean square: In ANOVA, an estimated population variance, symbolized by MS
F-distribution: The sampling distribution of all possible values of F that occur
when the null hypothesis is true and all conditions represent one population
Fisher’s protected t-test: The post hoc procedure performed with ANOVA to
compare means from a factor in which all levels do not have equal ns
F-ratio: In ANOVA, the ratio of the mean square between groups to the mean
square within groups






160

LESSON – 12

MEASURE OF RELATIONSHIP ---- CORRELATION AND REGRESSION


12.1 INTRODUCTION
In Social Study as well as Psychology there are times where it is needed to know
whether there exists any relationship between the different abilities of the individual
or they are independent of each other. Consequently, there are numerous questions
like the following which, have to be answered.
 Does scholastic achievement depend upon the general intelligence of a
child?
 Is it true that the height of the children i ncreases with the increase in
their age?
 Is there are relationship between the size of the skull and general
intelligence of the individuals?
 Is it true that Dull children tend to be more neurotic than the bright
children?
The questions and problems like the above in which there is a need of finding
out the relationship between two variables (Age and Height Intelligence and
Achievement etc.) can be tackled properly by the method of correlation.
There are many types of correlation like Linear, Curvilinear, Biserial, Partial or
Multiple correlations that are computed in Statistics. As we, in this text, aim to
have an elementary knowledge of the statistical methods we will take only the
Linear correlation in the following pages.
12.2 OBJECTIVES
After studying this lesson, you will be able to understand:
 Measure of relationship
 Correlation
 Linear correlation
 Scatter plot
 Regression
 Dependent and independent variable
12.3 CONTENTS
12.3.1 Factors influencing the size of correlation coefficient
12.3.2 Linear correlation
12.3.3 Coefficient of correlation
12.3.4 Computation of coefficient of correlation
12.3.5 Interpretation of Pearson’s correlation coefficient
12.3.6 Regression
161

12.3.1 Factors influencing the size of correlation Coefficient


1. The size of r is very much dependent upon the variability of measured
values in the correlated sample. The greater the variability, the higher will be the
correlation, everything else being equal.
2. The size of r is altered when researchers select extreme groups of subjects in
order to compare these groups with respect to certain behaviors. Selecting extreme
groups on one variable increases the size of r over what would be obtained with
more random sampling.
3. Combining two groups which differ in their mean values on one of the
variables is not likely to faithfully represent the true situation as far as the
correlation is concerned.
4. Addition of an extreme case (and conversely dropping of an extreme case)
can lead to changes in the amount of correlation. Dropping of such a case leads to
reduction in the correlation while the converse is also true.
12.3.2 Linear Correlation
This the simplest kind of correlation to be found between the two sets of scores
or variables. Actually when the relationship between two sets of scores or variables
can be represented graphically by a straight line, it is known as Linear Correlation.
Such type of correlation clearly reveals how the change in one variable is
accompanied by a change or to what extent increase or decrease in one is
accompanied by the increase or decrease in order.
The correlation between two sets of measures of variables can be positive or
negative. It is said to be positive when an increase (or decrease) in the corresponds
to an increase (or decrease) in the other. It is negative when increase corresponds to
decrease and decrease corresponds with increase. There is also possibility of third
type of correlation i.e. zero correlation between the two sets of measures of
variables if there exists no relationship between them.
Assumptions
We have in the past considered two types of assumptions:
 Validity assumptions
 Distributional assumptions
Validity assumptions require valid measurements, a good sample, and
uncompounded comparisons. There requirements are consistent throughout your
studies. The only distributional assumption needed to use r descriptively is
linearity. Correlation coefficients assume the relation can be described with a
straight line. The most direct way to determine linearity is to look at the scatterplot.
Inferential methods require that the joint distribution of X and Y is bivariate
Normal. This means the three-dimensional distribution of the scatter plot is bell -
shaped from all angles:
162

Scatter Plot
The first step is creating a scatter plot of the data. “There is no excuse for
failing to plot and look.” In general, scatter plots may reveal a
• Positive Correlation (high values of X associated with high values of Y)
• Negative Correlation (high values of X associated with low values of Y)
• No Correlation (values of X are not at all predictive of values of Y).
These patterns are demonstrated in the figure to the right.

12.3.3 Coefficient of Correlation


For expressing the degree of relationship quantitatively between two sets of
measures of variables we usually take the help of an index that is known as
coefficient of correlation. It is a kind of ratio which expresses the extent to which
changes in one variable are accompanied with changes in the other variable. It
involves no units and varies from -1 (indicating perfect negative correlation) to + 1
(indicating perfect positive correlation). In case the coefficient of correlation is zero
it indicates zero correlation between two sets of measures.
163

12.3.4 Computation of Coefficient of Correlation


There are two different methods of computing coefficient of correlation (linear).
There are–
(a) Rank Difference Method
(b) Product Moment Method
a) Rank Difference Method of Computing Coefficient of Correlation
In computing coefficient of correlation between two sets of scores achieved by
the individuals, with the help of this method we require ranks i.e. positions of
merits of these individuals in the possession of certain characteristics. The
coefficient of correlation computed by this method as it considers only the ranks of
the individuals in the characteristics A and B is known as Rank correlation
coefficient and is designated by Greek letter (Rho). Sometimes it is also known as
Spearman's coefficient of correlation after the name of its inventor.
In case where we do not have scores and have to work with data in which
differences between the individuals in the possession of certain characteristics can
be expressed only by ranks. Rank correlation coefficient is the only correlation
coefficient that can be computed. But this does not mean that it cannot be
computed from the usual data given in scores. In case the data contain scores of
individuals, we can compute by converting them into ranks. For example, if the
marks of a group of 5 students are given as 17, 25, 9, 35, 18, we will rank them as
4, 2, 5, 1 and 3. We determine the rank of position of the individuals in both the
given sets of scores. These ranks are then subjected to further calculation for the
determination of the coefficient of correlation.
164

How it is done can be understood properly through the following illustration -


Example
Marks in the Difference
Marks in subject o f Rank in in
Rank in Difference
Individual the subject Civics y signs History Rank
Civics squared D
o f History +ve or -ve d = Ri irrespective
Rj - R2 of
A 80 82 2 3 1 1
B 45 86 11 2 9 81
C 55 50 10 10 0 0
D 56 48 9 191 2 4
E 58 60 8 9 1 12
F 60 62 7 8 1 1
G 65 64 6 7 1 1
H 68 65 5 6 1 1
I 20 70 4 5 1 1
J 75 74 3 4 1 1
K 85 90 1 1 0 1
N = 11, d2 =92


Steps for Calculation of p
1. First of all it is required to assign position of merit or rank to each
individual joins either test. These ranks are put under column 3. (Designated as R1 )
and 4 (designated as R2 ) respectively. The task of assigning ranks in the cases like
example 1st is not difficult. But in the cases, like example 2nd, where two or more
individuals are found to achieve same score, some difficulty arises. In the above
example, in the first test X, B and F are two individuals who have the same score
i.e. 15. Therefore, score 15 occupies 6th position in order of merit. But now the
question arises which one of the two individuals B and F should be ranked as 6th
or 7th. In order to overcome this difficulty we equally share the rank 5th and 7th
between them and thus rank each one of them as 6.5.
165

1. Similarly, if there are three persons who have the same score and share the
same ranks, we take the average of the ranks claimed by these persons. For
example, we can take the score 20 in the 2nd example which is shared by three
individuals D, G and H. It is ranked third in the whole series and therefore the
ranks 3, 4 and 5 are shared equally by D, G and H and hence we attribute rank to
each of them.
2. After writing down the allotted rank to all the individuals 0 either of the two
tests, the differences in these ranks are calculated. In doing so we do not 'consider
the algebraic signs +ve or -ve of the difference. This difference is written under
column 5th (designated as d 2)
3 In the next column (designated as d 2) we square up the Rank difference or
the values of d written in the column 5th.
4. Now the value of p is calculated by the formula, where d2 stands for the sum
of the squares of differences between the ranks of the scores on two different tests
and N for the number of individuals whose scores are under consideration for
computing.

b) Product Moment Method of Computing Coefficient of Correlation


This method is also known as Pearson Moment method in the honour of the
English statesman Karl Pearson who is said to be the inventor of this method. The
coefficient of correlation computed by this method is known as Product Moment
coefficient of correlation symbolically represented by ‘r’.
(a) The calculation of ‘r’ from ungrouped data
The basic formula for the computation of ‘r’ for the ungrouped data by this
method is

Where x and y represent the deviation of scores in the tests X and Y from the
means of each distribution. The procedure of calculation r by this formula can be
understood by the following illustration
Individuals Scores in Scores in X Y Xy X2 2
Y

A 15 60 -10 10 -100 100 100


B 25 70 0 20 0 0 400
C 20 40 -5 -10 50 25 190
D 30 50 5 0 0 25 0
E 35 30 -10 -20 -200 100 400
∑xy = -250 ∑x2 = 250 ∑y2 = 1000
166

Computation of r directly from raw scores when deviations are take n from zero
(without calculating deviations from the means. Here we apply the formula.
Scores in 1st Scores in 2nd
Subject
Test (X) Test (Y) XY X 2 Y 2
A 5 12 60 25 144
B 3 15 45 2 225
C 2 11 22 4 121
D 8 10 80 64 100
E 6 18 108 36 324

12.3.5 Interpretation of Pearson’s Correlation Coefficient


The sign of the correlation coefficient determines whether the correlation is
positive or negative. The magnitude of the correlation coefficient determines the
strength of the correlation. Although there are no hard and fast rules for describing
correlation strength, many experts [hesitatingly] offer these guidelines:
0 < |r| < .3 weak correlation
.3 < |r| < .7 moderate correlation
|r| > 0.7 strong correlation
For example, r = -0.849 suggests a strong negative correlation.
12.3.6 Regression
It is quite often that the investigator likes to check how variables are related,
and to "find if he can" a relationship among the variables on hand. An interesting
question might be raised by a researcher such as: am I able to predict the value of a
random variable if I have the value of one or more variables available? This kind of
study is what we call regression analysis. As we know variables are two types,
either independent or dependent, based on how the value of that variable is
attained or obtained? For example the relationship between heights and weights of
individuals, temperature and pressure, the weight of a fish and its breathing
capacity, the annual food expenditure and the annual income of the family, and so
on, might be of interest to the scientist. In the above examples there were two
167

quantities, and we could label them as: input and output, independent and
dependent, or explanatory and response. For example, the family income is the
independent or explanatory variable while the food expenditure will be called the
dependent or response variable.
The response variable is that variable whose value can be explained by the
value of one or more explanatory variable or predictor variables. There are many
types of regression, based on the number of variables involved. When we have one
response and one predictor, this is the case for linear regression. In case we have
one response and many predictor variables, this is the case of multiple regressions.
In this chapter, we are interested in the case of two variables, and when the
relationship between them is assumed to be a linear one. In addition to linear
regression between two variables there might be logarithmic, exponential,
quadratic, cubic, and so on regression.
The purpose of regression analysis is to determine the existence, the kind, or
the extent of a relationship (in the form of a mathematical equation) between two or
more variables. For example, if x and y denote two variables under study, then, in
general, we want to determine the best equation (relation) y = f(x) that describes the
relationship between x and y.
The term "regression" goes back to Sir Francis Galton. In his work on
hereditary, Galton noted that the sons of very tall fathers tended to be tall, but n ot
quite as tall as their fathers. Also, sons of very short fathers tended to be short, but
not quite as short as their fathers. He called this tendency, of individuals to be not
quite tall or as short as their fathers, the tendency of regression toward mediocrity",
i.e., going back to the average.
In much of the experimental work, it is desired to investigate how the changes
in one variable affect another variable, or introduce some changes in that variable.
One can distinguish between exact and non -exact relationships.
Exact Relationship
Sometimes we find that two variables are linked by an exact relationship. For
example, if the resistance "R" of a simple circuit is kept constant, the current "I" is
related to the voltage "V" by Ohm's Law: I = V/R, which is an equation of first
degree and the graph of it, is a straight line. This is an exact law. However, if we
want to verify it empirically by making changes in V and observing I while R is kept
constant we notice that the plot of the empirical points (V, I) will not fall exactly on
a straight line. The reason is that the measurements maybe subject to slight errors
and thus the plotted points would not fall exactly on the straight line but would
vary randomly around it. For purposes of predicting the value for I for a particular
value for V (with R as a fixed constant) we should use the plotted straight line .
Non-Exact Relationship
Sometimes the relationship is not exact even apart from the errors in
measurements. For example, suppose we consider the height and weight of adult
168

males for some population. If we plot the ordered pair (weight, height) = (x, y), a
diagram, like Figure 1 below, will result. Such a representation is conventionally
called a scatter diagram

Note that for any given height there is a range of observed weights and vice
versa. This variation will be partially due to measurement errors but primarily due
to variation among individuals. If you consider two males with same height, their
weights will be different not only because we cannot me asure the weights exactly
correct, but also because the two males will have slight different weights. Thus no
unique relationship between the actual weight and height can be written. But we
can notice that the average observed weight, for a given observed height,
increases as height increases. Whether a relationship is exact or non -exact, in so
far as average values are concerned, it will be useful especially for prediction
purposes.
It is not always clear which variable should be considered the response
variable and which the explanatory variable. For example, does high school GPA
predict a student's SAT score or can the SAT score be used to predict the student's
GPA? The investigator should determine which variable plays the role of the
explanatory variable based on the question that needs to be answered.
Scatter Diagrams show the type of the relation that exists between two
variables. By plotting the explanatory variable along the horizontal axis and the
response along the vertical axis, our goal is to distinguish which scatter diagrams
will imply a linear relation from those that will imply a nonlinear relation and those
that imply no relation. Figure 2 and Figure 3 show some scatter diagrams and the
type of relation implied.
169

Figure 3 – Perfect Negative correlation

Figure – 4
Figure 4, above, shows the type of linear and nonlinear relationships between
two variables.
In Figure 4A we see a perfect negative linear relationship, and we say that the
two variables are negatively associated. In other words, two variables are negatively
associated if, whenever the value of one variable increases, the value of the other
variable decreases.
On the other hand, Figure 4B shows that there are two variables that are
linearly related and positively associated. In the same vein, two variables are
170

positively associated if, whenever the value of one variable increases, the value of
the other variable increases.
The situation is quite different in Figure 4C. There is the case where we see a
horizontal line, although it is linear, but the response variable, Y, is not affected by
the change in the explanatory variable, x. Thus Y = a constant.
Figure 4D shows a non-linear relationship between the two variables x and y.
Figure 4 displays ideal cases for positively and negatively associated variables
based on a linear relationship between the two. Later we will check on the strength
of that linear relationship between the explanatory variable x and the response
variable Y.
2.7.1 Regression Models
The simplest linear regression model is of the form

Where the response variable, y, depends on one regressor, or explanatory,


variable, x, and the equation is a polynomial of the first degree in x. This model is
called the simple first degree model. β0 and β1are unknown constants, called the
parameters of the model. Specifically, β0 is the y-intercept and β1 is called the
regression coefficient, or the slope of the line. The symbol E denotes a random error
(the model error), which will be assumed to have a normal distribution with me an 0
and variance xxxx. The first order model (or the straight line relationship between
the response and the explanatory variables) can be valuable in many situations.
The relationship may actually be a straight line relationship as in the example on
Ohm's law. Even if the relationship is not actually of the first order (or linear), it
may be approximated by a straight line at least over some range of the input data.
In the Figure 5 below, the relationship between x and y is obviously nonlinear over
the range 0 < x < 100. However, if we were interested primarily in the range 0 < x <
25, a straight line relationship evaluated, on observations in this range, might
provide a perfectly adequate representation of the function. Hence the relationship
that just got fitted would not apply to values of x beyond the restricted range and
could not be used for predictive purposes outside that range.

Figure 5
171

12.7.2 Fitting the straight line (first order model)


The first order model given by
y = 0 + 1 x + E
has the unknown parameters 0 and 1, that need to be estimated in order to
use the relationship for prediction. Thus a sample of size n of ordered observations
(x1, y 1), (x2, y 2), ... , (xn, y n), will be used to estimate those two parameters. The
estimators of those two parameters will be denoted by b0 and b1 respectively, and
hence we will have the following equation
yˆ  b0  b1 x. ,
In the notation, above, ŷ is used in the least squares regression line to serve as
a predicted value of y for a given value of x. It is worth noting that the least squares

line contains the point x, y  where


n

x i
x i 1
n
And
n

y i
y i 1
n
Each observation (xi, y i), i = 1, 2,..., n, in the sample, satisfies the equation
Yi = 0 + 1x1+i
Where i is the value assumed by Ei when Yi takes on the value y i. The above
equation can be viewed as the model for a single observation y. Similarly using the
estimated, or fitted regression line

Each pair of observations satisfies the relation

Where e i = y i - y1 is called a residual and describes the error in the fir of the
model at the ith data point or observation
We shall find b0 and b1, the estimates of 0 and 1, so that the sum of the
squares of the residuals is a minimum. The residual sum of squares is often called
the sum of squares of the errors about the regression line and denoted by SSE. This
minimization procedure for estimating the parameters is called the Method of Least-
Squares. Thus we shall find b0 and bl so as to minimize
172

n n n
SSE  e 2 i  ( yi  yˆ1 ) 2  ( yi  b0  b1 xi ) 2
i 1 i 1 i 1

(The interested reader, who is aware of differential calculus, can check the
procedure in R.E. Walpole, and R.H. Myers; Probability and Statistics for Engineers
and Scientists, 4 th Edition, p. 361.) The least -squares regression line is the line
that minimizes the square of the vertical distance between the observed values of Y

and those predicted by the line, ŷ Thus solving the normal equations resulting from
the differentiation of the SSE, we see that

 n n

 n ( xi )( yi ) 
 x , y  i 1 

i 1

i 1
i i
n 
 
b1   
  n
 
2
   x
 n 2  i 1  i 
x i  
 i 1 
 
  , and b0 = yˆ  b1x
EXAMPLE 5.1
The following measurements of the specific heat of a certain chemical were
made in order to investigate the variation in specific hear with temperature.

Tempo 0 10 20 30 40

Specific Hear 0.51 0.55 0.57 0.59 0.63


Estimate the regression line of specific hear on temperature, and predict the
o
value of the specific heat when the temperature is 25 C
Solution
n n
From the data we have: n  5,  X1  10, x - 200,  Y1  2.85 y  0.57 , and
i-l i-l
n

X
il
i Yi = 59.8. Thus by applying the above formulas for bo b1 , we have bo  0.514 ,

and b1  0.0028. Hence the fitted equation will be given by y = 0.514 0.0028x.
When temperature x  25 C the predicted specific heat is 0.514-00028*25 = 0.584.
0

EXAMPLE 5.2 (Method of coding)


When the values of the controlled (independent, explanatory) variable are
equally spaced, the calculations of the regression line become considerably
simplified by coding the data in integers symmetrically about zero to make
n
b xi 0
i 1 and thus the estimators of coefficients in the regression line become.
173

n n
bi   xi yi /  x 2i , andb0  y
i 1 i 1

If there are an odd number of evenly spaced data values for X, denote them by:
-3, -2, -1, 0, 1, 2, 3 ...
If there is an even number of evenly spaced data values for X, denote them by:
-5, -3, -1, 1,3, 5, ....
Consider the following data that stands for the output of a certain company for
the years 1960-1964. Predict the output of the company in the year 1965.

Solution
Based on the above setup we can see that the data can be coded to look like
the following table.
Output (1000 tons) 11.1 12.3 13.7 14.6 15.6
Year -2 -1 0 1 2
For manual calculations, we have the following Table.

Variable Data Total


Output y 11.1 12.3 13.7 14.6 15.6 67.3
Year x -2 -1 0 1 2 0

x2 4 1 0 1 4 10

XY -22.2 -12.3 0 14.6 31.2 11.3

Table
Hence the estimated values in the predicted equation will have the following

values: b0 = 13.46 and b =1.13 and the fitted equation is given by ŷ = 13.46 +
1.13x, where x is the coded year.
To predict the output for the year 1965, we take x = 3, which is its coded

value, in the predicting equation, to obtain ŷ = 13.46 + 1.13*3 = 16.85 or 16850


tons.
EXAMPLE 5.3: (Non-linear Model)
The importance of the linear relationship (or the first order model) goes beyond
the cases where the linearity is apparent. Some relationships are not linear to start
with, but they can be linearized easily. For example consider the experiment growth
(or Decay, when the rate is negative) model:
P = Ae rt.u,
174

Where P is the population size at time t, A is the population size at time zero (a
constant), "r" is a constant representing the rate (of growth when r>0 and decay
when r<0) and u represents a random error. This model describes situations when
the relative growth is constant in the population. It can describe the growth in
human or biological populations, and the growth in compound interest, when u =1.
Consider the following data for the population of the USA (in millions) during
the years 1860-1900:
Year 1860 1870 1880 1890 1900
Population 31.4 39.8 50.2 63.0 76.0
Which can be written as: y = 0 + 1x + E, where y = Ln P, 0 = r, x = t and E =
Ln u = 0. Now we have a linear mode l, we can fit using the data given above.
Moreover, by appealing to EXAMPLE 5.2, and coding the data, we have the
following generated data in Table 2 for manual calculations.

Variable Data Totals


X -2 -1 0 1 2 0
Y = Ln P 3.45 3.68 3.91 4.14 4.33 19.51
XY -6.9 -3.68 4.14 8.66 2.22
2 4 1 0 1 4 10
X
Thus we have
n n
bi   xi , yi /  x 2i  0.222, and
i 1 i 1

b0 = y = 3.902, and the liner model takes the form


ŷ = 3.902 + 0.222 x.

By encoding we reach at P̂ = 49.5e 0.222t


12.7.3 Hypothesis Testing in Regression Analysis
In this section we discuss the precision of the regression coefficients, the
construction of confidence limits, and testing the statistical hypotheses about the
regression coefficients. Before proceeding we need to make some basic assumptions
on the first order model as follows. While the model is given by
Yi = 0 +1xi + i, i = 1, 2, 3, ………….n,
Provided that the i are independent random variables which have a normal
distribution with mean zero and variance  2, i.e. E (i) = 0, and Var (i) =  2. This
implies that
E(Yi)= 0 +1xi,
and
Var (Yi) =  2 ,
175

with the Yi s taken as independent and normally distributed Random variables


Partitioning the Total Variation
The total variation in the response variable (the dependent variable), y, will be
decomposed into meaningful components for later use in testing the hypothesis
about the coefficients of the regression line. Consider the identity

Yi = y + ( i - y ) + (Yi -
ŷi ), for i = 1, 2, 3, ……….., n

Equivalently we have

Yi - y = ( i - y ) + (Yi -
ŷi ), for i = 1,2, 3, ….., n.
In other words, we can write
Total deviation = Deviation due to regression + Deviation about regression
Squaring both sides and summing over i = 1, 2, ....n, we have (interested
reader might need to check this identity by recalling the identity

 (x
i 1
1 x  0)

n n n

(Yi  y)2 ( yˆi  y)2  (Yi  yˆi )2


i 1 i 1 i 1

SStotal = SSregr + SSresiduals


This decomposition shows that, of the total variation in the y s about their
mean, some of the variation can be ascribed to the regression line and some to the
fact that the actual observations do not all lie exactly on a regression line. It is seen
that the value of SSregr can be used as a measure of the importance of the fitted
regression line. If SSreg is large in relative to SS residuals, this indicates the fitted
line is significant.
One measure for the importance of regression is the "Coefficient of
Determination" r2, which is given by r2 = SSreg / SStotal. The above sums of squares
can be computed easily using the following formulas:
n

n n Y 2
i

SStotal   (Yi  y )  Y i  2 2 i 1

i 1 i 1 n
n

n n n Y 2
i

SStotal   (Yi  y )  b0  Yi  b1  X iYi 


2 i 1
, and
i 1 i 1 i 1 n
n n n
SSresiduals  SStotal  SSreg  Yi  b0 Yi  b1  X iYi ,
2

i 1 i 1 i 1
176

What is that r2? It is not surprising to see that it is the square of the linear
correlation coefficient found above, when it needs to be calculated. If r 2 is found
first, clearly we can find r by taking the square root. Another que stion: is r > 0 or r
< 0? How do you decide? No doubt that finding r first is more enlightening than
finding r2 first. Here is another definition.
The coefficient of determination, r2 , measures the percentage of total variation
in the response variable that is explained by the least square regression line based
on the one explanatory variable x. it is a number between 0 and 1 inclusive, that is,
0 <r2 < 1. Recall that r2 is the estimated value of the random variable R 2. Practically
r2 is expressed as a percentage, and it is < 100% . What about that difference
percentage between the 100% and r 2? This difference is the unexplained variation
in Y due to not including all the variables that the response variable can get
affected by.
The results are summarized in the ANOVA Table below:

Source df SS Ms E(MS)
Regression 1 SSregr n
1
SSregr  - b 2  X X)2
1 i
i 1
Residual n-2 SSresidual SStotal - SSregr SSresidual
 -2
Total n-2 SStotal (See Bakir and Shayib
1990)
Testing for Significant Regression
In the model y = 0 + 1x + e, if 1 = 0, then Y is not related to x by the first
linear model, i.e., this first order model is incorrect. However if β1 ≠ 0, then there is
a linear relationship between x and y as described in the model. At this time we
cannot tell whether the relationship is positively related, i.e. y increases as x
increases, or negatively related, i.e., x increases then Y decreases. Thus to test for
significant linear (first order) relationship, between x and Y is equivalent to testing
the following hypotheses:

The test statistic for the above hypothesis is given by


F  MSregr /MSresidual' with

SS
regr
MSregr  , and
1
SS
MSresidual residuals
 -2
177

With the understanding of SSregr , SS , as described with 1 is the degree


residuals
of freedom for the numerator in the F expression, and n – 2 degrees of freedom for
the denominator. Thus the test statistic F, as defined above, will have an F -
distribution with (1, n – 2) degrees of freedom, when Ho is true. Hence the critical
region of size  to reject Ho using the F-distribution, is given by f
1 α
where

f is that value of the F-distribution such that P (F  f


1 α 1  0)  1 - 
12.7.4 Confidence Interval on β0 and β1
As it was the case when we constructed a confidence interval on one
parameter of the population (i.e. on one proportion, or one mean, or one variance)
or on two parameters of the population, we can construct a confidence interval on
the coefficients in the least squares regression line, namely, the parameters β0 and
β1. Clearly, those parameters needed to be estimated by b0 and b1 respectively. We
refer the interested reader, for the proofs and realization of the following formulas,
to Walpole and Myers. thus we have
n
 2  X12
E (bo ) Var (bo )  n
i 1

n (Xi - X) 2
i 1

2
E (b1 )  1 , Var (b1 )  n
n (Xi - X)2
i 1

It is to be noticed that the estimators b0 and b1 are unbiased, with σ2 to be


estimated by residuals MS . Hence the 100 (1- α) % C.I on β0 is given by:

 n



 X1
2


(b o )  t1 - /2 MSresiduals n i 1
 2 
 n (X i - X) 
 i 1 
And the 100 (1 -  ) % on 1 is:
 
 

b1  t1 - /2 n
MSresiduals 
 2 
 n (X i - X) 
 i 1 
When 1 /2 t −α is read from the t-table with n–2 degrees of freedom = residual
degrees of freedom.
Remember there are equivalent forms for the above intervals in terms of the
standard error.
178

For more C.I. on the mean of the response variable at a given value for the
explanatory variable, or for a C.I. for the regression line at a given x value or for a
C.I. on a predicted single y value at a specified x value, we refer the interested
reader to Bakir and Shayib 1990, pp. 172–176. In addition to the above the
interested reader can check that the random variable T2 has an F-distribution.
From the model: y = β0 + β1x + e, again for the sake of completion, with the
usual assumptions for the E’s that are independent random variables with mean
zero and variance σ2, let us give, without derivation, the 100(1 –α) % C.I. for a mean
response, i.e. on E(Y) = β0 + β1x, which depends on the x-value when it is given.
The following formula can be used to construct a confidence interval on μ y|x

When the specified value for the explanatory variable is x0, n is the sample
size, and 1 /2 t −α is the critical value, with n–2 degrees of freedom.
The procedure for obtaining a prediction interval for an individual response is
identical to that for finding a C.I. on the mean of that response. The difference is in
the standard error. More variability is associated with one value than with the
mean of some values. he following formula can be used to construct a prediction
interval for a predicted y value at a specified x-value, x = x0.
12.4 REVISION POINTS
1. The size of r is very much dependent upon the variability of measured values
in the correlated sample.
2. The relationship between two sets of scores or variables can be represented
graphically by a straight line, it is known as Linear Correlation.
3. Correlation coefficients assume the relation can be described with a straight
line.
4. Coefficient Of Correlation is a kind of ratio which expresses the extent to
which changes in one variable are accompanied with changes in the other
variable. It involves no units and varies from -1 (indicating perfect negative
correlation) to + 1 (indicating perfect positive correlation). In case the
coefficient of correlation is zero it indicates zero correlation between two sets
of measures.
5. The response variable is that variable whose value can be explained by the
value of one or more explanatory variable or predictor variables.
179

12.5 INTEXT QUESTIONS


1. Discuss the applications of correlation
2. Discuss the applications of regression
3. What are the various correlations?
12.6 SUMMARY
The product movement correlation coefficient measures the linear association
between two metric variables. Its square measures the proportion of one variable
explained by other. Bivariate regression drives a mathematical equation between a
single metric criterion variable and a single metric predictor variable. The equati on
is derived in the form of straight line by using least square procedure
The purpose of regression analysis is to determine the existence, the kind, or
the extent of a relationship (in the form of a mathematical equation) between two or
more variables.
12.7 TERMINAL EXERCISES
1. ---------- is used to be found between the two sets of scores or variables
2. For expressing the degree of relationship quantitatively between two sets of
measures of variables we usually take the help of an index that is known as
-----------
3. --------- is calculated with the help of rank
4. The ---------- is that variable whose value can be explained by the value of
one or more explanatory variable or predictor variables.
12.8 SUPPLEMENTARY MATERIALS
1. Sekaran, U., & Martin, H. J. (1982, Spring/Summer). An examination of the
psychometric properties of some commonly researched individual
differences, job, and organizational variables in two cultures. Journal of
International Business Studies, 51–66.
2. Wildstrom, S. H. (1998, Jan. 26). Web sites made simpler. Business Week
12.9 ASSIGNMENTS
1. Describe the various measures of relationships often used in context of
research studies. Explain the meaning of the following correlation
coefficients:
(i) ryx, (ii) ry1 × x2, (iii) Ry× x1x2
2. Write an equation in slope intercept form that goes through (2, 6) and (-12, 6).
3. Write an equation in slope intercept form that goes through (-5, 6), (2, 13)
and (4, 15).
180

12.10 SUGGESTED READINGS/ REFERENCE BOOKS


1. Harnett, D. L., & Horrell, J. F. (1998). Data, statistics, and decision models
with Excel. New York: John Wiley.
2. Hoel, P. G., & Jessen, R. J. (1971). Basic statistics for business and
economics. New York: John Wiley.
3. McClave, J. T., & Benson, P. G. (1988). Statistics for business and
economics (4th ed.). San Francisco: Dellen Publishing Co.
4. Kothari, C.R., 1985, Research Methodology-Methods and Techniques, New
Delhi, Wi ley Eastern Limited.
5. Kumar, Ranjit, 2005, Research Methodology-A Step-by-Step Guide for
Beginners, (2nd.ed.), Singapore, Pearson Education.
12.11 LEARNING ACTIVITIES
1. You have been hired as a marketing research analyst by burger king. Your
boss, the marketing manager, is wondering what statistical analysis should be
conducted to explain the patronage of burger king in terms of the consumers
evaluations of burger king on the factors of consumers choice criteria. Explain to
your boss (a student in your class) the analysis that you would conduct
12.12 KEYWORDS
Regression Effect: Sometimes people who place themselves on the extreme
positions of a measurement scale at the pre-test stage may, for a number of
reasons, shift towards the mean at the post-test stage. They might feel that they
have been too negative or too positive at the pre -test stage. Therefore, the mere
expression of the attitude in respon se to a questionnaire or interview has caused
them to think about and alter their attitude towards the mean at the time of the
post-test. This type of effect is known as the regression effect.
Region of Rejection: That portion of a sampling distribution containing values
considered too unlikely to occur by chance, found in the tail or tails of the
distribution
Regression Line: The line drawn through the long dimension of a scatter plot
that best fits the center of the scatter plot, thereby visually summari zing the scatter
plot and indicating the type of relationship that is present
Nonlinear Relationship: A relationship in which the Y scores change their
direction of change as the X scores change; also called a curvilinear relationship.



181

UNIT – IV
LESSON – 13

NON PARAMETRIC TEST


13.1 INTRODUCTION
In contrast to parametric tests, non-parametric tests do not require any
assumptions about the parameters or about the nature of population. Itis because
of this that these methods are sometimes referred to as the distribution free
methods. Most of these methods, however, are based upon the weaker assumptions
that observations are independent and that the variable under study is continuous
with approximately symmetrical distribution. In addition to this, the se methods do
not require measurements as strong as that required by parametric methods. Most
of the non-parametric tests are applicable to data measured in an ordinal or
nominal scale. As opposed to this, the parametric tests are based on data measured
at least in an interval scale. The measurements obtained on interval and ratio scale
are also known as high level measurements.
In this unit, meaning of NON –PARAMETRIC METNODS- SIGN TEST FOR
PAIRED TEST- Rank sum test – Mann – Whitney U test and Kruskal wallis test -
one sample run test – Chi-square test – intdependence of attributes and goodness
of fit are dealt
13.2 OBJECTIVE
The main aim of this Lesson is to study the Non parametric test. After going
through this Lesson you should be able to understan d Level of measurement,
Parametric test, Non parametric test, Advantage and disadvantage of non
parametric test
13.3 CONTENT
13.3.1 Level of measurement
13.3.2 Parametric test
13.3.3 Non parametric test
13.3.4 Advantage and disadvantage of non parametric test
13.3.1 Level of Measurement
1. Nominal scale: This scale uses numbers or other symbols to identify the
groups or classes to which various objects belong. These numbers or symbols
constitute a nominal or classifying scale. For example, classification of individuals
on the basis of sex (male, female) or on the basis of level of education (metric, senior
secondary, graduate, post graduate), etc. This scale is the weakest of all the
measurements.
2. Ordinal scale: This scale uses numbers to represent some kind of ordering or
ranking of objects. However, the differences of numbers, used for ranking, don’t
have any meaning. For example, the top 4 students of class can be ranked as 1, 2,
3, 4, according to their marks in an examination.
182

3. Interval scale: This scale also uses numbers such that these can be ordered
and their differences have a meaningful interpretation.
4. Ratio scale: A scale possessing all the properties of an interval scale along
with a true zero point is called a ratio scale. It may be pointed out that a zero point
in an interval scale is arbitrary. For example, freezing point of water is defined at
0° Celsius or 32° Fahrenheit, implying thereby that the zero on either scale is
arbitrary and doesn’t represent total absence of heat. In contrast to this, the
measurement of distance, say in metres, is done on a ratio scale. The term ratio is
used here because ratio comparisons are meaningful. For example, 100 kms of
distance is four times larger than a distance of 25 kms while 100°F may not mean
that it is twice as hot as 50°F. It should be noted here that a test that can be
performed on high level measurements can always be performed on ordinal or
nominal measurements but not vice-versa. However, if along with the high level
measurements the conditions of a parametric test are also met, the parametric test
should invariably be used because this test is most powerful in the given
circumstances. From the above, we conclude that a non -parametric test should be
used when either the conditions about the parent population are not met or the
level of measurements is inadequate for a parametric test.
A parametric statistical test is one that makes assumptions about the
parameters (defining properties) of the population distribution(s) from which one's
data are drawn.
A non-parametric test is one that makes no such assumptions. In this strict
sense, "non-parametric" is essentially a null category, since virtually all statistical
tests assume one thing or another about the properties of the source popu lation(s).
Which is more powerful?
Non-parametric statistical procedures are less powerful because they use less
information in their calculation. For example, a parametric correlation uses
information about the mean and deviation from the mean while a n on-parametric
correlation will use only the ordinal position of pairs of scores.
Parametric Assumptions
 The observations must be independent
 The observations must be drawn from normally distributed populations
 These populations must have the same variances
 The means of these normal and homoscedastic populations must be linear
combinations of effects due to columns and/or rows
Nonparametric Assumptions
Certain assumptions are associated with most nonparametric statistical tests,
but these are fewer and weaker than those of parametric tests.
183

Advantages of Nonparametric Tests


 Probability statements obtained from most nonparametric statistics are
exact probabilities, regardless of the shape of the population distribution
from which the random sample was drawn
 If sample sizes as small as N=6 are used, there is no alternative to using a
nonparametric test
 Easier to learn and apply than parametric tests
 Based on a model that specifies very general conditions.
 No specific form of the distribution from which the sample was drawn.
 Hence nonparametric tests are also known as distribution free tests.
Disadvantages of nonparametric tests
 Losing precision/wasteful of data
 Low power
 False sense of security
 Lack of software
 Testing distributions only
 Higher-ordered interactions not dealt with
 Parametric models are more efficient if data permit.
 It is difficult to compute by hand for large samples
 Tables are not widely available
 In cases where a parametric test would be appropriate, non -parametric tests
have less power. In other words, a larger sample size can be required to
draw conclusions with the same degree of confidence.
Few points
 The inferences drawn from tests based on the parametric tests such as t, F
and Chi-square may be seriously affected when the parent population’s
distribution is not normal.
 The adverse effect could be more when sample size is small.
 Thus when there is doubt about the distribution of the parent population, a
nonparametric method should be Non Parametric Tests: Hands on SPSS N. Uttam
Singh, Aniruddha Roy & A. K. Tripathi – 2013 3 used.
 In many situations, particularly in social and behavioral sciences,
observations are difficult or impossible to take on numerical scales and a
suitable nonparametric test is an alternative under such situations.
Measurement
The 41 levels of meausrement
184

1. Nominal or Classificatory Scale



In research activities a YES/NO scale is nominal. It has no order and there
is no distance between YES and NO.
2. Ordinal or Ranking Scale




There is no objective distance between any two points on your subjective
scale.
3. Interval Scale
It is an
interval scale because it is assumed to have equidistant points between each of the
scale elements.
4. Ratio Scale


Ratio data is interval data with a natural zero point
13.4 REVISION POINTS
1. Level of measurement
2. Parametric test
3. Non parametric test
4 Advantage and disadvantage of non parametric test
13.5 INTEXT QUESTIONS
1. List out the advantages and disadvantages of non parametric test
2. Explain the limitations of nonparametric tests
13.6 SUMMARY
After going through this Lesson you should be able to explain Level of
measurement, Parametric test, Non parametric test, Advantage and disadvantage of
non parametric test
13.7 TERMINAL EXERSICE
1. Compare and contrast parametric and nonparame tric tests
2. Identify multiple applications where nonparametric approaches are
appropriate
185

13.8 SUPPLEMENTARY MATERIALS


1. Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, rev. ed.,
New York: Harcourt, Brace & World, Inc., 1970.
13.9 ASSIGNMENT
1. Identify the appropriate nonparametric hypothesis testing procedure based
on type of outcome variable and number of samples
13.10 SUGGESTED READINGS/REFERENCE BOOKS
1. Ackoff, Russell L., The Design of Social Research, Chicago: University of
Chicago Press, 1961.
2. Ackoff, Russell L., Scientific Method, New York: John Wiley & Sons, 1962.
3. Allen, T. Harrell, New Methods in Social Science Research, New York:
Praeger Publishers, 1978.
4. Anderson, H.H., and Anderson, G.L., An Introduction to Projective
Techniques and Other Devices for
5. Understanding the Dynamics of Human Behaviour, New York: Prentice Hall,
1951.
6. Anderson, T.W., An Introduction to Multivariate Analysis, New York: John
Wiley & Sons, 1958.
7. Bailey, Kenneth D., “Methods of Social Research,” New York, 1978.
8. Baker, R.P., and Howell, A.C., The Preparation of Reports, New York: Ronald
Press, 1938.
9. Bartee, T.C., “Digital Computer Fundamentals,” 5th Ed., McGraw-Hill,
International Book Co., 1981.
10. Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, re v. ed.,
New York: Harcourt, Brace & World, Inc., 1970.
11. Bell, J.E., Projective Techniques: A. Dynamic Approach to the Study of
Personality, New York: Longmans Green, 1948.
12. Bellenger, Danny N., and Greenberg, Barnett A., Marketing Re search—A
Management Information Approach, Homewood, Illinois: Richard D. Irwin,
Inc., 1978.
13.11 LEARNING ACTIVITY
1. Which particular non-parametric test should be carried out in case of
multiple group comparisons? Which software should be used for this?
13.12 KEYWORD
1. Level of measurement, Parametric test, Non parametric test, Advantage and
disadvantage of non parametric test

186

LESSON – 14
SIGN TEST
14.1 INTRODUCTION
One of the easiest non-parametric tests is the sign test. The test is known as
the sign test as it is based on the direction of the plus or minus signs of
observations in a sample instead of their numerical values. There are two types of
sign tests: (a) One-sample sign test,and (b) Two-sample sign test.
14.2 OBJECTIVE
The main aim of this Lesson is to study the Non parametric test. After going
through this Lesson you should be able to understand Sign test, one sample sign
test and two sample sign test
14.3 CONTENT
14.3.1 Sign test
14.3.2 One sample sign test
14.3.3 Two sample sign test
14.3.1 one -sample sign test
The one-sample sign test is a very simple non -parametric test applicable on
the assumption that we are dealing with a population having a continuous
symmetrical distribution. As such, the probability of getting a value less than the
mean is 0.5. Likewise, the probability of getting a value greater than the mean
isalso 0.5. To test the null hypothesis µ = µ0against an appropriate alternative,
each sample value greater than µ0is replaced by plus (+) sign and each sample
value less than µ0with a minus (-) sign. Having done this, we can test the null
hypothesis that the probabilities of getting both plus and mi nus signs are 0.5. It
may be noted that if a sample value happens to be equal to µ0, it is simply
discarded.
To perform the actual test, we use either of the two methods. When the sample
is small, the test is performed by computing the binomial probabilities or by
referring to the binomial probabilities table. When the sample is large, the normal
distribution is used as an approximation of the binomial distribution. Let us take
an example to show how the onesample sign test is applied.
Example 1
We are required to test the hypothesis that the mean value µ of a continuous
distribution is 20 against the alternative hypothesis µ ≠20. Fifteen observations
were taken and the following results were obtained:
18, 19, 25, 21, 16, 15, 19, 22, 24, 21, 18, 17, 15, 26 and 24.
We may use α= 0.05 level of significance.
Solution: Replacing each value greater than 20 with a plus (+) sign and each
value less than
187

20 with a minus (-) sign, we get


--++---+++---++
Now, the question before us is whether 7 plus signs observed in 15 trials
support the null hypothesis p = 0.5 or the alternative hypothesis p ≠0.5. Using the
binomial probability tables or binomial probabilities, we find that the probability of
7 or more successes is 0.196 + 0.196 + 0.153 + 0.092 + 0.042 + 0.014 + 0.003 =
0.696* and p = 0.5 and since this value is greater than α/2 = 0.025, we find that
the null hypothesis will have to be accepted. We can also use normal approximation
to the binomial distribution when np ≥5. As here p = ½, the conditi on for the
normal approximation to the binomial distribution is satisfied asn > 10. As such,
we can use the Z statistic for which the following formula is to be used.

Since calculated Z = -0.26 lies between Z = - 1.96 and Z = 1.96 (the critical
value of Z at 0.05 level of significance), the null hypothesis is accepted.
The two-sample sign test
Example 1: Suppose we have the following table indicating the ratings
assigned to two brands of cold drink X and Y by 12 consumers. Each respondent was
asked to taste the two brands of cold drink and then rate them.
Table 15.1. Ratings of brands X and Y cold drinks
Brand X 26 30 44 23 18 50 34 16 25 49 37 20
Brand Y 22 27 39 71 56 30 14 18 51 33 16
Sign + + + + + - + + + - + +
We have to apply the two-sample sign test. H0being both brands enjoy equal
preference.
Solution: Row three of Table 15.1 shows + and – signs. When X’s rating is
higher than that of Y, then the third row shows the ‘+’ sign. As against this, when
X’s rating is lower than that of Y, then it shows the ‘-‘ sign. The table shows 10 plus
signs and 2 minus signs. Now, we have to examine whether ’10 successes in 12
trials’ supports the null hypothesis p = ½ or the alternative hypothesis p > ½. The
null hypothesis implies that both the brands enjoy equal preferences and none is
better than the other. The alternative hypothesis is that the brand X is better than
brand Y. Referring to the binomial probabilities table, we find that for n = 12 and p
= ½ the probability of ’10 or more successes’ is 0.016 + 0.003 = 0.019. It follows
188

that the null hypothesis can be rejected at α= 0.05 level of significance. We can,
therefore, conclude that brand X is a preferred brand as compared to brand Y.
Example 2: To illustrate the second case, which relates to two independent samples, let
us consider the following data pertaining to the down times (periods in which computers
were inoperative on account of failures, in minutes of two different computers. We have
to apply the two-sample sign test.

Solution: These data are shown in Table 15.2 along with+ or – sign as may be
applicable in case of each pair of values. A plus sign is assigned when the downtime
for computer A is greater than that for computer B and a minus sign is given when
the downtime for computer
B is greater than that for computer A.

It will be seen that there are 15 plus signs and 7 minus signs. Thus, we have
to ascertain whether ‘15 successes in 22 trials’ support the null hypothesis p = ½.
The null hypothesis implies that the true average downtime is the same for both the
computers A and B. The alternative hypothesis is p ≠½. The null hypothesis implies
that the true average downtime is the same for both the computers A and B.The
alternative hypothesis is p = ½.
Let us use in this case the normal approximation of the binomial distribution.
This can be done since np and n (1 – p) are both equal to 11 in this example.
Substituting n = 22 and p = ½ into the formulas for the mean and the standard
deviation of the binomial distribution, we get µ = np = 22 (½) = 11 and σ= np (1 - p)
= 22.½.½ = 2.345
Hence, Z = (X – µ)/σ= (15 – 11)/2.345 = 1.71
Since this value of 1.71 falls between – Z0.025 = - 1.96 and Z0.025= 1.96, we
find that the null hypothesis cannot be rejected. This means thatthe downtime in
the two computers is the same.
189

This seems to be surprising as we find that there are substantial differences.


The two sample means, for example, are 55.5 for A and 48.6 for B. This example
illustrates the point that at times the sign test can be quite a waste of information.
It may also be noted that had the continuity correction been used, we would have
obtained:
Z = 3.5/2.345 = 1.49
This would not have changed our earlier conclusion.
Friedman Rank Test
The Friedman test ranks the scores in each row of the data file independently
of every other row. The Friedman Chi-square tests the null hypothesis that the
ranks of the variables do not differ from their expected value. For a constant sample
size, the higher the value of this chi -square statistic, the larger the difference
between each variables rank sum and its expected value.
Friedman Test
Friedman test is used in this study to know about the attitude and perception
level of Khadi consumers towards the factors which motivates them to buy khadi
textile products through ranking methods. The various motivating facto rs
considered for this test are advertisements. Social status of Khadi, quality of Khadi
textile products, high market demand for Khadi products, price, customer service,
durability and comfortness of Khadi textile products.

Table 4.11 Mean Ranks of the Factors Which Motivates Khadi


Consumers To Buy Khadi Textile Products
Mean Rank
6.4
Advertisement
3.71
social status of khadi
2.38
quality of the products
6.5
high market demand
Price 5.17
4.2
better service from employees
3.9
Durability
3.74
comfort to wear
190

Table 4.11.1 Friedman test for Mean Rank Variation of the Factors Which Motivates
Khadi Consumers to Buy Khadi Textile Products
420
N
Chi-Square 995.064

Df 7

Asymp. Sig. 0.000


The above table indicates that quality of Khadi textile products (mean rank
2.38) is the leading factor among various other factors such as advertisement
(mean rank 6.40), social status (mean rank 3.71), high market demand (mean rank
6.50), price (mean rank 5.17), better employee service (mean rank 4.20), durability
(mean rank 3.90), comfort to wear (mean rank 3.74) which motivates them to
purchase Khadi textile products. This mean variation is statistically significant at 5
per cent level of significance. (P<.05 in the above cases)
MANN-WHITNEY TEST
Use this test when two different groups of participants perform both conditions
of the study: i.e., it is ppropriate for analysing the data from an independent-
measures design with two conditions. Use it when the data do not meet the
requirements for a parametric test (i.e. if the data are not normally distributed; if
the variances for the two conditions are markedly different; or if the data are
measurements on an ordinal scale). Otherwise, if the data meet the requirements
for a parametric test, it is better to use an independent-measures t-test (also known
as a "two-sample" t-test). The logic behind the Mann-Whitney test is to rank the
data for each condition, and then see how different the two rank totals are. If there
is a systematic difference between the two conditions, then most of the high ranks
will belong to one condition and most of the low ranks will belong to the other one.
As a result, the rank totals will be quite different. On the other hand, if the two
conditions are similar, then high and low ranks will be distributed fairly evenly
between the two conditions and the rank totals will be fairly similar. The Mann -
Whitney test statistic "U" reflects the difference between the two rank totals. The
SMALLER it is (taking into account how many participants you have in each group)
then the less likely it is to have occurred by chance. In this study this test is used
to know the rank preference of Khadi consumers based on gender wise
(male/female) towards Khadi textile products.
In the independent samples case, it is preferable to apply the independent
sample t test when there exists normality of the sample mean for each sample. If
there is a violation of this assumption, then a non parametric test for independent
samples is applied – the Wilcoxon Mann Whitney test. This test uses the relative
position of the data in a rank ordering, unlike the t test which uses the actual
values. To use the Wilcoxon Mann Whitney test (a) It must be reasonable to regard
the data as a random sample from their respective populations (b) Observations
within each sample must be independent of one another (c) The two samples must
be independent of one another
191

The following data indicate the lifetime (in hours) of samples of two kinds of
light bulbs in continuous use:

Brand A 603 625 641 622 585 593 660 600 633 580 615 648
Brand B 620 640 646 620 652 639 590 646 631 669 610 619
We are required to use the Mann-Whitney test to compare the lifetimes of
brands A and B Light bulbs.
Solution: The first step for performing the Mann-Whitney test is to rank the
given data jointly (as if they were one sample) in an increasing or decreasing order
of magnitude. For our data, we thus obtain the following array where we use the
letters A and B to denote whether the light bulb was from brand A or brand B.
Table 1 Ranking of light bulbs of brands A and B

As both the samples come from identical populations, it is reasonable to


assume that the means of the ranks assigned to the values of the two samples are more
or less the same. As such, our null hypothesis is:
H0: Means of ranks assigned to the values in the two groups are the same.
H1: Means are not the same.
However, instead of using the means of the ranks, we shall use rank sums for
which the following formula will be used.
192

U = n1n2+ [n1(n1+ 1)]/2 – R1Where n1and n2are the sample sizes of Group 1
and Group 2, respectively, and R1is the sum of the ranks assigned to the values of
the first sample. In our example, we have n1= 12, n2= 12 and R1= 1 + 2 + 4 + 5 + 6
+ 8 + 12 + 13 + 15 + 18 + 21 + 23 = 128. Substituting these values in the above
formula,
U = (12) (12) + [12 (12 + 1)]/2 – 128
= 144 + 78 – 128
= 94
From Appendix Table 9 for n1and n2, each equal to 12, and for 0.05 level of
significance is 37. Since the critical value is smaller than the calculated value of 94,
we accept the null hypothesis and conclude that there is no difference in the
average lifetimes of the two brands of light bulbs. The test statistic we have just
applied is suitable when n1and n2are less than or equal to 25. For larger values of
n1and/or n2, we can make use of the fact that the distribution of Ws approaches a
normal distribution as sample sizes increase. We can then use the Z test to test the
hypothesis. The normal approximation
Although our observations are limited, we may apply the normal
approximation to this problem. For this, we have to use the Z statistic.

3. (Statistic – Mean)/Standard deviation


= (94 – 72)/17.3 = 1.27
The critical value of Z at 0.05 level of significance is 1.64. Since the calculated
value of Z = 1.27 is smaller than 1.64, the null hypothesis is accepted. This shows
that there is no difference in average lifetimes of brands A and B bulbs. The Z test
is more dependable as compared to the earlier one. It may be noted that Mann-
Whitney test required fewer assumptions than the corresponding standard test. In
fact, the only assumption required is that the populations from which samples have
been drawn are continuous. In actual practice, even when this assumption turns
out to be wrong, this is not regarded serious.
e.g: 2
In the independent samples case, it is preferable to apply the independent
sample t test when there exists normality of the sample mean for each sample. If
there is a violation of this assumption, then a non parametric test for independent
samples is appied – the Wilcoxon Mann Whitney test. This test uses the relative
position of the data in a rank ordering, unlike the t test which uses the actual
193

values. To use the Wilcoxon Mann Whitney test (a) It must be reasonable to regard
the data as a random sample from their respective populations (b) Observations
within each sample must be independent of one another (c) The two samples must
be independent of one another

Gender wise Mean Ranks Of Khadi Textile Products


Mean Sum of
Gender N Rank Ranks

Male 198 207.21 41028


textile products Female 221 212.5 46962
419
Total
Asymp. Sig. (2- 0.152
tailed)
Mann-Whitney Test For Genderwise Mean Rank Variation Of The
Responses Of Khadi Consumers With Respect To Khadi Textile
Products
textile products
Mann-Whitney U 21327
Wilcoxon W 41028

Z -1.433
Asymp. Sig. (2-tailed) 0.152
The table above shows that majority of Khadi textile products are purchased
by male consumers (mean rank 207.21) than female consumers even if this mean
variation is statistically not significant at 5 per cent level od significance(P>.05 in
the above case).
Kruskal-Wallis Test
The Kruskal–Wallis one-way analysis of variance by ranks is a non-
parametric method for testing whether samples originate from the same
distribution. It is used for comparing two or more samples that are independent,
and that may have different sample sizes, and extends the Mann–Whitney Utest to
more than two groups. The parametric equivalent of the Kruskal -Wallis test is
the one-way analysis of variance (ANOVA). When rejecting the null hypothesis of
the Kruskal-Wallis test, then at least one of sample stochastically dominates at
least one other sample. The test does not identify where this stochastic dominance
occurs or for how many pairs of groups stochastic dominance obtains.It would help
to analyze the specific sample pairs for stochastic dominance. Since it is a non -
parametric method, the Kruskal–Wallis test does not assume a normal
distribution of the residuals, unlike the analogous one -way analysis of variance. If
the researcher can make the more stringent assumptions of an identically shaped
and scaled distribution for all groups, except for any difference in medians, then
the null hypothesis is that the medians of all groups are equal, and the alternative
hypothesis is that at least one population median of one group is different than the
194

population median of at least one other group. This test is used in this study to
analyze the preference of khadi consumers towards Khadi textile products based on
their place of residence (urban, semi -urban and rural) by assinging ranks to their
preferences.
This test is appropriate for use under the following circumstances: (a) when
there is three or more conditions that researcher want to compare; (b) each
condition is performed by a different group of parti cipants; i.e. researcher have an
independent-measures design with three or more conditions. (c) the data do not
meet the requirements for a parametric test. (i.e. use it if the data are not normally
distributed; if the variances for the different condition s are markedly different; or if
the data are measurements on an ordinal scale). If the data meet the requirements
for a parametric test, it is better to use one -way independent-measures Analysis of
Variance (ANOVA) because it is more powerful than the Kruskal-Wallis test.
F-test, used in analysis of variance, was that each of the k populations should
be normal with equal variance. In contrast to this, the Kruskal -Wallis test only
assumes that the k populations are continuous and have the same pattern
(symmetrical or skewed) of distribution. The null and the alternative hypotheses
ofthe Kruskal-Wallis test are:
H0: m1= m2= … = mk(i.e., means of the k populations are equal)
Ha: Not all mi’s are equal.
The Test Statistic: The computation of the test statisti cfollows a procedure that
is very similar to the Mann-Whitney Wilcoxon test.
(i) Rank all the n1+ n2+ … + nk= n observations, arrayed in ascending order.
(ii) Find R1, R2, … Rk, where Riis the sum of ranks of the ith sample. The test
statistic, denoted by H, is given by

It can be shown that the distribution of H is c2with k – 1 d.f., when size of


each sample is at least 5. Thus, if H > 21 k− χ , H0 is rejected.
Example 1: To compare the effectiveness of three types of weight-reducing
diets, a homogeneous groups of 22 women was divided into three sub-groups and
each sub-group followed one of these diet plans for a period of two months. The
weight reductions, in kgs, were noted as given below:

Use the Kruskal-Wallis test to test the hypothesis that the effectiveness of the
three weight reducing diet plans is same at 5% level of significance.
195

Solution:
It is given that n1= 6, n2= 7 and n3= 9.
The total number of observations is 6 + 7 + 9 = 22. These are ranked in their
ascending order as given below:

The tabulated value of χ2at 2 d.f. and 5% level of significance is 5.99. Since H
is greater than this value, H0is rejected at 5% level of significance. e.g.). If the data
meet the requirements for a parametric test, it is better to use one -way
independent-measures Analysis of Variance (ANOVA) because it is more powerful
than the Kruskal-Wallis test.

Area wise Mean Ranks Of Khadi Textile Products


place of residence Mean
N
Rank

Urban 192 209


Textile semi-urban 139 208.63
products
Rural 89 211.95
420
Total

Kruskal-Wallis Test For Mean Rank Variation Of The


Responses Of Khadi Consumers With Respect To Khadi Textile
Products
textile products
Chi-Square 0.481
2
Df
0.786
Asymp. Sig.
The above table indicates that the Khadi textile produc ts are frequently
purchased by the consumers from rural area (mean rank 211.95) compared to
consumers from urban (mean rank 209) and semi -urban areas (mean rank 208.63).
This mean variation is statistically insignificant at 5 per cent level of significance
(P>.05 in above cases)
196

14.4 REVISION POINTS


1. Sign test
2. One sample sign test
3. Two sample sign test
14.5 INTEXT QUESTIONS
1. List out the advantages and disadvantages of non parametric test
2. Explain the Importance of sign tests
14.6 SUMMARY
After going through this Lesson you should be able to Explain Sign test, one
sample sign test and two sample sign test
14.7 TERMINAL EXERCISES
It turns out that 5 of the participants like the new drink better, and the rest
prefer the old one. At .05 significance level can we reject the notion that the two
drinks are equally popular?
14.8 SUPPLEMENTARY MATERIALS
Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, rev. ed., New
York: Harcourt, Brace & World, Inc., 1970.
14.9 ASSIGNMENTS
The sign test is considered a weaker test, because it tests the pair value below
or above the median and it does not measure the pair difference - Comment
14.10 SUGGESTED READINGS/REFERENCE BOOKS
1. Ackoff, Russell L., the Design of Social Research, Chicago: University of
Chicago Press, 1961.
2. Ackoff, Russell L., Scientific Method, New York: John Wiley & Sons, 1962.
3. Allen, T. Harrell, New Methods in Social Science Research, New York:
Praeger Publishers, 1978.
4. Anderson, H.H., and Anderson, G.L., An Introduction to Projective
Techniques and Other Devices for Understanding the Dynamics of Human
Behaviour, New York: Prentice Hall, 1951.
5. Anderson, T.W., An Introduction to Multivariate Analysis, New York: John
Wiley & Sons, 1958.
6. Bailey, Kenneth D., “Methods of Social Research,” New York, 1978.
7. Baker, R.P., and Howell, A.C., The Preparation of Reports, New York: Ronald
Press, 1938.
8. Bartee, T.C., “Digital Computer Fundamentals,” 5th Ed., McGraw-Hill,
International Book Co., 1981.
9. Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, rev. ed.,
New York: Harcourt, Brace & World, Inc., 1970.
197

10. Bell, J.E., Projective Techniques: A. Dynamic Approach to the Study of


Personality, New York: Longmans Green, 1948.
11. Bellenger, Danny N., and Greenberg, Barn ett A., Marketing Research—A
Management Information Approach, Homewood, Illinois: Richard D. Irwin, Inc.,
1978.
14.11 LEARNING ACTIVITIES
The disadvantage is that nonparametric procedures throw away information!
The sign test, for example, uses only the si gns of the observations - Comment
14.12 KEY WORD
Sign test, one sample sign test and two sample sign test
198

LESSON – 15
CHI SQUARE TEST
15.1 INTRODUCTION
Any sample comprising numerical observations can be treated in the same
manner by using the letters a and b to denote, respectively, values above the
median and values below the median of the sample. In case an observation is equal
to the median, it is omitted. The resulting series of as and bs (representing the data
in their original order) can be tested for randomness on the basis of the total
number of runs above and below the median,respectively.
15.2 OBJECTIVE
After going through this Lesson you should be able to Understand Chi square
test, steps involved in chi square test and Goodness of fit.
15.3 CONTENT
15.3.1 Chi square test
15.3.2 Steps involved in chi square test
15.3.3 Goodness of fit
Let us take an example.
Example:
Suppose we have the following series of29 college students. After performing a
set of study exercises, increases in their pulse rate were recorded as follows:
22, 23, 21, 25, 33, 32, 25, 30, 17, 20, 26, 12, 21, 20, 27, 24, 28, 14, 29, 23,
22, 36, 25, 21, 23, 19, 17, 26 and 26. We have to test the randomness of these
data.
Solution:First, we have to calculate the median of this series. If we arrange
these values in an ascending order, we find that the size of (n +1)/2th item, that is,
15thitem is 24. Thus, the median is 24. As there is one value, which is 24 we omit it and
get the following arrangement of as and bs where a stands for an item greater than (or
above) the median and b stands for an item lower than (or below) the median:
bbb aaaaa bb a bbb aa b a bb aa bbbb aa
On the basis of this arrangement, we find that n1, (i.e. a) = 13, n2, (i.e. bs) =
15, and u = 12, we get
199

Since Z = -1.14 falls between –Z0.025= -1.96 and Z0.025= 1.96, the null
hypothesis cannot be rejected at the level of significance α= 0.05. We can, therefore,
conclude that the randomness of the original sample cannot be questioned. It may
be noted that this test is particularly useful in detecting trends and cyclic patterns
in a series. If there is a trend, there would be first mostly as and later mostly bs or
vice versa. In case of a repeated cyclic pattern, there would be a systematic
alternation of as and bs and probably, too many runs.
CHI-Square analysis
The Chi-square test is an important test amongst the several tests of
significance developed by statisticians. Chi -square, symbolically written as X 2, is a
statistical measure used in the context of sampling analysis for comparing a
variance to a theoretical variance. It is a non -parametric test, “it could be used to
determine if categorical data shows dependency or the two classifications are
independent. It can also be used to make comparisons between theoretical
populations and actual data when categories are used. Thus, the Chi -square test is
applicable in large number of problems. The test is a technique through the use of
which it is possible for all researchers to do the following 2.I) Test the goodness of fit,
ii) Test the significance of association between two attributes and i ii) Test the
homogeneity or the significance of population variance.
Steps involved in Applying Chi-Square Analysis
1. Calculate the expected frequencies on the basis of given hypothesis or on
the basis of null hypothesis.
2. Obtain the difference between observed and expected frequencies and find
out the squares of such differences i.e. calculate (Oij-Eij)2.
3. Divide the quantity (Oij-Eij)2 obtained as stated above by the corresponding
expected frequency to get (Oij-Eij)2 /Eij and this should be done for all the cell
frequencies or the group frequencies.
200

4. Find the summation of (Oij-Eij)2 / Eij values or what we call a (Oij-Eij)2 / Eij.
This is the required C2 value. The X2 value obtained as such should be compared
with relevant table value of X2 and then inference be drawn.
Goodness of Fit Test
This section begins with a short sketch of the method used to conduct achi
square goodness of fit test. An example of such a test is then provide d,followed by a
more detailed discussion of the methods used. Finall y, thereis a variety of examples
used to illustrate some of the common types ofgoodness of fit tests.The chi square
goodness of fit test begins by hypothesizing that the distribution of a variable
behaves in a particular manner. For example, inorder to determine daily staffing
needs of a retail store, the manager maywish to know whether there are an equal
number of customers each day ofthe week. To begin, an hypothesis of equal
numbers of customers on eachday could be assumed, and this would be the null
hypothesis. A student may observe the set of grades for a class, and suspect that
the professor allocatedthe grades on the basis of a normal distribution. Another
possibility is thata researcher claims that the sampleselected has a distribution
which is veryclose to distribution of the population. While no general statement can
beprovided to cover all these possibilities, what is common to all is that aclaim has
been made concerning the nature of the whole distribution.Suppose that a variable
has a frequency distribution with k categoriesinto which the data has been
grouped. The frequencies of occurrence of thevariable, for each category of the
variable, are called the observed values.
The manner in which the chi square goodness of fit test works is to
determinehow many cases there would be in each category if the sample data
weredistributed exactly according to the claim. These are termed the
expectednumber of cases for each category. The total of the expected number
ofcases is always made equal to the total of the observed number of cases.
The null hypothesis is that the observed number of cases in each categoryis
exactly equal to the expected number of cases in each category. Thealternative
hypothesis is that the observed and expected number of casesdiffer sufficiently to
reject the null hypothesis.
Let Oiis the observed number of cases in category i and Eiis the expected
number of cases in each category, for each of the k categoriesi = 1, 2, 3, · · · , k ,
into which the data has been grouped. The hypotheses are
H0: Oi = Ei
H1: Oi # Ei and the test statistic is

where the summation proceeds across all k categories, and there are k −
1degrees of freedom.Large values of this statistic lead the researcher to rejectthe
201

null hypothesis, smaller values mean that the null hypothesis cannotbe rejected.
The chi square distribution with k − 1 degrees of freedom inAppendix ?? is used to
determine the critical χ2value for the test.This sketch of the test should allow you
to follow the first example of achi square goodness of fit test. After Example 10.3.1,
the complete notationfor a chi square goodness of fit test is provided. This notation
will becomemore meaningful after you have seen how a goodness of fit test is
conducted.
Median test for two independent samples
In order to perform this test, let us use our previous example, which pertains
to the Down times of the two computers. The median of the combined data is 52,
which can easily be checked. There are 5 values below 52 and 15 values above it, in
case of computer A. As regards computer B, the corresponding figures are 16 and
6. All this information is Summarized in Table 15.3, which also indicates the totals
of the rows and columns.
Example 1
Table 1. Classification of downtime for computers A and B

Our null hypothesis H0is that there is no difference inthe median downtime for
the two computers. The alternative hypothesis H1is that there is difference in the
downtime of the two computers.
We now calculate the expected frequencies by the formula (Rowi ×
Columni)/Grand total. Thus, Table 4 shows both the observed and the expected
frequencies. Of course, we could have obtained these results by arguing that half
the values in each sample can be expected to fall above the median and the other
half below it.
Table 2. Calculation of chi-square
202

The critical value of χ2at 0.05 level of significance for (2 – 1) (2 – 1) = 1 degree


of freedom is 3.841(χ2-Table). Since the calculated value of χ2exceeds the critical
value, the null hypothesis has to be rejected. In other words,there is no evidence to
suggest that the downtime is the same in case of the two computers. It may be
recalled that in the previous example having the same data, the null hypothesis
could not be rejected. In contrast, we find here that the two-sample median test has
led to the rejection of the null hypothesis. This may be construed as evidence that
the median test is not quite as wasteful of the information as the sign test.
However, in general, it is very difficult to make a meaningful comparison of the
merits of two or more non-parametric tests, which can be used for the same
purpose.
2.e.g
GENDER AND LEVEL OF PREFERENCE TOWARDS PURCHASING OF CAUSE-
RELATED MARKETING PRODUCTS (TWO-WAY TABLE)
LEVEL OF PREFERENCE TOWARDS PURCHASING OF
CAUSE RELATED MARKETING PRODUCTS
S.NO. GENDER
Very high High Neutral Low Very low TOTAL
preference preference preference preference
15 147 77 0 8 247
1 Male
6.1% 59.5% 31.2% .0% 3.2% 100.0%
77 240 154 16 16 503
2 Female
15.3% 47.7% 30.6% 3.2% 3.2% 100.0%
92 387 231 16 24 750
TOTAL
12.3% 51.6% 30.8% 2.1% 3.2% 100.0%

Source: Primary Data


The majority of male respondents feel that the high preference towards
purchasing pattern of cause-related marketing products more 59.5 per cent high
preference and 6.1 per cent agree very high preferen ce (Table 4.2.2). Only about 8
respondents (3.2 per cent) give very low preference. A minimum percentage of
respondents (31.2 per cent) remain neutral in their opinion about the purchasing
pattern of cause-related marketing products. Almost the same trend can be noticed
among the female respondents. About 47.7 per cent of female respondents high
preference towards purchasing pattern of cause -related products (15.3)per cent
very high preference, and 30.6 per cent neutral about the purchasing pattern of
cause related marketing. Meager percentage about 6.4 per cent of female
respondents low preference and a very low preference about the purchasing pattern
of cause related marketing. In order to find out the relationship between gender and
level of preference towards purchasing pattern of cause-related marketing products,
a chi-square test has been employed to test the hypothesis given below.
203

TABLE - 4.2.2(A)
CHI-SQUARE TEST

DF S/NS Remarks
Pearson Chi-Square Calculated Chi-square Value Table value

GENDER 23.864 4 .000** S Rejected

** - Significant at 1% level S-Significant


It has been divulged from the chi -square test that the p-value (.000) has been
less than 0.01 and the result has significance at 1 per cent level. Hence, the null
hypothesis (H0) has been rejected and the alternative hypothesis (H 1 ) has been
accepted. From the analysis it has been concluded that there is a relationship
between gender and level of preference towards purchasing pattern of cause -related
marketing products.
The K-sample median test
The median test can easily be generalised so that it can be applied to K-
samples. In accordance with the earlier procedure, first find the median of the
combined data. We then determine how many of the values in each sample fall
above or below the median. Finally, we analyse the resulting contingency table by
the method of chi-square. Let us take an example.
Example 2: Suppose that we are given the following data relating to marks
obtained by students in Statistics in the three different sections of a MBA class in G.J.U.
Hisar. The maximum marks were 100.

Test whether the differences among the three sample means are significant.
Solution: In case of such problems, analysis of variance is ordinarily
performed. However, here we find that the data for Section C have much more
variability as compared to the data for the other two sections. In view of this, it
would be wrong to assume that the three population standard deviations are the
same. This means that the method of one-way analysis of variance cannot be used.
In order to perform a median test, we should first determine the median of the
combined data. This comes out to 63.5, as can easily be checked. Then we count
how many of the marks in each sample fall below or above the median. Thus, the
results obtained are shown in Table 1
204

Table 1 Worksheet for calculating chi-square

Since the corresponding expected frequencies for Section A are 10 and 10, for
Section B are 8 and 8, and for Section C 10 and 10, we can obtain the value of chi-
square. These calculations are shown below:

Now, we have to compare this value with the critical value of χ2at 5 per cent
level of significance. This value is 5.991 for 2 (K – 1 = 3 – 1) degrees of freedom (Chi-
square Table). As the calculated value of χ2is less than the critical value, the null
hypothesis cannot be rejected. In other words, we cannot conclude that there is a
difference in the true average (median) marks obtained by the students in Statistics
test from the three sections.
EXAMPLE
SIZE OF FARMER AND LEVEL OF SATISFACTION TOWARDS PRODUCTION OF
BANANA: (TWO WAY TABLE)
Level of satisfaction Total
HDS DS N S HS
No.of.Respondents
Small 42 42 28 16 22 150
Percentage 28.0% 28.0% 18.7% 10.7% 14.7% 100.0%
SIZE OF No.of.Respondents 12 19 20 30 19 100
FARMERS medium Percentage 12.0% 19.0% 20.0% 30.0% 19.0% 100.0%
No.of.Respondents 6 7 10 14 13 50
Large
Percentage 12.0% 14.0% 20.0% 28.0% 26.0% 100.0%
No.of.Respondents 60 68 58 60 54 300
Total
Percentage 20.0% 22.7% 19.3% 20.0% 18.0% 100.0%
Source: Primary Data
Null hypothesis: There is no relationship between the size of farmers and
level of satisfaction towards production of banana among the small, medium, large
farmers.
Alternative hypothesis: There is a relationship between the size of farmers
level of satisfaction towards production of banana among the small, medium, large
farmers.
The relationship between size of farmers and level of satisfaction towards the
production of banana. Out of the 300 respondents, 150 respondent up to three
members in these factions, 14.7 percent of the respondents fall under highly
satisfied and 28 percent of the respondents fall under highly dissatisfied about the
205

banana.100 respondents 4 to 6 members, 19 percent of the respondents fall under


highly satisfied and 12 percent of the responde nts fall under highly dissatisfied
about the banana. 50 respondents fall under from 7 and above members,
26percent of the respondents fall under highly satisfied, 12percent of the
respondents fall under highlydissatisfied about the production of banana.
CHI-SQUARE TEST

Pearson Chi-Square Calculated chi-square value Df P value s/ns Remarks

Size of farmers 29.694 8 .000 S Rejected


SOURCE: Primary Data
Weighted Average Score Analysis
The weighted average score analysis deals with the respondents’ opinion abo ut
the level of awareness towards cause -related marketing products. The factors
considered in the group as opinion factors are:
1. Quality seal/labels, 2. CRM campaign, 3. Promotional Activity, 4. Price, 5.
Availability of choice, 6. Brand name, and 7. Name of the alliance or non- profit
partner
For the purpose of analysis five point scaling technique was used to convert
the qualitative information into quantitative one. For the opinion about awareness
level of the score assigned are as follows:
1. Highly aware, 2. Aware, 3. Neutral, 4. Partially aware, Not aware.
In this section, for each product factor of the product, the level of opinion
about awareness level was determined and the results are given in the tables with
suitable interpretations.
Hindusthan
Colgate ITC P&G Tata Tea
Unilever
S.N TOTA TOTA TOTA TOTA TOTA
FACTOR
O. L WA L WA L WA L WA L WA
SCOR S SCOR S SCOR S SCOR S SCOR S
E E E E E
Quality
1 seal/ 3020 4.03 3005 4.01 3041 4.05 3236 4.31 3025 4.03
labels
(CRM)
2 Campaig 3072 4.10 2028 2.70 2770 3.69 2997 4.00 2804 3.74
n
Promotio
3 nal 2139 2.85 2162 2.88 2889 3.85 2863 3.82 2747 3.66
Activity
4 Price 2180 2.91 2186 2.91 3084 4.11 3132 4.18 2859 3.81
Availabilit
5 y of 2256 3.01 2158 2.88 2968 3.96 2984 3.98 2825 3.77
choice
206

Brand
6 2327 3.10 3119 4.16 3110 4.15 3096 4.13 3171 4.23
Name
Name of
the
alliance
7 1792 2.39 1762 2.35 2456 3.27 2711 3.61 2330 3.11
or non-
profit
partner
Source: Primary Data
15.4 REVISION POINTS
1. Chi square test
2. Steps involved in chi square test
3. Goodness of fit
15.5 INTEXT QUESTION
Does the new math program improve student performance? Suppose you take
a random sample of 20 students who are using a new algebra text which features
group work and unit summaries and a second sample of 30 students who are using
a more traditional text. You compare student achievement on the state test given to
all students at the end of the course. Use the frequency table to determine if the
proportions from each group are equal at each performance level.

Below grade level At grade level Advanced


New text 8 6 6
Old text 6 15 9
15.6 SUMMARY
After going through this Lesson you should be able to Explain Chi square test,
steps involved in Chi square test and Goodness of fit.
15.7 TERMINAL EXERCISES
A sample of 100 voters are asked whi ch of four candidates they would vote for
in an election. The number supporting each candidate is given below:
Higgins Reardon White Charlton
41 19 24 16
Do the data suggest that all candidates are equally popular?
15.8 SUPPLEMENTARY MATERIALS
Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, rev. ed., New
York: Harcourt, Brace & World, Inc., 1970.
15.9 ASSIGNMENTS
A study of conformity using the Asch paradigm involved two conditions: one
where one confederate supported the true judgement, and another where no
confederate gave the correct response
207

Support: No Support:
Conform: 18 40
Not Conform: 32 10
Is there a significant difference between the "support" and "no support"
conditions in the frequency with which individuals are likely to conform?
15.10 SUGGESTED READINGS / REFERENCE BOOKS
1. Ackoff, Russell L., The Design of Social Research, Chicago: University of
Chicago Press, 1961.
2. Ackoff, Russell L., Scientific Method, New York: John Wiley & Sons, 1962.
3. Allen, T. Harrell, New Methods in Social Science Research, New York:
Praeger Publishers, 1978.
4. Anderson, H.H., and Anderson, G.L., An Introduction to Projective
Techniques and Other Devices for Understanding the Dynamics of Human
Behaviour, New York: Prentice Hall, 1951.
5. Anderson, T.W., An Introduction to Multivariate Analysis, New York: John
Wiley & Sons, 1958.
6. Bailey, Kenneth D., “Methods of Social Research,” New York, 1978.
7. Baker, R.P., and Howell, A.C., The Preparation of Reports, New York: Ronald
Press, 1938.
8. Bartee, T.C., “Digital Computer Fundamentals,” 5th Ed., McGraw-Hill,
International Book Co., 1981.
9. Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, rev. ed.,
New York: Harcourt, Brace & World, Inc., 1970.
10. Bell, J.E., Projective Techniques: A. Dynamic Approach to the Study of
Personality, New York: Longmans Green, 1948.
11. Bellenger, Danny N., and Greenberg, Barnett A., Marketing Research—A
Management Information Approach, Homewood, Illinois: Richard D. Irwin, Inc.,
1978.
15.11 LEARNING ACTIVITIES
A researcher is interested in sex differences in social play in cats. She observes
sixteen kittens (eight male and eight females), and records the sex of the
participants in 500 play-bouts, taking a note of the sex of the animal initiating a
bout and the sex of the animal who responds to the initiation. Are there sex -
differences in the kittens' play? Perform a Chi -Square test on these data. Is there
anything wrong with doing this test?
Respondent Male Female
initiator of bout: Male: 200 150
Female: 100 50
15.12 KEY WORDS
Chi square test, steps involved in Chi square test and Goodness of fit.
208

LESSON – 16
MANN-WHITNEY TEST
16.1 INTRODUCTION
Statistics is a set of tools used to organize and analyze data. Data must either
be numeric in origin or transformed by researchers into numbers. Employing
statistics serves two purposes, (1) description and (2) prediction. Statistics are used
to describe the characteristics of groups. These characteristics are referred to as
variables. Data is gathered and recorded for each variable. Descriptive statistics can
then be used to reveal the distribution of the data in each variable. Statistics is also
frequently used for purposes of prediction. Prediction is based on the concept of
generalizability: if enough data is compiled about a particular context (e.g.,
students studying writing in a specific set of classrooms), the patterns revealed
through analysis of the data collected about that context can be generalized (or
predicted to occur in) similar contexts. The prediction of what will happen in a
similar context is probabilistic. That is, the researcher is not certain that the same
things will happen in other contexts; instead, the researcher can only reasonably
expect that the same things will happen.
16.2 OBJECTIVE
The main aim of this Lesson is to study the Basics of statistics. After going
through this Lesson you should be able to:
 Describe the origin of Statistics
 Explain about the importance of Statistics
Use this test when two different groups of participants perform both conditions
of the study: i.e., it is ppropriate for analysing the data from an independent-
measures design with two conditions. Use it when the data do not meet the
requirements for a parametric test (i.e. if the data are not normally distributed; if
the variances for the two conditions are markedly different; or if the data are
measurements on an ordinal scale). Otherwise, if the data meet the requirements
for a parametric test, it is better to use an independent-measures t-test (also known
as a "two-sample" t-test). The logic behind the Mann-Whitney test is to rank the
data for each condition, and then see how different the two rank totals are. If there
is a systematic difference between the two conditions, then most of the high ranks
will belong to one condition and most of the low ranks will belong to the other one.
As a result, the rank totals will be quite different. On the other hand, if the two
conditions are similar, then high and low ranks will be distributed fairly evenly
between the two conditions and the rank totals will be fairly similar. The Mann -
Whitney test statistic "U" reflects the difference between the two rank totals. The
SMALLER it is (taking into account how many participants you have in each group)
then the less likely it is to have occurred by chance. In this study this test is used
to know the rank preference of Khadi consumers based on gender wise
(male/female) towards Khadi textile products.
209

16.3 CONTENT
16.3.1 Meaning
16.3.2 Definition
16.3.3 Uses of statistics
16.3.1 MEANING
“Statistics”, that a word is often used, has been derived from the Latin word
‘Status’ that means a group of numbers or figures; those represent some
information of our human interest. We find statistics in everyday life, such as in
books or other information papers or TV or newspapers. Although, in the beginning
it was used by Kings only for collecting information about states and other
information which was needed about their people, their numbe r, revenue of the
state etc. This was known as the science of the state because it was used only by
the Kings. So it got its development as ‘Kings’ subject or ‘Science of Kings’ or we
may call it as “Political arithmetic’s”. It was for the first time, perhaps in Egypt to
conduct census of population in 3050 B.C. because the king needed money to erect
pyramids. But in India, it is thought, that, it started dating back to Chandra Gupta
Maurya’s kingdom under Chankya to collect the data of births and deaths. T M has
also been stated in Chankya’s Arthshastra. But now-a-days due to its pervading
nature, its scope has increased and widened. It is now used in almost in all the
fields of human knowledge and skills like Business, Commerce, Economics, Social
Sciences, Politics, Planning, Medicine and other sciences, Physical as well as
Natural.
16.3.2 DEFINITION
The term ‘Statistics’ has been defined in two senses, i.e. in Singular and in
Plural sense.
a) In the Plural Sense
“Statistics are numerical statements of facts in any department of enquiry
placed in relation to each other.” —A.L. Bowley
“The classified facts respecting the condition of the people in a state —
especially those facts which can be stated in numbers or in tables of numbers or in
any tabular or classified arrangement.” —Webster
“By statistics we mean aggregates of facts affected to a marked extent by
multiplicity of causes, numerically expressed, enumerated or estimated according
to reasonable standard of accuracy, collected in a systematic manner for a
predetermined purpose, and placed in relation to each other.” —Horace Sacrist
b) In the Singular Sense
“Statistics refers to the body of technique or methodology, which has been
developed for the collection, presentation and analysis of quantitative data and for
the use of such data in decision making.” —Ncttor and Washerman
“Statistics may rightly be called the science of averages.” —Bowleg
210

“Statistics may be defined as the collection, presentation, analysis, and


interpretation of numerical data.” —Croxton and Cowden
Stages of Investigations
1. Collection of Data: It is the first stage of investigation and is regarding
collection of data. It is determined that which method of collection is needed in this
problem and then data are collected.
2. Organisation of Data: It is second stage. The data are simplified and made
comparative and are classified according to time and place.
3. Presentation of Data: In this third stage, organised data are made simple and
attractive. These are presented in the form of tabl es diagrams and graphs.
4. Analysis of Data: Forth stage of investigation is analysis. To get correct
results, analysis is necessary. It is often undertaken using Measures of central
tendencies, Measures of dispersion, correlation, regression and interpolation etc.
5. Interpretation of Data: In this last stage, conclusions are enacted. Use of
comparisons is made. On this basis, forecasting is made.
16.3.3 USES OF STATISTICS
1. Statistics helps in providing a better understanding and exact description of
a phenomenon of nature.
2. Statistical helps in proper and efficient planning of a statistical inquiry in
any field of study.
3. Statistical helps in collecting an appropriate quantitative data.
4. Statistics helps in presenting complex data in a suitable tabular,
diagrammatic and graphic form for an easy and clear comprehension of the
data.
5. Statistics helps in understanding the nature and pattern of variability of a
phenomenon through quantitative obersevations.
6. Statistics helps in drawing valid inference, along with a measure of their
reliability about the population parameters from the sample data.
THE BIRTH OF PROBABILITY AND STATISTICS
The original idea of "statistics" was the collection of information about and for
the "state". The word statistics derives directly, n ot from any classical Greek or
Latin roots, but from the Italian word for state.
The birth of statistics occurred in mid-17th century. A commoner, named John
Graunt, who was a native of London, began reviewing a weekly church publication
issued by the local parish clerk that listed the number of births, christenings, and
deaths in each parish. These so called Bills of Mortality also listed the causes of
death. Graunt who was a shopkeeper organized this data in the form we call
descriptive statistics, which was published as Natural and Political Observations
Made upon the Bills of Mortality. Shortly thereafter he was elected as a member of
Royal Society. Thus, statistics has to borrow some concepts from sociology, such as
211

the concept of Population. It has be en argued that since statistics usually involves
the study of human behavior, it cannot claim the precision of the physical sciences.
Probability has much longer history. Probability is derived from the verb to
probe meaning to "find out" what is not too easily accessible or understandable .
The word "proof" has the same origin that provides necessary details to understand
what is claimed to be true.
Probability originated from the study of games of chance and gambling during
the 16 th century. Probability theory was a branch of mathematics studied by Blaise
Pascal and Pierre de Fermat in the seventeenth century. Currently in 21 st century,
probabilistic modeling is used to control the flow of traffic through a highway
system, a telephone interchange, or a computer processor; find the genetic makeup
of individuals or populations; quality control; insurance; investment; and other
sectors of business and industry.
New and ever growing diverse fields of human activities are using statistics;
however, it seems that this field itself remains obscure to the public. Professor
Bradley Efron expressed this fact nicely:
During the 20 th Century statistical thinking and methodology have become the
scientific framework for literally dozens of fields including education, agriculture,
economics, biology, and medicine, and with increasing influence recently on the
hard sciences such as astronomy, geology, and physics. In other words, we have
grown from a small obscure field into a big obscure field.
GARRETT RANKING TECHNIQUE
This technique was used to rank the preference of the respondents on different
aspects of the study. The order of merit given by the respondents were converted
into ranks by using the following formula.

100 (R  0.5)
ij
Percentage Position =
H
j

Where R = Rank given for ith for jth individual


ij

H Rank given for for jth individual


j

The percentage position of each rank thus obtained is converted into sc ores by
referring to the table given by Henry Garrett. Then for each factor the scores of
individual respondents are added together and divided by the total number of
respondents for whom the scores were added. These mean scores for all the factors
were arranged in the descending order, ranks given and most important aspects
identified.
212

EG. 1 PROBLEMS FACED BY THE FARMERS IN BANANA PRODUCTION


TOTAL MEAN
SL.NO PROBLEMS RANK
SCORE SCORE
1 Heavy Damage by wind 18013 60.04 I
2 Severity of pests 17441 58.14 II
3 Deteriorating soil quality 17134 57.11 III
4 Severity of diseases 16975 56.58 IV
5 Deteriorating water quality 16616 55.39 V
6 Heavy weed occurrences 13795 45.98 VI
7 Non availability of water 13697 45.66 VII
8 Environment 13418 44.73 VIII
9 Lack of labour supply 13136 43.79 IX
10 Affect the growth from natural calamities 12507 41.69 X
11 Lack of availability of quality fertilizer
and pesticide 11609 38.70 XI
12 No aware about crop insurance 11045 36.82 XII
SOURCE: PRIMARY DATA
The table reveals that the problems faced by the farmers in banana
production. “Heavy damage by wind” was ranked first by the selected sample
respondents with a total score of 18013 and mean score of 60.04. “Severity of
pests” was ranked second with a total score of 17441 and mean score of 58.14.
“Deteriorating soil quality” occupied third position with a total score of 17134and
mean score of 57.11. “Severity of diseases “was ranked fourth with a total score of
16975 and mean score of 56.58. “Deteriorating water quali ty” was ranked with fifth
position with a total score of 16616 and mean score of 55.39. “Heavy weed
occurrences” was ranked with sixth position with a total score of 13795 and mean
score of 45.98. “Non availability of water” was ranked seventh by the selec ted
sample respondents with a total score of 13697and mean score of 45.66.
“Environment” was ranked eighth with a total score of 13418 and mean score of
44.73. “Lack of labour supply” occupied ninth position with a total score of
13136and mean score of 43.79. “Affect the growth from natural calamities “was
ranked with tenth with a total score of 12507 and a mean score of 41.69. “Lack of
availability of quality fertilizer and pesticide” was ranked with eleventh position
with a total score of 11609and mean score of 38.70. The last occupied by “no
aware about crop insurance” with total score of 11045 and mean score of 36.82.It is
evident that most of the respondents gave top priority to “Heavy damage by wind”
for supporting problems faced by the farmers in banana production.
213

E.G 2 Problems Faced by the Farmers In Banana Production


TOTAL MEAN
SL.NO PROBLEMS RANK
SCORE SCORE
1 Indebtedness to traders 16834 56.11 I
2 Heavy Commission charges 16669 55.56 II
3 Inadequate finance 15315 51.05 III
4 Price fluctuations 15240 50.80 IV
5 Absence of grading processing 15119 50.40 V
6 High transport cost 14954 49.85 VI
7 Lack of storage facilities 14819 49.40 VII
8 No regular payment 14685 48.95 VIII
9 Lack of market information 14367 47.89 IX
10 Seasonal glut 13849 46.16 X
11 Malpractices by the intermediaries 13418 44.73 XII
12 Absence of Co-operative marketing and
regulated marketing 12663 42.21 XII
SOURCE: PRIMARY DATA
The table reveals that the problems faced by the farmers in banana
production. “Indebtedness to traders” was ranked first by the selected sample
respondents with a total score of 16834 and mean score of 56.11. “Heavy
commission charges” was ranked second with a total score of 16669 and mean
score of 55.56. “Inadequate finance” occupied thi rd position with a total score of
15315and mean score of 51.05. “Price fluctuations “was ranked fourth with a total
score of 15240 and mean score of 50.80. “Absence of grading, processing” was
ranked with fifth position with a total score of 15119 and mean score of 50.40.
“High transport cost” was ranked with sixth position with a total score of 14954
and mean score of 49.85. “Lack of storage facilities” was ranked seventh by the
selected sample respondents with a total score of 14819and mean score of 49.40 .
“No regular payment” was ranked eighth with a total score of 14685 and mean
score of 48.95. “Lack of market information” occupied ninth position with a total
score of 14367and mean score of 47.89. “Seasonal glut “was ranked with tenth with
a total score of 13849 and mean score of 46.16. “Malpractices by the
intermediaries” was ranked with eleventh position with a total score of 13418and
mean score of 44.73. The last occupied by “absence of co-operative marketing and
regulated marketing” with total score of 12663 and mean score of 42.21.
It is evident that most of the respondents gave top priority to “indebtedness to
traders” for supporting problems faced by the farmers in banana production.
In the independent samples case, it is preferable to apply the in dependent
sample t test when there exists normality of the sample mean for each sample. If
there is a violation of this assumption, then a non parametric test for independent
samples is applied – the Wilcoxon Mann Whitney test. This test uses the relative
214

position of the data in a rank ordering, unlike the t test which uses the actual
values. To use the Wilcoxon Mann Whitney test (a) It must be reasonable to regard
the data as a random sample from their respective populations (b) Observations
within each sample must be independent of one another (c) The two samples must
be independent of one another
The following data indicate the lifetime (in hours) of samples of two kinds of
light bulbs in continuous use:
Brand A 603 625 641 622 585 593 660 600 633 580 615 648
Brand B 620 640 646 620 652 639 590 646 631 669 610 619
We are required to use the Mann-Whitney test to compare the lifetimes of
brands A and B
Light bulbs
Solution:The first step for performing the Mann-Whitney test is to rank the
given data jointly(as if they were one sample) in an increasing or decreasing order of
magnitude. For our data, we thus obtain the following array where we use the
letters A and B to denote whether the light bulb was from brand A or brand B.
Table 1 Ranking of light bulbs of brands A and B
Sample score Group Rank Sample score Group Rank
580 A 1 625re A 13
585 A 2 631 B 14
590 B 3 633 A 15
593 A 4 639 B 16
600 A 5 640 B 17
603 A 6 641 A 18
610 B 7 646 B 19.5
615 A 8 646 B 19.5
619 B 9 648 A 21
620 B 10.5 652 B 22
620 B 10.5 660 A 23
622 A 12 669 B 24

As both the samples come from identical populations, it is reasonable to


assume that the means of the ranks assigned to the values of the two samples are
more or less the same.
As such, our null hypothesis is:
H0: Means of ranks assigned to the values in the two groups are the same.
H1: Means are not the same.
However, instead of using the means of the ranks, we shall use rank sums for
which the following formula will be used.
215

U = n1n2+ [n1(n1+ 1)]/2 – R1Where n1and n2are the sample sizes of Group 1
and Group 2, respectively, and R1is the sum of the ranks assigned to the values of
the first sample. In our example, we have n1= 12, n2= 12 and R1= 1 + 2 + 4 + 5 + 6
+ 8 + 12 + 13 + 15 + 18 + 21 + 23 = 128. Substituting these values in the above
formula,
U = (12) (12) + [12 (12 + 1)]/2 – 128
= 144 + 78 – 128
= 94
From Appendix Table 9 for n1and n2, each equal to 12, and for 0.05 level of
significance is 37. Since the critical value is smaller than the calcul ated value of 94,
we accept the null hypothesis and conclude that there is no difference in the
average lifetimes of the two brands of light bulbs. The test statistic we have just
applied is suitable when n1and n2are less than or equal to 25. For larger values of
n1and/or n2, we can make use of the fact that the distribution of Ws approaches a
normal distribution as sample sizes increase. We can then use the Z test to test the
hypothesis. The normal approximation
Although our observations are limited, we may apply the normal approximation
to this problem. For this, we have to use the Z statistic.

1. Mean μu  [(N1 N2/2)  [(12 x 12)2] 72

n1 n2 (n1 n2  1
2. Standard error =
12

12  12 (12  12) 1
 300  17.3
12
3. (Statistic – Mean)/Standard deviation
= (94 – 72)/17.3 = 1.27
The critical value of Z at 0.05 level of significance is 1.64. Since the calculated
value of Z = 1.27 is smaller than 1.64, the null hypothesis is accepted. This shows
that there is no difference in average lifeti mes of brands A and B bulbs. The Z test
is more dependable as compared to the earlier one. It may be noted that Mann -
Whitney test required fewer assumptions than the corresponding standard test. In
fact, the only assumption required is that the populations from which samples have
been drawn are continuous. In actual practice, even when this assumption turns
out to be wrong, this is not regarded serious.
e.g:2
In the independent samples case, it is preferable to apply the independent
sample t test when there exists normality of the sample mean for each sample. If
there is a violation of this assumption, then a non parametric test for independent
samples is appied – the Wilcoxon Mann Whitney test. This test uses the relative
216

position of the data in a rank ordering, unlike the t test which uses the actual
values. To use the Wilcoxon Mann Whitney test (a) It must be reasonable to regard
the data as a random sample from their respective populations (b) Observations
within each sample must be independent of one another (c) The two samples must
be independent of one another
Gender wise Mean Ranks of Khadi Textile Products
Mean Sum of
Gender N Rank Ranks

Male 198 207.21 41028


textile products Female 221 212.5 46962
Total 419
Asymp. Sig. (2- 0.152
tailed)
Mann-Whitney Test For Genderwise Mean Rank Variation of
The Responses of Khadi Consumers With Respect to Khadi Textile Products
Textile Products
Mann-Whitney U 21327
Wilcoxon W 41028
Z -1.433
Asymp. Sig. (2-tailed) 0.152
The table above shows that majority of Khadi textile products are purchased
by male consumers (mean rank 207.21) than female consumers even if this mean
variation is statistically not significant at 5 per cent level od significance(P>.05 in
the above case).
16.4 REVISION POINT
Statistics refers to the methodology, which has been developed for the
collection, presentation and analysis of quantitative data and for the use of such
data in decision making.
16.5 INTEXT QUESTIONS
1. Define Statistics
2. Briefly explain the uses of statistics
3. Highlight the importance of statistics
16.6 SUMMARY
By the end of this lesson, you should be able to un derstand:
1. What is Statistics
2. Importance of the statistics
3. Uses of the statistics
217

16.7 TERMINAL EXERCISES


1. Explain the importance of statistics in formulating economic policies
2. Give examples of problem situations in which statistics was used
16.8 SUPPLEMENTARY MATERIALS
1. Gupta, S.P., ’Statistics Methods’, Sultan Chand and Co., New Delhi
16.9 ASSIGNMENTS
1. Provide at least three examples of problem situations in which statistics was
used or could be used
2. Why is statistics important in the real world
16.10 REFERENCE
1. Agarwal, B.L. ‘Basic Statistics’, Wiley Eastern Ltd., New Delhi.
2. Chance, W., ‘Statistical methods for Decision’ Irwin Inc., Homewood.
3. Gopikuttam, G., ‘Quantitative Methods and Operations Research’, Himalaya
Publishing House, Bombay.
4. Gupta, S.P., ’Statistics Methods’, Sultan Chand and Co., New Delhi
5. Levin, R., ‘Statistics for Management’, Prentice – Hall of India, New Delhi,
1984.
6. Reddy, C.R., ‘Quantitative Methods for Management Decision’, Himalaya
Publishing House, Bombay, 1990.
16.11 LEARNING ACTIVITIES
1. Explain the Importance of statistics in day to day life
2. Explain the scope of statistics in different fields
16.12 KEYWORDS
1. Collection of Data
2. Organization of Data
3. Presentation of Data
4. Analysis of Data
5. Interpretation of Data


218

UNIT – V
LESSON – 17

MULTIVARIATE STATISTICAL ANALYSIS


17.1 INTRODUCTION
This chapter focuses the Multivariate analysis and Factor analysis in detail
with MANOVA and Discriminant analysis. This chapter also deals the Structural
equation modeling, Conjoint analysis, Multidimensional scaling, theory of Cluster
Analysis and Role of statistics software in business decision.
Multivariate analysis consists of a collection of methods that can be used when
several measurements are made on each individual or object in one or more
samples. We will refer to the measurements as variables and to the individuals or
objects as units(research units, sampling units, or experimental units) or
Observations. In practice, multivariate data sets are common, although they
are not always analyzed as such. But the exclusive use of univariate procedures
with such data is no longer excusable, given the availability of multivariate
techniques and inexpensive computing power to carry them out.
Historically, the bulk of applications of multivariate techniques have been in
thebehavioral and biological sciences. However, interest in multivariate methods
has now spread to numerous other fields of investigation. For example, I have
collaborated on multivariate problems with researchers in education, chemistry,
physics, geology, engineering, law, business, literature, religion, public
broadcasting, nursing, mining, linguistics, biology, psychology, and many other
fields. Table 1.1 showssome examples of multivariate observations.
The reader will notice that in some cases all the variables are measured in the
same scale (see 1 and 2 in Table 1.1). In other cases, measurements are in different
scales (see 3 in Table 1.1). In a few techniques, such as profile analysis (Sections
5.9 and 6.8), the variables must be commensurate, that is, similar in scale of
measurement; however, most multivariate methods do not require this.
Ordinarily the variables are measured simultaneously on each sampling unit.
Typically, these variables are correlated. If thi s were not so, there would be little use
for many of the techniques of multivariate analysis. We need to untangle the
overlapping information provided by correlated variables and peer beneath the
surface to see the underlying structure. Thus the goal of many multivariate
approaches is simplification. We seek to express what is going on in terms of a
reduced set of dimensions.Such multivariate techniques are exploratory; they
essentially generate hypotheses rather than test them.On the other hand, if our
goal is a formal hypothesis test, we need a technique that will (1) allow several
variables to be tested and still preserve the significance level12
Multivariate statistical analysis refers to multiple advanced techniques for
examining relationships among mul tiple variables at the same time. Researchers
use multivariate procedures in studies that involve more than one dependent
219

variable (also known as the outcome or phenomenon of interest), more than one
independent variable (also known as a predictor) or both. Upper-level
undergraduate courses and graduate courses in statistics teach multivariate
statistical analysis. This type of analysis is desirable because researchers often
hypothesize that a given outcome of interest is effected or influenced by more than
one thing.
16.2 OBJECTIVE
After going through this Lesson you should be able to understand the types of
multivariate analysis, multiple regression, factor analysis, Path analysis and
MANOVA
17.3 CONTENT
17.3.1 Types
17.3.2 Regression Analysis
17.3.3 Factor analysis
17.3.4 Path analysis
17.3.5 MANOVA
17.3.1 Types
There are many statistical techniques for conducting multivariate analysis,
and the most appropriate technique for a given study varies with the type of study
and the key research questions. Four of the most common multivariate techniques
are multiple regression analysis, factor analysis, path analysis and multiple
analysis of variance, or MANOVA.
17.3.2 Multiple Regressions
Multiple regression analysis, often referred to simply as regression analysis,
examines the effects of multiple independent variables (predictors) on the value of a
dependent variable, or outcome. Regression calculates a coefficient for each
independent variable, as well as its statistical significance, to estimate the effect of
each predictor on the dependent variable, with other predictors held constant.
Researchers in economics and other social sciences often use regression analysis to
study social and economic phenomena. An example of a regression study is to
examine the effect of education, experience, gender, and ethnicity on income.
17.3.3 Factor Analysis
Factor analysis is a data reduction technique in which a researcher reduces a
large number of variables to a smaller, more manageable, number of factors. Factor
analysis uncovers patterns among variables and then clusters highly interrelated
variables into factors. Factor analysis has many applications, but a common use is
in survey research, where researchers use the technique to see if lengthy series of
questions can be grouped into shorter sets.
17.3.4 Path Analysis
This is a graphical form of multivariate statistical analysis in which graphs
known as path diagrams depict the correlations among variables, as well as the
220

directions of those correlations and the "paths" alon g which these relationships
travel. Statistical software programs calculate path coefficients, the values of which
estimate the strength of relationships among the variables in a researcher's
hypothesized model.
The term ‘path analysis’ was first introduce d by the biologist Sewall Wright in
1934 in connection with decomposing the total correlation between any two
variables in a causal system. The techniqueof path analysis is based on a series of
multiple regression analyses with the added assumption ofcausal relationship
between independent and dependent variables. This technique lays relativelyheavier
emphasis on the heuristic use of visual diagram, technically described as a path
diagram. Anillustrative path diagram showing interrelationships between Fathe rs’
education, Fathers’ occupation,Sons’ education, Sons’ first and Sons’ present
occupation can be shown in the Fig. 13.2.
Path analysis makes use of standardized partial regression coefficients (known
as beta weights) as effect coefficients. In linear additive effects are assumed, then
through path analysis a simple setof equations can be built up showing how each
variable depends on preceding variables. “The mainprinciple of path analysis is that
any correlation coefficient between two variables, or a gross oroverall measure of
empirical relationship can be decomposed into a series of parts: separate paths
ofinfluence leading through chronologically intermediate variable to which both the
correlated variables have links.”
The merit of path analysis in comparison to correlational analysis is that it
makes possible theassessment of the relative influence of each antecedent or
explanatory variable on the consequent orcriterion variables by first making explicit
the assumptions underlying the causal connection s andthen by elucidating the
indirect effect of the explanatory variables.

“The use of the path analysis technique requires the assumption that there
are linear additive, a symmetric relationships among a set of variables which can be
measured at least on a quasi-intervalscale. Each dependent variable is regarded as
determined by the variables preceding it in the pathdiagram, and a residual
221

variable, defined as uncorrelated with the other variables, is postulated toaccount


for the unexplained portion of the variance in the dependent variable. The
determiningvariables are assumed for the analysis to be given (exogenous in the
model).”
We may illustrate the path analysis technique in connection with a simple
problem of testing a causal model with three explicit variables as shown in the
following path diagram: Fig. 13.3
Fig not clear
The structural equation for the above can be written as:

 x1   e 
 x    p X1  e   pX  e
  
2 21 1 2 
x 3  p31X2  X2  e3 
where the Xvariables are measured as deviations from their respective means.
p21may be estimated from the simple regression of X 2on X1i.e., X2 = b21 Xl and p31
and p32may be estimated from the regression of X 3 on X2 and X1 as under:

X3  b31.2 X1  b21 X2

where b31.2 means the standardized partial regression coefficient for predicting
variable 3 from variable 1 when the effect of variable 2 is held constant.
In path analysis the beta coefficient i ndicates the direct effect of Xj (j = 1, 2, 3,
..., p) on the dependent variable. Squaring the direct effect yields the proportion of
the variance in the dependentvariable Ywhich is due to each of the pnumber of
independent variables Xj (i = 1, 2, 3, ..., p). After calculating the direct effect, one
may then obtain a summary measure of the total indirect effect ofXjon the
dependent variable Yby subtracting from the zero correlation coefficient ryxj, the
beta coefficient bji.e.,

Indirect effect of X1 on Y  c  r  b For all j = 1, 2 = …p


yx 1

Such indirect effects include the unanalysed effects and spurious relationships
due to antecedent variables. In the end, it may again be emphasised that the main
virtue of path analysis lies in making explicitthe assumptions underlying the causal
connections and in elucidating the indirect effects due toantecedent variables of the
given system.
17.3.5 MANOVA
Multiple Analysis of Variance, or MANOVA, is an advanced form of the more
basic analysis of variance, or ANOVA. MANOVA extends the technique to studies
with two or more related dependent variables while controlling for the correlations
among them. An example of a study for which MANOVA would be an appropriate
technique is a study of health among three groups of teens: those who exercise
222

regularly, those who exercise on occasion, and those who never exercise. A
MANOVA for this study would allow multiple health-related outcome measures
such as weight, heart rate, and respiratory rates.
Benefits
Multivariate statistical analysis is especially important in social science
research because researchers in these fields are often unable to use randomized
laboratory experiments that their counterparts in medicine and natural sciences
often use. Instead, many social scientists must rely on quasi -experimental designs
in which the experimental and control groups may have initial differences that
could affect or bias the outcome of the study. Multivariate techniques try to
statistically account for these differences and adjust outcome measures to control
for the portion that can be attributed to the differences.
Example
A research team wants to test the user acceptance with a new onli ne travel
booking tool. The team conducts a study where they assign 30 randomly chosen
people into 3 groups. The first group needs to book their travel through an
automated online-portal; the second group books over the phone via a hotline; the
third group sends a request via the online-portal and receives a call back. The
team measures user acceptance as the behavioral intention to use the system, they
do measure the latent construct behavioral intention with 3 variables – ease of use,
perceived usefulness, effort to use.
In the example, some statisticians argue that the MANOVA can only find the
differences in the behavioral intention to use the system. However, some
statisticians argue that you can establish a causal relationship between the
channel they used and the behavioral intention for future use. It is generally
assumed that the MANOVA is an ‘analysis of dependencies’. It is referred to as
such because it proves an assumed cause -effect relationship between two or more
independent variables and two or more dependent variables. In more statistical
terms, it tests the effect of one or more independent variables on one or more
dependent variables.
Other things you may want to try
When faced with a question similar to the one in our example you cou ld also
try to run a 3 factorial ANOVAs, testing the influence of the three independent
variables (the three channels) on each of the three dependent variables (ease of use,
perceived usefulness, effort to use) individually. However running multiple factorial
ANOVAs does not account for the full variability in all three dependent variables
and thus the test has lesser power than the MANOVA.
Another thing you might want to try is running a factor analysis on the three
dependent variables and then running a factorial ANOVA. The factor analysis
reduces the variance within the three dependent variables to one factor, thus this
procedure does have lesser power than the MANOVA.A third approach would be to
conduct a discriminant analysis and switch the dependent and independent
223

variables. That is the discriminant analysis uses the three groups (online, phone,
call back) as the dependent variable and identifies the significantly discriminating
variables from the list of continuous-level variables (ease of use, perceived
usefulness, effort to use).
Mathematically, the MANOVA is fully equivalent to the discriminant analysis.
The difference consists of a switching of the independent and dependent variables.
Both the MANOVA and the discriminant analysis are a series of canonical
regressions. The MANOVA is therefore the best test use when conducting
experiments with latent variables. This is due to the fact that it only requires a
nominal scale for the independent variables which typically represent the
treatment. This includes multiple continuous-level independent variables – which
typically measure one latent (not directly observable) construct.
The MANOVA is much like the one-way ANOVA and the factorial ANOVA in
that the one-way ANOVA has exactly one independent and one dependent variable.
The factorial ANOVAs can have one or more independent variables but always has
only one dependent variable. On the other hand the MANOVA can have two or
more dependent variables.
Considerations
Statistical software programs such as SAS, Stata, and SPSS can perform
multivariate statistical analyses. These programs are frequently used by university
researchers and other research professionals. Spreadsheet programs can perform
some multivariate analyses, but are intended for more g eneral use and may have
limited abilities than a specialized statistical software package.
One-Way MANOVA
MANOVA is short for Multivariate ANalysis Of Variance. The main purpose of
a one-way ANOVA is to test if two or more groups differ from each other
significantly in one or more characteristics. A factorial ANOVA compares means
across two or more variables. Again, a one-way ANOVA has one independent
variable that splits the sample into two or more groups whereas the factorial
ANOVA has two or more independent variables that split the sample in four or more
groups. A MANOVA now has two or more independent variables and two or more
dependent variables.
For some statisticians the MANOVA doesn't only compare differences in mean
scores between multiple groups but also assumes a cause effect relationship
whereby one or more independent, controlled variables (the factors) cause the
significant difference of one or more characteristics. The factors sort the data
points into one of the groups causing the difference in the mean value of the
groups.
FACTOR ANALYSIS
Factor Analysis is a set of technique which by analyzing correlations between
variables reduces their numbers into fewer factors which explain much of the
original data, more economically. Even though a subjective interpretation can result
224

from a factor analysis output, the procedure often provides an insight into relevant
psychographic variables, and results in economic use of data collection efforts. The
subjective element of factor analysis is reduce d by splitting the sample randomly
into two and extracting factors separately from both parts. If similar factors result,
the analysis is assumed as reliable or stable.
A. METHODS OF CONDUCTING FACTOR ANALYSIS5, 6, 7

There are two stages in factor analysis.


Stage 1: Factor Extraction process,- this process is primarily used to
determine how many factors will be extracted from data.
Stage 2: Rotation of Principal Components.-This is actually optional, but
highly recommended. In this study, the rotation of principal component is used.
After extracting the factors, the next task is to interpret and title the relevant
factors. This is done by the process of identifying which factors are associated with
which original variables. The factor matrix is used for this purpose. The rotated
factor matrix comes about in stage 2, these rotated factor matrices is used to
anlayse and interpret the factors.
The factor matrix gives us the loading of each variable on each of the extracted
factors. This is similar to a correlation matrix with ‘loadings’ are having values
between 0 and 1. Values close to 1 represent high loadings and those close to 0,
denotes low loadings. The normal procedure is to identify the high loading factors
and provide a suitable title.
B. STEPS INVOLVED IN CONDUCTING FACTOR ANALYSIS
The steps involved in conducting factor analysis are as follows,
 The first step is to formulate the factor analysis problem and identify the
variables to be factor analyzed
 The next step is to develop a correlation matrix of the variables and the
method of factor analysis is selected.
 The next step involves the researcher's decision on extraction of number of
factors and the method of rotation.
 Next, the rotated factors should be interpreted
 Depending on the objective, the factor scores may be calculated, or
surrogate variables selected to represent the factors in subsequent
multivariate analysis
 Dimensionality of the multi-scale items (factor analysis)
 Factor Analysis is a set of technique which by analyzing correlation s
between variables reduces their numbers into fewer factors which explain
much of the original data, more economically. Even though a subjective
interpretation can result from a factor analysis output, the procedure
often provides an insight into relevant psychographic variables, and
results in economic use of data collection efforts. The subjective element of
225

factor analysis is reduced by splitting the sample randomly into two and
extracting factors separately from both parts. If similar factors result, the
analysis is assumed as reliable or stable.
TABLE - 4.4.1.1
KMO AND BARTLETT’S TEST FOR FACTORS RELATED TO INFLUENCE IN
CHOOSING CRM PRODUCTS
Kaiser-Meyer-Olkin Measure of Sampling Adequacy .681
Bartlett’s Test of Sphericity: Approx. Chi -Square 2130.534
Sig 0.000**
S/NS S

**p<0.01 S-Significant
From the above table, two tests, namely Kaiser-Meyer-Olkin Measure of
Sampling Adequacy (KMO) and Bartlett’s Test of Sphericity have been applied to
test whether the relationship among the variables has been significant or not. The
Kaiser-Meyer-Olkin Measure of sampling adequacy shows that the value of test
statistics is 0.681, which means the factor analysis for the selected variable is
found to be appropriate or good to the data. Bartlett’s test of sphericity is used to
test whether the data are statistically significant or not with the value of test
statistics and the associated significance level. It shows that there exists a high
relationship among variables.
Table – 4.4.1.2
EIGEN VALUES AND PROPORTION OF TOTAL VARIANCE OF EACH NDERLYING
FACTORS RELATED TO INFLUENCE IN CHOOSING CRM PRODUCTS
Initial Eigen v alues Ex traction Sums of Squared Loadings Rotation Sums of Squared loadings
Component Total % of Cumulativ e % Total % of Cumulativ e Total % of Cumulativ e %
Variance Variance % Variance
1 3.418 28.486 28.486 3.418 28.486 28.486 2.080 17.331 17.331
2 1.570 13.081 41.567 1.570 13.081 41.567 1.881 15.679 33.010
3 1.254 10.450 52.018 1.254 10.450 52.018 1.854 15.449 48.459
4 1.070 8.915 60.933 1.070 8.915 60.933 1.497 12.474 60.933

Extraction Method: Principal Component Analysis


The results of the factor analysis presented in the table – 4.4.1.2 regarding
factors related to influence in choosing CRM products, have revealed that there are
nineteen factors that had Eigen value exceeding “one”. Among those four factors,
the first factor accounted for 28.486 percent of the variance, the second 13.081
percent, the third factor 10.45 percent, and last factor 8.915 per cent of the
variance in the data set. The first four factors are the final factors solution and they
all together represent 60.93 percent of the total variance in the scale items
measuring the factors related to influence in choosing CRM products. Hence from
the above results, it is certain these are the factors that are related to level of
satisfaction about various products3 .
TABLE ---4.4.1.3
226

COMMUNALITIES FOR FACTORS RELATED TO INFLUENCE IN CHOOSING CRM


PRODUCTS
Sl.No. Items Initial Extraction(h2)
X1 Price of the product 1.000 .729
X2 Quality of the product 1.000 .537
X3 Nature of Cause 1.000 .460
X4 Availability of the product 1.000 .630
X5 Brand familiarity 1.000 .647
X6 CRM Campaign 1.000 .607
X7 Previous experience 1.000 .712
X8 Package of the product 1.000 .521
X9 Contribution level to cause 1.000 .619
X10 Attractive sales offers 1.000 .534
X11 Sales promotional activities 1.000 .678
X12 Cause- Corporate- Brand fit 1.000 .637

The above table (Communalities) represents the application of the Factor


Extraction Process, it was performed by Principal Component Analysis to identify
the number of factors to be extracted from the data and by specifying the most
commonly used Varimax rotation method 4. In the principal component analysis,
total variance in the data is considered. The proportion of the variance is explained
by the fourteen factors in each variable. The proportion of variance is explained by
the common factors called communalities of the variance. Principal Component
Analysis works on initial assumption that all the variance is common. Therefore,
before extraction the communalities are all 1.000. Then the most common
approach for determining the number of factors to retain 5 i.e. examining Eigen
values was done.
TABLE- 4.4.1.4
ROTATED COMPONENT MATRIX FOR FACTORS RELATED TO INFLUENCE IN
CHOOSING CRM PRODUCTS
Component
Variable code
I II III IV
X6 .713 .191 .241 .052
X5 .648 .256 -.268 .299
X12 .647 -.259 .372 .114
X9 .639 .169 .416 -.101
X1 -.016 .741 .367 -.212
X2 -.029 .643 .109 .333
X3 .281 .613 -.031 .074
227

X4 .426 .532 -.150 .378


X11 .172 .017 .793 .136
X10 .120 .141 .703 .079
X7 .068 .015 .031 .840
X8 .115 .204 .346 .589

Extraction Method: Principal Component Analysis.


The above table represents the Rotated Component Matrix, which is an
important output of principal component analysis. The coefficients are the factor
loadings which represent the correlation between the factors and the twelve
variables (X1 to X12). From the above factor matrix it is found that coefficients for
factor-I have high absolute correlations with variable X6 (CRM Campaign), X5
(Brand familiarity), X12 (Cause- Corporate- Brand fit ) and X9 (Contribution level to
cause) that is, .713, .648, .647 and.639 respectively. Similarly factor-II has high
absolute correlation with variable X1 (Price of the product), X2 (Quality of the
product), X3 (Nature of Cause), and X4 (Availability of the product) that is, .741,
.643, .613 and.532 respectively. Next, factor III has high absolute correlation with
variable X11 (Sales promotional activities) and X10 (Attractive sales offers) that is,
.793 and .703 respectively. Factor-IV has high absolute correlation with variable X7
(Previous experience), and X8 (Package of the product ) that is, .840 and .589
respectively. For example in this study, factor I is at least somewhat correlated
with twelve variable out of the twelve variables with absolute value of factor loading
greater than or equal to 0.5. In such a complex matrix it is difficult to interpret the
factor. So proceed to compute the rotated factor matrix.
TABLE 4.4.1.5
Component Transformation Matrix
Component 1 2 3 4
1 .637 .501 .454 .371
2 -.109 .530 -.747 .386
3 -.530 .664 .340 -.403
4 -.549 -.165 .346 .742

The above table reveals the factor correlation matrix. If the factors are
uncorrelated among themselves, then in the factor correlation matrix, the diagonal
elements will be 1’s and off diagonal elements will be 0’s. Since matrix was rotated
with Varimax, barring some variables all other variables are found to have, even i f
not zero correlations but fairly, low correlation.

CHART
Component Plot228
in Rotated Space

1.0
Q3.1.1
Q3.1.3

Component 2
0.5 Q3.1.2
Q3.1.4 Q3.1.5
Q3.1.8
Q3.1.10
0.0 Q3.1.6
Q3.1.7 Q3.1.9
Q3.1.11

-0.5 Q3.1.12

-1.0
-1.0
-1.0 -0.5 -0.5
0.0
0.0 0.5
0.5 1.0 1.0 t3
Componen p onen
t1 Com

CONCLUSION
Thus the twelve variables in the data were reduced to four Component factors
and each factor may be identified with the corresponding variables as follows:
TABLE -4.4.1.6
SHOWING THE FACTORS IDENTIFIED THE FACTORS RELATED TO
INFLUENCE IN FOR CHOOSING CRM PRODUCTS
X6 CRM Campaign 50.84 %
X5 Brand familiarity 41.99%
Factor I
X12 Cause- Corporate- Brand fit 41.86%
X9 Contribution level to cause 40.83%
X1 Price of the product 54.91%
X2 Quality of the product 41.34%
Factor II
X3 Nature of Cause 37.58%
X4 Availability of the product 28.30%
X11 Sales promotional activities 62.88%
Factor III
X10 Attractive sales offers 49.42%
X7 Previous experience 70.56%
Factor IV
X8 Package of the product 34.69%
17.4 REVISION POINTS
229

1. Types
2. Multiple Regression Analysis
3. Factor analysis
4. Path analysis
5. MANOVA
17.5 INTEXT QUESTIONS
1. List out the advantages and disadvantages of Factor analysis
2. Explain the limitations of Path Analysis
17.6 SUMMARY
After going through this Lesson you should be able to explain the types of
multivariate analysis, multiple regression, factor analysis, Path analysis and
MANOVA
17.7 TERMINAL EXERSICE
1. Explain the importance of factor analysis in research
2. Explain the importance of Multiple Regression analysis in research
17.8 SUPPLEMENTARY MATERIALS
1. Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, rev. ed.,
New York: Harcourt, Brace & World, Inc., 1970.
17.9 ASSIGNMENT
1. Briefly Explai how to conduct and interpret Factor analysis?
17.10 SUUGESTED READINGS/REFERENCE BOOKS
1. Ackoff, Russell L., The Design of Social Research, Chicago: University of
Chicago Press, 1961.
2. Ackoff, Russell L., Scientific Method, New York: John Wiley & Sons, 1962.
3. Allen, T. Harrell, New Methods in Social Science Research, New York:
Praeger Publishers, 1978.
4. Anderson, H.H., and Anderson, G.L., An Introduction to Projective
Techniques and Other Devices for Understanding the Dynamics of Human
Behaviour, New York: Prentice Hall, 1951.
5. Anderson, T.W., An Introduction to Multivariate Analysis, New York: John
Wiley & Sons, 1958.
6. Bailey, Kenneth D., “Methods of Social Research,” New York, 1978.
7. Baker, R.P., and Howell, A.C., The Preparation of Re ports, New York: Ronald
Press, 1938.
8. Bartee, T.C., “Digital Computer Fundamentals,” 5th Ed., McGraw-Hill,
International Book Co., 1981.
9. Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, rev. ed.,
New York: Harcourt, Brace & World, Inc., 1970.
230

10. Bell, J.E., Projective Techniques: A. Dynamic Approach to the Study of


Personality, New York: Longmans Green, 1948.
11. Bellenger, Danny N., and Greenberg, Barnett A., Marketing Re search—A
Management Information Approach, Homewood, Illinois: Richard D. Irwin, Inc.,
1978.
17.11 LEARNING ACTIVITY
Briefly Explai how to conduct and interpret Path analysis?
17.12 KEYWORD
Types of multivariate analysis, multiple regression, factor analysis, Path
analysis and MANOVA


231

LESSON – 18

KRUSKAL-WALLIS TEST
18.1 INTRODUCTION
Respondent’s opinion towards factors influencing for choosing Cause related
marketing products. In the study area out of seven hundred and fifty respondents
were divided into two groups
18.2 OBJECTIVE
After going through this Lesson you should be able to explain the types of
Discriminant function analysis Canonical Discriminant function, Relative
Discriminant function
18.3 CONTENT
1. Discriminant function analysis
2. Canonical Discriminant function,
3. Relative Discriminant function
DISCRIMINANT FUNCTION ANALYSIS
Low level of factors influencing for choosing Cause related marketing products
and the high level of factors influencing for choosing Cause related marketing
products. The difference of opinion of the respondents in one group from the other
is studied with the help of discriminant function analysis. For the purpose of the
study, the following variables were selected.
1. Gender
2. Age
3. Marital status
4. Educational Qualification
5. Occupational status
6. Monthly Income
7. Family size
8. Nature of the family
9. Number of earning members in the family
10. Name of the Zone
The discriminant function analysis attempts to construct a function with these
and other variables so that the respondents belonging to these two groups are
differentiated at the maximum. The linear combination of variables is known as
discriminant function and its parameters are called discriminant function
coefficients. In constructing this discriminant function all the variables which
contribute to differentiate these three groups are examined.
Mahalanobis minimum D2 method is based on the generalized squared
Euclidean distance that adjusts for unequal variances in the variables. The major
232

advantage of this procedure is that it is computed in the original space of the


predictor (independent) variables rather than as a collapsed version which is used
in the other method.
Generally, all the variables selected will not contribute to explain the
maximum discriminatory power of the function. So a selection rule is applied based
on certain criteria to include those variables which best discriminate. Stepwise
selection method was applied in constructing discriminant function which selects
one variable at a time to include in the function. Before entering into the function
the variables are examined for inclusion in the function.
The variables which could have maximum D 2 value, if entered into the
function is selected for inclusion in the function. Once entered any variable already
in the equation is again considered for removal based on certain removal criteria.
Likewise, at each step the next best discriminating variable is selected and included
in the function and any variable already included in the function is considered for
removal based on the selection and removal criteria respectively.
Discriminant function analysis involved classification problem also to
ascertain the efficiency of the discriminant function analysis and all the variables
which satisfy the entry and removal criteria were entered into the function.
Normally the criterion used to select the variables for inclusion in the function is
minimum F to enter into the equation (i.e.) F statistic calculated for the qualified
variable to enter into the function is fixed as ≥ 1. Similarly any variable entered in
the equation will be removed from the function if F statistic for the variable
calculated is < 1. The two groups are defined as
Group 1 - Low level
Group 2 - High level
The mean and standard deviation for these groups and for the entire samples
are given for each variable considered in the analysis.
TABLE - 4.4.3.1
GROUP MEANS (BETWEEN LOW AND HIGH GROUPS)
Sl. Low High Total
Factor
No. Mean SD Mean SD Mean SD
1 Gender 1.52 0.50 1.82 0.38 1.67 0.47
2 Age 2.87 1.29 2.95 1.33 2.91 1.31
3 Marital status 1.32 0.47 1.22 0.42 1.27 0.45
4 Educational Qualification 3.46 0.73 3.68 0.54 3.57 0.65
5 Occupational status 3.60 0.81 3.64 0.66 3.62 0.74
6 Monthly Income 2.38 1.45 2.36 1.36 2.37 1.40
7 Family size 2.22 1.04 1.93 0.91 2.08 0.99
8 Nature of the family 1.55 0.50 1.70 0.46 1.63 0.48
9 Number of earning members
2.11 0.64 2.10 0.64 2.10 0.64
in the family
10 Zone 3.02 1.26 2.98 1.56 3.00 1.42
233

The overall stepwise D.F.A results after all significant discriminators have been
included in the estimation of discriminated function are given in the following table .
TABLE -4.4.3.2
SUMMARY TABLE BETWEEN LOW LEVEL AND HIGH LEVEL GROUPS
Step Variables entered Wilk’s Lamda F-value Significance
1 Gender 0.897 85.729 .000**
2 Marital status 0.988 8.989 .003**
3 Educational Qualification 0.970 22.898 .000**
4 Family size 0.978 17.014 .000**
5 Nature of the family 0.976 18.740 .000**
*Significant at 1% level
The summary table indicates that variable gender entered in step one. The
variables such as Gender, Marital status, Educ ational Qualification, Family size
and Nature of the family are significant at one per cent significance level. All the
variables are significant discriminators based on their Wilk’s Lambda and F-value.
The multivariate aspect of this model is given in the following table.
TABLE -4.4.3.3
CANONICAL DISCRIMINANT FUNCTION
(BETWEEN LOW AND HIGH GROUPS)
Canonical S/NS
Wilk's Lamda Chi -square p-value
correlation
0.420 .823 144.438 .000** S

The canonical correlation in the discriminant group can be accounted for by


this model, Wilk's Lamda and Chi-square value suggest that D.F is significant at
one percent level.
The variables given above are identified finally by the D.F.A as the eligible
discriminating variables. Based on the selected variables the correspo nding D.F
coefficients are calculated. They are given in the following table.
TABLE -4.4.3.4
DISCRIMINANT FUNCTION COEFFICIENT
(BETWEEN LOW LEVEL AND HIGH LEVEL)
Gender 1.685
Age .064
Marital Status -.778
Educational Level .506
Occupational status .343
Monthly Income .033
Family size -.195
234

Nature of the family 1.011


Number of earning
.408
members in the family
Zone -.059
(Constant) -7.062
Using this D.F coefficients and variables discriminating scores for 2 groups
are found out and are called group centroids or group means
For low level user (Z 1) = -.459
For High level user (Z2) = .466
Discriminating factor is the weighted average of Z 1, Z2
(378x Z1 ) + (372 xZ2 )
(.i.e.) Z =
378+372
It is represented diagrammatically
Z1 Z Z2

-0.459 0 +0.466

Low level High level


Thus to classify any respondent as to low or high user the Z score for the
respondent is found out by using the equation. If the score found out for any
respondent is Z0 and if the value is > Z (i.e. Z0> Z) then it is classified into high user
and if Z0 <Z then (i.e. Z0<Z) it is classified into low user.
Now the questions remain to be answered are :
1. How efficient are the discriminating variables in the D.F.A?
2. How efficient the D.F itself is?
The first equation cannot be answered directly however the discriminating
power or the contribution of each variable to the function can sufficiently answer
the question. For this consider the following table .
TABLE - 4.4.3.5
RELATIVE DISCRIMINATING INDEX
(BETWEEN LOW LEVEL GROUP AND HIGH LEVEL GROUP)
Group
Group I II Unstandardised Ij=ABS (Kj) Rj=Ij/sum
Mean X1 Mean coefficient Mean (Xj0-Xji) Ijj*100
X2
Gender 1.52 1.82 1.685 0.51 61.00
Marital Status 2.87 2.95 -.778 0.06 7.18
235

Educational
3.46 3.68 .506 0.11 13.16
Level
Family size 2.22 1.93 -.195 0.006 0.72
Nature of the
1.55 1.70 1.011 0.15 17.94
family
TOTAL 100

The Kruskal–Wallis one-way analysis of variance by ranks is a non


parametric method for testing whether samples originate from the same
distribution. It is used for comparing two or more samples that are independent,
and that may have different sample sizes, and extends the Mann–Whitney Utest to
more than two groups. The parametric equivalent of the Kruskal -Wallis test is
the one-way analysis of variance (ANOVA). When rejecting the null hypothesis of
the Kruskal-Wallis test, then at least one of sample stochastically dominates at
least one other sample. The test does not identify where this stochastic dominance
occurs or for how many pairs of groups stochastic dominance obtains.It would help
to analyze the specific sample pairs for stochastic dominance. Since it is a non -
parametric method, the Kruskal–Wallis test does not assume a normal
distribution of the residuals, unlike the analogous one -way analysis of variance. If
the researcher can make the more stringent assumptions of an identically shaped
and scaled distribution for all groups, except for any difference in medians, then
the null hypothesis is that the medians of all groups are equal, and the alternative
hypothesis is that at least one population median of one group is different than the
population median of at least one other group. This test is used in this study to
analyze the preference of khadi consumers towards Khadi textile products based on
their place of residence (urban, semi -urban and rural) by assinging ranks to their
preferences.
This test is appropriate for use under the following circumstances: (a) when
there is three or more conditions that researcher want to compare; (b) each
condition is performed by a different group of participants; i.e. researcher have an
independent-measures design with three or more conditions. (c) the data do not
meet the requirements for a parametric test. (i.e. use it if the data are not normally
distributed; if the variances for the different conditions are markedly different; or if
the data are measurements on an ordinal scale). If the data meet the requirements
for a parametric test, it is better to use one -way independent-measures Analysis of
Variance (ANOVA) because it is more powerful than the Kruskal -Wallis test.
F-test, used in analysis of variance, was that each of the k populations should
be normal with equal variance. In contrast to this, the Kruskal -Wallis test only
assumes that the k populations are continuous and have the same pattern
(symmetrical or skewed) of distribution. The null and the alternative hypotheses
ofthe Kruskal-Wallis test are:
236

H0: m1= m2= … = mk(i.e., means of the k populations are equal)


Ha: Not all mi’s are equal.
The Test Statistic: The computation of the test statisticfollows a procedure that
is very similar to the Mann-Whitney Wilcoxon test.
(i) Rank all the n1+ n2+ … + nk= n observations, arrayed in ascending order.
(ii) Find R1, R2, … Rk, where Riis the sum of ranks of the ith sample. The test
statistic, denoted by H, is given by

 2 R 2 2
12  R1 2
R
k 
H    ...  - 3 (n  1)
n(n  1)  n n n 
1 2 k
 
It can be shown that the distribution of H is c2with k – 1 d.f., when size of
each sample is at least 5. Thus, if H > 21 k− χ , H0 is rejected.
Example 1: To compare the effectiveness of three types of weight-reducing
diets, a homogeneous groups of 22 women was divided into three sub-groups and
each sub-group followed one of these diet plans for a period of two months. The
weight reductions, in kgs, were noted as given below:

I 4.3 3.2 2.7 6.2 5.0 3.9


Diet Plan II 5.3 7.4 8.3 5.5 6.7 7.2 8.5
II 1.4 2.4 2.7 3.1 1.5 0.7 4.3 3.5 0.3
Use the Kruskal-Wallis test to test the hypothesis that the effectiveness of the
three weight reducing diet plans is same at 5% level of significance.
Solution
It is given that n1= 6, n2= 7 and n3= 9.
The total number of observations is 6 + 7 + 9 = 22. These are ranked in their
ascending order as given below:

I 12.5 9 6.5 17 14 11 70
Diet II 15 20 21 16 18 19 22 131
Pla II 3 5 6.5 8 4 2 12.5 10 1 52

The tabulated value of χ2at 2 d.f. and 5% level of significance is 5.99. Since H
is greater than this value, H0is rejected at 5% level of significance.
e.g.
If the data meet the requirements for a parametric test, it is better to use one -
way independent-measures Analysis of Variance (ANOVA) because it is more
powerful than the Kruskal-Wallis test.
237

Area wise Mean Ranks of Khadi Textile Products


Mean
Place of Residence N
Rank
Urban 192 209

textile Semi-urban 139 208.63


products Rural 89 211.95
Total 420
Kruskal-Wallis Test For Mean Rank Variation of The Responses of
Khadi Consumers With Respect to Khadi Textile Products
Textile Products
Chi-Square 0.481
Df 2

Asymp. Sig. 0.786

The above table indicates that the Khadi textile products are frequently
purchased by the consumers from rural area (mean rank 211.95) compared to
consumers from urban (mean rank 209) and semi -urban areas (mean rank 208.63).
This mean variation is statistically insignificant at 5 per cent level of significance
(P>.05 in above cases)
18.4 REVISION POINTS
1. Sign test
2. One sample sign test
3. Two sample sign test
18.5 INTEXT QUESTIONS
1. List out the advantages and disadvantages of non parametric test
2. Explain the Importance of sign tests
18.6 SUMMARY
After going through this Lesson you should be able to Explain Sign test, one
sample sign test and two sample sign test
18.7 TERMINAL EXERSICE
It turns out that 5 of the participants like the new drink better, and the rest
prefer the old one. At .05 significance level, can we reject the notion that the two
drinks are equally popular?
18.8 SUPPLEMENTARY MATERIALS
Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, rev. ed., New
York: Harcourt, Brace & World, Inc., 1970.
18.9 ASSIGNMENT
The sign test is considered a weaker test, because it tests the pair value below
or above the median and it does not measure the pair difference - Comment
238

18.10 SUGGESTED READINGS/REFERENCE BOOKS


1. Ackoff, Russell L., The Design of Social Research, Chicago: University of
Chicago Press, 1961.
2. Ackoff, Russell L., Scientific Method, New York: John Wiley & Sons, 1962.
3. Allen, T. Harrell, New Methods in Social Science Research, New York:
Praeger Publishers, 1978.
4. Anderson, H.H., and Anderson, G.L., An Introduction to Projective
Techniques and Other Devices for Understanding the Dynamics of Human
Behaviour, New York: Prentice Hall, 1951.
5. Anderson, T.W., An Introduction to Multivariate Analysis, New York: John
Wiley & Sons, 1958.
6. Bailey, Kenneth D., “Methods of Social Research,” New York, 1978.
7. Baker, R.P., and Howell, A.C., The Preparation of Reports, New York: Ronald
Press, 1938.
8. Bartee, T.C., “Digital Computer Fundamentals,” 5th Ed., McGraw-Hill,
International Book Co., 1981.
9. Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, rev. ed.,
New York: Harcourt, Brace & World, Inc., 1970.
10. Bell, J.E., Projective Techniques: A. Dynamic Approach to the Study of
Personality, New York: Longmans Green, 1948.
11. Bellenger, Danny N., and Greenberg, Barnett A., Marketing Re search—A
Management Information Approach, Homewood, Illinois: Richard D. Irwin,
Inc., 1978.
18.11 LEARNING ACTIVITIES
The disadvantage is that nonparametric procedures throw away information !
The sign test, for example, uses only the signs of the observations - Comment
18.12 KEYWORD
Sign test, one sample sign test and two sample sign test


239

LESSON – 19

STRUCTURAL EQUATION MODELING


19.1 INTRODUCTION
Structural equation modeling, or SEM, is a very general, chiefly linear, chiefly
cross-sectional statistical modeling technique. Factor analysis, path analysis and
regression all represent special cases of SEM.
SEM is a largely confirmatory, rather than exploratory, technique. That is, a
researcher are more likely to use SEM to determine whether a certain model is
valid., rather than using SEM to "find" a suitable model --although SEM analyses
often involve a certain exploratory element.
19.2 OBJECTIVE
The main aim of this Lesson is to study the SEM. After going through this
Lesson you should be able to understand Structural Equation Modelling, form of
Structural Equation Modelling, latent construct, Structural model, structural error,
Manifest Variables, CFA and PLS
19.3 CONTENT
19.3.1 Structural Equation Modelling
19.3.2 Form of Structural Equation Modelling
19.3.3 Latent construct
19.3.4 Structural model
19.3.5 Structural error
19.3.6 Manifest Variables
19.3.7 CFA
19.3.8 PLS
19.3.1 Structural Equation Modelling
In SEM, interest usually focuses on latent constructs--abstract psychological
variables like "intelligence" or "attitude toward the brand"--rather than on the
manifest variables used to measure these constructs. Measurement is recognized as
difficult and error-prone. By explicitly modeling measurement error, SEM users
seek to derive unbiased estimates for the relations between latent constructs. To
this end, SEM allows multiple measures to be associated with a single latent
construct.
A structural equation model implies a structure of the covariance matrix of the
measures (hence an alternative name for this field, "analysis of covariance
structures"). Once the model's parameters have been estimated, the resultin g
model-implied covariance matrix can then be compared to an empirical or data-
based covariance matrix. If the two matrices are consistent with one another, then
the structural equation model can be considered a plausible explanation for
relations between the measures.
240

Compared to regression and factor analysis, SEM is a relatively young field,


having its roots in papers that appeared only in the late 1960s. As such, the
methodology is still developing, and even fundamental concepts are subject to
challenge and revision. This rapid change is a source of excitement for some
researchers and a source of frustration for others.
19.3.2 The Form of Structural Equation Models
Structural equation modeling incorporates several different approaches or
frameworks to representing these models. In one well -known framework
(popularized by Karl Jöreskog, University of Uppsala), the general structural
equation model can be represented by three matrix equations:

 m  1   m  m  m  1   m  n  n 1   m 1


y p 1  y p  m m  1    p 1
X q 1  xq  n  n 1   q 1

However, in applied work, structural equation models are most often


represented graphically. Here is a graphical example of a structural equation
model:
Graph to be write in this page
1  2 3

Y1 Y2 Y3
1 X1
1
2 X2 1 11
1 31 3
Y6 6
3 X3 21 2
Y7 7
4 X4 2 2  32
 22
5 X5 2
Y4 Y5

 4 5

For more information, click on an element of this diagram, or choose from this
list: Latent Constructs Structural Model Structural Error Manifest Variables
Measurement Model Measurement Error
This diagram uses the dominant symbolic language in the SEM world.
However, there are alternate forms, including the "RAM," reticular action model.
Latent Constructs
In structural equation modeling, the key variables of interest are usually
"latent constructs"--abstract psychological concepts such as "intelligence" or
241

"attitude." We can observe the behavior of latent variables only indirectly, and
imperfectly, through their effects on manifest variables.
A structural equation model may include two types of latent
constructs--exogenous and endogenous. In the most
traditional system, exogenous constructs are indicated by
the Greek character "ksi" (at left) and endogenous constructs are
indicated by the Greek character "eta" (at right). These two types of
constructs are distinguished on the basis of whether or not they are dependent
variables in any equation in the system of equations represented by the model.
Exogenous constructs are independent variables in all equations in which they
appear, while endogenous constructs are dependent variables in at least one
equation--although they may be independent variables in other equations in the
system. In graphical terms, each endogenous construct is the target of at least one
one-headed arrow, while exogenous constructs are only targeted by two-headed
arrows.
Structural Model
In SEM, the structural model includes the relationships among the latent
constructs. These relationships are chiefly linear, although flexible extensions to
the basic SEM system allow for the inclusion of nonlinear relations, as well. In the
diagram, one-headed arrows represent regression relationships, while two-headed
arrows represent correlational relations--that is, shared variation that is not
explained within the model.
Parameters representing regression relations between latent
constructs are typically labeled with the Greek character "gamma"
(at left) for the regression of an endogenous construct on an exogenous
construct, or with the Greek character "beta" (at right) for the regression of
one endogenous construct on another endogenous construct. Typically in
SEM, exogenous constructs are allowed to covary freely. Parameters
labeled with the Greek character "phi" (at left) represent these covariances.
This covariance comes from common predictors of the exogenous constructs which
lie outside the model under consideration.
Structural Error
Few SEM researchers expect to perfectly predict their dependent
constructs, so model typically include a structural error term, labeled with
the Greek character "zeta" (at right). To achieve consistent parameter estimation,
these error terms are assumed to be uncorrelated with the model's exogenous
constructs. (Violations of this assumption come about as a result of the excluded
predictor problem.) However, structural error terms may be modeled as being
correlated with other structural error terms. Such a specification indicates that the
endogenous constructs associated with those error terms share common variation
that is not explained by predictor relations in the model.
242

Manifest Variables
SEM researchers use manifest variables--that is, actual
measures and scores--to ground their latent construct models with
real data. Manifest variables associated with exogenous constructs
are labeled X, while those associated with endogenous constructs are labeled Y.
Otherwise, there is no fundamental distinction between these measures, and a
measure that is labeled X in one model may be labeled Y in another.
Measurement Model
In SEM, each latent construct is usually
associated with multiple measures. SEM
researchers most commonly link the latent
constructs to their measures through a factor
analytic measurement model. That is, each
latent construct is modeled as a common
factor underlying the associated measures. These "loadings" linking constructs to
measures are labeled with the Greek character "lambda" (at left). Structural
equation models can include two separate "lambda matrices, one on the X side and
one on the Y side. In SEM applications, the most common measurement model is
the congeneric measurement model, where each measure is associated with only
one latent construct, and all covariation between measures is a consequence of the
relations between measures and constructs.
(Sometimes, however, it makes more sense to model a latent construct as the
result or consequence of its measures. This is the causal indicators model. This
alternative measurement model is also central to Partial Least Squares, a
methodology related to SEM.)
Measurement Error
SEM users typically recognize that their measures are
imperfect, and they attempt to model this imperfection. Thus,
structural equation models include terms representing measurement error. In the
context of the factor analytic measurement model, these measurement error terms
are uniquenesses or unique factors associated wi th each measure. Measurement
error terms associated with X measures are labeled with the Greek character "delta"
(at left) while terms associated with Y measures are labeled with "epsilon" (at right).
Conceptually, almost every measure has an associated error term. In other words,
almost every measure is acknowledged to include some error.
However, when a construct is associated with only a single measure, it is
usually impossible (due to the limits of identification) to estimate the amount of
measurement error within the model. In such cases, the researcher must prespecify
the amount of measurement error before attempting to estimate model parameters.
In this situation, researchers may be tempted to simply assume that there is no
measurement error. However, if this assumption is false, then model parameter
estimates will be biased.
243

Methodological Alternatives to SEM/CFA


While structural equation modeling (SEM), based on a confirmatory factor
analysis (CFA) measurement model, is a very general methodology, it is not the
solution to every problem. Here are some methodologies that are "related" to SEM
and CFA, but which are appropriate to rather different research problems.

Alternatives to SEM Alternatives to CFA

Partial Least Squares (PLS) Exploratory Factor Analysis

Tetrad Analysis Unrestricted Factor Analysis

Latent Class Analysis Item Response Theory (IRT)

Principal Components Analysis (PCA)

Formative (or "Causal") Indicators


Partial least squares
Partial least squares (PLS) was invented by Herman Wold (mentor to Karl
Jöreskog, a founder of SEM), as an analytical alternative for situations where
theory is weak and where the available manifest variables or measures would likely
not conform to a rigorously specified measurement model. For these reasons, Wold
labeled his approach, "soft modeling." In recent years, active PLS researchers,
including Wold's son, Svante, and Fred Bookstein, have developed and even
redefined PLS, so there is room for some confusion. The material presented here
adheres most closely to Wold's original formulation, but it is generally applicable to
the other forms, as well.
Rather than using the factor analytic measurement model associated with
SEM, PLS more often uses a principal components measurement model, where the
"latent constructs" are defined as linear composites of the measures associated with
them. In restricted cases, these linear composites are equivalent to principal
components.
Also, rather than seek overall optimization in parameter estimates through a
full information estimation technique (such as maximum likelihood), Wold opted for
limited information methods that provide statistically inferior estimates but which
make minimal demands on the data. Thus, PLS may represent a pragmatic
alternative to SEM. In addition, the PLS method is designed to maximize prediction
rather than fit. That is, the method is optimized to maximize the proportion of
variance of the dependent "construct" that is explained by the predictor
"constructs." SEM, by contrast, is designed to maximize and then test the degree of
consistency between model and data.
While many researchers are devoted to PLS as a practical method, others find
it objectionable. Some argue that the statistical properties of the method are not
well understood, and that PLS users rely too much on mere assertions that are not
244

supported by rigorous analysis. Other researchers argue that the "latent


constructs" in PLS are not really "latent" at all, since they are strict linear
composites of observed variables.
For more information about PLS, see Wynne Chin's PLS Web site or Wynne's
overview chapter in Marcoulides (1998). Falk and Miller's (1992) short text provides
a readable introduction to the practice and phil osophy of PLS. McDonald's (1996)
paper is a fascinating look at exactly how PLS and SEM differ, and showing how
standard SEM programs can be used to mimic PLS estimation methods.
You can download PLS-PC, perhaps the most widely known PLS program, from
Jack McArdle's web site.
Tetrad analysis
(This section is essentially taken from Bollen & Ting 1993, Ting 1995, and
Spirtes, Glymour and Scheines 1993.)
While structural equation modeling works generally at the level of variances
and covariances, tetrad analysis works with relations between sets of four
covariances at once. A tetrad (from the Greek word for "four") for four variables g, h,
i, j is defined as:
where indicates, of course, the covariance between the subscripted variables.
A tetrad that is equal to 0 is called a "vanishing tetrad."
Tetrad analysis actually encompasses two very different methodologie s.
Confirmatory tetrad analysis (Bollen & Ting 1993) evaluates a proposed structural
equation model by determining whether the vanishing tetrads implied by the model
are really present, within sampling error. Spirtes, Glymour and Scheines' (1993)
TETRAD II package explores a covariance matrix, looking for models that are
consistent with it. Thus, their work presents an exception to the generally
confirmatory character of SEM.
Confirmatory tetrad analysis may be useful to researchers who are facing
identification problems. A structural equation model that is not identified in terms
of its parameters may still be identified in terms of vanishing tetrads. So a
researcher may be able to derive fit information for the model based on tetrad
analysis. In addition, two models that are not nested in terms of their parameters
may be nested in terms of vanishing tetrads. That is, one model may imply a set of
vanishing tetrads that is a subset of those implied by the other model, facilitating a
comparative analysis of the two models. However, performing this tetrad analysis
by hand is tricky. Many of the tetrads implied by a model will be redundant with
other tetrads, but an analysis must be based on a unique set. It can be difficult to
determine which tetrads are unique. Ting (1995) describes a SAS macro, called
CTA-SAS, which performs this analysis. The program and documentation are
available via ftp.
The exploratory tools developed by Spirtes, et al. may be useful to modelers
facing a covariance matrix with little prior information. Indeed, some related work
245

by Glymour and others has appeared in the literature on "data mining," a field
which deals with the increasingly common situation of an analyst being
overwhelmed by the quantity of available data.
However, these tools may also be useful in a confirmatory context. Researchers
may wish to evaluate fit information for one model in light of fit information for
other models. If many models fit a data set well, then the good fit of one model
would seem to offer less compelling evidence in favor of the theory that inspired the
model. This procedure might also be sued to find equivalent models empirically.
Spirtes et al. have suggested that augmenting "traditional" SEM analysis with
their exploratory approach is more likely to produce the "true" model for a data set
(if one exists) than is SEM analysis by itself.
In their work, Spirtes et al. use a distinctive, graphically oriented language. A
link between two constructs or variables is called an "edge." If the link represents a
covariance between the two, it is called an "undirected edge"; otherwise, it is called
a "directed edge." Depending on the algorithm used, TETRAD II may begin with a
model where all variables are correlated (an "undirected g raph") and then proceed
to eliminate edges where possible and turn undirected edges into directed edges.
The result is a "partially direct acyclic graph" which fits the data set to a tolerance
specified by the user. The word "acyclic" indicates that the re sulting model will not
involve reciprocal or feedback relations--a limitation of the method.
Tetrad analysis
Latent class analysis (LCA) is appropriate when the dependent variable is
intrinsically categorical--that is, when it boils down to distinguishing between those
who take an action and those who do not, or between those who vote for different
political candidates. In SEM, ordinal measures may be used to reflect continuous
latent variables, but in LCA both the measure and the underlying latent variable
are categorical. For more information, see Keith Markus' LCA Web site, or Karen
Schmidt McCollam's overview chapter in Marcoulides (1998).
Tetrad analysis
Unrestricted factor analysis
(This section consists of edited SEMNET comments from Stan Mulaik.)
Researchers can use unrestricted factor analysis to test only the hypothe sis
that relations between a set of measures can be explained by a certain number of
common factors, without specifying a particular factor structure. This allows the
researcher to determine whether any p-factor model will provide acceptable fit. If
this hypothesis fails, then there is no point in testing more constrained CFA
models.
The unrestricted model was described by Karl Jöreskog in an addendum to a
reprint of the first paper on confirmatory factor analysis that appeared in
Psychometrika in 1972. This addendum is to be found in:
246

Jöreskog, K. G. & Sörbom, D. (1979), In Jay Magidson (Ed.), Advances in


Factor Analysis and Structural Equation Models. Cambridge, MA: Abt Books.
On pp.40-43, Jöreskog describes how to specify an unrestricted model for k
factors:
 Fix the k variances of the latent common factors to unity, and free the
covariances between the common factors.
 For each factor, pick a manifest variable that one expects to have the
highest or near highest loading on that
 factor. Free its loading in the factor loading matrix. Fix the k - 1 remaining
loadings of that variable on the remaining factors to 0. You will now have k
rows of Lambda with k - 1 zeros in each of them
 Free all remaining loadings.
For a model with two factors and four measures per factor, the factor
covariance and loading matrices should look like this (where 0 indicates an element
fixed to 0 and ? indicates a parameter to be estimated):
? 0
? ?
? ?
? ?
0 ?
? ?
? ? 1
? ? ? 1
Lambda Phi
Experience has shown that a problem you may have with convergence will be
cleared up with good starting values. Use an exploratory factor analysis as the
basis for picking what variables to choose to specify the above model and use the
corresponding parameter values as starting values for iterations. The result will fit
as well as a two-factor exploratory factor analysis estimated in the same way.
An alternative method for specifying the above model is to free all elements of
Phi while fixing the supposed highest loading item on a factor to unity and its
loading on the other factors to zero. This is often the way to proceed in doing
multiple group analyses where you want to set the metric without explicitly fixing
the variances of the common factors so they can differ across groups while the
Lambda loadings are constrained to be equal across groups.
If you decide that k is the number of factors, then you can impose additional
restrictions on the loadings (e.g. setting some to zero) to specify a measurement
model. The measurement model may not fit, and you might then use modification
247

indicies to free up zero loadings, one at a time, to get an optimal measurement


model (at the expense of lost degrees of freedom).
Once you have a good measurement model, you can then test your structural
model involving relations between latent variables.
As Roger Millsap later pointed out, in a message to SEMNET:
In the addendum, Jöreskog attempted to lay out sufficient conditions for
"identification" in factor analysis through fixing various loadings to zero or nonzero
values. Joöreskog himself noted in a later paper that these conditions were not
sufficient in general to identify the model, and were only sufficient to achieve
rotational uniqueness. This is all contained in the paper:
Bollen & Jöreskog (1985), Uniqueness does not imply identification. Sociological
Methods and Research, 14, 155-163.
The distinction is also mentioned in Bollen's (1989) text. I was also not fully
aware of the distinction until a few years ago when I was writing a paper on
identification in MTMM factor models, and found the above paper. The basic point
is that using Jöreskog's restrictions, one can show that they are sufficient to define
a unique rotational position for the factors. They are not sufficient in general to
define a unique solution for ALL of the factor model parameters however, which
include the unique variances as well as the loadings and common factor
covariances.
Stan Mulaik then asked whether a UFA solution would be equivalent to a EFA
solution for the same data and same number of factors. Catharina Hartmann
provided an answer, originally due to Karl Jöreskog, and referencing Lawley and
Maxwell:
Jöreskog indicated that, even if two solutions are equivalent [i.e. the EFA
solution vs the CFA unrestricted model so- lution], the chi-square from Lisrel will
be slightly higher than the one in EFA because of different multiplicative factors in
front of the ML fit function. For example, Lawley and Maxwell (1971), pp. 35-36,
provided the following multiplicative factor:
n-(2p+5)/6, where p=number of variables and n=N-1, N=number of
observations
Roger Millsap replied
It turns out that Lawley and Maxwell had advocated a slightly different
multiplier:
n-(2p+4r+5)/6 where r=number of factors
248

Item response theory


For more information about Rasch modeling and/or item response theory
(IRT), there are two online sources to recommend. The web site for the journal,
Rasch Measurement Transactions, includes the journal's past contents and
instructions for subscribing to a listserv devoted to the Rasch approach. A newer
mailing list, maintained by R. J. Harvey, solicits more general discussion of IRT. To
subscribe to the latter, send email with body: subscribe IRT firstname lastname to
address listserv@listserv.vt.edu.
References
Bollen, K. A. & Ting, K. F. (1993). Confirmatory tetrad analysis. In P. Marsden
(Ed.), Sociological methodology (pp. 147-175), Washington, DC: American
Sociological Association.
Chin, W.W. (1998). The partial least squares approach for structural equation
modeling. In G.A. Marcoulides (Ed.), Modern methods for business research (pp.
295-336), Mahwah, NJ: Lawrence Erlbaum.
McDonald, R.P. (1996). Path analysis with composite variables. Multivariate
Behavioral Research, 31(2), 239-270. Schmidt McCollam, K.M. (1998). Latent trait
and latent class models. In G.A. Marcoulides (Ed.), Modern methods for business
research (pp. 23-46), Mahwah, NJ: Lawrence Erlbaum.
Spirtes, P., Glymour, C., & Scheines, R. (1993). Causation, prediction and
search. New York: Springer-Verlag.
Ting, K. F. (1995). Confirmatory tetrad analysis with SAS. Structural Equation
Modeling, 2(2), 163-171.
Statistical Power in Structural Equation Models David Kaplan
Department of Educational Studies
University of Delaware
The concept of power in statistical theory is defin ed as the probability of
rejecting the null hypothesis given that the null hypothesis is false. In the context
of structural equation modeling, the null hypothesis is defined by the specification
of fixed and free elements in relevant parameter matrices of the model equations.
The specification of fixed and free elements represents the researchers' initial
hypothesis concerning the putative direct and/or indirect effects among the latent
variables. The null hypothesis is assessed by forming a discrepancy function
between the model-implied set of moments (mean vector and/or covariance matrix)
and the sample moments. Various discrepancy functions can be formed depending
on the particular minimization algorithm being used (e.g. maximum likelihood), but
the goal remains the same - namely to derive a test statistic that has a known
distribution, and then compare the obtained value of the test statistic against
tabled values in order to render a decision vis-a-vis the null hypothesis.
In the framework of structural equation modeling the assessment of power is
complicated. Unlike simple procedures such as the t-test or ANOVA wherein
alternative hypotheses pertain to only a few parameters, in structural equation
249

modeling there are considerably more parameters. Each fixed parameter in the
model is potentially false and each can take on, in principle, an infinite number of
alternative values. Thus, each fixed parameter needs to be evaluated, in principle,
one at a time.
A method for evaluating the power of the likelihood ratio test in structural
equation modeling was developed by Satorra and Saris (1985). Their method can be
outlined as follows. First, estimate the model of interest. Second, choose a fixed
parameter whose power is desired. Third, re -estimate the initial model with each
estimated parameter fixed at their estimated value and choose an "alternative" fixed
value for the parameter of interest. Note that if the null hypothesis is true for that
parameter, then the likelihood ratio chi -square for the model would be zero with
degrees-of-freedom equaling the degrees-of-freedom of the model. If the null
hypothesis is false for that parameter, then the l ikelihood ratio chi-square will be
some positive number reflecting the specification error incurred by fixing that
parameter to the value chosen in the initial model. This number is the
noncentrality parameter (NCP) of the noncentral chi-square distribution, which is
the distribution of the test statistic when the null hypothesis is false. This number
can be compared to tabled values of the noncentral chi -square distribution to
assess power.
Clearly, the method originally proposed by Satorra and Saris (1985) would be
tedious to carry out in practice. Recen tly, Satorra (1989) recognized that the
modification indices (in LISREL) or the Lagrange multiplier tests (in EQS) are
actually one degree-of-freedom noncentrality parameters. Thus, for any estimated
model, it is a simple matter to look at these indices in relation to tabled value s of
the noncentral chi-square distribution in order to assess power. It should be noted
that power can be assessed for free parameters as well. That is, the square of the T -
value (in LISREL) or the Wald test (in EQS) can be used to assess the power of an
estimated parameter in the model, against a null hypothesis that the value of the
parameter is zero.
Consideration of the power associated with the likelihood ratio test (or other
asymptotically equivalent tests) led to an approach for conducting model
modification. Specifically, Saris, Satorra, and Sörbom (1987) advocated a
combination of the modification index and the expected change statistic (EC) to
guide model modifications. The EC is an estimate of the value that a fixed
parameter will take if that parameter is freed. Thus it represents the "distance"
between the value of the parameter under the null hypothesis (usually zero) and the
value of the parameter under the alternative hypothesis. Therefore, with the
modification index (NCP) and EC in hand one has all the ingredients to engage in
model modification with an eye toward power considerations. This power-based
approach to model modification was advocated by Kaplan (1990, with subsequent
commentary).
250

Under the approach advocated by Kaplan (1990) and Saris, Satorra, and
Sörbom (1987) parameters with large MIs and large ECs should be freed first
provided it makes substantive sense to do so because these parameters are
associated with higher probabilities of being false. Parameters associated with large
MIs and small ECs need to be studied carefully because , as shown by Kaplan
(1989), this might reflect a sample size sensitivity problem. Parameters associated
with small MIs and large ECs also need to be studied carefully in that it may also
reflect a sample size problem. Nevertheless, there may be substantive value in re -
estimating the model with this parameter freed because of the large contribution it
would be expected to make. Finally, parameters associated with small MIs and
small ECs could be ignored. Note that relative comparisons across parameters can
be made by employing standardized or completely standardized ECs (see Kaplan,
1989; Chou and Bentler, 1990).
With the advent of multivariate methods of model modification (see e.g.
Bentler, 1992) the question naturally arises how this might pertain to assessin g
power in more than one parameter simultaneously. Saris, Satorra, and Sörbom
(1987) recognized that power of two parameters of equal "true value" can differ
depending on the location of the parameters in the model. That is, if two
parameters have the same true alternative value, the power associated with each
may be different because of their different statistical associations with other
parameters in the model. The mechanism underlying these associations pertain to
the pattern of zero and non-zero values in the covariance matrix of the parameter
estimates. This matrix and its role in specification error and power have been
discussed recently in Kaplan and Wenger (1993). Suffice to say that multivariate
methods of model modification can, in some specific cases, lead one to miss the
subtle changes that can take place in the underlying system of relationships that
can be observed by a univariate approach to model modification. Thus, while it is
true that using multivariate methods minimizes Type II (or Type I) errors incurred
by model modification relative to univariate approaches, one does so at the expense
of monitoring subtle changes that might impact the substantive interpretation of
the model.
Finally, an issue of some practical importance is the role that power plays in
multiple group structural equation modeling. Recently, Kaplan and George (1995)
studied power in the multiple group confirmatory factor analysis setting.
Specifically, Kaplan and George (1995) examined the power of the Wald test of
factor mean differences under violations of factorial invariance. Using an approach
similar to that used in power studies of Hotelling's T-square, Kaplan and George
(1995) found that power was most affected by the degree of true factor mean
differences. The size of the model was also found to affect the power of the test with
larger models giving rise to increased power. Finally, with equal sample sizes,
Kaplan and George (1995) found that the power of the Wald test of factor mean
differences is relatively robust to violations of factorial invariance. With unequal
251

sample sizes, large changes in the power of the test were observed even under
conditions of factorial invariance.
When framed in terms of decisions errors, the results of Kaplan and George
(1995) suggest that the marginal effect of non -invariance was to decrease the
probability of Type II errors. In contrast, the marginal effect of inequality of sample
size was to increase Type II errors. Kaplan and George (1995) concluded with
practical guidelines suggesting that if the hypothesis of factor mean differences was
rejected under conditions of unequal sample size, then it was probably the case
that the null hypothesis was false. On the other hand, if the hypothesis was not
rejected, then it could be the effect of unequal sample siz es, lack of substantively
large differences, or both.
19.4 REVISION POINTS
1. Structural Equation Modelling
2. Form of Structural Equation Modelling
3. Latent construct
4. Structural model
5. Structural error
6. Manifest Variables
7. CFA
19.5 INTEXT QUESTIONS
1. Write short notes on
a) Latent Construct
b) Structural Model
c) Structural error
d) Manifest Variables
e) CFA
19.6 SUMMARY
After going through this Lesson you should be able to Explain Structural
Equation Modelling, form of Structural Equation Modelling, late nt construct,
Structural model, structural error, Manifest Variables, CFA and PLS
19.7 TERMINAL EXERSICE
1. Kindly explain the importance of structural equation modeling
19.8 SUPPLEMENTARY MATERIALS
Saris, W. E., & Satorra, A. (1987). Characteristics of structural equation
models which affect the power of the likelihood ratio test. In W.E. Saris & I. N.
Gallhofer (Eds.), Sociometric research (Vol. 2). London: Macmillan.
19.9 ASSIGNMENT
1. Briefly explain when to use structural equation modeling. Discuss with
examples.
252

19.10 SUGGESTED READING/REFERENCES BOOKS


1. Chou, C.-P., & Bentler, P. M. (1990). Model modification in covariance
structure modeling: A comparison among likelihood ratio, Lagrange
multiplier, and Wald tests. Multivariate Behavioral Research, 25, 115-136.
2. Kaplan, D. (1989a). Model modification in covariance structure analysis:
Application of the expected parameter change statistic. Multivariate
Behavioral Research, 24, 285-305.
3. Kaplan, D. (1995). Statistical power in structural equation modeling. In R. H.
Hoyle (ed.), Structural Equation Modeling: Concepts, Issues, and
Applications (pp. 100-117). Newbury Park, CA: Sage Publications, Inc.
4. Kaplan, D., & George, R. A study of the power associated with testing factor
mean differences under violations of factorial invariance. Structural Equation
Modeling: A Multidisciplinary Journal , 2, 101-118. Kaplan, D., & Wenger,
R.N. (1993). Asymptotic independence and separability in covariance
structure models: Implications for specification error, power, and model
modification. Multivariate Behavioral Research, 28, 483-498.
5. Saris, W. E., & Satorra, A. (1987). Characteristics of structural equation
models which affect the power of the likelihood ratio test. In W. E. Saris & I.
N. Gallhofer (Eds.), Sociometric research (Vol. 2). London: Macmillan.
6. Saris, W. E., Satorra, A., & Sörbom, D. (1987). The detection and correction
of specification errors in structural equation models. In C. C. Clogg (Ed.),
Sociological methodology (pp. 105-129). San Francisco: Jossey-Bass.
7. Satorra, A. (1989). Alternative test criteria in covariance structure analysis:
A unified approach. Psychometrika, 54, 131-151.
8. Satorra, A., & Saris, W. E. (1985). Power of the likelihood ratio test in
covariance structure analysis. Psychometrika, 50, 83-90.
19.11 LEARNING ACTIVITY
1. Explain briefly about the advantages of structural equation modeling. Enlist
the limitation of SEM
19.12 KEYWORD
Structural Equation Modelling, form of Structural Equation Modelling, latent
construct, Structural model, structural error, Manifest Variable s, CFA and PLS


253

LESSON – 20

CLUSTER ANALYSIS
20.1 INTRODUCTION
The statistical procedure used to form groups that share similar
characteristics is called cluster analysis. The choice of variables is crucial to cluster
analysis since only those used will determine the clusters or groups. The concepts
of DISTANCE and SIMILARITY are key ingredients in this statistical procedure.
Distance is the measure of how far apart cases are, whereas, similarity measures
the closeness of cases within a specific group or cluster. Distance measures are
SMALL while similarity measures are LARGE. Within cluster analysis, cases are
grouped together on the basis of their “nearness”. Cluster Analysis usually employs
the DISTANCE measure (how far apart cases are) in defining the clusters or groups.
20.2 OBJECTIVE
After going through this Lesson you should be able to understand Cluster
analysis, K-Mean Cluster analysis, Multi dimensional scaling and logic of Multi
dimensional scaling
20.3 CONTENT
20.3.1 Cluster analysis,
20.3.2 K-Mean Cluster analysis,
20.3.3 Multi dimensional scaling
20.3.4 Logic of Multi dimensional scaling
20.3.1 Cluster Analysis
Cluster analysis technique is employed in this study to segment the
respondents in to two or more similar groups. The factors derived from the Factor
Analyses in relation to Shopping Behaviour and Purchase Decision and the
satisfaction scores were used for this analysis. The main purpose of this
classification is to see whether respondents can be grouped into similar patterns.
These groups will be further used to analyze whether there is association between
respondents of different segments based on personal variables of respondents. At
this juncture K- means cluster analysis was applied to classify the consumer’s
perception and purchase decision. Hierarchical cluster method with agglomeration
schedule decides the number of clusters. It was found that three large coefficients
in the agglomeration schedule were significantly large in their differences.
Therefore, it was concluded that there was an existence of three clusters which was
presented as follows:
20.3.2 K-Means Cluster Analysis
There are different classification techniques to perform cluster analysis. K-
means cluster attempts to identify relatively homogeneous groups of cases based on
selected characteristics, using an algorithm that can handle large numbers of
cases. However, the algorithm requires specifying the number of clusters. Number
of initial cluster centers can be specified before if this information is known.
254

Alternatively, the procedure is repeated by increasing the number of clusters from 2


to the required number of clusters, until the procedure, differentiates well between
the clusters, and the cases within each cluster are homogenous as far as possible.
There are two methods for classifying cases, either updating cluster centers
iteractively or classifying only. For the study, the cluster centers are updated
iteratively till meaningful clusters arrived at. Once the numbers of clusters are
arrived at, then analysis of variance was applied to find whether the clusters
significantly differ from their group means. Since the cases or respondents were
forced into similar groups, the ANOVA conducted is seen as method of verification
of the cluster procedure.
For the purpose of cluster analysis the 14 factors (satisfaction score, 7 factors
from Purchase Decision and 9 factors from Shopping Behaviour) were used for
analysis. The classification procedure was repeated to find some meaningful
clusters. After repeated iteractions and inc reasing the cluster groups one by one,
finally 3 cluster groups were formed. The initial cluster centers are formed by
selecting the means of the groups of each variable as centers. The table given below
gives the initial cluster center values for each variable selected for this purpose. The
scores calculated for each factor were used for this purpose.
Cluster analysis consists of methods of classifying variables into clusters.
Technically, a cluster consists of variables that correlate highly with one another
and have comparatively low correlations with variables in other clusters. The basic
objective of cluster analysis is to determine how many mutually and exhaustive
groups or clusters, based on the similarities of profiles among entities, really exist
in the population and then to state the composition of such groups. Various groups
to be determined in cluster analysis are not predefined as happens to be the case in
discriminant analysis. Steps: In general, cluster analysis contains the following
steps to be performed:
(i) First of all, if some variables have a negative sum of correlations in the
correlation matrix, one must reflect variables so as to obtain a maximum sum of
positive correlations for the matrix as a whole.
(ii) The second step consists in finding out the highest correlation in the
correlation matrix and the two variables involved (i.e., having the highest
correlation in the matrix) form the nucleus of the first cluster.
(iii) Then one looks for those variables that correlate highly with the said two
variables and includes them in the cluster. This is how the first cluster is formed.
(iv) To obtain the nucleus of the second cluster, we find two vari ables that
correlate highly but have low correlations with members of the first cluster.
Variables that correlate highly withthe said two variables are then found. Such
variables along the said two variables thus constitute the second cluster.
(v) One proceeds on similar lines to search for a third cluster and so on.
255

From the above description we find that clustering methods in general are
judgemental and are devoid of statistical inferences. For problems concerning large
number of variables, various cut-andtry methods have been proposed for locating
clusters. McQuitty has specially developed a number of rather elaborate
computational routines* for that purpose.
In spite of the above stated limitation, cluster an alysis has been found useful
in context of market research studies. Through the use of this technique we can
make segments of market of a product on the basis of several characteristics of the
customers such as personality, socio-economic considerations, psychological
factors, purchasing habits and like ones.
8.3.4 Multidimensional Scaling
Multidimensional scaling (MDS) allows a researcher to measure an item in
more than one dimension at a time. The basic assumption is that people perceive a
set of objects as being more or less similar to one another on a number of
dimensions (usually uncorrelated wi th one another) instead of only one.
There are several MDS techniques (also known as techniques for dimensional
reduction) often used for the purpose of revealing patterns of one sort or another in
interdependent data structures. If data happen to be non-metric, MDS involves
rank ordering each pair of objects in terms of similarity. Then the judged
similarities are transformed into distances through statistical manipulations and
are consequently shown in n-dimensional space in a way that the interpoint
distances best preserve the original interpoint proximities. After this sort of
mapping is performed, the dimensions are usually interpreted and labeled by the
researcher.
The significance of MDS lies in the fact that it enables the rese archer to study
“The perceptual structure of a set of stimuli and the cognitive processes underlying
the development of this structure.... MDS provides a mechanism for determining
the truly salient attributes without forcing the judge to appear irrational.” With
MDS, one can scale objects, individuals or both with a minimum of information.
The MDS analysis will reveal the most salient attributes which happen to be the
primary determinants for making a specific decision.
Multidimensional scaling (MDS) can be considered to be an alternative to factor
analysis (see Factor Analysis). In general, the goal of the analysis is to detect
meaningful underlying dimensions that allow the researcher to explain observed
similarities or dissimilarities (distances) between the investigated objects. In factor
analysis, the similarities between objects (e.g., variables) are expressed in the
correlation matrix. With MDS, you can analyze any kind of similarity or
dissimilarity matrix, in addition to correlation matrices.
8.3.5 Logic of MDS
The following simple example may demonstrate the logic of an MDS analysis.
Suppose we take a matrix of distances between major US cities from a map. We
then analyze this matrix, specifying that we want to reproduce the distances based
256

on two dimensions. As a result of the MDS analysis, we would most likely obtain a
two-dimensional representation of the locations of the cities, that is, we would
basically obtain a two-dimensional map.
In general then, MDS attempts to arrange "objects" (major cities in this
example) in a space with a particular number of dimensions (two-dimensional in
this example) so as to reproduce the observed distances. As a result, we can
"explain" the distances in terms of underlying dimensions; in our example, we could
explain the distances in terms of the two geographical dimensions: north/south
and east/west.
Orientation of axes: As in factor analysis, the actual orientation of axes in the
final solution is arbitrary. To return to our e xample, we could rotate the map in any
way we want, the distances between cities remain the same. Thus, the final
orientation of axes in the plane or space is mostly the result of a subjective decision
by the researcher, who will choose an orientation that can be most easily explained.
To return to our example, we could have chosen an orientation of axes other than
north/south and east/west; however, that orientation is most convenient because it
"makes the most sense" (i.e., it is easily interpretable).
Computational Approach
MDS is not so much an exact procedure as rather a way to "rearrange" objects
in an efficient manner, so as to arrive at a configuration that best approximates the
observed distances. It actually moves objects around in the space defi ned by the
requested number of dimensions, and checks how well the distances between
objects can be reproduced by the new configuration. In more technical terms, it
uses a function minimization algorithm that evaluates different configurations with
the goal of maximizing the goodness-of-fit (or minimizing "lack of fit").
Measures of goodness-of-fit: Stress. The most common measure that is used
to evaluate how well (or poorly) a particular configuration reproduces the observed
distance matrix is the stress measure. The raw stress value Phi of a configuration is
defined by:

Phi = [dij - f (ij)]2


In this formula, dij stands for the reproduced distances, given the respective
number of dimensions, and ij (deltaij) stands for the input data (i.e., observed
distances). The expression f (ij) indicates a nonmetric , monotone transformation of
the observed input data (distances). Thus, it will attempt to reproduce the general
rank-ordering of distances between the objects in the analysis.
There are several similar related measures that are commonly used; however,
most of them amount to the computation of the sum of squared deviations of
observed distances (or some monotone transformation of those distances) from the
reproduced distances. Thus, the smaller the stress value, the better is the fit of the
reproduced distance matrix to the observed distance matrix.
257

Shepard diagram. You can plot the reproduced distances for a particular
number of dimensions against the observed input data (distances). This scatterplot
is referred to as a Shepard diagram. This plot shows the reproduced distances
plotted on the vertical (Y) axis versus the original similarities plotted on the
horizontal (X) axis (hence, the generally negative slope). This plot also shows a step-
function. This line represents the so- called D-hat values, that is, the result of the
monotone transformation f( ) of the input data. If all reproduced distances fall
onto the step-line, then the rank-ordering of distances (or similarities) would be
perfectly reproduced by the respective solution (dimensional model). Deviations
from the step-line indicate lack of fit.
How Many Dimensions to Specify?
If you are familiar with factor analysis, you will be quite aware of this issue. If
you are not familiar with factor analysis, you may want to read the Factor Analysis
section in the manual; however, this is not necessary in order to understand the
following discussion. In general, the more dimensions we use in order to reproduce
the distance matrix, the better is the fit of the reproduced matrix to the observed
matrix (i.e., the smaller is the stress). In fact, if we use as many dimensions as
there are variables, then we can perfectly reproduce the observed distance matrix.
Of course, our goal is to reduce the observed complexity of nature, that is, to
explain the distance matrix in terms of fewer underlying dimensions. To return to
the example of distances between cities, once we have a two-dimensional map it is
much easier to visualize the location of and navigate between cities, as compared to
relying on the distance matrix only.
Sources of misfit. Let's consider for a moment why fewer factors may produce
a worse representation of a distance matrix than would more factors. Imagine the
three cities A, B, and C, and the three cities D, E, and F; shown below are their
distances from each other.

A B C D E F

0
0
A 9 D
0 90 0
B 0 E
9 18 9
C 9 0 F 0
0 0 0
0
In the first matrix, all cities are exactly 90 miles apart from each other; in the
second matrix, cities D and F are 180 miles apart. Now, can we arrange the three
cities (objects) on one dimension (line)? Indeed, we can arrange cities D, E, and F
on one dimension:
D---90 miles---E---90 miles---F
D is 90 miles away from E, and E is 90 miles away from F; thus, D is
90+90=180 miles away from F. If you try to do the same thing with cities A, B, and
C you will see that there is no way to arrange the three cities on one line so that the
258

distances can be reproduced. However, we can arrange those cities in two


dimensions, in the shape of a triangle:

90 90
miles miles

B 90 miles C
Arranging the three cities in this manner, we can perfectly reproduce the
distances between them. Without going into much detail, this small example
illustrates how a particular distance matrix implies a particular number of
dimensions. Of course, "real" data are never this "clean," and contain a lot of noise,
that is, random variability that contributes to the differences between the
reproduced and observed matrix.
Scree test: A common way to decide how many dimensions to use is to plot
the stress value against different numbers of dimensions. This test was first
proposed by Cattell (1966) in the context of the number-of-factors problem in factor
analysis (see Factor Analysis); Kruskal and Wish (1978; pp. 53-60) discuss the
application of this plot to MDS.
Cattell suggests to find the place where the smooth decrease of stress values
(eigenvalues in factor analysis) appears to level off to the right of the plot. To the
right of this point, you find, presumably, only "factorial scree" - "scree" is the
geological term referring to the debris which collects on the lower part of a rocky
slope.
Interpretability of Configuration: A second criterion for deciding how many
dimensions to interpret is the clarity of the final configuration. Sometimes, as in
our example of distances between cities, the resultant dimensions are easily
interpreted. At other times, the points in the plot form a sort of "random cloud," and
there is no straightforward and easy way to interpret the dimensions. In the latter
case, you should try to include more or fewer dimensions and examine the
resultant final configurations. Often, more interpretable solutions emerge. However,
if the data points in the plot do not follow any pattern, and if the stress plot does
not show any clear "elbow," then the data are most likely random "noise."
Interpreting the Dimensions
The interpretation of dimensions usually represents the final step of the
analysis. As mentioned earlier, the actual orientations of the axes from the MDS
analysis are arbitrary, and can be rotated in any direction. A first step is to produce
scatterplots of the objects in the different two-dimensional planes.
In addition to "meaningful dimensions," you should also look for clusters of
points or particular patterns and configurations (such as circles, manifolds, etc.).
For a detailed discussion of how to interpret final configurations, see Borg and
Lingoes (1987), Borg and Shye (in press), or Guttman (1968).
259

Use of multiple regression techniques: An analytical way of interpreting


dimensions (described in Kruskal & Wish, 1978) is to use multiple regression
techniques to regress some meaningful variables on the coordinates for the different
dimensions. Note that this can easily be done via Multiple Regression
Applications
The "beauty" of MDS is that we can analyze any kind of distance or similarity
matrix. These similarities can represent people's ratings of similarities between
objects, the percent agreement between judges, the number of times a subjects fails
to discriminate between stimuli, etc. For example, MDS methods used to be very
popular in psychological research on person perception where similarities between
trait descriptors were analyzed to uncover the underlying dimensionality of people's
perceptions of traits (see, for example Rosenberg, 1977). They are also very popular
in marketing research, in order to detect the number and nature of dimensions
underlying the perceptions of different brands or products & Carmone, 1970).
In general, MDS methods allow the researcher to ask relatively unobtrusive
questions ("how similar is brand A to brand B") and to derive from those questions
underlying dimensions without the respondents ever knowing what is the
researcher's real interest.
MDS and Factor Analysis
Even though there are similarities in the type of research questions to which
these two procedures can be applied, MDS and factor analysis are fundamentally
different methods. Factor analysis requires that the underlying data are distributed
as multivariate normal, and that the relationships are linear. MDS imposes no such
restrictions. As long as the rank-ordering of distances (or similarities) in the matrix
is meaningful, MDS can be used. In terms of resultant differences, factor analysis
tends to extract more factors (dimensions) than MDS; as a result, MDS often yields
more readily, interpretable solutions. Most importantly, however, MDS can be
applied to any kind of distances or similarities, while factor analysis requires us to
first compute a correlation matrix. MDS can be based on subjects' direct
assessment of similarities between stimuli, while factor analysis requires subjects
to rate those stimuli on some list of attributes (for which the factor analysis is
performed).
In summary, MDS methods are applicable to a wide variety of research designs
because distance measures can be obtained in any number of ways (for different
examples, refer to the references provided at the beginning of this section).
Introduction to Statistical Thinking for Decision Making
This site builds up the basic ideas of business statistics systematically and
correctly. It is a combination of l ectures and computer-based practical, whereby
theory is firmly placed into practice. It provides an introduction to techniques for
summarizing and presenting data, estimation, confidence intervals and hypothesis
testing. The presentation is centered around focusing more on conceptual
understanding of key concepts, and statistical thinking, and less on formulas and
260

calculations, which can now be left to PCs using accessible and user-friendly
Statistical JavaScript Applets.
Today's good decisions are driven by data. In all aspects of our lives, and
importantly in the business context, an amazing diversity of data is available for
inspection and enlightenment. Moreover, business managers and professionals are
increasingly encouraged to justify decisions on the basis of data.
Business managers need statistical model-based decision support systems.
Statistical skills enable you to intelligently collect, analyze and interpret data
relevant to their decision-making. Statistical concepts and statistical thinking
enable you:
 to solve problems in a diversity of contexts
 to add substance to decisions.
This Web site is a course in statistics appreciation; i.e., to acquire a feel for the
statistical way of thinking. Appreciation of statistics is wonderful: it makes what is
excellent in statistical thinking belong to you as well. It is an introductory course in
statistics that is designed to provide you with the basic concepts, and methods of
statistical analysis for processes and products. Materials in this Web site are
tailored to meet your needs in making good decision and they promote one to think
statistically. The cardinal objective for this Web site is to increase the extent to
which statistical thinking is embedded with managerial thinking for decision
making under uncertainties.
In competitive environment, business managers must design quality into
products, and into the processes of making the products. They must facilitate a
process of never-ending improvement at all stages of manufacturing and service.
This is a strategy that employs statistical methods, particularly statistically
designed experiments, and produces processes that provide high yield an d products
that seldom fail. Moreover, it facilitates development of robust products that are
insensitive to changes in the environment and internal component variation.
Carefully planned statistical studies remove hindrances to high quality and
productivity at every stage of production. This saves time and money. It is well
recognized that quality must be engineered into products as early as possible in the
design process. One must know how to use carefully planned, cost-effective
statistical experiments to improve, optimize and make robust products and
processes.
20.4 REVISION POINTS
1. Cluster analysis,
2. K-Mean Cluster analysis,
3. Multi dimensional scaling
4. Logic of Multi dimensional scaling
261

20.5 INTEXT QUESTIONS


1. What do you mean by multi dimension scaling?
Explain briefly about the logic of Multi dimension scaling
20.6 SUMMARY
After going through this Lesson you should be able to Explain Cluster
analysis, K-Mean Cluster analysis, Multi dimensional scaling and logic of Multi
dimensional scaling
20.7 TERMINAL EXERSICE
1. Kindly explain the importance of Cluster analysis
20.8 SUPPLEMENTARY MATERIALS
Saris W.E., & Satorra, A. (1987). Characteristics of structural equation models
which affect the power of the likelihood ratio test. In W. E. Saris & I. N. Gallhofer
(Eds.), Sociometric research (Vol. 2). London: Macmillan.
20.9 ASSIGNMENT
1. Briefly explain when to use Cluster analysis. Discuss with examples
20.10 SUGGESTED READING REFERENCES BOOKS
Chou, C.-P.,& Bentler, P.M. (1990). Model modification in covariance structure
modeling: A comparison among likelihood ratio, Lagrange multiplier, and Wald
tests. Multivariate Behavioral Research, 25, 115-136.
Kaplan, D. (1989a). Model modification in covariance structure analysis:
Application of the expected parameter change statistic. Multivariate Behavioral
Research, 24, 285-305.
Kaplan, D. (1995). Statistical power in structural equation modeling. In R. H.
Hoyle (ed.), Structural Equation Modeling: Concepts, Issues, and Applications (pp.
100-117). Newbury Park, CA: Sage Publications, Inc.
Kaplan, D., & George, R. A study of the power associated with testing factor
mean differences under violations of factorial invariance. Structural Equation
Modeling: A Multidisciplinary Journal , 2, 101-118. Kaplan, D., & Wenger, R. N.
(1993). Asymptotic independence and separability in covariance structure models:
Implications for specification error, power, and model modification. Multivariate
Behavioral Research, 28, 483-498.
Saris, W.E., & Satorra, A. (1987). Characteristics of structural equation models
which affect the power of the likelihood ratio test. In W. E. Saris & I. N. Gallhofer
(Eds.), Sociometric research (Vol. 2). London: Macmillan.
Saris, W. E., Satorra, A., & Sörbom, D. (1987). The detection and correction of
specification errors in structural equation models. In C. C. Clogg (Ed.), Sociological
methodology (pp. 105-129). San Francisco: Jossey-Bass.
Satorra, A. (1989). Alternative test criteria in covariance structure analysis: A
unified approach. Psychometrika, 54, 131-151.
262

Satorra, A., & Saris, W.E. (1985). Power of the likelihood ratio test in
covariance structure analysis. Psychometrika, 50, 83-90.
20.11 LEARNING ACTIVITY
Explain briefly about the advantages of Cluster analysis. Enlist the limitation
of Cluster analysis
20.12 KEYWORD
Cluster analysis, K-Mean Cluster analysis, Multi-dimensional scaling and logic
of Multi-dimensional scaling

263

UNIT – VI
LESSON – 21

RESEARCH REPORT
21.1 INTRODUCTION
In this chapter deals with Research Report - Structure of Research Report – Title
of a Report – Table of contents - Synopsis – Introduction and and body of a report -
Table and diagrams – Results and Discussions – Recommendations, conclusion, and
scope for further research-reference –Appendices.
There are people who do not consider writing of report as an integral part ofthe
research process. But the general opinion is in favour of treating the presentation of
research results or the writing of report as part and parcel of the research project.
Writing of report is the last step in a research study and requires a set of skills
somewhat different from those called for in respect of the earlier stages of research.
This task should be accomplished by the researcher with utmost care; he may seek
the assistance and guidance of experts for the purpose.
21.2 OBJECTIVE
After going through this Lesson you should be able to understand what is
report, report writing Steps in report writing and lay out of research report
21.3 CONTENT
9.3.1 WHAt is report,
9.3.2 Report writing
9.3.3 Steps in report writing and
9.3.4 Lay out of research report
DIFFERENT STEPS IN WRITING REPORT
Research reports are the product of slow, painstaking, accurate inductive
work. The usual steps involved in writing report are:
(a) logical analysis of the subject-matter; (b) preparation of the final outline;
(c) preparation of the rough draft; (d) rewriting and polishing; (c) preparation of the
final bibliography; and (f) writing the final draft. Though all these steps are self
explanatory, yet a brief mention of each one of these will be appropriate for better
understanding.
a) Logical analysis of the subject matter
It is the first step which is primarily concerned with the development of a
subject. There are two ways in which to develop a subject (a) logically and (b)
chronologically. The logical development is made on the basis of mental
connections and associations between the one thing and another by means of
analysis. Logical treatment often consists in developing the material from the
simple possible to the most complex structures. Chronological development is
264

based on a connection or sequence in time or occurrence. The directions for doing


or making something usually follow the chronological order.
b) Preparation of the final outline
It is the next step in writing the research report “Outlines are the framework
upon which long written works are constructed. They are an aid to the logical
organization of the material and a reminder of the points to be stressed in the
report.”
c) Preparation of the rough draft:
This follows the logical analysis of the subject and the preparationof the final
outline. Such a step is of utmost importanc e for the researcher now sits to write
down what he has done in the context of his research study. He will write down the
procedure adopted by him in collecting the material for his study along with various
limitations faced by him, the technique of analysis adopted by him, the broad
findings and generalizations and the various suggestions he wants to offer
regarding the problem concerned.
d) Rewriting and polishing of the rough draft
This step happens to be most difficult part of all formal writing. Usuall y this
step requires more time than the writing of the rough draft. The careful revision
makes the difference between a mediocre and a good piece of writing. While
rewriting and polishing, one should check the report for weaknesses in logical
development or presentation. The researcher should also “see whether or not the
material, as it is presented, has unity and cohesion; does the report stand upright
and firm and exhibit a definite pattern, like a marble arch? Or does it resemble an
old wall of moldering cement and loose brick.” In addition the researcher should
give due attention to the fact that in his rough draft he has been consistent or not.
He should check the mechanics of writing—grammar, spelling and usage.
e) Preparation of the final bibliography
Next in order comes the task of the preparation of the final bibliography. The
bibliography, which is generally appended to the research report, is a list of books
in some way pertinent to the research which has been done. It should contain all
those works which the researcher has consulted. The bibliography should be
arranged alphabetically and may be divided into two parts; the first part may
contain the names of books and pamphlets, and the second part may contain the
names of magazine and newspaper articles. Generally, this pattern of bibliography
is considered convenient and satisfactory from the point of view of reader, though it
is not the only way of presenting bibliography. The entries in bibliography should
be made adopting the following order:
For books and pamphlets the order may be as under:
1. Name of author, last name first.
2. Title, underlined to indicate italics.
3. Place, publisher, and date of publication.
265

4. Number of volumes.
Example
Kothari, C.R., Quantitative Techniques, New Delhi, Vikas Publishing House
Pvt. Ltd., 1978.
For magazines and newspapers the order may be as under:
1. Name of the author, last name first.
2. Title of article, in quotation marks.
3. Name of periodical, underlined to indicate italics.
4. The volume or volume and number.
5. The date of the issue.
6. The pagination.
Example
Robert V. Roosa, “Coping with Short-term International Money Flows”, The
Banker, London, September, 1971, p. 995.
The above examples are just the samples for bibliography entries and may be
used, but oneshould also remember that they are not the only acceptable forms.
The only thing important is that, whatever method one selects, it must remain
consistent.
f) Writing the final draft
This constitutes the last step. The final draft should be written in a concise
and objective style and in simple language, avoiding vague expressions such as “it
seems”, “there may be”, and the like ones. While writing the final draft, the
researcher must avoid abstract terminology and technical jargon. Illustrations and
examples based on common experiences must be incorporated in the final draft as
they happen to be most effective in communicating the research findings to others.
A research report should not be dull, but must enthuse people and maintain
interest and must show originality. It must be remembered that every report should
be an attempt to solve some intellectual problem and must contribute to the
solution of a problem and must add to the knowledge of both the researcher and
the reader.
LAYOUT OF THE RESEARCH REPORT
Anybody, who is reading the research report, must necessarily be conveyed
enough about the study so that he can place it in its general scientific context,
judge the adequacy of its methods and thus form an opinion of how seriously the
findings are to be taken. For this purpose there is the need of proper layout of the
report. The layout of the report means as to what the research report should
contain. A comprehensive layout of the research report should comprise (A)
preliminary pages; (B) the main text; and (C) the end matter. Let us deal with them
separately.
266

a) Preliminary Pages
In its preliminary pages the report should carry a title and date, followed by
acknowledgements in the form of ‘Preface’ or ‘Foreword’. Then there should be a
table of contents followed by list of tables and illustrations so that the decision -
maker or anybody interested in reading the report can easily locate the required
information in the report.
b) Main Text
The main text provides the complete outline of the research report along with
all details. Title of theresearch study is repeated at the top of the first page of the
main text and then follows the other details on pages numbered consecutively,
beginning with the second page. Each main section of the report should begin on a
new page. The main text of the report should have the following sections: (i)
Introduction; (ii) Statement of findings and recommendations; (iii) The results;
(iv)The implications drawn from the results; and (v) The summary.
i) Introduction
The purpose of introduction is to introduce the research project to the readers.
It should contain a clear statement of the objectives of research i.e., enough
background should be given to make clear to the reader why the problem was
considered worth investigating. A brief summary of other relevant research may
also be stated so that the present study can be seen in that context. The
hypotheses of study, if any, and the definitions of the major concepts employed in
the study should be explicitly stated in the introduction of the report. The
methodology adopted in conducting the study must be fully explained. The
scientific reader would like to know in detail about such thing: How was the study
carried out? What was its basic design? If the study was an experimental one, then
what were the experimental manipulations? If the data were collected by means of
questionnaires or interviews, then exactly what questions were asked (The
questionnaire or interview schedule is usually given in an appendix)? If
measurements were based on observation, then what instructions were given to the
observers? Regarding the sample used in the study the reader should be told: Who
were the subjects? How many were there? How were they selected? All these
questions are crucial for estimating the probable limits of generalizability of the
findings. The statistical analysis adopted must also be clearly stated. In addition to
all this, the scope of the study should be stated and the boundary lines be
demarcated. The various limitations, under which the research project was
completed, must also be narrated.
ii) Statement of findings and recommendations
After introduction, the research report must contain a statement of findings
and recommendations in non-technical language so that it can be easily
understood by all concerned. If the findings happen to be extensive, at this point
they should be put in the summarized form.
267

iii) Results
A detailed presentation of the findings of the study, with supporting data in
the form of tables and charts together with a validation of results, is the next step
in writing the main text of the report. This generally comprises the main body of the
report, extending over several chapters. The result section of the report should
contain statistical summaries and reductions of the data rather than the raw data.
All the results should be presented in logical sequence and splitted into readily
identifiable sections. All relevant results must find a place in the report. But how
one is to decide about what is relevan t is the basic question. Quite often guidance
comes primarily from the research problem and from the hypotheses, if any, with
which the study was concerned. But ultimately the researcher must rely on his own
judgement in deciding the outline of his report. “Nevertheless, it is still necessary
thathe states clearly the problem with which he was concerned, the procedure by
which he worked onthe problem, the conclusions at which he arrived, and the
bases for his conclusions.”
iv) Implications of the results
Toward the end of the main text, the researcher should again put down the
results of his research clearly and precisely. He should, state the implications that
flow from the results of the study, for the general reader is interested in the
implications for understanding the human behaviour. Such implications may have
three aspects as stated below:
(a) A statement of the inferences drawn from the present study which may be
expected to apply in similar circumstances.
(b) The conditions of the present study whi ch may limit the extent of legitimate
generalizations of the inferences drawn from the study.
(c) The relevant questions that still remain unanswered or new questions
raised by the study along with suggestions for the kind of research that would
provide answers for them.
It is considered a good practice to finish the report with a short conclusion
which summarises andrecapitulates the main points of the study. The conclusion
drawn from the study should be clearly related to the hypotheses that were stated
in the introductory section. At the same time, a forecast of the probable future of
the subject and an indication of the kind of research which needs to be done in that
particular field is useful and desirable.
v) Summary
It has become customary to conclude the research report with a very brief
summary, resting in brief the research problem, the methodology, the major
findings and the major conclusions drawn from the research results.
c) End Matter
At the end of the report, appendices should be enlisted in respect of all
technical data such as questionnaires, sample information, mathematical
derivations and the like ones. Bibliography of sources consulted should also be
given. Index (an alphabetical listing of names, places and topics along with the
268

numbers of the pages in a book or report on which they are mentioned or


discussed) should invariably be given at the end of the report. The value of index
lies in the fact that it works as a guide to the reader for the contents in the report.
21.4 REVISION POINTS
1. What is report,
2. Report writing
3. Steps in report writing and
4. Lay out of research report
21.5 INTEXT QUESTION
Briefly explain how to write a research report?
21.6 SUMMARY
After going through this Lesson you should be able to explain what is report,
report writingSteps in report writing and lay out of research report
21.7 TERMINAL EXERSICE
1. Explain the importance of Research report
2. Explain the limitations of research report
21.8 SUPPLEMENTARY MATERIALS
1. Roscoe, John T., Fundamental Research Statistics for the Behavioral
Sciences, New York: Holt, Rinehart and Winston, Inc., 1969.
21.9 ASSIGNMENT
1. What are the main advantage and dis advantage of research report writing?
21.10 SUGGESTED READING Reference
1. Redman, L.V., and Mory, A.V.H., The Romance of Research, 1923.
2. Roscoe, John T., Fundamental Research Statistics for the Behavioral
Sciences, New York: Holt, Rinehart and Winston, Inc., 1969.
3. Runyon, Richard P., Inferential Statistics, Philippines: Addison -Wesley
Publishing Company, Inc., 1977.
4. Sadhu, A.N., and Singh, Amarjit, Research Methodology in Social
Sciences,Bombay: Himalaya Publishing House, 1980.
5. Seboyar, G.E., Manual for Report and Thesis Writing, New York: F.S. Crofts
& Co., 1929.
6. Selltiz, Claire: Jahoda, Marie, Deutsch, Morton, and Cook, Stuart W.,
Research Methods in Social Relations, rev. ed. New York: Holt, Rinehart and
Winston, Inc., 1959.
7. Sharma, B.A.V., et al., Research Methods in Social Sciences, New Delhi:
Sterling Publishers Pvt. Ltd., 1983.
8. Sharma, H.D., and Mukherji, S.P., Research in Economics and Commerce,
Methodology and Surveys, Varanasi: Indian Biographic Centre, 1976.
21.11 LEARNING ACTIVITY
Briefly Explain the steps involved in research report writting?
21.12 KEYWORD
Report, report writingSteps in report writing and l ay out of research report

269

LESSON – 22

TYPES OF REPORT
22.1 INTRODUCTION
Research reports vary greatly in length and type. In each individual case, both
the length and theform are largely dictated by the problems at hand. For instance,
business firms prefer reports in theletter form, just one or two pages in length.
Banks, insurance organisations and financial institutionsare generally fond of the
short balance-sheet type of tabulation for their annual reports to theircustomers
and shareholders. Mathematicians prefer to write the results of their investigations
in theform of algebraic notations. Chemists report their results in symbols and
formulae. Students ofliterature usually write long reports presenting the critical
analysis of some writer or period or the likewith a liberal use of quotations from the
works of the author under discussion. In the field of educationand psychology, the
favourite form is the report on the results of experimentation accompanied bythe
detailed statistical tabulations. Clinic al psychologists and social pathologists
frequently find itnecessary to make use of the case -history form.
22.2 OBJECTIVE
After going through this Lesson you should be able to understand what are the
Types of report, Technical report, Popular report and Oral presentation
22.3 CONTENT
22.3.1 Types of report,
22.3.2 Technical report,
22.3.3 Popular report and
22.3.4 Oral presentation
22.3.1 Types of Report
News items in the daily papers are also forms of report writing. They represent
firsthand on-the-scene accounts of the events described or compilations of
interviews with persons who were on thescene. In such reports the first paragraph
usually contains the important information in detail and thesucceeding paragraphs
contain material which is progressively less and less important.
Book-reviews which analyze the content of the book and report on the author’s
intentions, hissuccess or failure in achieving his aims, his language, his style,
scholarship, bias or his point of view.Such reviews also happen to be a kind of
short report. The reports prepared by governmental bureaus,special commissions,
and similar other organisations are generally very comprehensive reports onthe
issues involved. Such reports are usually considered as important research
products. Similarly,Ph.D. theses and dissertations are also a form of report-writing,
usually completed by students inacademic institutions.
The above narration throws light on the fact that the results of a research
investigation can bepresented in a number of ways viz., a technical report, a
popular report, an article, a monograph or attimes even in the form of oral
270

presentation. Which method(s) of presentation to be used in a particularstudy


depends on the circumstances under which the study arose and the nature o f the
results. Atechnical report is used whenever a full written report of the study is
required whether for record-keeping or for public dissemination. A popular report is
used if the research results have policyimplications. We give below a few details
about the said two types of reports:
22.3.2 TECHNICAL REPORT
In the technical report the main emphasis is on (i) the methods employed, (it)
assumptions made inthe course of the study, (iii) the detailed presentation of the
findings including their limitations andsupporting data.A general outline of a
technical report can be as follows:
1. Summary of results: A brief review of the main findings just in two or three
pages.
2. Nature of the study: Description of the general objectives of study,
formulation of the problem inoperational terms, the working hypothesis, the type of
analysis and data required, etc.
3. Methods employed: Specific methods used in the study and their limitations.
For instance, insampling studies we should give details of sample design viz .,
sample size, sample selection, etc.
4. Data: Discussion of data collected, their sources, characteristics and
limitations. If secondarydata are used, their suitability to the problem at hand be
fully assessed. In case of a survey, themanner in which data were collected should
be fully described.
5. Analysis of data and presentation of findings: The analysis of data and
presentation of thefindings of the study with supporting data in the form of tables
and charts be fully narrated. This, infact, happens to be the main body of the report
usually extending over several chapters.
6. Conclusions: A detailed summary of the findings and the policy implications
drawn from theresults be explained.
7. Bibliography: Bibliography of various sources consulted be prepared and
attached.
8. Technical Appendices: Appendices be given for all technical matters relating
to questionnaire,mathematical derivations, elaboration on particular technique of
analysis and the like ones.
9. Index: Index must be prepared and be given in variably in the report at the
end.
The order presented above only gives a general idea of the nature of a technical
report; the orderof presentation may not necessarily be the same in all the technical
reports. This, in other words,means that the presentation may vary in different
reports; even the different sections outlined abovewill not always be the same, nor
will all these sections appear in any particular report.
271

It should, however, be remembered that even in a technical report, simple


presentation and readyavailability of the findings remain an important
consideration and as such the liberal use of charts anddiagrams is considered
desirable.
22.3.3 POPULAR REPORT
The popular report is one which gives emphasis on simplicity and attractiveness.
The simplificationshould be sought through clear writing, minimization of technical,
particularly mathematical, detailsand liberal use of charts and diagrams. Attractive
layout along with large print, many subheadings,even an occasional cartoon now and
then is another characteristic feature of the popular report.Besides, in such a report
emphasis is given on practical aspects and policy implications.
We give below a general outline of a popular report.
1. The Findings and their Implications: mphasis in the report is given on the
findings of mostpractical interest and on the implications of these findings.
2. Recommendations for Action: Recommendations for action on the basis of
the findings of thestudy is made in this section of the report.
3. Objective of the study: A general review of how the problem arise is
presented along with the specific objectives of the project under study.
4. Methods Employed: A brief and non-technical description of the methods
and techniques used,including a short review of the data on which the study is
based, is given in this part of the report.
5. Results: This section constitutes the main body of the report wherein the
results of the study are presented in clear and non -technical terms with liberal use
of all sorts of illustrations such as charts, diagrams and the like ones.
6. Technical Appendices: More detailed information on methods used, forms,
etc. is presented in the form of appendices. But the appendices are often not
detailed if the report is entirely meant for general public.
There can be several variations of the form in which a popular report can be
prepared. The onlyimportant thing about such a report is that it gives emphasis on
simplicity and policy implications from the operational point of view, avoiding the
technical details of all sorts to the extent possible.
22.3.4 ORAL PRESENTATION
At times oral presentation of the results of the study is considered effective,
particularly in caseswhere policy recommendations are indicated by project results.
The merit of this approach lies in the fact that it provides an opportunity for give -
and-take decisions which generally lead to a better understanding of the findings
and their implications. But the main demerit of this sort of presentation is the lack
of any permanent record concerning the research details and it may be just
possible that the findings may fade away from people’s memory even before an
action is taken. In order to overcome this difficulty, a written report may be
circulated before the oral presentation and referred to frequently during the
272

discussion. Oral presentation is effective when supplemented by various visual


devices. Use of slides, wall charts and blackboards is quite helpful in contributing
to clarity and in reducing the boredom, if any. Distributing a board outline, with a
few important tables and charts concerning the research results, makes the
listeners attentive who have a ready outline on which to focus their thinking. This
very often happens in academic institutions where the researcher discusses his
research findings and policy implications with others either in a seminar or in a
group discussion.
Thus, research results can be reported in more than one ways, but the usual
practice adopted, inacademic institutions particularly, is that of writing the
Technical Report and then preparing several research papers to be discussed at
various forums in one form or the other. But in practical field and with problems
having policy implications, the technique followed is that of writing a popular
report. Researches done on governmental account or on behalf of some major
public or private organisations are usually presented in the form of technical
reports.
22.4 REVISION POINTS
1. Types of report,
2. Technical report,
3. Popular report and
4. Oral presentation
22.5 INTEXT QUESTION
Briefly explain how to write a Technical report?
22.6 SUMMARY
After going through this Lesson you should be able to explain what are the
Types of report, Technical report, Popular report and Oral presentation
22.7 TERMINAL EXERSICE
1. Explain the importance of Popular report
2. Explain the limitations of Technical report
22.8 SUPPLEMENTARY MATERIALS
1. Roscoe, John T., Fundamental Research Statistics for the Behavioral
Sciences, New York: Holt, Rinehart and Winston, Inc., 1969.
22.9 ASSIGNMENT
1. What are the main advantage and dis advantage of Oral presentation?
22.10 SUGGESGTED READING REFERENCE BOOKS
1. Redman, L.V., and Mory, A.V.H., The Romance of Research, 1923.
2. Roscoe, John T., Fundamental Research Statistics for the Behavioral
Sciences, New York: Holt, Rinehart and Winston, Inc., 1969.
3. Runyon, Richard P., Inferential Statistics, Philippines: Addison -Wesley
Publishing Company, Inc., 1977.
273

4. Sadhu, A.N., and Singh, Amarjit, Research Methodology in Social


Sciences,Bombay: Himalaya Publishing House, 1980.
5. Seboyar, G.E., Manual for Report and Thesis Writing, New York: F.S. Crofts
& Co., 1929.
6. Selltiz, Claire: Jahoda, Marie, Deutsch, Morton, and Cook, Stuart W.,
Research Methods in Social Relations, rev. ed. New York: Holt, Rinehart and
Winston, Inc., 1959.
7. Sharma, B.A.V., et al., Research Methods in Social Sciences, New Delhi:
Sterling Publishers Pvt. Ltd., 1983.
8. Sharma, H.D., and Mukherji, S.P., Research in Economics and Commerce,
Methodology and Surveys, Varanasi: Indian Biographic Centre, 1976.
22.11 LEARNING ACTIVITY
1. Briefly Explain the types of report with suitable examples?
22.12 KEYWORD
Types of report, Technical report, Popular report and Oral presentation

274

LESSON – 23

MECHANICS OF WRITING A REPORT


23.1 INTRODUCTION
There are very definite and set rules which should be followed in the actual
preparation of theresearch report or paper. Once the techniques are finally decided,
they should be scrupulously adhered to, and no deviation permitted. The criteria of
format should be decided as soon as the materials for the research paper have been
assembled.
23.2 OBJECTIVE
After going through this Lesson you should be able to understand what are the
Mechanics of writing a report
23.3 CONTENT
23.3.1 Size and Physical design
23.3.2 Procedure
23.3.3 Layout
23.3.4 Treatment of quotation
23.3.5 Foot note
23.3.6 Documentation style
23.3.7 Punctuations
23.3.8 Use of Statistical analysis
23.3.9 Final draft
23.3.10 Bibliography
23.3.11 Index
Mechanics of writing a report
The following points deserve mention so far as the mechanics of writing a
report are concerned:
1. Size and physical design: The manuscript should be written on unruled
paper 8 1/2′′ × 11′′ in size. If it is to be written by hand, then black or blue -black
ink should be used. A margin of at leastone and one -half inches should be allowed
at the left hand and of at least half an inch at the right hand of the paper. There
should also be one-inch margins, top and bottom. The paper should be neat and
legible. If the manuscript is to be typed, then all typing should be double -spaced on
one side of the page only except for the insertion of the long quotations.
2. Procedure: Various steps in writing the report should be strictly adhered (All
such steps have already been explained earlier in this chapter).
3. Layout: Keeping in view the objective and nature of the problem, the layout
of the report should be thought of and decided and accordingly adopted (The layout
of the research report and various types of reports have been described in this
275

chapter earlier which should be taken as a guide for report-writing in case of a


particular problem).
4. Treatment of quotations: Quotations should be placed in quotation marks
and double spaced, forming an immediate part of the text. But if a quotation is of a
considerable length (more than four or five type written lines) then it should be
single-spaced and indented at least half an inch to the right of the normal text
margin.
5. The footnotes: Regarding footnotes one should keep in view the followings:
(a) The footnotes serve two purposes viz., the identification of materials used
in quotations in the report and the notice of materials not immediately necessary to
the body of the research text but still of supplemental value. In other words,
footnotes are meant for cross references, citation of authorities and sources,
acknowledgement and elucidation or explanation of a point of view. It should
always be kept in view that footnote is not an end nor a means of the display of
scholarship. The modern tendency is to make the minimum use of footnotes for
scholarship does not need to be displayed.
(b) Footnotes are placed at the bottom of the page on whic h the reference or
quotation which they identify or supplement ends. Footnotes are customarily
separated from the textual material by a space of half an inch and a line about one
and a half inches long.
(c) Footnotes should be numbered consecutively, usually beginning with 1 in
each chapter separately. The number should be put slightly above the line, say at
the end of a quotation. At the foot of the page, again, the footnote number should
be indented and typed a little above the line. Thus, consecutive nu mbers must be
used to correlate the reference in the text with its corresponding note at the bottom
of the page, except in case of statistical tables and other numerical material, where
symbols such as the asterisk (*) or the like one may be used to preven t confusion.
(d) Footnotes are always typed in single space though they are divided from
one another bydouble space.
6. Documentation style: Regarding documentation, the first footnote reference
to any given work should be complete in its documentation, g iving all the essential
facts about the edition used. Such documentary footnotes follow a general
sequence. The common order may be described as under:
i) Regarding the single-volume reference
1. Author’s name in normal order (and not beginning with the l ast name as in
a bibliography) followed by a comma;
2. Title of work, underlined to indicate italics;
3. Place and date of publication;
4. Pagination references (The page number).
276

Example
John Gassner, Masters of the Drama, New York: Dover Publications, Inc.
1954, p. 315.
ii) Regarding multivolumed reference
1. Author’s name in the normal order;
2. Title of work, underlined to indicate italics;
3. Place and date of publication;
4. Number of volume;
5. Pagination references (The page number).
iii) Regarding works arranged alphabetically
For works arranged alphabetically such as encyclopedias and dictionaries, no
pagination reference is usually needed. In such cases the order is illustrated as
under:
Example 1
“Salamanca,” Encyclopaedia Britannica, 14th Edition.
Example 2
“Mary Wollstonecraft Godwin,” Dictionary of national biography.
But if there should be a detailed reference to a long encyclopedia article,
volume and pagination reference may be found necessary.
iv) Regarding periodicals reference
1. Name of the author in normal order
2. Title of article, in quotation marks
3. Name of periodical, underlined to indicate italics
4. Volume number
5. Date of issuance
6. Pagination
v) Regarding anthologies and collections, reference
Quotations from anthologies or collections of literary works must be
acknowledged not only by author, but also by the name of the collector.
vi) Regarding second-hand quotations, reference
In such cases the documentation should be handled as follows:
1. Original author and title;
2. “quoted or cited in,”;
3. Second author and work.
Example
J.F. Jones, Life in Ploynesia, p. 16, quoted in History of the Pacific Ocean area,
by R.B. Abel, p. 191.
277

vii) Case of multiple authorship


If there are more than two authors or editors, then in the documentation the
name of only the first is given and the multiple authorship is indicated by “et al.” or
“and others”. Subsequent references to the same work need not be so detailed as
stated above. If the work is cited again without any other work in tervening, it may
be indicated as ibid, followed by a comma andthe page number. A single page
should be referred to as p., but more than one page be referred to aspp. If there are
several pages referred to at a stretch, the practice is to use often the pag e number,
for example, pp. 190ff, which means page number 190 and the following pages; but
only for page 190 and the following page ‘190f’. Roman numerical is generally used
to indicate the number of the volume of a book. Op. cit. (opera citato, in the work
cited) or Loc. cit. (loco citato, in the place cited) are two of the very convenient
abbreviations used in the footnotes. Op. cit. or Loc. cit. after the writer’s name
would suggest that the reference is to work by the writer which has been cited in
detail in an earlier footnote but intervened by some other references.
7. Punctuation and abbreviations in footnotes
The first item after the number in the footnote is the author’s name, given in
the normal signature order. This is followed by a comma. After the comma, the title
of the book is given: the article (such as “A”, “An”, “The” etc.) is omitted and only
the first word and proper nouns and adjectives are capitalized. The title is followed
by a comma. Information concerning the edition is given next. This entry is followed
by a comma. The place of publication is then stated; it may be mentioned in an
abbreviated form, if the place happens to be a famous one such as Lond. for
London, N.Y. for New York, N.D. for New Delhi and so on. This entry is followed by a
comma. Then the name of the publisher is mentioned and this entry is closedby a
comma. It is followed by the date of publication if the date is given on the title page.
If the date appears in the copyright notice on the reverse side of the title page or
elsewhere in the volume, the comma should be omitted and the date enclosed in
square brackets [c 1978], [1978]. The entry is followed by a comma. Then follow the
volume and page references and are separated by a comma if both are given. A
period closes the complete documentary reference. But one should remember that
the documentation regarding acknowledgement from magazine articles and
periodical literature follow a different form as stated earlier while explaining the
entries in the bibliography. Certain English and Latin abbreviations are quite often
used in bibliographies and footnotes to eliminate tedious repetition. The following is
a partial list of the most common abbreviations frequently used in report-writing
(the researcher should learn to recognise them as well as he should learn touse
them):
anon., anonymous
ante., before
art., article
aug., augmented
278

bk., book
bull., bulletin
cf., compare
ch., chapter
col., column
diss., dissertation
ed., editor, edition, edited.
ed. cit., edition cited
e.g., exempli gratia: for example
eng., enlarged
et.al., and others
et seq., et sequens: and the following
ex., example
f., ff., and the following
fig(s)., figure(s)
fn., footnote
ibid., ibidem: in the same place (when two or more successive footnotes refer
to the same work, it is not necessary to repeat complete reference for the second
footnote. Ibid. may be used. If different pages are referred to, pagination must be
shown).
id., idem: the same
ill., illus., orillust(s) illustrated, illustration(s)
Intro., intro., introduction
l, or ll, line(s)
loc. cit., in the place cited; used as op.cit., (when new reference
loco citato: is made to the same pagination as cited in the previous note)
MS., MSS., Manuscript or Manuscripts
N.B., nota bene: note well
n.d., no date
n.p., no place
no pub., no publisher
no(s)., number(s)
o.p., out of print
op. cit: in the work cited (If reference has been made to a work
279

opera citato and new reference is to be made, ibid., may be used, if intervening
reference has been made to different works, op.cit. must be used. Thename of the
author must precede.
p. or pp., page(s)
passim: here and there
post: after
rev., revised
tr., trans., translator, translated, translation
vid or vide: see, refer to
viz., namely
vol. or vol(s)., volume(s)
vs., versus: against
8. Use of statistics, charts and graphs
A judicious use of statistics in research reports is oftenconsidered a virtue for
it contributes a great deal towards the clarification and simplification of the
material and research results. One may well remember that a good picture is often
worth more than a thousand words. Statistics are usually presented in the form of
tables, charts, bars and line-graphsand pictograms. Such presentation should be
self explanatory and complete in itself. It should be suitable and appropriate
looking to the problem at hand. Finally, statistical presentation should be neatand
attractive.
9. The final draft
Revising and rewriting the rough draft of the report should be done with great
care before writing the final draft. For the purpose, the researcher should put to
himself questionslike: Are the sentences written in the report clear? Are they
grammatically correct? Do they say what is meant’? Do the various points
incorporated in the report fit together logically? “Having atleast one colleague read
the report just before the final revision is extremely helpful. Sentences that seem
crystal-clear to the writer may prove quite confusing to other people; a connection
that had seemed self evident may strike others as a non -sequitur. A friendly critic,
by pointing out passages that seem unclear or illogical, and perhaps suggesting
ways of remedying the difficulties, can be an invaluable aid in achieving the goal of
adequate communication.”
10. Bibliography
Bibliography should be prepared and appended to the research report as
discussed earlier.
11. Preparation of the index
At the end of the report, an index should invariably be given, the value of
which lies in the fact that it acts as a good guide, to the reader. Index may be
prepared both as subject index and as author index. The former gives the names of
280

the subject-topics or conceptsalong with the number of page s on which they have
appeared or discussed in the report, where as the latter gives the similar
information regarding the names of authors. The index should always be arranged
alphabetically. Some people prefer to prepare only one index common for names of
authors, subject-topics, concepts and the like ones.
23.4 REVISION POINTS
1. Size and Physical design
2. Procedure
3. Lay out
4. Treatment of quotation
5. Foot note
6. Documentation style
7. Punctuations
8. Use of Statistical analysis
9. Final draft
10. Bibliography
11. Index
23.5 INTEXT QUESTIONS
Briefly explain the role of statistics in a research report?
23.6 SUMMARY
After going through this Lesson you should be able to explain what are the
Mechanics of writing a report
23.7 TERMINAL EXERSICE
Explain the importance of Lay out while preparing a research report
23.8 SUPPLEMENTARY MATERIALS
 Roscoe, John T., Fundamental Research Statistics for the Behavioral
Sciences, New York: Holt, Rinehart and Winston, Inc., 1969.
23.9 ASSIGNMENT
 Explain the importance of Size and physical design in the mechanics of the
report?
23.10 SUGGESTED READING REFERENCE BOOKS
 Redman, L.V., and Mory, A.V.H., The Romance of Research, 1923.
 Roscoe, John T., Fundamental Research Statistics for the Behavioral
Sciences, New York: Holt, Rinehart and Winston, Inc., 1969.
 Runyon, Richard P., Inferential Statistics, Philippines: Addison -Wesley
Publishing Company, Inc., 1977.
 Sadhu, A.N., and Singh, Amarjit, Research Methodology in Social
Sciences,Bombay: Himalaya Publishing House, 1980.
281

 Seboyar, G.E., Manual for Report and Thesis Writing, New York: F.S. Crofts
& Co., 1929.
 Selltiz, Claire: Jahoda, Marie, Deutsch, Morton, and Cook, Stuart W.,
Research Methods in Social Relations, rev. ed. New York: Holt, Rinehart
and Winston, Inc., 1959.
 Sharma, B.A.V., et al., Research Methods in Social Sciences, New Delhi:
Sterling Publishers Pvt. Ltd., 1983.
 Sharma, H.D., and Mukherji, S.P., Research in Economics and Commerce,
Methodology and Surveys, Varanasi: Indian Biographic Centre, 1976.
23.11 LEARNING ACTIVITY
Briefly Explain the Mechanics of a good Report?
23.12 KEYWORD
Size and Physical design, Procedure, Lay out, Treatment of quotation, Foot
note, Documentation style, Punctuations, Use of Statistical analysis, Final draft,
Bibliography, Index


282

LESSON – 24

PRECAUSIONS FOR WRITING A REPORT


24.1 INTRODUCTION
There are people who do not consider writing of report as an integral part
ofthe research process. But the general opinion is in favour of treating the
presentation of research results or the writing of report as part and parcel of the
research project. Writing of report is the last step in a research study and requires
a set of skills somewhat different from those called for in respect of the earlier
stages of research. This task should be accomplished by the researcher with utmost
care; he may seek the assistance and guidance of experts for the purpose.
24.2 OBJECTIVE
After going through this Lesson you should be able to understand what is
report, report writing and precautions for writing report
24.3 CONTENT
12.3.1 Precautions for writing report
PRECAUTIONS FOR WRITING RESEARCH REPORTS
1. Research report is a channel of communicating the research findings to
the readers of the report. Agood research report is one which does this
task efficiently and effectively. As such it must beprepared keeping the
following precautions in view:
2. While determining the length of the report (since research reports vary
greatly in length),one should keep in view the fact that it should be long
enough to cover the subject but short enough to maintain interest. In fact,
report-writing should not be a means to learning moreand more about less
and less.
3. A research report should not, if this can be avoided, be dull; it should be
such as to sustainreader’s interest.
4. Abstract terminology and technical jargon should be avoided in a research
report. The report should be able to convey the matter as simply as
possible. This, in other words,means that report should be written in an
objective style in simple language, avoidingexpressions such as “it seems,”
“there may be” and the like.
5. Readers are often interested in acquiring a quick knowledge of the main
findings and as such the report must provide a ready availability of the
findings. For this purpose, charts,graphs and the statistical tables may be
used for the various results in the main report inaddition to the summary
of important findings.
6. The layout of the report should be well thought out and must be
appropriate and in accordance with the objective of the rese arch problem.
283

7. The reports should be free from grammatical mistakes and must be


prepared strictly in accordance with the techniques of composition of
report-writing such as the use of quotations,footnotes, documentation,
proper punctuation and use of abbreviations in footnotes and thelike.
8. The report must present the logical analysis of the subject matter. It must
reflect a structurewherein the different pieces of analysis relating to the
research problem fit well.
9. A research report should show originality and should necessarily be an
attempt to solve some intellectual problem. It must contribute to the
solution of a problem and must add tothe store of knowledge.
10. Towards the end, the report must also state the policy implications
relating to the problem under consideration. It is usually considered
desirable if the report makes a forecast of theprobable future of the
subject concerned and indicates the kinds of research still needs tobe
done in that particular field.
11. Appendices should be enlisted in respect of all the technical data in the
report.
12. A Bibliography of sources consulted is a must for a good report and must
necessarily be given.
13. The Index is also considered an essential part of a good report and as
such must be prepared and appended at the end.
14. A Report must be attractive in appearance, neat and clean, whether typed
or printed.
15. Calculated confidence limits must be mentioned and the various
constraints experienced in
a) Conducting the research study may also be stated in the report.
16. The Objective of the study, the nature of the problem, the methods
employed and the analysis techniques adopted must all be clearly stated in
the beginning of the report in the form ofintroduction.
24.4 REVISION POINTS
Precautions for writing a Report
24.5 INTEXT QUESTION
Briefly explain how to write a good research report?
24.6 SUMMARY
After going through this Lesson you should be able to explain what is report,
report writingSteps in report writing and Precautions for writing a Report
24.7 TERMINAL EXERSICE
Explain the importance of a good Research report writing
284

24.8 SUPPLEMENTARY MATERIALS


Roscoe, John T., Fundamental Research Statistics for the Behavioral Sci ences,
New York: Holt, Rinehart and Winston, Inc., 1969.
24.9 ASSIGNMENT
What are the main advantage and dis advantage of a good research report
writing?
24.10 SUGGESTED READINGS REFERENCE BOOKS
1. Redman, L.V., and Mory, A.V.H., The Romance of Research, 1923.
2. Roscoe, John T., Fundamental Research Statistics for the Behavioral
Sciences, New York: Holt, Rinehart and Winston, Inc., 1969.
3. Runyon, Richard P., Inferential Statistics, Philippines: Addison -Wesley
Publishing Company, Inc., 1977.
4. Sadhu, A.N., and Singh, Amarjit, Research Methodology in Social
Sciences,Bombay: Himalaya Publishing House, 1980.
5. Seboyar, G.E., Manual for Report and Thesis Writing, New York: F.S. Crofts
& Co., 1929.
6. Selltiz, Claire: Jahoda, Marie, Deutsch, Morton, and Cook, Stuart W.,
Research Methods in Social Relations, rev. ed. New York: Holt, Rinehart
and Winston, Inc., 1959.
7. Sharma, B.A.V., et al., Research Methods in Social Sciences, New Delhi:
Sterling Publishers Pvt. Ltd., 1983.
8. Sharma, H.D., and Mukherji, S.P., Research in Economics and Commerce,
Methodology and Surveys, Varanasi: Indian Biographic Centre, 1976.
24.11 LEARNING ACTIVITY
Briefly Explain the precautionary measures that has to taken while preparing
a good report?
24.12 KEYWORD
Report, report writingSteps in report writing and Precautions for writing a
report

285

Selected References and Recommended Readings


1. Ackoff, Russell L., The Design of Social Research, Chicago: University of
Chicago Press, 1961.
2. Ackoff, Russell L., Scientific Method, New York: John Wiley & Sons, 1962.
3. Allen, T. Harrell, New Methods in Social Science Research, New York:
Praeger Publishers, 1978.
4. Anderson, H.H., and Anderson, G.L., An Introduction to Projective
Techniques and Other Devices for
Understanding the Dynamics of Human Behaviour, New York: Prentice Hall,
1951.
5. Anderson, T.W., An Introduction to Multivariate Analysis, New York: Jo hn
Wiley & Sons, 1958.
6. Bailey, Kenneth D., “Methods of Social Research,” New York, 1978.
7. Baker, R.P., and Howell, A.C., The Preparation of Reports, New York: Ronald
Press, 1938.
8. Bartee, T.C., “Digital Computer Fundamentals,” 5th Ed., McGraw-Hill,
International Book Co., 1981.
9. Barzun, Jacques, and Graff, Henery, F., The Modern Researcher, rev. ed.,
New York: Harcourt, Brace &
World, Inc., 1970.
10. Bell, J.E., Projective Techniques: A. Dynamic Approach to the Study of
Personality, New York: Longmans Green, 1948.
11. Bellenger, Danny N., and Greenberg, Barnett A., Marketing Re search—A
Management Information Approach, Homewood, Illinois: Richard D. Irwin, Inc.,
1978.
12. Berdie, Douglas R., and Anderson, John F., Questionnaires: Design and
Use, Metuchen N.J.: The Scarecrow Press, Inc., 1974.
13. Berelson, Bernard, Content Analysis in Communication Research, New
York: Free Press, 1952.
14. Berenson, Conard, and Colton, Raymond, Research and Report Writing for
Business and Economics, New York: Random House, 1971.
15. Best, John W., and Kahn, James V., “Research in Education,” 5th Ed.,
New Delhi: Prentice-Hall of India Pvt. Ltd., 1986.
16. Bhattacharya, Srinibas, Psychometrics & Behavioural Research, New
Delhi: Sterling Publishers Pvt. Ltd., 1972.
17. Boot, John C.G., and Cox, Edwin B., Statistical Analysis for Managerial
Decisions, 2nd ed. New Delhi: McGraw-Hill Publishing Co. Ltd., (International
Student Edition), 1979.
286

18. Bowley, A.L., Elements of Statistics, 6th ed. London: P.S. King and Staples
Ltd., 1937.
19. Burgess, Ernest W., “Research Methods in Sociology” in Georges Gurvitch
and W.E. Moore (Ed.), Twentieth Century Sociology, New York: New York
Philosophical Library, 1949.
20. Chance, William A., Statistical Methods for Decision Making, Bombay:
D.B. Taraporevala Sons & Co. Pvt. Ltd., 1975.
21. Chaturvedi, J.C., Mathematical Statistics, Agra: Nok Jhonk Karyalaya,
1953.
22. Chou, Ya-Lun, Statistical Analysis with Business and Economic
Applications, 2nd ed. New York: Holt, Rinehart & Winston, 1974.
23. Clover, Vernon T., and Balsley, Howard L., Business Research Methods,
Columbus, O.: Grid, Inc., 1974.
24. Cochran, W.G., Sampling Techniques, 2nd ed. New York: John Wiley &
Sons., 1963.
25. Cooley, William W., and Lohnes, Paul R., Multivariate Data A nalysis, New
York: John Wiley & Sons.,1971.
26. Croxton, F.E., Cowden, D.J., and Klein, S., Applied General Statistics, 3rd
ed., New Delhi: Prentice-Hall ofIndia Pvt. Ltd., 1975.
27. Dass, S.L., Personality Assessment Through Projective Movie Pictures, New
Delhi: S. Chand & Co. (Pvt.) Ltd., 1974.
28. Davis, G.B., “Introduction to Computers,” 3rd ed., McGraw-Hill
International Book Co., 1981.
29. Deming, W. Edwards., Sample Design in Business Research, New York:
John Wiley & Sons., Inc., 1960.
30. Dennis, Child, The Essentials of Factor Analysis, New York: Holt, Rinehart
and Winston, 1973.
31. Denzin, Norman, The Research Act, Chicago: Aldine, 1973.
32. Edwards, Allen, Statistical Methods, 2nd ed., New York: Holt, Rinehart &
Winston, 1967.
33. Edwards, Allen L., Techniques of Attitude Scale Construction, New York:
Appleton-Century-Crofts, 1957.
34. Emory, C. William, Business Research Methods, Illinois: Richard D. Irwin,
Inc. Homewood, 1976.
35. Ferber, Robert (ed.), Handbook of Marketing Research, New York: McGraw-
Hill, Inc., 1948.
36. Ferber, R., and Verdoorn, P.J., Research Methods in Economics and
Business, New York: The Macmillan Company, 1962.
287

37. Ferguson, George A., Statistical Analysis in Psychology and Education, 4th
ed., New York: McGrawHill Book Co., Inc., 1959.
38. Festinger, Leon and Katz, Daniel (Eds.), Research Methods in the
Behavioral Sciences, New Delhi: Amerind Publishing Co. Pvt. Ltd., Fourth Indian
Reprint, 1976.
39. Fiebleman, J.K., Scientific Method, Netherlands: Martinus Nijhoff, The
Hague, 1972.
40. Fisher, R.A., Statistical Methods for Research Workers, 13th ed., New
York: Hafner Publishing Co., 1958.
41. Fisher, R.A., The Design of Experiments, 7th rev. ed., New York: Hafner
Publishing Co., 1960.
42. Fox, James Harold, Criteria of Good Research, Phi Delta Kappa, Vol. 39
(March 1958).
43. Freedman, P., The Principles of Scientific Research, 2nd ed., New York:
Pergamon Press, 1960.
44. Fruchter, Benjamin, Introduction to Factor Analysis, Princeton, N.J.:
D.Van Nostrand, 1954.
45. Gatner, Elliot S.M., and Cordasco, Francesco, Research and Report
Writing, New York: Barnes & Noble, Inc., 1956.
46. Gaum, Carl G., Graves, Harod F., and Hoffman, Lyne, S.S., Report Writing,
3rd ed., New York: Prentice-Hall, 1950.
47. Ghosh, B.N., Scientific Methods and Social Research, New Delhi: Sterling
Publishers Pvt. Ltd., 1982.
48. Gibbons, J.D., Nonparametric Statistical Inference, Tokyo: McGraw-Hill
Kogakusha Ltd., (International Student Edition), 1971.
49. Giles, G.B., Marketing, 2nd ed., London: Macdonal d & Evans Ltd., 1974.
50. Glock, Charles Y., Survey Research in the Social Sciences, New York:
Russell Sage Foundation, 1967.
51. Godfrey, Arthur, Quantitative Methods for Managers, London: Edward
Arnold (Publishers) Ltd., 1977.
52. Good, Carter V., and Douglas, E. Scates, Methods of Research—
Educational, Psychological, Sociological, New York: Appleton-Century-Crofts, Inc.,
1954.
53. Goode, William J., and Hatt, Paul K., Methods in Social Research, New
York: McGraw-Hill, 1952.
54. Gopal, M.H., An Introduction to Research Procedure in Social
Sciences,Bombay: Asia Publishing House, 1964.
55. Gopal, M.H., Research Reporting in Social Sciences, Dharwar: Karnatak
University, 1965.
288

56. Gorden, Raymond L., Interviewing: Strategy, Techniques and Tactics, rev.
ed., Homewood, Ill.: Dorsey Press, 1975.
57. Green, Paul E., Analyzing Multivariate Data, Hinsdale, Ill.: Dryden Press,
1978.
58. Green, Paul E., and Carmone, F.J., Multidimensional Scaling in Marketing
Analysis, Boston: Allyn&Bacon, Inc., 1970.
59. Guilford, J.P., Psychometric Methods, New York: McGraw Hill, Inc., 1954.
60. Harnett, Donald L., and Murphy, James L., Introductory Statistical
Analysis,Philippines: AddisonWesley Publishing Co., Inc., 1975.
61. Hillway, T.,Introduction to Research, 2nd ed., Boston: Houghton Mifflin,
1964.
62. Hollander, Myles, and Wolfe, Douglas A., Nonparametric Statistical
Methods, New York: John Wiley, 1973.
63. Hunt, R., and Shelley, J., “Computers and Common Sense,” 3rd ed., New
Delhi: Prentice-Hall of India Ltd., 1984.
64. Hyman, Herbert H., et al., Interviewing in Social Research, Chicago:
University of Chicago Press, 1975.
65. John, Peter W.M., Statistical Design and Analysis of Experiments, New
York: The Macmillan Co., 1971.
66. Johnson, Ellen, The Research Report: A Guide for the Beginner, New York:
Ronald Press, 1951.
67. Johnson, Rodney D., and Siskin, Bernard R., Quantitative Techniques for
Business Decisions, New Delhi: Prentice-Hall of India Pvt. Ltd., 1977.
68. Kahn, Robert L. and Cannell, Charles F., The Dynamics of In terviewing,
New York: John Wiley & Sons, 1957.
69. Karson, Marvin J., Multivariate Statistical Methods, Ames, Iowa: The Iowa
State University Press, 1982.
70. Kendall, M.G., A Course in Multivariate Analysis, London, Griffin, 1961.
71. Kerlinger, Fred N. and Pedhazur, Elazar J., Multiple Regression in
Behavioral Research, New York: Holt, Rinehart and Winston, 1973.
72. Kerlinger, Fred N., Foundations of Behavioral Research, 2nd ed., New
York: Holt, Reinhart and Winston, 1973.
73. Kish, Leslie., Survey Sampling, New York: John Wiley & Sons, Inc., 1965.
74. Kothari, C.R., Quantitative Techniques, 2nd ed., New Delhi: Vikas
Publishing House Pvt. Ltd., 1984.
75. Lastrucci, Carles L., The Scientific Approach: Basic Principles of the
Scientific Method, Cambridge, Mass.: Schenkman Publishing Co., Inc., 1967.
289

76. Lazersfeld, Paul F.,“Evidence and Inference in Social Research,” in David


Lerher, Evidence and Inference, Glencoe: The Free Press, 1950.
77. Leonard Schatzman, and Anselm L. Strauss, Field Research, New Jerse y:
Prentice-Hall Inc., 1973.
78. Levin, Richard I.,Statistics for Management, New Delhi: Prentice -Hall of
India Pvt. Ltd., 1979.
79. Levine, S. and Elzey, Freeman F., A Programmed Introduction to Research,
California: Wods Worth Publishing Co., 1968.
80. Maranell, Gary M. (ed.), Scaling: A Source Book for Behavioral Scientists,
Chicago: Aldine, 1974.
81. Maxwell, Albert E., Analyzing Qualitative Data, New York: John Wiley &
Sons, 1961.
82. Meadows, R., and Parsons, A.J., “Microprocessors: Essentials,
Components and Systems,” Pitman, 1983.
83. Meir, Robert C., Newell, William T., and Dazier, Harold L., Simulation in
Business and Economics, Englewood Cliffs, N.J: Prentice Hall, Inc., 1969.
84. Miller, Delbert C., Handbook of Research Design & Social Measurement,
3rd ed., New York: David Mckay Company, Inc., 1977.
85. Moroney, M.J., Facts from Figures, Baltimore: Penguin Books, 1956.
86. Morrison, Donald F., Multivariate Statistical Methods, New York: McGraw-
Hill, 1967.
87. Nagel, Stuart S., and Neef, Marian, Policy Analysis in Social Science
Research, London: Sage Publications, 1979.
88. Nie, N.H., Bent, D.H., and Hull, C.H., Statistical Package for the Social
Sciences, New York: McGrawHill, 1970.
89. Noether, G.E., Elements of Nonparametric Statistics, New York: John Wiley
& Sons, Inc., 1967.
90. Nunnally, Jum C.,Psychometric Theory, 2nd ed., New York: McGraw-Hill,
1978.
91. Odum, H.W., and Jocher, Katharine, An Introduction to Social Research,
New York: Henry Holt and Co.,1929.
92. Oppenheim, A.N., Questionnaire Design and Attitude Measurement, New
York: Basic Books, 1966.
93. Ostle, Bernard, and Mensing, Richard W., Statistics in Research, 3rd ed.,
Ames Iowa: The Iowa State University Press, 1975.
94. Payne, Stanley, The Art of Asking Questions, Princeton: Prin ceton
University Press, 1951.
290

95. Pearson, Karl, The Grammar of Science, New York: Meridian Books, Inc.,
1957.
96. Phillips, Bernard S., Social Research, Strategy and Tactics, 2nd ed., New
York: The Macmillan Company, 1971.
97. Piaget, Jean, Main Trends in Interdisciplinary Research, London: George
Allen and Unwin Ltd., 1973.
98. Popper, Karl R., The Logic of Scientific Discovery, New York: Basic Books,
1959.
99. Rajaraman, V., “Fundamentals of Computers,” New Delhi: Prentice -Hall of
India Pvt. Ltd., 1985.
100. Ramchandran, P.,Training in Research Methodology in Social Sciences in
India, New Delhi: ICSSR 1971.
101. Redman, L.V., and Mory, A.V.H., The Romance of Research, 1923.
102. Roscoe, John T., Fundamental Research Statistics for the Behavioral
Sciences, New York: Holt, Rinehart and Winston, Inc., 1969.
103. Runyon, Richard P., Inferential Statistics, Philippines: Addison -Wesley
Publishing Company, Inc., 1977.
104. Sadhu, A.N., and Singh, Amarjit, Research Methodology in Social
Sciences,Bombay: Himalaya Publishing House, 1980.
105. Seboyar, G.E., Manual for Report and Thesis Writing, New York: F.S.
Crofts & Co., 1929.
106. Selltiz, Claire: Jahoda, Marie, Deutsch, Morton, and Cook, Stuart W.,
Research Methods in Social Relations, rev. ed. New York: Holt, Ri nehart and
Winston, Inc., 1959.
107. Sharma, B.A.V., et al., Research Methods in Social Sciences, New Delhi:
Sterling Publishers Pvt. Ltd., 1983.
108. Sharma, H.D., and Mukherji, S.P., Research in Economics and
Commerce, Methodology and Surveys, Varanasi: Indian Biographic Centre, 1976.
109. Siegel, S., Nonparametric Statistics for the Behavioral Sciences, New
York: McGraw-Hill Publishing Co., Inc., 1956.
110. Subramanian, N., “Introduction to Computers,” New Delhi: Tata McGraw-
Hill Publishing Co. Ltd., 1986.
111. Summers, Gene F., (Ed.), Attitude Measurement, Chicago: Rand McNally
& Co., 1970.
112. Takeuchi, K., Yanai, H. and Mukherjee, B.N., The Foundations of
Multivariate Analysis, New Delhi: Wiley Eastern Ltd., 1982.
113. Tandon, B.C., Research Methodology in Social Sciences, Allahabad:
Chaitanya Publishing House, 1979.
291

114. Thorndike, Robert L. and Hagen, Elizabeth P., Measurement and


Evaluation in Psychology and Education, 4th ed., New York: John Wiley & Sons,
1977.
115. Thurstone, L.L., The Measurement of Values, Chicago: University of
Chicago Press, 1959.
116. Torgerson, W., Theory and Methods of Scaling, New York: John Wiley &
Sons, 1958.
117. Travers, Robert M.W., An Introduction to Educational Research, 4th ed.,
New York: Macmillan Publishing Co., Inc., 1978.
118. Tryon, R.C., and Bailey, D.E., Cluster Analysis, New York: McGraw-Hill,
1970.
119. Ullman, Neil R., Elementary Statistics, New York: John Wiley & Sons,
Inc., 1978.
120. Whitney, F.L., The Elements of Research, 3rd ed., New York: Prentice -
Hall, 1950.
121. Wilkinson, T.S. and Bhandarkar, P.L., Methodology and Techniques of
Social Research, Bombay: Himalaya Publishing House, 1979.
122. Willemsen, Eleanor Walker, Understanding Statistical Reasoning, San
Francisco: W.H. Freeman and Company, 1974.
123. Yamane, T.,Statistics: An Introductory Analysis, 3rd ed., New York:
Harper and Row, 1973.
124. Young, Pauline V., Scientific Social Surveys and Research, 3rd ed., New
York: Prentice-Hall, 1960


You might also like