You are on page 1of 8

Special Section on Scaling and Measurement

Note from Section Editor James Comer: Professor Churchill has long been acknowledged as
one of the pioneers of measurement research in sales management. His willingness to initiate
this section with a portrayal of the key measurement issues is much appreciated. Thank you
Gil.
Manuscript Submissions: The reader who wishes to have a manuscript reviewed for possible
publication in this section should examine page ix of the Fall 1991 issue of the Journal of Per-
sonal Selling and Sales Management for details on section objectives andl guidelines or you can
send for guidelines to: Professor James M. Comer, University of Cincinnati, College of Business
Administration, Department of Marketing, Linder Hall (#145), Cincinnati, Ohio 45221.

Better M:easurement Practices are Critical to Better


Und1erstanding of Sales ManagemeJnt Issues
Gilbert A. Churchill, Jr.

The study of personal selling and sales management shares a characteristic


common to other areas of scientific inquiry. If knowledge and understanding of
its phenomena are to advance, then its measurement tools must also. This
article overviews the conceptual foundations of measurement in terms of how
they apply to sales research. It reviews the notions of reliability and validity and
the types of evidence they provide regarding the quality of measures. It also
offers suggestions on how sales researchers should proceed when working with
new or borrowed measures.

The new section on "Scaling and Measurement" The various scientifie disciplines differ from one
is a welcome addition to the Journal of Personal another in a variety of ways ... One such [way] ...
Selling and Sales Management, in that the study of is according to the degree to wlrich theoretical
procedures or explanations are used as con-
personal selling and sales management shares a
trasted with correlational procedures or expla-
characteristic common to other areas of scientific nations ... The distinction here is between a sci-
inquiry. If knowledge and understanding of its ence that consists largely of statements describ-
phenomena are to advance, then its measurement ing the degree of relationship among more or
tools must also. 'I'he two go hand-in-hand. One less directly observable variables and a science
only has to think of the value of the telescope to that attempts (succe~ssfully) to derive, account
astronomy and the microscope to biology and chem- for, or explain these r~lationships from principles
istry to appreciate the relationship between knowl- that are not immediately given, but lie beyond
edge development and good measurement. As straight empirical knowledge. Although no sci-
ence is all correlational or entirely theoretical in
Torgerson suggested more than thirty years ago:
this sense, .. .it is clear that sciences do differ in
the degree to which they rely on one or the other
Gilbert A. Churchill, .Jrr., is the Arthur C. Nielsen, Jr., Chair of level of explanation. It may be also noted that
Marketing Research, University of Wisconsin-Madison, Univer-
sity ofWisconsin-Madison, 1155 Observatory Drive, Madison, WI Journal of Personal Selling & Sales Man.agement,
53706. Volume :XU, Number 2 (Spring 1992).
74 Journal of Personal Selling & Sales Management

Figure 1 tion may take the form of an equation that pre-


Schematic Diagram Illustrating the Structure of cisely expresses the interrelationship of the con-
Science and the Problem of Measurement struct to the other constructs, such as the equation
in mechanics that suggests that force equals mass
time acceleration. Alternatively, the relationship
may be only imprecisely stated, which is typically
the case in the social sciences.
The double lines represent operational definitions.
An operational definition describes how the con-
Observable struct is to be measured. It specifies the activities
Data that the researcher must complete in order to assign
a value to the construct (e.g., sum the scores on the
ten individual statements to generate a total score).
In essence, the operational definition tells the in-
vestigator what to do and in what manner to mea-
sure the concept. Conceptual definitions logically
precede operational definitions and guide their de-
all sciences begin as largely correlational, and velopment, for we must specify what a construct is
progress toward the theoretical. In this sense, before we can develop rules for assessing its magni-
we might say that the social sciences are largely tude. When scientists test scientific propositions,
correlational, as are most of biology and psy- they are in essence testing their conjectures about
chology, though possibly less so. Physics and the relationships among the C's in the diagram.
chemistry would be examples standing at the They carry out these tests by assessing the strength,
other extreme.
direction, and form of the relationships between
It is more than a mere coincidence that the sci- the observable variables. Suppose an observed re-
ences would order themselves in largely the same lationship is inconsistent with the relationship that
way if they were classified on the basis of the
has been hypothesized or that there is no relation-
degree to which satisfactory measurement of
their important variables has been achieved. ship at all among the observable variables. Is the
(Torgerson, 1958, pp. 1-2, emphasis added) conceptual structure or theory wrong or are the
measures at fault? It is difficult to say. It depends
The situation in sales management parallels that on the confidence scientists have in the measures
in the social sciences in general. The studies are used to assess the relationships.
largely correlational. Moreover, very little attention Technically, measurement is nothing more than
can be devoted to assessing the functional form "rules for assigning numbers to objects to represent
relating variables because the measurement of the quantities of attributes." The rigor with which these
discipline's most important variables has not been rules are defined and the skill with which they are
achieved. implemented determine whether the construct has
been captured by the measure.
Nature and Importance of You would undoubtedly scoff at the following
Measurement measurement procedure: John has blue eyes and
Bill has brown eyes; therefore, John is taller than
The essential nature of the measurement task is Bill. You might reply that the color of a person's
captured in Figure 1. The basic researcher or sci- eyes has nothing to do with the person's height and,
entist uses theories in an attempt to explain phe- further, that if you wanted to see who was taller,
nomena. These theories or models consist of con- the best procedure would be to measure them with
structs (denoted by the circles with C's in them), a yardstick or to stand them side by side and com-
linkages among and between the constructs (single pare their heights. You would be right on both
lines connecting the C's), and data that connect the counts. Ifi measured both John and Bill by asking
constructs with the empirical world (double lines). them how tall they were, you would have probably
The single lines represent conceptual or constitu- voiced less objection to my procedure-unless John
tive definitions in that a given construct is defined said he was taller while your observation of the two
in terms of other constructs in the set. The defini- men showed that Bill was definitely the taller.

t n,; . #t AIM. 14 IAULMd .q&I_I,ZJ4dMP.i g J d


'"" """\1"'"' 1%if!I!.HMM iii\MMMM21UU&P lt.M1# ,5$14&QIA .. JQh#IA&M4 r M a ,;q 1\kMt&q 4 I t 44Jilk , A b.WR Mh @,
Spring1992 75

Now the interesting thing about most psychologi- tency of repeated or equivalent measurements when
cal constructs is that we cannot rely on visual com- the measurements are made on the same object or
parisons to either confirm or refute a measure. We person. An example would be the use of an elastic
cannot see a salesp1erson's attitude, a rep's person- ruler to measure a man's height. It is unlikely that
ality characteristic, a salesperson's knowledge about on two successive measurements the observer would
a particular product, or other psychological charac- stretch the elastic ruler to the same degree of taut-
teristics such as intelligence, mental anxiety, or ness, and, therefore, the two measures would not
whatever. These characteristics are all part of the agree although the man's height had not changed.
representative's black box. Their magnitude must The distinction between systematic error and
be inferred from our measurements. Since we can- random error is critical because of the way the va-
not resort to a visual check on the accuracy of our lidity of measures is assessed. Validity is synony-
measures, we must rely on evaluating the proce- mous with accuracy or correctness. The validity of
dures used to determine the measure. Eye color is a measuring instrument is defined as "the extent to
certainly not height, but can we capture sales rep- which differences in scores on it reflect true differ-
resentatives' satisfaction with their job if we ences among individuals on the characteristic we
ask them directly how satisfied they are? Probably seek to measure, rather than constant or random
not, for reasons that will become obvious as we errors" (Selltiz, Wrightsman, and Cook 1976, p. 169).
continue our discussion. The ability to conclude When a measurement is valid, X0 =Xr, since there
that a salesperson's job satisfaction or any other is no error.
characteristic of the person has indeed been cap- The basic measurement problem is to develop
tured by the measurement depends on understand- measures in which the score we observe and record
ing measurement error and its assessment using actually represents the true score of the object on
evidence of the reliability and validity of the mea- the characteristic we are attempting to measure.
sure. This is much harder to do than to say. It is not
accomplished by simply making up a set of ques-
Classification and Assessment tions or statements to measure a salesperson's job
of Error satisfaction. Rather, the burden is on the researcher
to establish that the measure accurately captures
The ideal in measurement is to generate a score the characteristic of interest. The relationship be-
that reflects true differences in the characteristic tween measured score and true score is never es-
one is attempting to measure and nothing else. tablished unequivocally but is always inferred. The
What we in fact obtain, though, is something else. bases for such inferenees are two: (1) direct as-
A measurement, calli it X0 , for what is observed can sessment employing validity measures and (2) in-
be written as a function of several components: direct assessment via reliability measures (Peter
Xo=Xr+Xs+XR and Churchill 1986). Let us consider each of these
where sources of evidence in turn.
Xr represents the true score of the character-
istic being measured; Validity Measures
X8 represents systematic error; and
XR represents random error. As mentioned, a measuring instrument is valid to
The total error of a measurement is given by the the extent that differences in scores among objects
sum ofXg and~ and it is important to note that it reflect the objects' true differences on the charac-
has two components. Systematic error is also known teristic that the instrument tries to measure. We
as constant error, because it affects the measure- normally do not know the true score of an object
ment in a constant way. An example would be the with respect to a given characteristic. If we did
measurement of a man's height with a poorly cali- know it, there wouldl be no need to measure the
brated wooden yardstick. Random error is not object on that characteristic. What we do, therefore,
constant error but, rather, is due to transient as- is infer the validity of the measure by looking for
pects of the person or measurement situation. A evidence of its predictive, content, and construct
random error manifests itself in the lack of consis- validity.
76 Journal of Personal Selling & Sales Management

Predictive Validity spelling test. Further, the basis for your objection
probably would be the fact that all the words relate
The predictive approach to validation focuses on to the sport of football. Therefore, you could argue
the usefulness of the measuring instrument as a that an individual who is basically a very poor speller
predictor of some other characteristic or behavior of could do well on this test simply because he or she
the individual; it is thus sometimes called crite- is a football enthusiast. You would be right, of
rion-related validity. Predictive validity is ascer- course. A person with a basic capacity for spelling
tained by how well the measure predicts the crite- but with little interest in football might, in fact, do
rion, be it another characteristic or a specific be- worse on this spelling test than one with less native
havior. An example would be the Graduate Man- ability but a good deal more interest in football.
agement Admissions Test. The fact that this test is The test could be said to lack content validity, since
required by most of the major schools of business it does not properly sample the domain of all possible
attests to its predictive validity; it has proven to be words that could be used but is very selective in its
useful in predicting how well a student with a par- emphasis.
ticular score on the exam will do in an accredited The preceding example illustrates how content
MBA program. The test score is used to predict the validity is assessed, although not how it is estab-
criterion of performance. An example of an attitude lished. Content validity is sometimes known as
scale might be using scores that sales representa- "face validity" because it is assessed by examining
tives achieved on an instrument designed to assess the measure with an eye toward ascertaining the
their job satisfaction to predict who might quit. domain being sampled. If the included domain is
The attitude score would again be used to predict a decidedly different from the domain of the variable
behavior-the likelihood of quitting. Predictive as conceived, the measure is said to lack content
validity is determined strictly by the correlation validity. Theoretically, to capture a person's spell-
between the two measures; if the correlation is high, ing ability, we should ask the person to spell all the
the measure is said to have predictive validity. words in the language. The person who spelled the
Predictive validity is relatively easy to assess. It greatest number of words correctly would be said to
requires, to be sure, a reasonably valid measure of have the most spelling ability. This is a completely
the criterion with which the scores on the measur- unrealistic procedure. It would take several life-
ing instrument are to be compared. Given that times to complete. We, therefore, resort to sampling
such scores are available (for example, the grades the domain of the characteristic by constructing
the student actually achieves in an MBA program, spelling tests that consist of samples of all the pos-
the sales representative's quitting or not), all that sible words that could be used. Different samplings
the researcher needs to do is to establish the degree of items can produce different comparative perfor-
of relationship, usually in the form of some kind of mances by individuals, and consequently whether
correlation coefficient, between the scores on the a particular test assesses "true" spelling ability de-
measuring instrument and the criterion variable. pends on how well the test samples the domain of
Although easy to assess, predictive validity is rarely the characteristic. This is not only true for spelling
the most important kind of validity. We are often ability, but also holds for psychological character-
concerned with "what the measure in fact measures" istics in which we have an interest.
rather than simply whether it predicts accurately How can we ensure that a measure will possess
or not. content validity? We can never guarantee it because
it is partly a matter of judgement. We may feel
Content Validity quite comfortable with the items included in a
Content validity focuses on the adequacy with measure, for example, while a critic may argue that
which the domain of the characteristic is captured we have failed to sample from some relevant domain
by the measure. Consider, for example, the char- of the characteristic. Although we can never guar-
acteristic "spelling ability" and suppose that the antee the content validity of a measure, we can
following list of words was used to assess an severely diminish the objections of the critics. The
individual's spelling ability: quarterback, guard, key to content validity lies in evaluating the proce-
tackle, end, pass, fumble, punt, touchdown, ru~. dures that are used to develop the instrument
Now, you would probably take issue with this (Churchill1979).

"""""' UL,QJXid UU!4U:: ; t4 a, ; . tt!M 4. MBM4Mhk iiutt; 4t fl ZS rb+MW


'h4!llhtil1i&l,iiiW$RYA&;ibMbM .Jt04Zii%t ., AMMZ.S&MQ&?IJIAIM MM AMI
Spring 1992 77

Figure 2 to get at the same thing. Technically, this means it


Diagram Rela1ting the Constructs "Job should have convergent validity. It should also
Satisfaction" and "Intention to Quit" have discriminant vallidity which means it should
not correlate too highly with measures designed to
assess different things (Campbell and Fiske 1959).
Finally, the construct should behave as expected
X with respect to other constructs to which it is theo-
retically related which technically means it should
display nomological validity (Peter 1981).
For example, consider Figure 2 which depicts
the relationship between the constructs "job satis-
faction" and "intention to quit" and suppose that we
had developed the measure X to assess a sales
y representative's job satisfaction. Now, the construct
validity of the measure could be assessed by ascer-
taining the relationship between job satisfaction as
measured by X and intention to quit as measured
by Y. Those individuals for whom the X-scores are
Construct Validity low, indicating less job satisfaction, should express
higher likelihood of quitting than those with high
Construct validity is most directly concerned with
X-scores. If they do not, one would question the
the question of what the instrument is, in fact,
construct validity of the job satisfaction measure.
measuring. 'What eonstruct, concept, or trait un-
In other words, the construct validity of a measure
derlies the performance or score achieved on that
is assessed by whether the measure confirms or
test? Does the measure of attitude measure atti-
denies the hypotheses predicted from the theory
tude or some other underlying characteristic of the
based on the constructs. Is the evidence consistent
individual that affects his or her score? Construct
with the hypothesized linkages among the con-
validity lies at the very heart of scientific progress.
structs as captured in a model like Figure 2?
Scientists need constructs with which to communi-
Note though there is a circular logic in the argu-
cate. So do you and I. Thus, in sales management
ment. The argument rests on four separate propo-
we speak of salespeople's motivation, their person-
sitions:
alities, their attitudes, and so on. These are all
constructs that we use as we try to explain their 1. The constructs job satisfaction and inten-
behaviors. And although vital, they are also tion to quit are negatively related.
unobservable. We ean observe behavior related to 2. The scale X provides a measure of job satis-
these constructs, but we cannot observe the con- faction.
structs themselves. Rather, we operationally de- 3. The scale Y provides a measure of intention
fine the constructs in terms of a set of"observables." to quit.
When we agree on the operational definitions, pre- 4. X and Y correlate negatively .
cision in communication is advanced. Instead of Only the fourth proposition is directly examined
saying that what is measured by 25 items on a scale with empirical data. To establish that X measures
is the salesperson's role ambiguity, we can speak of job satisfaction, one must assume that propositions
the notion of role ambiguity. 1 and 3 are correct in addition to observing that X
While the measurement of constructs is vital to andY correlate negatively. Thus, one must have a
scientific progress, construct validity is the most good measure for intention to quit and the theory
difficult type of validity to establish. We need to relating job satisfaction and intention to quit must
ensure, through the plans and procedures used in be true. In sum, the fallacy in the argument is that
developing the instrument, that we have adequately the failure to observe the hypothesized relationship
sampled the domain of the construct and that there may be due to a lack of construct validity or incorrect
is internal consistency among the items of the do- theory. We often try to establish the construct
main. We also need to assess whether the measure validity of a measure, therefore, lby relating it to a
behaves as expected. That means it should correlate number of other constn1cts rather than simply one,
with or behave similarly to other measures designed and we also try to use those theories and hypoth-
78 Journal of Personal Selling & Sales Management

eses that are sufficiently well founded to inspire Stability


confidence in their probable correctness.
One of the more popular ways of establishing the
reliability of a measure is to measure the same
Reliability Measures
objects or individuals at two different points in time
The similarity of results provided by independent and to correlate the obtained scores. Assuming
but comparable measures of the same object, trait, that the objects or individuals have not changed in
or construct is called reliability. It is distinguished the interim, the two scores should correlate per-
from validity in that validity is represented in the fectly. To the extent that they do not, random
agreement between two attempts to measure the disturbances were operating in either one or both of
same trait through maximally different methods, the test situations to produce random error in the
whereas reliability is the agreement between two measurement. The procedure is known as test-
efforts to measure the same trait through maxi- retest reliability assessment.
mally similar methods (Campbell and Fiske 1959,
p. 83). If a measure was valid, there would be little Equivalence
need to worry about its reliability. If a measure is
valid, it reflects the characteristic that it is sup- The basic assumption in constructing an attitude
posed to measure and is not distorted by other fac- scale, say, is that when several items are summed
tors, either systematic or transitory. into a single attitude score, the items are measuring
Evaluating the reliability of any measuring in- the same underlying attitude. Each item can, in
strument consists of determining how much of the one sense, be considered a measure of the attitude,
variation in scores is due to inconsistencies in and the items should be consistent (or equivalent)
measurement (Peter 1979). Before discussing how in what they indicate about the attitude. Within a
this evidence is generated, one point needs empha- given scale, then, the equivalence measure of reli-
sis. Reliability involves determining the consis- ability focuses on the internal consistency or inter-
tency of independent or comparable measures of nal homogeneity of the set of items forming the
the same object, group, or situation. To the extent scale.
that a measure is reliable, it is not influenced by The earliest measure of the internal consistency
transitory factors. In other words, the more reliable of a set of items was the split-half reliability of the
the measure, the lower is XR in the equation for scale. In assessing split-half reliability, the total
observed scores. Note what this implies. A mea- set of items is divided into two equivalent halves;
sure could be reliable and still not be valid, since the total scores for the two halves are correlated,
x x
even ifXR equals zero, 0 =~ + 8 . The converse and this is taken as the measure of reliability of the
x
is, of course, not true. If the measure is valid, 0 = instrument. Sometimes the division of items is
~. because the measure is simply reflecting true made randomly, whereas at other times the even
scores without error. Thus, it is often said that (1) items are assumed to form one-half and the odd the
if a measure is valid, it is reliable; (2) if it is not other half of the instrument. The total score on the
reliable, it cannot be valid, since at a very minimum even items is then correlated with the total score
Xo =~ + XR; and (3) if it is reliable, then it may or obtained from the odd items.
may not be valid, because reliability does not account The necessarily arbitrary division of the items
for systematic error. In other words, although lack into equivalent halves produces an inherent weak-
of reliability provides negative evidence of the va- ness in split-half reliability. Each of the many
lidity of a measure, the mere presence of reliability possible divisions can produce different correlations
does not mean that the measure is valid. Reliabil- between the two forms or different reliabilities.
ity is a necessary, but not a sufficient, condition for Which division is correct or, alternatively, what is
validity. Reliability is more easily measured than then the reliability of the instrument? For example,
validity, though, and this accounts for the emphasis a ten-item scale has 126 possible splits or 126 pos-
on it over the years. The empirical evidence indi- sible reliability coefficients (Bohrnstedt 1970, p. 86).
cates that reliability is the most frequently investi- A more appropriate way to assess the internal
gated measurement property (Churchill and Peter homogeneity of a set of items is to look at all of the
1984). items simultaneously, using coefficient alpha. One
Spring 1992 79

reason is that coeffilcient alpha has a direct rela- using the Eviderwe
tionship to the most accepted and conceptually ap-
pealing measurement model, the domain sampling A logical question is how does the
model, which holds that the purpose of any par- nonmeasurement expert use the above information
ticular measurement is to estimate the score that to make judgments. The researcher's burden is
would be obtained if all the items in the domain were different depending on whether the measure is a
employed (Nunnally 1978, p. 194). The score that new one or one borrowed from another study.
any subject would obtain over the whole sample
domain is the person's true score Xfr
In practice, though, we do not use all of the items
New Measure
that could be used but rather only a sample of them. If the measure is new, then the burden on the
To the extent that the sample of items correlates researcher is to demonstrate to the scientific and
with true scores, it is good. According to the domain practitioner communiti,es that the measure has de-
sampling model then, a primary source of mea- sirable reliability and validity properties. One way
surement error is the inadequate sampling of the to do this is to follow a generally accepted paradigm
domain of relevant items. when developing the measure including: (1) defin-
Basic to the domain sampling model is the concept ing the construct and specifying its domain, (2)
of a very large correlation matrix showing all cor- generating items to sample the domain, (3) collect-
relations among the items in the domain. No single ing data and calculating the appropriate indexes by
item is likely to provide a perfect representation of which to assess the measure's reliability and validity
the concept just as no single word can be used to properties (Churchill1B79).
test for differences in subjects' spelling abilities and
no single question c:an measure a person's intelli-
gence. Rather, each item can be expected to have a Borrowed Measures
certain amount of distinctiveness or specificity even
though it relates to the concept. When borrowing measures from other studies,
The average correlation in this large matrix in- the researcher needs to at least look at the reliability
dicates the extent to which some common core exists and validity evidence that has been gathered in
in the items. The dispersion of correlations about support of the measure. The researcher also needs
the average indicates the extent to which items to weigh carefully that evidence. Many researchers
vary in sharing the common core. The key assump- borrow scales and simply assume that because the
tion in the domain sampling model is that all items, measure had attractive properties in other contexts
if they belong to the domain of the concept, have an that it can be used as i:s. They fail to realize that
equal amount of common core. This implies that scales developed in other situations may not display
the average correlation in each column of the hy- the same desirable properties in the new context.
pothetical matrix is the same and, in turn, equals For example, scales developed in other disciplines
the average correlation in the whole matrix. That or in classroom-like situations with student subjects
is, if all the items in a measure are drawn from the may not translate welll at all to corporate survey
domain of a single construct, responses to those environments involving salespeople. When they
items should be internally consistent or highly cor- borrow scales, researchers need to check their
related. Low inter-item correlations, on the other properties by calculating the appropriate reliabil-
hand, indicate that some items are not drawn from ity and validity coefficients. Only by doing so can
the appropriate domain and are producing error they avoid executing GIGO (garbage in, garbage
and unreliability. Coefficient alpha provides a out) routines and determine whether the conceptual
summary measure of the internal homogeneity that structures relating the eonstructs make sense given
exists among a set of items. Moreover, its square the empirical evidence . The new section on mea-
root equals the estimated correlation of the test with surement can only help in improving our under-
true scores (Nunnally 1978, p. 214). Thus, it is, or standing of the measures with which sales man-
should be, routinely calculated. agement practitioners and researchers work.
80 Journal of Personal Selling & Sales Management

References Peter, J. Paul (1979), "Reliability: A Review of Psychometric


Basics and Recent Marketing Practices," Journal of Mar-
Bohrnstedt, George W. (1970), "Reliability and Validity Assess- keting Research, 16 (February), 6-17.
ment in Attitude Measurement," in Attitude Measurement, - - - - - (1981), "Construct Validity: A Review of Basic
Gene F. Summers, eds., Chicago: Rand McNally. Issues and Marketing Practices," Journal of Marketing Re-
Campbell, Donald T. and Donald W. Fiske (1959), "Convergent search, 18 (May), 133-145.
and Discriminant Validation by the Multitrait-Multimethod - - - - - and Gilbert A. Churchill, Jr. (1986), "The Rela-
Matrix," Psychological Bulletin, 56, 81-105. tionship Among Research Design Choices and Psychomet-
Churchill, Jr., Gilbert A. (1979), "A Paradigm for Developing ric Properties of Rating Scales: A Meta-Analysis," Journal
Better Measures of Marketing Constructs," Journal of of Marketing Research, 23 (February), 1-10.
Marketing Research, 16 (February), 64-73. Selltiz, Claire, Lawrence L. Wrightsman, and Stuart W. Cook
- - - - - and J. Paul Peter (1984), "Research Design Ef- (1976), Research Methods in Social Relations, 3rd ed., New
fects on the Reliability of Rating Scales: A Meta-Analysis," York: Holt, Rinehart and Winston.
Journal of Marketing Research, 21 (February), 360-375. Torgerson, WarrenS. (1958), Theory and Methods of Scaling,
Nunnally, Jum C. (1978), Psychometric Theory, 2nd ed., New New York: John Wiley.
York: McGraw-Hill.

A qJM;;' ,.,;;;AiM II JZU t CJ 1M At&.$ ,$ . t l, 4fti.l$AJI& .A$ UiftA i I ikt I(ZU!Mi Qi,U& t 04 bit, M..-AMM41MA14 RW,QMZP Att 5a us t i a
!

You might also like