Professional Documents
Culture Documents
CONTENT
Importance of valid instrument What is validity? Types of validity evidence What is reliability? Difference between validity Types of reliability Measurement error
VALIDITY Defined as referring to the appropriateness, correctness, meaningfulness, and usefulness of specific inference researcher make based on data they collect. Validation is the process of collecting and analyzing evidence to support such inference.
TYPES OF VALIDITY
Face validity: Face validity simply means the validity at face value. As a check on face validity, test/survey items are sent to teachers to obtain suggestions for modification. Because of its vagueness and subjectivity, psychometricians have abandoned this concept for a long time.
CONTENT VALIDITY
Checking to make sure that you have picked questions that cover the areas you want to cover, thoroughly and well. Difficulties: important to ensure that all major aspect that cover by test item and in the correct proportion.
CONSTRUCT VALIDITY
Construct validity defines how well a test or experiment measures up to its claims. Convergent validity and discriminate validity are commonly regarded as subsets of construct validity. Convergent validity tests that constructs that are expected to be related are, in fact, related. Discriminate validity (or divergent validity) tests that constructs that should have no relationship do, in fact, not have any relationship.
CRITERION VALIDITY
Criterion validity assesses whether a test reflects a certain set of abilities. Concurrent validity measures the test against a benchmark test and high correlation indicates that the test has strong criterion validity. Predictive validity is a measure of how well a test predicts abilities. It involves testing a group of subjects for a certain construct and then comparing them with results obtained at some point in the future.
THREATS TO VALIDITY
Inappropriate selection of constructs or measures. Insufficient data collected to make valid conclusions. Too great a variation in data (can't see the wood for the trees). Inadequate selection of target subjects. Complex interaction across constructs. Subjects giving biased answers or trying to guess what they should say. Experimental method not valid. Operation of experiment not rigorous.
RELIABILITY
Reliability is the consistency of your measurement, or the degree to which an instrument measures the same way each time it is used under the same condition with the same subjects. In short, it is the repeatability of your measurement. Example: A measure is considered reliable if a person's score on the same test given twice is similar. It is important to remember that reliability is not measured, it is estimated.
ERROR OF MEASUREMENT
Error of measurement refer to variation in scores obtained by the same individuals on the same instrument.
TEST-RETEST
Test/retest reliability is evaluated by giving the same set of questions on two different occasions. If the results are consistent, then the measure is said to be reliable.
a class of children are given several tests that are intended to assess the same abilities. A week and a month later, they are given the same tests. With allowances for learning, the variation in the test and retest results are used to assess which tests have better test-retest reliability.
When equivalent form method is used, two different but equivalent forms of an instrument are administered to the same group of individual during the same time period.
INTERNAL-CONSISTENCY METHODS
KUDER-RICHARDSON APPROACH
Most frequently employed method to determined the internal consistency. Alpha coefficient: one split-half reliability and then randomly divide the items into another set of split halves and recomputed, and keep doing this until we have computed all possible split half estimates of reliability. Cronbach's Alpha is mathematically equivalent to the average of all possible split-half estimates, although that's not how we compute it. Limitation: not use full for heterogeneous test.
It shows the index the extend to which a measurement vary under changed circumstance. For example it will be smaller if it include only error due to different content (internal consistency or equivalent) or passage of time (test retest)