You are on page 1of 17

HISTORY OF PSYCHOLOGICAL TESTING

Rise of Civil Service Examination (2200 B.C - Chan Dynasty) Civil Service Examination first introduced in China; used to know ones proficiency for government positions. Areas tested: music, archery, horsemanship, writing, arithmetic, geography. RISE OF BRASS INSTRUMENT 1862 Wilhelm Wundt used a calibrated pendulum or thought meter (brass instruments) to measure the speed of thought. 1884 Francis Galton (Father of Mental Tests) administered the first anthropometric & psychometric test battery to thousands of citizens at the International Health Exhibit in London. Anthropometric and psychometric tests used to measure individual differences. 1890 James McKeen Cattell used the term mental test in a publication. 1892 Psychiatrist Emil Kraepelin who studied with Wundt, published his work regarding the use of a test that involves word association.

USE OF TESTS IN MENTALLY RETARDED & THE ADVENT OF IQ 1904 Charles Spearman proposed that intelligence consists of a single general factor g and numerous factors S1, S2, S3 & from it begins to lay the foundation for the concept of the test reliability and later for factor analysis. At the same time, Karl Pearson formulated the theory of correlation. 1905 Alfred Binet and Theodore Simon invented the first modern intelligence test. (Binet-Simon Scales Stanford-Binet Intelligence Test) 1908 Henry H. Goddard translated the Binet-Simon Scales from French into English. 1913 John Watson published psychology as the behaviorist views. It making behavioral observation a key tool in assessment. 1914 Stern and Terman introduced the formula: IQ = MA/CA x 100 Lewis Terman revised the Binet-Simon Scales and published the Stanford-Binet Intelligence test. Revisions were 1937, 1960, and 1986.

INVENTION OF NONVERBAL TESTS, USE OF IQ/GROUP TESTS IN MILITARY 1917 Robert Yerkes spearheaded the Army Alpha and Beta examinations used for testing WWI recruits. Army Alpha consists of native Americans, fluent in English, literate (written test) Army Beta consists of immigrants, illiterate (abstract test) Robert Woodsworth developed the Personal Data Sheet, the premiere personality test.

RISE OF PROJECTIVE TESTS, EDUCATIONAL TESTS, VOCATIONAL TESTS, AND OTHER LANDMARKS 1920 The Rorschach Inkblot Test is published by Hermann Rorschach. 1921 Psychological Corporation (the first major test publishing company is founded by Cattell, Thorndike, & Woodsworth with the goal of useful application of psychology). 1926 Florence Goodenough published the Draw-A-Man test. 1927 Kurt Goldstein developed neurodiagnostic tests (neuropsychological tests) with soldiers who suffered brain injury during WWI. 1931 L.L Thurnstone Multiple Factor Analysis 1938 Lauretta Bender Bender Gestalt Test (a neuropsychological test that uses 9 stimulus cards) 1939 Weschler Bellevue Intelligence Scale Weschler Adult Intelligence Scale (David Weschler) 1942 Minnesota Multiphasic Personality Inventory (MMPI) a clinical test used to know if one has a mental disorder; a paper and pencil test. 1943 Henry Murray his theory revolves around the needs of a person. He made the Thematic Apperception Test (TAT), a test that involves storytelling. 1948 Office of Strategic Services (OSS) used situational techniques for selection of officers. Situational test/techniques officers were given different situations and was asked to enact what they have to do; roleplaying 1949 WISC Rotter Incomplete Sentences Blank

Psychological Testing process of measuring psychology-related variables by means of devices or procedures designed to obtain a sample of behavior (Cohen & Swerdik, 2010). Variable what we want to measure Psychological Assessment gathering and integration of psychology-related data for the purpose of making psychological evaluation, accomplished through the use of tools (such as tests, interviews, case studies, behavioral observation) A problem solving process that begins with referral source and culminates into a psychological report.

Assessment involves judgment; general; broader than testing; entails complex/comprehensive approach. It is the evaluation through the use of instruments/tools Testing actual administering of psychological tests Testing is a part of assessment. Psychometrician/Psychometrist specialist in psychology and education who develops and evaluates psychological tests. Also known as test users. Test taker one who takes the test. Test user one who develops tests; administers/evaluates test data Tests measuring devices or procedures Features of a test: Content Format Administration Procedures this is usually found on the test manual Scoring & Interpretation Procedures Technical Quality & Utility

Assessment Tools Interview communication as a method of data gathering (e.g. face-to-face, telephone, panel/board). It requires note-taking as to the content. It follows a guide and is formal in nature. Transcription writing word-for-word the response of the client

Portfolio Assessment samples of ones ability and accomplishments composed of different medium as a tool for evaluation. It consists of the work samples of the client (such as canvas, paper, film, video, audio, etc.) Case History Data records, transcripts, and other accounts in written, pictorial, or other form, archival information and other data relevant to an assessee (such as medical records, TOR, etc.). Behavioral Observation monitoring the actions of others or oneself by visual or electronic means while recording quantitative and/or qualitative information regarding the assessees actions. 2 types of behavioral observation: 1. Naturalistic 2. Controlled/Laboratory Environment

Role-Play Tests assesses are directed to act as if they were in a particular situation and they are evaluated through expressed thoughts, behaviors, abilities, and other variables. This is also called situational performance test Computers used in test administration, scoring, and interpretation. Computer Assisted Psychological Assessment (CAPA) Computer Adaptive testing (CAT)

USES OF TESTING/ASSESSMENT 1. Selection/Screening 2. Placement 3. Classification categorization of people according to their skills 4. Description/Profiling 5. Decision-making 6. Tracking/Monitoring 7. Counseling/Self-Knowledge 8. Diagnosis/Prognosis 9. Treatment Planning/Intervention 10. Research & Evaluation 11. Program Development

12. Case Conferencing the professionals talk about the clients case 13. Expert Testimony/Forensic Use 14. Credentialing/Certification

MAIN TYPES OF PSYCHOLOGICAL TESTS 1. Intelligence Tests measure an individuals ability in relatively global areas and thereby help determine potential for scholastic work or certain occupations. 2. Aptitude Tests measure the capability for a relatively specific task or type of skill. It is a narrower form of ability testing. 3. Achievement Tests measure a persons degree of learning, success, or accomplishment in a subject or task. (e.g. board exams) 4. Personality Tests measure the traits, qualities, or behaviors that determine our individuality. (includes checklists, inventories, & projective techniques) 5. Career/Vocational Tests measure an individuals preference for a certain activity or topic and thereby help determine occupational choice. 6. Behavioral Procedures objectively describe and count the frequency of a behavior, identifying the antecedents and consequences of the behavior. It includes checklists, rating scales, interviews, structured observations. 7. Neuropsychological Tests measure cognitive, sensory, perceptual, and motor performance to determine the extent, locus, and behavioral consequences of brain damage.

TESTING PROCESS Standardized Procedures uniformity of procedures; usually found in the test manual WHY CONTROL THE USE OF PSYCHOLOGICAL TEST? To ensure that the test is used by a qualified examiner To prevent general familiarity with the test content which would invalidate the test.

LEVELS OF TEST USERS Level A paper & pencil test; can be utilized by psychology and education graduates

Level B for degree holders (MS/MA); with exposure to assessment; those who have background in psychometrics & statistics; covers Level A test and paper & pencil personality test. Level C projective tests; MA/PhD; had training and background in psychological testing & assessment.

GUIDELINES ON TEST ADMINISTRATION 1. Advance preparation of the examiner (reading the manual before the test) 2. Physical conditions 3. Rapport and test-taker orientation 4. Sensitivity to disabilities 5. Giving directions/manner of administration 6. Proper Timing

EXAMINER & SITUATIONAL VARIABLES 1. Examiner Personality 2. Rapport 3. Test Anxiety 4. Motivation 5. Test takers activities immediately preceding the test 6. Faking/Guessing 7. Test Sophistication 8. Training/Coaching

SOURCES OF INFORMATION ON TEST 1. Reference Books 2. Publishers Catalogues 3. Journals 4. Test Manuals

ETHICS 1. Right of informed consent Test takers have the right to know why they are evaluated, how the test data will be used, and what information will be released to whom. The consent must be in written form. 2. Right to be informed of test findings Test takers are also entitles to know the recommendations. Providing feedback also serves as therapeutic (oral feedbacking) 3. Right to privacy and confidentiality withholding information; keeping the clients disclosures confidential 4. Right to the least stigmatizing label least stigmatizing labels should always be assigned when reporting test results. Psychologists must take responsible ways to avoid harming their clients/patients and to minimize harm.

OTHER APPLICABLE ETHICAL CONSIDERATIONS 1. Bases for assessment Psychologists opinions must be based on actual data. All things written on the report should be data-driven. 2. Use of assessments assessment instruments must have validity and reliability. Assessment methods must be appropriate to the clients language. 3. Interpreting assessment results Take into account the purpose of assessment. 4. Assessment by unqualified persons 5. Maintaining test security test materials should be secured Test data refers to raw and scaled scores

PSYCHOMETRIC PROPERTIES OF PSYCHOLOGICAL TESTS Reliable test = Consistent test (dependable) Properties of a Good test Clear instructions for administration, scoring, and interpretation Economy in time and money in terms of administration, scoring, and interpretation One that measures what it purports to measure Having psychometric soundness, that is reliable and valid. *A test is considered unethical if it is not reliable and valid. A test with test manual is reliable and valid.

ASSUMPTIONS UNDER THE CLASSICAL TEST THEORY A score on a test is presumed to reflect not only the test takers true score on the attribute being measured but also the error. Error refers to the component of the observed test score that does not have to do with the test takers ability. X=T+E - formula to represent the presence of error

X = observed score (raw score) T = true score E = error (external/environmental factors) Environmental factors affects the test takers performance (e.g. lights, humidity, noise, etc)

SOURCES OF ERROR VARIANCE 1. Test Construction it is the way the test was developed. It involves item/content sampling. 2. Test Administration it is the ways the test was delivered/handled. It involves the test environment, test taker variables, and examiner-related variables. (The test taker must be relaxed before taking the test.) 3. Test Scoring & Interpretation always refer to the test manual. 4. Other potential sources of error

Reliability dependability or consistency. A test is reliable if it involves the test itself and the test score, the consistency of the measuring tool, and if it generates consistent test scores. Reliability Coefficient an index of reliability that shows the degree of error present. High reliability coefficient = good test Reliability Estimates: 1. Test-Retest Reliability (Coefficient of Stability) an estimate of reliability obtained by correlating pairs of scores from the same people on 2 different administration of similar tests.

CONDITIONS: Is appropriate when evaluating the reliability of a test that purports to measure something relatively stable over time such as personality test. The passage of time can be a source of error variance. *Avoid a very long time interval between the tests. The ideal time interval is 1-3 weeks. 2. Parallel Form/Alternate Forms Reliability (Coefficient of Equivalence) Parallel Form scores obtained on parallel or each form of the test correlate equally; 2 test sessions. Should have same validity (e.g. both IQ tests) and same length of test. Alternate Forms scores obtained on a different versions of the tests are correlated; 1 or 2 testing sessions. Equivalent tests by the way of content and level of difficulty. Same type of test but different authors. 3. Split-Half Reliability estimate of it is obtained by correlating two pairs of scores obtained from equivalent halves of a single test administered once. 3 steps in computing the estimate: a. Divide the test into equivalent halves. b. Calculate a Pearson r between scores on the 2 halves of a test c. Adjust the half-test reliability using the Spearman formula. Spearman Brown (SB) formula can estimate internal consistency reliability from a correlation of 2 halves of a test. 4. Inter-item/Internal Consistency refers to the degree of correlation among all items on a scale. An indicator of inter-item consistency is useful in assessing test homogeneity. Homogeneity refers to the degree to which the test measures a single factor (unifactorial) Heterogeneity refers to the degree to which a test measures different items/factors 2 formulas that measure test homogeneity & heterogeneity: KR 20 / KR 21 Cronbachs - Kuder Richardson - Lee Cronbach

(Coefficient Alpha)

5. Interscorer/Interrater Reliability is the degree of agreement or consistency between 2 or more scorers/judges/raters with regard to a particular measure.

*The way of determining the degree of consistency among scorers in the scoring of a test is to calculate a coefficient of correlation. (Pearson r)

SUMMARY OF RELIABILITY TYPES TYPES OF RELIABILITY NO. OF TESTING SESSIONS 2 NO. OF TEST FORMS SOURCES OF ERROR VARIANCE Administration, changes over time Test construction or administration Test construction; nature of split Item sampling; test heterogeneity Scoring & Interpretation; Scorer differences STATISTICAL PROCEDURES Pearson r/ Spearman rho Pearson r/ Spearman rho Pearson r/ SB Formula Coefficient Alpha

Test-Retest

Parallel Form/ Alternate Forms

1 (A) or 2 (P)

Split-Half Inter-item/ Internal Consistency Interscorer/ Interrater

Pearson r/ Spearman rho

INTERPRETING A COEFFICIENT RELIABILITY 95 90 = High Acceptability 89 80 = Moderate Acceptability 79 70 & below = Unacceptable

Validity an important attribute of a test for it refers to the degree to which it measures what it claims to measure (Gregory, 2000) A test is valid to the extent that inferences made from it are appropriate, meaningful, and useful.

Also concerns to what the test measures and to what a test score really means. High validity coefficient = valid test Major Categories of Validity: 1. Content Validity involves essentially the systematic examination of the test content to determine whether it covers a representative sample of the behavior domain to be measured. It is done when test items are judged by experts in the field to be suitable for their purpose. The test should be checked/reviewed by professionals to know if it is valid. Face Validity is not validity in the technical sense nor to what extent the test actually measures, but to what it appears to measure. Judgment based on tests appearance (did not undergo technical examination) Ex: quizzes/tests on magazines 2. Criterion Validity is demonstrated when a test is shown to be effective in estimating an examinees performance on some outcome measure or the criterion. It refers to the Concurrent Validity is measured by its correlations with other similar tests taken at the same time. (e.g IQ test is correlated with another IQ test) Predictive Validity tests scores are used to estimate outcome measures obtained at a later date. 3. Construct Validity a test designed to measure a construct must estimate the existence of an inferred, underlying characteristic based on a limited sample of behavior. Construct a theoretical, intangible quality or trait in which individuals differ. (e.g IQ, personality, aptitude, self-esteem, etc.) Approaches to construct validity: Analysis to determine whether the test items or subtests are homogenous and therefore measure a single construct. Study of developmental changes to determine whether they are consistent with the theory of the construct. Research to ascertain whether group differences on test scores are theory consistent.

Analysis to determine whether intervention effects on test scores are theory consistent. Correlation of the test with other related and unrelated tests and measures. Factor analysis of test scores in relation to other sources of information Factor analysis statistical procedure to check tests construct validity

Types of Construct Validity: a. Convergent Validity a test correlates highly with other variables or tests with which it shares an overlap of constructs. b. Discriminant validity a test doesnt correlate with variables or tests from which it should differ.

NORMS AND TEST STANDARDIZATION Norms based upon the distribution of scores obtained by a representative sample of examinees Scores on psychological tests are interpreted with reference to norms.

Norm-referenced test a test that follows a norm (e.g psychological tests) Criterion-referenced test score is compared on a certain criterion (e.g classroom tests) Norm group consists of a sample of examinees who are representative of the population for whom the test is intended. Sample representatives of the whole population Varieties of norms: Percentile rank Age equivalents Grade equivalents Standard scores

*Norms indicate the examinees standing on the test relative to the performance of other persons of the same age, grade, sex, etc. Norm-referenced tests refers to a method of evaluation and a way of deriving meaning from test scores by evaluating an individual test takers score and comparing it to scores of a group of test takers on the same test.

RAW SCORE TRANSFORMATIONS 1. Percentile & Percentile Ranks expresses the percentage of people in the standardization sample who scored below a specific raw score It indicates only how an examinee is being compared to the standardization sample and does not convey the percentage of questions answered correctly. Ranks in a group of 100 representative subjects

2. Standard Scores (z score) uses the standard deviation of the total distribution of raw scores as the unit of measurement. It expresses the distance from the mean in standard deviation units. It always retains the relative magnitude of differences found in the original raw scores.

3. T scores a standardized score with mean of 50 and standard deviation of 10. It eliminates fractions and negative signs. T score scales are common among personality tests. 4. Stanine (standard nine scale) all raw scores are converted to a single-digit system of scores ranging from 1-9. Mean = always 5 Variations of stanine: Sten scale with 5 units above and below the mean (expressed in 10 units) C scale consisting of 11 units

SELECTING A NORM GROUP 1. Random Sampling used in research and testing *Fishbowl technique 2. Stratified Random Sampling clusters, groupings Nonprobability sampling is NOT used in testing

Age Norms depicts the level of test performance for each separate age group in the normative sample Purpose: to facilitate same-aged comparisons The examinees performance is interpreted in relation to standardization of subjects of the same age.

Grade Norms - depicts the level of test performance for each separate grade group in the normative sample Very useful in school settings when reporting the achievement levels of school children

Local and Subgroup Norms derived from representative local examinees as opposed to a national sample. Subgroup norms consist of the scores obtained from an identified subgroup (e.g African American, Hispanics, Tagalog, Ilocano, etc.)

Criterion-referenced tests compares an examinees accomplishments to a well-defined content domain or standards (e.g classroom tests) The focus is on what the test taker could do rather than in comparison to the performance levels of others. Identify on examinees absolute mastery or non-mastery of specific behaviors or expectations.

PSYCHOLOGICAL TEST REPORT WRITING Psychological Report a document written as a means of understanding certain features about the person and current life circumstances in order to make decisions and to intervene positively in a problem situation It is the culmination of the practice of psychological testing and usually is the product of assessment procedures. It results from a process that starts with a referral source.

Psychological report writing is an important skill because of the potential, lasting impact of the written document. Psychological test report is a communication. Therefore, it must be written in a way that corresponds to the readers level of understanding and training. It must meet the criteria of clarity, meaningfulness, analysis/synthesis, and good clinical judgment. The ultimate goal is to provide helpful perspectives on the client, not to impress the referral source.

USES OF TESTING/ASSESSMENT 1. Selection/Screening 2. Placement

3. Classification categorization of people according to their skills 4. Description/Profiling 5. Decision-making 6. Tracking/Monitoring 7. Counseling/Self-Knowledge 8. Diagnosis/Prognosis 9. Treatment Planning/Intervention 10. Research & Evaluation 11. Program Development 12. Case Conferencing the professionals talk about the clients case 13. Expert Testimony/Forensic Use 14. Credentialing/Certification POSSIBLE REFERRAL SOURCES/RECIPIENTS OF PSYCHOLOGICAL REPORTS 1. Client themselves 2. Parents of minors/mentally deficient 3. Regular class/Special education teachers 4. Counselors and school administrators 5. Company employers/Human resource officers 6. Psychologists/Psychotherapists 7. Medical doctors 8. Court and correctional personnel 9. Social/Community/Mental health volunteers

CONDITIONS IN WRITING THE REPORT The report should be written with the needs of the referring person in mind and should explicitly target referral questions. Reports should provide information that is relevant to the readers work with the client but that may have not been requested.

Reports should communicate in a way that is appropriate to the reports intended recipient. Reports should affect the way the reader/recipient works with the client through its recommended specific course of actions. Effective reports stay within the bounds of expertise of the examiner. The best way to meet the needs of the client is to recommend immediate, consultation with appropriate professionals.

ORGANIZATION OF THE REPORT 1. Identifying Information/Demographic Information 2. Reason for Referral (e.g The client was referred to assessment as part of the requirements in the subject Psychological Testing 1) 3. Psychological Tests Given and the dates 4. Psychological Test Results/Interpretation Intellectual Functioning Career Interest Personality Purpose in Life

5. Summary of the Findings and Recommendation 6. Prepared by: *Name, Course/Year/Section*

Noted:

*Name of faculty-in-charge* Faculty-in-Charge

Identifying Information/Demographic Information Name of client, chronological age, sex, date of birth, place of birth, home address, current address, tel./contact number, educational background, school, occupation, company, marital status, citizenship, date of report, name of examiner

Reason for Referral This should cover the 1-2 statements/sentences of the reason for referral and specific statement of the question/s to be answered.

Psychological Test Results/Interpretation This should cover the list of assessment tolls with abbreviations and date when the test was taken. Results should be data-driven, should be based most of all from the test manual, but may also include the examiners analysis of the clients test results or direct responses. The test protocols must also be attached to the psychological test report. Words that may be used: tends, has the tendency, shows, likely, may, might, possibly, suggests, indicates, appears, seems, etc.

Summary and Recommendation This should cover the integration or interplay of all things and statements linking summary/conclusions to recommendations. It should also highlight or emphasize the overall psychological functioning of the client based on all the tests given (with the examiners own analysis as well)

You might also like