Professional Documents
Culture Documents
Assessments as Tools
• Assessment is a process of observing a sample of students’ behavior
and drawing inferences about their knowledge and abilities.
• We use a sample of student behavior to draw inferences about
student achievement.
Promoting learning
• Assessments as motivators
• Assessments as mechanisms for review
• Assessments as influences on cognitive processing- studying more
effectively for types of test items.
• Assessments as learning experiences
• Assessments as feedback
Reliability
• The extent to which the instrument gives consistent information about
the abilities being measured.
• Reliability coefficient- correlation coefficient +1 to -1
Standardization
• The concept that assessment instruments must have similar, consistent
content, format, & be administered & scored in the same way for
everyone.
• Standardized tests reduce error in assessment results & are considered
to be more reliable.
Validity
• The extent an instrument measures what it is designed to measure.
• Content validity- items are representative of skills described
• Predictive validity- how well an instrument predicts future
performance. SAT, ACT
• Construct validity- how well an instrument measures an abstract,
internal characteristic- motivation, intelligence, visual-spatial ability.
Essentials of testing
• An assessment tool may be more valid for some purposes than for
others.
• Reliability is necessary to produce validity.
• But reliability doesn’t guarantee validity.
Practicality
• The extent to which instruments are easy to use.
• How much time will it take?
• How easily is it administered to a group of children?
• Are expensive materials needed?
• How much time will it take?
• How easily can performance be evaluated?
Standardized tests
• Criterion-referenced scores show what a student can do in accord with
certain standards.
• Norm-referenced scores compare a student’s performance with other
students on the same task.
• Norms are derived from testing large numbers of students.
Norm-referenced scores
Grade-equivalents and age-equivalents compare a student’s
performance to the average performance of students at the same age/ grade.
Percentile ranks- show the percentage of students at the same age/
grade who made lower scores than the individual.
Standard scores- show how far the individual performance is from the
mean by standard deviation units.
Standard scores
• Normal distribution- bell curve
• Mean
• Standard deviation- variability of a set of scores.
• IQ scores
• ETS scores
• Stanines
• Z-scores
Standard deviation
• IQ scores- mean of 100, SD of 15
• ETS scores- (Educational Testing Service tests- SAT, GRE)
mean of 500, SD of 100
• Stanines- for standardized achievement tests- mean- 5, SD- 2
• z-scores- mean of 0, SD of 1- used statistically
definition -
In general, testing is finding out how well something works. In terms of human beings,
testing tells what level of knowledge or skill has been acquired. In computer hardware
and software development, testing is used at key checkpoints in the overall process to
determine whether objectives are being met. For example, in software development,
product objectives are sometimes tested by product user representatives. When the design
is complete, coding follows and the finished code is then tested at the unit or module
level by each programmer; at the component level by the group of programmers
involved; and at the system level when all components are combined together. At early or
late stages, a product or service may also be tested for usability.
At the system level, the manufacturer or independent reviewer may subject a product or
service to one or more performance tests, possibly using one or more benchmarks.
Whether viewed as a product or a service or both, a Web site can also be tested in various
ways - by observing user experiences, by asking questions of users, by timing the flow
through specific usage scenarios, and by comparing it with other sites.
Meaning of Evaluation
• problem.
• Meaning of evaluation:
• Any information obtained by any means on either the conduct or the outcome of
interventions, treatment or of social change projects is considered to be
evaluation.
Purpose of Evaluation
• -Verifying whether the benefits reached the people for whom the program was
meant .
Principles of Evaluation
• The following are some of the principles, which should be kept in view in
evaluation.
• 4. Evaluation must be done on a co-operative basis in which the entire staff and
the board members should participate (total participation).
• 5. As far as possible, the agency should itself evaluate its program but
occasionally outside evaluation machinery should also be made use of (external
evaluation).
• 6. Total overall examination of the agency will reveal strength and weaknesses.
(agency / program totality).
• 7. The result of evaluation should be shared with workers of the agency (sharing).
Stages in Evaluation.
• Formative evaluation or
• Exploratory evaluation or
• Need assessment.
• Concurrent evaluation
• Impact evaluation or
• Final evaluation.
• Formative Evaluation
• C) By Stages
• Grant giving bodies in order to find out how the money given is utilized by the
agency or how the program is implemented sent experienced and qualified
evaluators (inspectors) to assess the work E.g. Central social welfare Board
• Some donors may send consultants in order to see how far the standards laid
down are put into practice.
• Inter agency evaluation. In this type two agencies mutually agree to evaluate their
program by the other agency.
• One of the simplest and easiest methods of evaluation by getting first hand
information about the progress, performance, problem areas etc,. of a project from
a host of staff, line officers, field personnel, other specialists and public who
directly associated with the project. Direct observation & hearing about the
performance and pitfalls further facilitate the chances of an effective evaluation.
• From this one can understand the current status, performance, schedule, cost and
hold ups, deviations from the original schedule.
• This indicates the time schedule for implementation of the project. From this one
can understand any delay, the cost of delay and the ultimate loss.
• Informal Reports:
• Graphic presentations:
• Project Profiles:
• Areas of evaluation:
• Evaluation may be split into various aspects, so that each area of the work of the
agency, or of its particular project is evaluated. These may be,
• 1.Purpose 2.Programs 3.Staff 4.Financial Administration 5.General.
• Purpose:
• The review the objectives of the agency / project and how far these are being
fulfilled.
• Programs:
• Staff:
• The success of any welfare program / agency depends upon the type of the staff
an agency employs. Their attitude, qualifications, recruitment policy, pay and
other benefits and organizational environment. These are the areas which help to
understand the effectiveness of the project / agency.
• Financial Administration:
• The flow of resources and its consumption is a crucial factor in any project /
agency. Whether the project money is rightly consumed any over spending in
some headings, appropriation and misappropriation. These are some of the
indicators that reveal the reasons for the success or failures of any project.
• General:
• Factors like public relations strategies employed by the project / agency, the
constitution of the agency board or project advisory committee and their
contribution future plans of the agency are important to understand the success or
failures of any project.
Evaluation ……
• Relevance
• Possible questions:
• still valid?
• Efficiency
• Possible questions:
Effectiveness = A measure of the extent to which an aid intervention attains its objectives
Possible questions: To what extent were the objectives achieved/are likely to be
achieved? What were the major factors influencing the achievement or non-achievement
of the objectives? Impact = The positive and negative changes produced by an
intervention, directly or indirectly, intended or unintended. Possible questions: What has
happened as a result of the programme or project? What real difference has the activity
made to the beneficiaries? How many people have been affected? Sustainability =
Sustainability is concerned with measuring whether the benefits of an activity are likely
to continue after donor funding has been withdrawn . Possible questions: To what extent
did the benefits of a programme or project continue after donor funding ceased? What
were the major factors which influenced the achievement or non-achievement of
sustainability of the program or project?