Professional Documents
Culture Documents
WHAT IS A RUBRIC?
Rubrics (or "scoring tools") are a way of describing evaluation criteria (or "grading
standards") based on the expected outcomes and performances of students. Typically,
rubrics are used in scoring or grading written assignments or oral presentations;
however, they may be used to score any form of student performance. Each rubric
consists of a set of scoring criteria and point values associated with these criteria. In
most rubrics the criteria are grouped into categories so the instructor and the student
can discriminate among the categories by level of performance. In classroom use, the
rubric provides an "objective" external standard against which student performance
may be compared
WHAT IS INVOLVED?
Instructor Preparation
Medium to High.
Time:
Continuous; but students catch on fairly
Preparing Your Students:
quickly.
Variable. As students use rubrics, they
become better writers and oral presenters;
Class Time:
hence the time instructors spend evaluating
students' work is reduced.
Disciplines: All.
Class Size: All. Rubrics are easy to use in small classes,
and are particularly useful in large classes to
facilitate scoring large numbers of written or
oral assignments
Special
Classroom/Technical None.
Requirements:
Individual or Group
Both.
Involvement:
The level of analysis depends on the
instructor's intended goal of the assessment
task and the type of data desired about
students' performance. For detailed analysis
Analyzing Results: of students' responses, each section of the
rubric can be scored independently then
totaled. For a holistic analysis of students'
responses, all sections of the rubric can be
blended and an overall score assigned.
Rubrics must be readily available to students
before they begin an assignment or written
Other Things to Consider: test. Posting rubrics on the web and including
them in the course pack for in-class writing
promotes their usefulness.
Description
Assessment Purposes
Limitations
Step-by-Step Instructions
There are many routes to developing a useful scoring rubric, however, all of them
involve the following five steps:
Develop the goals for your course and daily class meetings
Select the assessment tasks that provide data aligned with the goals
Develop performance standards for each of the goals
Differentiate performances (categories) based on well-described criteria
Rate (assign weight or value to) the categories
To provide a useful example of how these steps "play out" in a real world context, I will
describe how I developed rubrics for my own introductory biology and ecology courses.
Examine the rubric for Quizzes and Homework. The criteria for responses fall
into two major categories: general approach and comprehension. Although
these two categories are not discrete as indicated by the dotted line between
them, students can see all of the itemized components of an exemplary
answer. These categories can be divided further. For example,
comprehension could be divided into content knowledge, conceptual
understanding, and reasoning and critical thinking skills (Freeman 1994).
Freeman (1994) includes communication skills as a category in rubrics.
Essentially, my rubrics cover the same categories; the difference is in the
number of columns used.
Notice, when it is possible to quantify the categories, I did so. So, for
example, the criteria for acceptable style and grammar in an exemplary
answer is based on no errors.
Variations
Student Roles
I involve students in a dialogue about criteria that we use in any rubric. Students
gain a keen sense of my expectations for assessment by explicitly understanding the
criteria and by contributing to the modification of criteria in a rubric to enhance clarity.
Consider letting students develop class rubrics for certain assessments. When
students understand and describe criteria for their own performance, they are often
better at attaining those standards. My students developed the rubric for the poster
displays of their laboratory research. To accomplish this, they walked around the
biology department looking at the variety of posters displayed on the walls and then
determined their own criteria for what makes a quality poster presentation. We
collated this information and designed a rubric for content as well as format.
Students use rubrics when completing any assessment task for the course such as
writing in class, writing on an exam, designing homework, completing and
investigation, preparing a research paper.
Faculty Roles
The critical factor for faculty to consider is that assessments must be linked to the
goals of the course. For example, if the goal is for students to demonstrate their ability
to design a testable hypothesis in a particular content area - asking students to
actually write a testable hypothesis would provide meaningful feedback. The recurring
question we must ask is, "Does this evidence convince us that students understand
how to write a testable hypothesis?"
Include rubrics on your web site and in your course packs. Students should refer to
rubrics while they are completing any assessment task.
Rubrics are dynamic and involve a major up-front investment of time.
You must provide students repeated instruction on how to use rubrics as well as
how to achieve each of the criteria.
The advantage of rubrics is that you and the students have well defined pathways
for gathering evidence that they have achieved the goals of the course. If either you or
your students are dissatisfied with the evidence or criteria, the rubrics should be
revised.
Analysis
Rubrics are scoring tools that enable me to assign points to students' assignments and
tests. Students' accumulation of points determines their grade in the course. Each
assignment, quiz, or test is weighted in terms of value in the overall course evaluation.
For example, daily writing samples (quizzes) are worth 5 points, twice weekly, 15
weeks per semester; hence a student can earn a maximum of 75 points for daily
performance. The pattern of students' performance is consistent from semester to
semester. At the beginning of each semester, many students' responses are below
college-level. As students begin to understand the criteria and practice writing, they
attain college-level work or exemplary performance on short, five-point assignments or
quizzes. A key strategy in promoting improvement by all students is peer review within
their cooperative groups.
However:
Time - rubric development requires time up front, but the payoff is in increased
performance by the majority of students and which ultimately leads to less
instructor time in assessment.
Criteria - qualitative scales are more difficult to define than quantitative scales.
Practice - both students and faculty need to practice and refine the use of
rubrics for multiple types of assignments. Time to do this in class will affect
"coverage."
We need confidence in the quality of the data we gather about our students if we want
to justify our subsequent decisions about teaching. Many of us who teach introductory
science courses are dissatisfied with the type of evidence we are collecting about our
students' learning. We admit that data from multiple choice tests measure inert bits of
knowledge and some comprehension but provide us incomplete and inadequate
feedback about our students' learning. We would like to use alternative forms of
assessment to gather multiple, substantive forms of data about active student
learning, such as understanding, analysis, reasoning, and synthesis (Ebert-May et al
1997). These kinds of assessments include short answer items, essays, minute papers,
oral communication, poster presentations, laboratory projects and research papers,
but because of large class sizes and individual research priorities we have limited time
to evaluate extended responses from students.
Assessment is learning. We and our students both benefit from meaningful assessment
information about the achievement of the broader course goals. Multiple assessment
strategies can be implemented to provide evidence that students have or have not
learned, have or have not accomplished the goals of the course. Rubrics help us set
well-defined standards for our students, provide students guidelines for achieving
those standards and facilitate grading extended written and oral responses. This
feedback provides us data to interpret and make informed decisions about our
students' learning and our own teaching practice, similar to the process of data
evaluation that we use daily in our scientific research.
Links
Diane Ebert-May, Professor, Department of Plant Biology, Michigan State
University, East Lansing, Michigan.
email: ebertmay@pilot.msu.edu
Sources
Ebert-May D, Brewer C, Allred S. 1997. Innovation in large lectures-teachings for
active learning. Bioscience 47: 601-607.
Freeman, RHL. 1994. Open-ended questioning: a handbook for educators. Menlo Park,
California; Reading, Massachusetts; New York; Don Mills, Ontario; Wokingham,
England; Amsterdam; Bonn; Sydney; Singapore; Tokyo; Madrid; San Juan; Paris;
Seoul; Milan; Mexico City; Taipei. Addison-Wesley Publishing Company (The book is
published by Innovative Learningª, an imprint of the Addison-Wesley Alternative
Publishing Group.)
King PM, Kitchener KS. 1994. Developing reflective judgement: understanding and
promoting intellectual growth and critical thinking in adolescents and adults. San
Francisco (CA): Jossey-Bass Publishers.
Novak JD, Gowin DB. 1984. Learning how to learn. New York: Cambridge University
Press.
WHAT IS AN INTERVIEW?
A formal interview consists of a series of well-chosen questions (and often a set of
tasks or problems) which are designed to elicit a portrait of a student's understanding
about a scientific concept or set of related concepts (Southerland, Smith & Cummins,
2000). The interview may be videotaped or audiotaped for later analysis.
WHAT IS INVOLVED?
Several hours required to develop a set
Instructor Preparation ofgood questions, tasks and problems sets.
Time: Additional time to locate appropriate props
and recording equipment, if desired.
Interviews are most fruitful when the student
has developed a good rapport with you. It is
Preparing Your Students:
essential that the student feels relaxed and at
ease.
One-on-one or small group interviews may be
conducted in less than an hour in your office
Class Time: or other convenient "private space." Some
practice will reduce the time required to
conduct a good interview.
No disciplinary constraints. Appropriate for all
Disciplines:
STEM fields.
Normally, structured interviews are conducted
outside of class. It is important that subjects
Class Size: be carefully selected to represent a range of
ability and interest levels among students
enrolled in a course.
Special
Interview protocol, props, recording
Classroom/Technical
equipment and small private space.
Requirements:
The most useful interviews are those
conducted with individuals or small groups
Individual or Group
outside of class. Sometimes this is done well
Involvement:
in laboratory sections, but TAs will need
special training or assistance.
For "formative" assessment, the instructor
may want to review taped interviews with
special attention to potential
Analyzing Results:
"misconceptions." If used for "summative"
evaluation, a type of "scoring rubric" may be
developed.
Other Things to Consider: None.
Description
In its simplest form a structured interview is simply one person asking another person
a series of questions about a carefully selected concept/topic or asking her to perform
a task. Any materials to be used (props, problems, etc.), many of the questions to be
asked, and some responses from the interviewer to expected statements or actions of
the interviewee are carefully planned in advance. Importantly, however, the
interviewer is free to ask additional questions that focus on issues arising during the
course of the interview. It is this freedom to follow the interviewee, to ask for
clarifications, and to focus on errors, misconceptions, and gaps in knowledge, that
makes the interview so much more fruitful than more traditional methods of
assessment.
During a structured interview, the instructor uses a set of questions, called "probes"
(and sometimes selected photographs or other props) designed in advance of the
interview to elicit a portrait of the learner's understanding about a specific
concept/topic (e.g., evolution; molecular/kinetic theory; plate tectonics; binary stars;
Newton's laws). The student may be asked to use her own words to explain a concept
(e.g., "What is natural selection?") but is typically required to go beyond simple
recognition of a concept to construct a detailed personal explanation. Typically the
student is also asked to use that concept to solve a problem or other application task
(e.g., "Explain why cave fish have no color"). Valuable information is often obtained
not only from listening to what the interviewee says, but also from observing what she
does, including facial expressions and other body language.
Assessment Purposes
Structured interviews may serve many functions, among them:
It is also important to note that the goal of the interview is to describe how a student
understands a scientific concept or phenomenon, and not simply to provide a
measurement of the degree to which this understanding approximates the scientific
explanation. Thus, interviews are typically used to provide the instructor with insight
about students' understandings in order to refine and target instruction ("formative
assessment") rather than to evaluate the knowledge of individual students for
purposes of assigning a grade ("summative assessment").
Limitations
Structured interviews are used to describe individual student's understandings, and are
best conducted individually with students; thus time is a major inhibiting factor in
using structured interviews to inform teaching. To prevent this issue from being
prohibitive, selective sampling of a broad range of students in a classroom may be
employed to make the technique more practical, yet still provide a portrait of how
different students in a class are engaging with course material.
Teaching Goals
Structured interviews are very powerful tools for gaining insight into students' thinking.
They are especially useful in diagnosing "learning errors", "misconceptions", and
limitations in reasoning and critical thinking. With some training and practice, teaching
assistants may be encouraged to use interviewing strategies in small groups and
laboratory sections of large classes. Students themselves often find that knowledge of
interviewing is useful in collaborative learning environments.
Step-by-Step Instructions
Several types of interview strategies have been developed for use in STEM disciplines
(Southerland, Smith & Cummins, 2000). However, the "Interview about Instances and
Events" (White & Gunstone, 1992) is possibly the most widely used format for probing
understanding about single concepts. In this interview, the student is presented with a
set of 10-20 line-drawings, photographs or diagrams that depict examples and
counterexamples of natural objects (e.g., a mammal; a volcano; a planetary system; a
molecule) or events (eg. a burning candle; a moving automobile; a girl throwing a
baseball).
Revealing one drawing at a time, the student is asked to indicate whether it depicts an
example of the concept in question, and to provide a rationale or justification. For
example, consider a baseball in flight: Is there a "force" on the ball? What makes you
say that? Tell me more about that. Or consider a burning candle: Is this "sublimation"?
Why do you think that? Can you say some more about that? After each question, the
instructor gently probes further into the reasoning the student uses and encourages
him/her to elaborate on the responses to provide as complete a picture as possible of
the student's understanding. A few general suggestions for conducting successful
interviews:
1. The interview should begin with a focus question that requires application of the
concept to be investigated, without forcing the student into an explicit
definition. A more traditional assessment might ask the student to choose the
correct definition of the concept from among four choices or to write down a
definition of the concept. The more indirect approach of a structured interview
is usually more productive because it allows the student to evince her
understanding rather than relying on memorized, rote definitions. This also
enables the instructor to gain an idea of how the student applies the implicit
concept.
2. Do not force the student into a specific response to each graphic. If the student
needs to "waffle" in her answer, she should be allowed to do so. If the student
does not have an understanding of the concept that allows her to make a
decision about a specific instance, do not force her to choose. This lack of
understanding is an important piece of her "conceptual framework".
3. Specific definitions of the concept, if needed, should be sought only after
understanding the student's response to the focusing questions. Again, this
prevents students from early closure on a rote definition. Thus, in our example,
it would be inappropriate to ask, "Well, what is a force (or sublimation)?"
4. It is important for the interviewer to wait at least 3 to 5 seconds after each
prompt before trying to interpret the question or ask another. Classroom
research has shown that when this "wait time" is observed, both the length of
the student responses and the cognitive level of the responses increases (Rowe,
1974).
Variations
As mentioned previously, structured interviews are used to describe individual
students' understandings of specific scientific concepts and the degree to which they
can apply that understanding. Different interview probes allow for the investigation of
different degrees of student understanding.
Instances Interviews
In Interviews about Instances, a student is presented with a specific set of examples
and counterexamples of the concept of interest and is asked to identify which cases
are examples of the concept, and then to explain that decision. For practical reasons
the examples are usually graphics such as line pictures, drawings, or diagrams.
Prediction Interviews
Prediction Interviews require students to anticipate an outcome of a situation and
explain or justify that prediction. The strength of this kind of interview is that it focuses
on the ways a student can apply her personal meanings of the concept. And because
they require application, prediction interviews are very useful in teasing out what has
been learned by rote with minimal understanding from what is meaningful knowledge.
Sorting Interviews
In a Sorting Interview, the student is presented with a group of objects and asked to
sort them according to specific instructions. This exercise can be structured in many
different ways to match the purpose of the assessment. For example, the interviewer
may present a series of graphics depicting some natural phenomenon. The student
may then be asked to select any number of cards to be used in any order to explain
the phenomenon. Alternatively, a student may be presented with a set of genetics,
physics or chemistry problem cards and asked to sort them according to similarity
(e.g., Smith, 1992). As with other kinds of interviews described in this CAT, the
student is encouraged to talk about her reasoning as she attempts to construct an
explanation for her sorting.
Analysis
Note-taking during an interview can be beneficial, but it generally provides only a
superficial picture of a student's meaning. Instead, it is usually beneficial to record the
interviews, allowing for more intensive data analysis. As with most classroom
assessment activities, analysis of interview data may be accomplished in a variety of
ways, with some methods capturing a richer and more multilayered perspective than
others.
In order to analyze the results of structured interviews, we suggest that the instructor
attempt to put her expectations aside to the extent possible, and instead review the
tape or read the transcript with a fresh "eye," allowing important trends from the
learner's responses to emerge. Ideally, a sample of interview transcripts should be
reviewed several times, so that ideas emerging from one review can inform
subsequent readings. As strong trends are noted throughout several interviews,
negative examples (occasions for which the tentative trend fails to hold true) should be
searched for. This inductive approach to data analysis, i.e., looking for similarities and
differences in sets of data, allows for a more informative and reliable portrait of
learners to emerge.
However,
WHAT IS INVOLVED?
Description
A concept map is a two-dimensional, hierarchical node-link diagram that depicts the
structure of knowledge within a scientific discipline as viewed by a student, an
instructor or an expert in a field or sub-field. The map is composed of concept labels,
each enclosed in a box or oval; a series of labeled linking lines, and an inclusive,
general-to-specific organization. By reading the map from top to bottom, an instructor
can:
In addition to these applications in assessment, faculty have also used concept maps
to organize their ideas in preparation for instruction, as a graphic organizer during
class, and as a way to encourage students to reflect on their own knowledge and to
work together and share their understandings in collaborative group settings.
Figure 1: Concept Map Of Concept Maps
Click here to see a larger version of this graph.
Assessment Purposes
To investigate how well students understand the disciplinary-acceptable connections
among concepts in a subject, to document the nature and frequency of students'
misconceptions, and to capture the development of students' ideas over time.
Limitations
Concept maps provide a useful and visually appealing way of depicting the structure of
conceptual knowledge that people have stored in long-term memory. As a result, they
offer a readily accessible way of assessing how well students see "the big picture."
They are not designed to tap into the kind of process knowledge that students also
need to solve novel problems or for the routine application of algorithmic solutions.
Because they probe an individual's or a group's cognitive organization, they are very
idiosyncratic and difficult to compare, either among individuals or groups, or across
time for the same individuals or groups.
Figure 2: Jason's Concept Map on the Human Circulatory System
[From Mintzes, Wandersee & Novak, 1998]
Teaching Goals
Instructional Tool.
The instructor can present "expert" concept maps to the whole class to highlight key
concepts and connections. These should be more detailed and flow from the global
maps executed for the course design. Concept maps can then serve as "advanced
organizers" (to preview material) and also for review. An instructor can continuously
refer to a concept map in class to show how to "grow" the connections, and to keep
the instruction focused. Caveat: At first, students will find concept maps very strange
and may even try to memorize them, rather than use them as a thinking tool.
Learning Tool.
Ask students to construct their own concept maps covering a section of the course
material from class or the textbook. Most (if not all!) of them will probably never have
seen a concept map before, and many have not developed the learning skills needed to
construct them. As a result, the instructor will need time (either in class, or perhaps in
the lab) to work with groups and individuals. The impact of student-created concept
maps is so powerful that it is worth the investment of time!
Step-by-Step Instructions
Variations
Collaborative Concept Mapping.
Sometimes the frustration levels can be very high when concept mapping is first
introduced, especially in large classes of relative novices. To counter some of this
anxiety and to encourage students to reflect on their own thinking, ask groups of 3 or
4 students to work together on a concept map. This exercise is often a very rewarding
and rich learning experience as peers argue, debate, and cajole each other. The result
is a genuine effort to negotiate the meaning of scientific concepts, attempting (as
scientists do) to reach consensus, or to stake out different points of view. The power of
the process resides in the interpersonal sharing of ideas, which are made explicit to the
instructor.
Analysis
To start, we suggest that you focus primarily on the qualitative aspects of students'
concept maps with emphasis on the accuracy or validity of the knowledge students
represent. Among the questions you might ask yourself are the following:
Concept maps help students focus on the "big picture", enabling them to devote
more of their time to conceptual understanding rather than rote learning
Concept maps force students (and instructors!) to make valid connections
among concepts
They provide a low tech (cheap!) vehicle that enables students to represent
graphically their knowledge, and to share it with the instructor and other
students
They shift the emphasis from inert, static knowledge to contextually-embedded
knowledge; from isolated facts to theoretical frameworks of related concepts
In addition to their role as assessment tools, concept maps offer a useful way to
help students "learn how to learn"; they also serve as useful vehicles for course
development and as graphic organizers before, during and after instruction
However:
Comparisons among students are more difficult because concept maps tend to
reveal the idiosyncratic way that students view a scientific explanation, as a
result...
Evaluation can become more time-consuming for the instructor, especially in
large classes, unless some variation (such as Select & Fill-in) is adopted
If you score maps, you must use a consistent (and tested) scheme
Students who have developed a strong facility for rote learning of verbal
knowledge sometimes find concept maps intimidating
Constructing concept maps is a demanding cognitive task that requires training
Portfolios Defined
Portfolios are personalized long term documentation of student mastery of course material. An
essential element of portfolios are student reflections on their own learning and progression
towards the mastery of the material documented in the portfolio. As such, portfolios are windows
on the metacognitive process of students.
Types of Portfolios
In order to be useful as assessment tools the portfolio should include entries that demonstrate a
progression of student understandings and ultimately mastery of the concepts. Huba and Freed
(2000) identify two different portfolio types.
All-Inclusive Portfolios that contain a complete record of all work done by a student in a
course or program.
Selection Portfolios that are focused on documenting the achievement of mastery of
specific course goals/objectives set by the course instructor.
Description
Student portfolios are a collection of evidence, prepared by the student and evaluated
by the faculty member, to demonstrate mastery, comprehension, application, and
synthesis of a given set of concepts. Accordingly, portfolio assessment strategies
substantially increase the rigor of an introductory science or mathematics course. For
example, in a physics course, this might include quantitative analysis of a video
showing motion. In a geology course, this might include an analysis of the impact of
agriculture on the community's water quality using locally acquired data. Students
must organize, synthesize, and clearly describe their achievements and effectively
communicate what they have learned. The evidence can be presented in a three-ring
binder, as a multimedia tour, or as a series of short papers.
Assessment Purposes
The overall goal of the preparation of a portfolio is for the learner to demonstrate and
provide evidence that he or she has mastered a given set of learning objectives. More
than just thick folders containing student work, portfolios are typically personalized,
long-term representations of a student's own efforts and achievements. Whereas
multiple-choice tests are designed to determine what the student doesn't know,
portfolio assessments emphasize what the student does know.
Limitations
Portfolio assessments provide students and faculty with a direct view of how students
organize knowledge into overarching concepts. As such, portfolios are inappropriate for
measuring students' levels of factual knowledge (i.e., recall knowledge) or for drill-
and-skill activities and accordingly should be used in concert with more conventional
forms of assessment. Similarly, student work completed beyond the context of the
classroom is occasionally subject to issues of academic dishonesty.
Teaching Goals
1. The size of the human population, and the causes for change in its
size in various areas of the world.
2. The source, use, pollution and cleanup of the worlds water resources.
3. The origin and evolution of soils, and the way soils are affected by
agriculture.
4. Current and alternative sources of food.
5. The origin, advantages and disadvantages of current sources of
energy.
6. The origin, operation and potential for alternative sources of energy.
7. The causes of extinction and the processes which control the rate of
extinction.
8. Factors which control the use of land by people.
9. The geologic processes which cause earthquakes, and the potential for
predicting and preventing such events.
10. The origin, extraction and importance of ores.
11. The composition, management and recycle potential for solid &
hazardous waste material.
12. The origin, evolution and productivity of coastal areas.
13. The impact of human activities on coastal areas.
14. The origin, effect and remediation of atmospheric pollution.
15. How humans affect the earth's environment.
Step-by-Step Instructions
Carefully construct and distribute 12-25 overarching learning objectives for the
course.
Decide if a portfolio supports student learning and assessment for these
objectives.
Determine if the portfolio is primarily a learning activity or an assessment tool.
Inform students of your expectations that students have the opportunity to
clearly demonstrate to the professor that course learning objectives have been
attained.
Require that each piece of evidence must be clearly labeled as to which
objective the evidence pertains.
Require that each piece of evidence must be accompanied by a written
paragraph of rationale and a separate written paragraph of self-reflection.
Emphasize to students that it is their responsibility to clearly demonstrate
mastery of the learning objectives for this course.
Score each item of evidence in the portfolio according to a scheme that has
been distributed to the students when the portfolios are initially assigned.
Variations
Showcase Portfolios
A showcase portfolio is a limited portfolio where a student is only allowed to present a
few pieces of evidence to demonstrate mastery of learning objectives. Especially useful
in a laboratory course, a showcase portfolio might ask a student to include items that
represent: (i) their best work; (ii) their most interesting work; (iii) their most improved
work; (iv) their most disappointing work; (v) and their favorite work. Items could be
homework assignments, examinations, laboratory reports, news clippings, or other
creative works. An introductory letter that describes why each particular item was
included and what it demonstrates makes this type of portfolio especially insightful to
the instructor.
Checklist Portfolios
A checklist portfolio is composed of a predetermined number of items. Often, a course
syllabus will have a predetermined number of assignments for students to complete. A
checklist portfolio takes advantage of such a format and gives the students the choice
of a number of different assignment selections to complete in the course of learning
science. For example, instead of assigning exactly 12 sets of problems from the end of
each text chapter, students could have the option of replacing several assignments
with relevant magazine article reviews or laboratory reports that clearly demonstrate
mastery of a given learning objective. Additionally, class quizzes and tests can become
part of the portfolio if that is what is on the checklist of items to be included. A sample
checklist might require a portfolio to have 10 correctly worked problem sets, two
magazine article summaries, two laboratory reports, and two examinations in addition
to self-reflection paragraphs where the student decides which objectives most closely
fit which assignments.
Open-Format Portfolios
An open-format for a portfolio generally provides the most insightful view of a
student's level of achievement. In an open-format portfolio, students are allowed to
submit anything they wish to be considered as evidence for mastery of a given list of
learning objectives. In addition to the traditional items like exams and assignments,
students can include reports on museum visits, analysis of amusement park rides,
imaginative homework problems, and other sources from the "real world". Although
these portfolios are more difficult for the student to create and for the instructor to
score, many students report that they are very proud of the time spent on such a
portfolio.
Analysis
Because each portfolio is individualized, student assessment must be compiled by
looking at the portfolio's contents relative to the course learning objectives. Each piece
of evidence should be graded according to a predetermined scheme. The items can be
scored discretely as a 0, 1, 2, or 3 based on the grader's judgment about the student's
presentation as related to the stated learning goals. (A larger scale can be used, but
the reliability of different faculty giving the student the same score decreases.)
Grading Criteria
Each individual piece of evidence will be graded according to the following
scale:
Grading Rubric
The overall portfolio is scored as follows as an indication of the extent to
which the portfolio indicates that the student has mastered the 15 course
objectives listed elsewhere in the syllabus:
Grade:
Rubric:
Viewing student portfolios from this perspective drastically changes the emphasis from
collections of facts to encompassing concepts. Such a grading procedure also shifts
responsibility for demonstrating competence from the instructor to the student.
Effectively shifting this responsibility affects comments placed in the portfolio by the
grader; comments are directed toward improving the next submission as well as
indicating the inadequacies of the current evidence.
However:
Many teachers dislike preparing and grading exams, and most students dread taking
them. Yet tests are powerful educational tools that serve at least four functions. First,
tests help you evaluate students and assess whether they are learning what you are
expecting them to learn. Second, well-designed tests serve to motivate and help
students structure their academic efforts. Crooks (1988), McKeachie (1986), and
Wergin (1988) report that students study in ways that reflect how they think they will
be tested. If they expect an exam focused on facts, they will memorize details; if they
expect a test that will require problem solving or integrating knowledge, they will
work toward understanding and applying information. Third, tests can help you
understand how successfully you are presenting the material. Finally, tests can
reinforce learning by providing students with indicators of what topics or skills they
have not yet mastered and should concentrate on. Despite these benefits, testing is
also emotionally charged and anxiety producing. The following suggestions can
enhance your ability to design tests that are effective in motivating, measuring, and
reinforcing learning.
A note on terminology: instructors often use the terms tests, exams, and even quizzes
interchangeably. Test experts Jacobs and Chase (1992), however, make distinctions
among them based on the scope of content covered and their weight or importance in
calculating the final grade for the course. An examination is the most comprehensive
form of testing, typically given at the end of the term (as a final) and one or two times
during the semester (as midterms). A test is more limited in scope, focusing on
particular aspects of the course material. A course might have three or four tests. A
quiz is even more limited and usually is administered in fifteen minutes or less.
Though these distinctions are useful, the terms test and exam will be used
interchangeably throughout the rest of this section because the principles in planning,
constructing, and administering them are similar.
General Strategies
Spend adequate amounts of time developing your tests. As you prepare a test,
think carefully about the learning outcomes you wish to measure, the type of items
best suited to those outcomes, the range of difficulty of items, the length and time
limits for the test, the format and layout of the exam, and your scoring procedures.
Match your tests to the content you are teaching. Ideally, the tests you give will
measure students' achievement of your educational goals for the course. Test items
should be based on the content and skills that are most important for your students to
learn. To keep track of how well your tests reflect your objectives, you can construct
a grid, listing your course objectives along the side of the page and content areas
along the top. For each test item, check off the objective and content it covers.
(Sources: Ericksen, 1969; Jacobs and Chase, 1992; Svinicki and Woodward, 1982)
Try to make your tests valid, reliable, and balanced. A test is valid if its results
are appropriate and useful for making decisions about an aspect of students'
achievement (Gronlund and Linn, 1990). Technically, validity refers to the
appropriateness of the interpretation of the results and not to the test itself, though
colloquially we speak about a test being valid. Validity is a matter of degree and
considered in relation to specific use or interpretation (Gronlund and Linn, 1990). For
example, the results of a writing test may have a high degree of validity for indicating
the level of a student's composition skills, a moderate degree of validity for
predicting success in later composition courses, and essentially no validity for
predicting success in mathematics or physics. Validity can be difficult to determine.
A practical approach is to focus on content validity, the extent to which the content of
the test represents an adequate sampling of the knowledge and skills taught in the
course. If you design the test to cover information in lectures and readings in
proportion to their importance in the course, then the interpretations of test scores are
likely to have greater validity An exam that consists of only a few difficult items,
however, will not yield valid interpretations of what students know.
If you are interested in learning more about psychometric concepts and the technical
properties of tests, here are some books you might review:
Gronlund, N. E., and Linn, R. Measurement and Evaluation in Teaching. (6th ed.)
New York: Macmillan, 1990.
Use a variety of testing methods. Research shows that students vary in their
preferences for different formats, so using a variety of methods will help students do
their best (Jacobs and Chase, 1992). Multiple-choice or shortanswer questions are
appropriate for assessing students' mastery of details and specific knowledge, while
essay questions assess comprehension, the ability to integrate and synthesize, and the
ability to apply information to new situations. A single test can have several formats.
Try to avoid introducing a new format on the final exam: if you have given all
multiple-choice quizzes or midterms, don't ask students to write an all-essay final.
(Sources: Jacobs and Chase, 1992; Lowman, 1984; McKeachie, 1986; Svinicki,
1987)
Write questions that test skills other than recall. Research shows that most tests
administered by faculty rely too heavily on students' recall of information (Milton,
Pollio, and Eison, 1986). Bloom (1956) argues that it is important for tests to measure
higher-learning as well. Fuhrmann and Grasha (1983, p. 170) have adapted Bloom's
taxonomy for test development. Here is a condensation of their list:
Many faculty members have found it difficult to apply this six-level taxonomy, and
some educators have simplified and collapsed the taxonomy into three general levels
(Crooks, 1988): The first category knowledge (recall or recognition of specific
information). The second category combines comprehension and application. The
third category is described as "problem solving," transferring existing knowledge and
skills to new situations.
True-false tests. Because random guessing will produce the correct answer half the
time, true-false tests are less reliable than other types of exams. However, these items
are appropriate for occasional use. Some faculty who use true-false questions add an
"explain" column in which students write one or two sentences justifying their
response.
Matching tests. The matching format is an effective way to test students' recognition
of the relationships between words and definitions, events and dates, categories and
examples, and so on. See "Multiple-Choice and Matching Tests" for suggestions
about developing this type of test.
Essay tests. Essay tests enable you to judge students' abilities to organize, integrate,
interpret material, and express themselves in their own words. Research indicates that
students study more efficiently for essay-type examinations than for selection
(multiple-choice) tests: students preparing for essay tests focus on broad issues,
general concepts, and interrelationships rather than on specific details, and this
studying results in somewhat better student performance regardless of the type of
exam they are given (McKeachie, 1986). Essay tests also give you an opportunity to
comment on students' progress, the quality of their thinking, the depth of their
understanding, and the difficulties they may be having. However, because essay tests
pose only a few questions, their content validity may be low. In addition, the
reliability of essay tests is compromised by subjectivity or inconsistencies in grading.
For specific advice, see "Short-Answer and Essay Tests." (Sources: Ericksen, 1969,
McKeachie, 1986)
A variation of an essay test asks students to correct mock answers. One faculty
member prepares a test that requires students to correct, expand, or refute mock
essays. Two weeks before the exam date, he distributes ten to twelve essay questions,
which he discusses with students in class. For the actual exam, he selects four of the
questions and prepares well-written but intellectually flawed answers for the students
to edit, correct, expand, and refute. The mock essays contain common
misunderstandings, correct but incomplete responses, or absurd notions; in some
cases the answer has only one or two flaws. He reports that students seem to enjoy
this type of test more than traditional examinations.
They also give you some opportunity to see how well students can express their
thoughts, though they are not as useful as longer essay responses for this purpose.
See "Short-Answer and Essay Tests" for detailed guidelines.
Problem sets. In courses in mathematics and the sciences, your tests can include
problem sets. As a rule of thumb, allow students ten minutes to solve a problem you
can do in two minutes. See "Homework: Problem Sets" for advice on creating and
grading problem sets.
Oral exams. Though common at the graduate level, oral exams are rarely used for
undergraduates except in foreign language classes. In other classes they are usually
time-consuming, too anxiety provoking for students, and difficult to score unless the
instructor tape-records the answers. However, a math professor has experimented
with individual thirty-minute oral tests in a small seminar class. Students receive the
questions in advance and are allowed to drop one of their choosing. During the oral
exam, the professor probes students' level of understanding of the theory and
principles behind the theorems. He reports that about eight students per day can be
tested.
Specify the criteria to be used for rating or scoring (for example, the level of
accuracy in performing the steps in sequence or completing the task within a
specified time limit).
State the problem so that students know exactly what they are supposed to do
(if possible, conditions of a performance test should mirror a real-life
situation).
Give students a chance to perform the task more than once or to perform
several task samples.
"Create-a-game" exams. For one midterm, ask students to create either a board
game, word game, or trivia game that covers the range of information relevant to
your course. Students must include the rules, game board, game pieces, and whatever
else is needed to play. For example, students in a history of psychology class created
"Freud's Inner Circle," in which students move tokens such as small cigars and toilet
seats around a board each time they answer a question correctly, and "Psychogories,"
a card game in which players select and discard cards until they have a full hand of
theoretically compatible psychological theories, beliefs, or assumptions. (Source:
Berrenberg and Prosser, 1991)
Alternative Testing Modes
Take-home tests. Take-home tests allow students to work at their own pace with
access to books and materials. Take-home tests also permit longer and more involved
questions, without sacrificing valuable class time for exams. Problem sets, short
answers, and essays are the most appropriate kinds of take-home exams. Be wary,
though, of designing a take-home exam that is too difficult or an exam that does not
include limits on the number of words or time spent (Jedrey, 1984). Also, be sure to
give students explicit instructions on what they can and cannot do: for example, are
they allowed to talk to other students about their answers? A variation of a take-home
test is to give the topics in advance but ask the students to write their answers in
class. Some faculty hand out ten or twelve questions the week before an exam and
announce that three of those questions will appear on the exam.
Open-book tests. Open-book tests simulate the situations professionals face every
day, when they use resources to solve problems, prepare reports, or write memos.
Open-book tests tend to be inappropriate in introductory courses in which facts must
be learned or skills thoroughly mastered if the student is to progress to more
complicated concepts and techniques in advanced courses. On an open-book test,
students who are lacking basic knowledge may waste too much of their time
consulting their references rather than writing. Open-book tests appear to reduce
stress (Boniface, 1985; Liska and Simonson, 1991), but research shows that students
do not necessarily perform significantly better on open-book tests (Clift and Imrie,
1981; Crooks, 1988). Further, open-book tests seem to reduce students' motivation to
study. A compromise between open- and closed-book testing is to let students bring
an index card or one page of notes to the exam or to distribute appropriate reference
material such as equations or formulas as part of the test.
Group exams. Some faculty have successfully experimented with group exams,
either in class or as take-home projects. Faculty report that groups outperform
individuals and that students respond positively to group exams (Geiger, 1991;
Hendrickson, 1990; Keyworth, 1989; Toppins 1989). For example, for a fifty-minute
in-class exam, use a multiple-choice test of about twenty to twenty-five items. For the
first test, the groups can be randomly divided. Groups of three to five students seem
to work best. For subsequent tests, you may want to assign students to groups in ways
that minimize differences between group scores and balance talkative and quiet
students. Or you might want to group students who are performing at or near the
same level (based on students' performance on individual tests). Some faculty have
students complete the test individually before meeting as a group. Others just let the
groups discuss the test, item by item. In the first case, if the group score is higher
than the individual score of any member, bonus points are added to each individual's
score. In the second case, each student receives the score of the group. Faculty who
use group exams offer the following tips:
Ask students to discuss each question fully and weigh the merits of each
answer rather than simply vote on an answer.
If you assign problems, have each student work a problem and then compare
results.
If you want students to take the exam individually first, consider devoting two
class periods to tests; one for individual work and the other for group.
Show students the distribution of their scores as individuals and as groups; in
most cases group scores will be higher than any single individual score.
A variation of this idea is to have students first work on an exam in groups outside of
class. Students then complete the exam individually during class time and receive
their own score. Some portion of the test items are derived from the group exam. The
rest are new questions. Or let students know in advance you will be asking them to
justify a few of their responses; this will keep students from blithely relying on their
work group for all the answers. (Sources: Geiger, 1991; Hendrickson, 1990;
Keyworth, 1989; Murray, 1990; Toppins, 1989)
Paired testing. For paired exams, pairs of students work on a single essay exam, and
the two students turn in one paper. Some students may be reluctant to share a grade,
but good students will most likely earn the same grade they would have working
alone. Pairs can be self-selected or assigned. For example, pairing a student who is
doing well in the course with one not doing well allows for some peer teaching. A
variation is to have students work in teams but submit individual answer sheets.
(Source: Murray, 1990)
Prepare new exams each time you teach a course. Though it is timeconsuming to
develop tests, a past exam may not reflect changes in how you have presented the
material or which topics you have emphasized in the course. If you do write a new
exam, you can make copies of the old exam available to students.
Make up test items throughout the term. Don't wait until a week or so before the
exam. One way to make sure the exam reflects the topics emphasized in the course is
to write test questions at the end of each class session and place them on index cards
or computer files for later sorting. Software that allows you to create test banks of
items and generate exams from the pool is now available.
Ask students to submit test questions. Faculty who use this technique limit the
number of items a student can submit and receive credit for. Here is an example
(adapted from Buchanan and Rogers, 1990, p. 72):
You can submit up to two questions per exam. Each question must be typed or
legibly printed on a separate 5" x 8" card. The correct answer and the source (that is,
page of the text, date of lecture, and so on) must be provided for each question.
Questions can be of the short-answer, multiple-choice, or essay type.
Students receive a few points of additional credit for each question they submit that is
judged appropriate. Not all students will take advantage of this opportunity. You can
select or adapt student's test items for the exam. If you have a large lecture class, tell
your students that you might not review all items but will draw randomly from the
pool until you have enough questions for the exam. (Sources: Buchanan and Rogers,
1990; Fuhrmann and Grasha, 1983)
Cull items from colleagues' exams. Ask colleagues at other institutions for copies of
their exams. Be careful, though, about using items from tests given by colleagues on
your own campus. Some of your students may have previously seen those tests.
Prepare clear instructions. Test your instructions by asking a colleague (or one of
your graduate student instructors) to read them.
Include a few words of advice and encouragement on the exam. For example,
give students advice on how much time to spend on each section or offer a hint at the
beginning of an essay question or wish students good luck. (Source: "Exams:
Alternative Ideas and Approaches," 1989)
Put some easy items first. Place several questions all your students can answer near
the beginning of the exam. Answering easier questions helps students overcome their
nervousness and may help them feel confident that they can succeed on the exam.
You can also use the first few questions to identify students in serious academic
difficulty. (Source: Savitz, 1985)
Challenge your best students. Some instructors like to include at least one very
difficult question -- though not a trick question or a trivial one -- to challenge the
interest of the best students. They place that question at or near the end of the exam.
Try out the timing. No purpose is served by creating a test too long for even well-
prepared students to finish and review before turning it in. As a rule of thumb, allow
about one-half minute per item for true-false tests, one minute per item for multiple-
choice tests, two minutes per short-answer requiring a few sentences, ten or fifteen
minutes for a limited essay question, and about thirty minutes for a broader essay
question. Allow another five or ten minutes for students to review their work, and
factor in time to distribute and collect the tests. Another rule of thumb is to allow
students about four times as long as it takes you (or a graduate student instructor) to
complete the test. (Source: McKeachie, 1986)
Give some thought to the layout of the test. Use margins and line spacing that
make the test easy to read. If items are worth different numbers of points, indicate the
point value next to each item. Group similar types of items, such as all true-false
questions, together. Keep in mind that the amount of space you leave for short-
answer questions often signifies to the students the length of the answer expected of
them. If students are to write on the exam rather than in a blue book, leave space at
the top of each page for the student's name (and section, if appropriate). If each page
is identified, the exams can be separated so that each graduate student instructor can
grade the same questions on every test paper, for courses that have GSIs.
References
Brown, I. W. "To Learn Is to Teach Is to Create the Final Exam." College Teaching,
1991, 39(4), 150-153.
Clift, J. C., and Imrie, B. W. Assessing Students, Appraising Teaching. New York:
Wiley, 1981.
Ericksen, S. C. "The Teacher-Made Test." Memo to the Faculty, no. 35. Ann Arbor:
Center for Research on Learning and Teaching, University of Michigan, 1969.
"Exams: Alternative Ideas and Approaches." Teaching Professor, 1989, 3(8), 3-4.
Gronlund, N. E., and Linn, R. Measurement and Evaluation in Teaching. (6th ed.)
New York: Macmillan, 1990.
Jacobs, L. C., and Chase, C. I. Developing and Using Tests Effectively: A Guide for
Faculty. San Francisco: Jossey-Bass, 1992.
Jedrey, C. M. "Grading and Evaluation." In M. M. Gullette (ed.), The Art and Craft
of Teaching. Cambridge, Mass.: Harvard University Press, 1984.
Milton, O., Pollio, H. R., and Eison, J. A. Making Sense of College Grades: Why the
Grading System Does Not Work and What Can Be Done About It. San Francisco:
Jossey-Bass, 1986.
Murray, J. P. "Better Testing for Better Learning." College Teaching, 1990, 38(4),
148-152.
Multiple choice testing is an efficient and effective way to assess a wide range of
knowledge, skills, attitudes and abilities (Haladyna, 1999). When done well, it allows
broad and even deep coverage of content in a relatively efficient way. Though often
maligned, and though it is true that no single format should be used exclusively for
assessment (American Educational Research Association, the American Psychological
Association, and the National Council on Measurement in Education , 1999), multiple
choice testing still remains one of the most commonly used assessment formats
(Haladyna, 1999; McDougall, 1997).
The multiple-choice test is a very flexible assessment format that can be used to
measure knowledge, skills, abilities, values, thinking skills, etc. Such a test usually
consists of a number of items that pose a question to which students must select an
answer from among a number of choices. Items can also be statements to which
students must find the best completion. Multiple-choice items, therefore, are
fundamentally recognition tasks, where students must identify the correct response.
WHAT IS INVOLVED?
Description
A multiple choice test is constructed of multiple choice items, which consist of two,
sometimes three, parts as shown below.
The Stem of a multiple choice item is the part to which the student is to respond. In
everyday language, we'd call this the question, but since it could be a statement or
even an analogy or an equation, we'll use the technical term, Stem.
The Options are the choices from which examinees are to choose. There are two kinds
of options: the Key is the correct or best choice; the Distracters are incorrect or less
appropriate choices.
Sometimes Stimulus Materials can also be used with a multiple choice item, something
like a bar graph, a table, a map, a short text, etc. There will be more about stimulus
materials in a moment.
Assessment Purposes
Multiple choice tests are for grading only —This myth arises from the
misapprehension that assessment and instruction are separate stages of the
learning process. Indeed, there can be no instruction without sound
assessment, and, more importantly, both are critical for learning to occur. As
Crooks (1988) succinctly put it: "Too much emphasis has been placed on the
grading function of evaluation, and too little on its role in assisting students to
learn" (p. 468). There are a number of ways in which multiple choice items can
be designed and used to promote and refine learning, to inform instruction, as
well as to assign grades. Some of these will be highlighted throughout this CAT.
Teaching Goals
Step-by-Step Instructions
In content assessment, you should aim at finding out which students have mastered
the material and which have not, and to what extent they have or have not. The test
score should reflect that and only that. The "rules" of test development, then, work in
fundamentally two ways: to maximize the potential for the student to show you what
they know about the content, and to minimize the possibility that other factors will
influence the score. Such factors range from test forms that are hard to read and
navigate to fatigue to reading ability or computer literacy. In writing items, you must
eliminate "guessing" and "trick" questions. In test layout, that means avoiding student
confusion about how to complete the test.
Planning the Test: The critical issue at this stage is matching the test to your
teaching so that what was taught gets tested and what is tested gets taught. This is
fundamentally an issue of fairness as well as matching assessments to learning
outcomes. A great many of the complaints about testing stem from a mismatch here
("I didn't know that was going to be on the test!", "We didn't talk about that in
class!"). Constructing a test blueprint is an excellent planning tool for testing and for
teaching.
A test blueprint [sometimes called a Table of Specifications (e.g. Gronlund & Linn,
1990)] is a table that records the content areas for the test on one axis, which may be
listed by topic or text chapter or other divisions. The other axis of the table categorizes
the ways you expect your students to know that content. These can be traditional
taxonomies like Bloom's Taxonomy of Educational Objectives or your own categories.
Finally, each cell in the table records the relative weight you'll give to the intersection
of each content and way of knowing. These can be expressed as proportions or as
numbers of items. Here's an example of a test blueprint:
Once you have the Test Blueprint constructed, you can use it to plan your teaching as
well as to write test items. It's useful in planning teaching because it records what
content you feel is most important and how you expect your students to know that
content. It also informs decisions relevant to your instructional planning. And, since
the goal is a match between teaching, testing and outcomes, the same emphases used
in teaching should be used in testing. The Test Blueprint thus serves as a powerful way
to align teaching and assessment (Nitko, 2001).
Writing the Items: Perhaps the first issue here is whether you will actually write your
own items or use items from test or item banks that accompany textbooks. Using
publishers' test banks comes with a number of risks. First, your test blueprint and the
publishers' may look different, so you'll need to ensure that the test still retains your
emphases. Second, it's doubtful that published item banks consist of tried and true,
high quality items (e.g. Hansen & Dexter, 1997; Sims, 1997). Neither authors nor
publishers usually receive compensation for providing item banks, and therefore little if
any item development work is likely conducted. Therefore, you should use published
items sparingly or, at least, use them very carefully with as much scrutiny as you
would your own items.
These tips and "rules" for item writing are synthesized from research-based sources
and/ or consensus-of-the-field sources (e.g. Haladyna & Dowling, 1989). For an
extensive treatment of this topic with many examples and checklists, see Nitko's
chapter 8 (2001).
Determine how many total items you want. Considerations here include how
much material the test will cover, how deep the coverage needs to be, how
complex the items are, and how long students will have to take the test. [The
rule of thumb is one minute per question or possibly more if the items are
complicated (Gronlund, 1988; Oosterhof, 2001)].
Use the Test Blueprint to determine how many items you need to write in each
cell.
Avoid being overly general or overly specific with the content for each item.
This is somewhat dependent on your learning objectives, but you don't want to
be asking about broad, sweeping issues nor do you want to be asking about
minutia.
Make sure each item tests one and only one concept. If the item is "double-
barreled", that is tests two or more concepts, you won't know which of the two
the student truly understands if she gets the item correct. (See Pros and Cons
for more on this issue.)
Decide what item variations you're going to use (see below).
Either write the stem in the form of a question or, if a statement is in the stem
with its completion in the options, put the "blank" at the end of the stem, not in
the middle.
Poor: When looking at liquid in a test tube, the ____________ is the name of
the curved surface of the liquid.
(Answer: meniscus)
Put the "main idea" of the item in the stem, not in the options.
Streamline the stem to avoid extraneous language, but try to put as much of
the text in the stem leaving the options shorter.
Avoid negatives like "except" or "not" or highlight them in italics, boldface or
underline if you use do use them.
Highlight important words like "not", "only", "except" etc. if you use them at all.
Rules for Writing Options
(Note that a and b both include 0 and a and c both include 100 -- they overlap.)
Key-writing Tips
o Make sure the key is, in fact, correct according to the text and/ or
consensus in the field (Which, we hope, is also what's been taught in
class!).
Avoid "linking" items where the answer to one item is found in or dependent on
another item. This is something to check for near final test assembly when you
are proofreading. It's very easy to do inadvertently when you're writing items
across several sessions.
Constructing the Test: Here are a few tips for assembling the test.
Provide very clear, printed instructions and directions. If you're using a variety
of Multiple Choice formats, you may wish to write separate directions for each
section.
Order the items by content, by item format, and then by increasing difficulty of
items (Gronlund, 1988). This rule is based on information processing principles.
It is easier mentally for students to answer all of the items about one content
before moving to another. They also perform similar mental tasks on similar
items before changing mental tasks with other formats. Finally, putting easy
items before hard items helps students gain some success early on.
Randomize the ordering of keyed responses to overcome many of the rules of
thumb for guessing, like "always choose B or C". This also eliminates students
looking for or using patterns like AABBCCDD, ABADABA, etc.
Make sure students don't get "lost" in the test because then their score doesn't
reflect what they know, but how well they navigated the test.
o Don't crowd items on the page.
o Avoid using double-sided pages because students may miss the back
page.
o Use navigational cues throughout like, "continue on the next page" or
"page 4 of 12".
Scoring the Test: You may use a full-credit model, where the student gets the item
correct or incorrect. You may also use a partial credit model, where the key receives
full credit and some distracters receive partial credit. More information on scoring
appears in the "Analysis" section.
Variations
This is actually one of the great strengths of multiple-choice items. There are several
good variations that make multiple choice a very flexible item format.
Correct Answer: The key is clearly right and the distracters are clearly wrong.
For example, computation items often fit this category.
(Answer is a)
Best Answer: Here, the key is the best choice (most complete, most
comprehensive) while distracters, though "correct", may be incomplete.
What are the key elements that define Work in physics?
a) force and cause
b) force and displacement
c) force, cause and displacement
(Answer is C)
Problem A: You are asked to design a hoist system to move sacks of grain
from the entrance to a grain warehouse to a loft five meters above the
entrance. Each grain sack weighs 60 kg, and the warehouse operator wants to
move 20 bags per minute.
1. What formula would you use to compute the amount of work required to
move one bag of grain from the ground floor to the loft?
a) ½*m*v2
b) 1N*m
c) m*g*h
d) F*D*cos q
(Answer is a)
2. How much work would be required to move the twenty bags of grain to
the loft?
a) 60000 Joules
b) 3000 Joules
c) 12000 Joules
d) 1000 Joules
(Answer is a)
(Answer is a)
Analysis
There are essentially two kinds of analysis possible with multiple choice questions. The
first kind of analysis, scoring models, deals with scoring the test to determine student
achievement. The second kind of analysis, item analysis, deals with analyzing how well
each test item functioned.
Scoring Models: With multiple-choice questions, there are several potential ways to
score students' responses. Here are two:
One is the full credit model where the response is either correct or incorrect.
The student's total score is then the sum of correct responses.
A second is a partial credit model where some responses receive full credit and
others fractional credit. This works best with “best answer “questions or ones
where students may choose more than one answer. You can reward students
who pick a correct but not best answer, for example.
A final issue in scoring is the weighting of items. The easiest approach is to let each
item equal one point so all are equally weighted. However, based on the Test Blueprint
or other considerations, you may decide to prioritize some items over others and
weight them accordingly in computing the final score for the test.
Item Analysis: Once you have administered the test, it is possible to conduct several
analyses to ensure the quality of the questions for both this and future administrations
of the test. Many optical scanning systems common in university testing offices have
many of these options available for you automatically.
Item Difficulty—How hard or easy was the item for this group of students? This is
typically computed as the proportion of students who got the item correct. So a low
value means a hard question and a high value means an easy question on a scale from
0.00 to 1.00. It is best to have a mix of difficulties: some hard ones to challenge top
students; some easy ones so low-performing students will persist; and the bulk of
items at a moderate difficulty level. The average item difficulty across a test, for most
multiple-choice types, should be between 0.6 and 0.7 (Gronlund & Linn, 1990).
Item Discrimination—How well does the item differentiate among students who have
mastered the content and students who have not? This is calculated either as a
correlation coefficient between the item score and the total score or as a proportion of
high-scoring student who got the item right to low-scoring students who got the item
right. Either way, it is expressed on a scale from -1.00 to +1.00. Negative 1 means all
low scorers got the item right and all high scorers got the item wrong. Given that we
want students who have mastered the content getting each item correct, that's bad. A
positive 1 means the item worked exactly as it should. A zero means the item doesn't
distinguish between mastery and non-mastery students. That's also bad. There are a
number of statistical issues at work that cause +1.00 to be a rare occurrence, so that a
reasonable expectation for item discrimination indices is between 0.3 and 0.5
(Oosterhof, 2001). Again, a mix of different values in an exam is acceptable.
Distracter Analysis—this approach looks to see who is choosing each option for an
item. Usually the examinees are divided into low-scoring and high-scoring groups, and
the proportion of each choosing each option is reported. High-scoring students will
usually pick the correct response and low scoring students will usually pick a distracter.
If the opposite happens, that's a cue to revise the item. Something about a distracter
is attracting high performance students (Oosterhof, 2001).
One final statistic that is useful is the reliability coefficient. Reliability is the degree of
accuracy present in the score. For multiple-choice tests, this is indexed using a
reliability coefficient like Cronbach's Alpha or KR-21. Both range from 0.00 to 1.00 with
higher values indicating higher reliability. Though there are not strict cutoffs for
"acceptable" reliability coefficients because reliability is influenced by many diverse
factors, the consensus in the measurement field is that 0.60 or 0.70 would be an
acceptable lower value (Oosterhof, 2001). Having most item discrimination values at
or near .50, having a variety of item difficulties, and having more items will all increase
estimates of reliability (Gronlund, 1988; Nitko, 2001; Oosterhof, 2001). In essence,
you are calculating internal correlations of student responses between individual items.
These analyses help to identify items that are unfair, unclear, or poor in other ways.
This information can be used in a variety of ways.
In the current administration of the test. You can use this information to discard
items in the current administration of the test, another state-of-the-art practice
(McDougall, 1997). I find students really respect the scores more when I can
explain to them that certain poor items or unfair items were removed based on
item analysis. For example, if I find many high-scoring students choosing a
certain distracter, that cues me to rethink whether it could be correct. Often I'll
give all students who chose that distracter credit also.
In future administrations of the test. Reusing items is a very "cost-effective"
procedure. The overall quality of the test, however, will continue to increase if
you target poor items -- ones that are too easy or too difficult, that have zero
or negative discrimination, or ones with odd response patterns -- to be
reworked or discarded. I find it helpful when going over exams with classes or
individual students to have the item analyses handy so that I can listen for
reasons why those values came out as they did. Sometimes I'll even probe to
get explanations as to why someone chose a distracter.
In addition to the varieties of item types, there is also variety of ways to use multiple-
choice questions and administer multiple-choice tests. Here are some ideas:
Item Writing as a Study Strategy —A good study strategy for students is to have
them write multiple-choice items. This is especially good if you also want to help
students learn strategies for taking multiple-choice tests. This puts them "behind the
scenes" and helps them identify material that might be on the test and how it might be
asked. It also is a great way for the instructor to gauge their depth of understanding
and what issues/ topics they are focusing on. This can be very useful, for example, if
you notice that the things they're focusing on aren't what you would focus on.
Share the Test Blueprint —I usually give my students my test blueprint as a study
guide. After all, it tells me what's important, what I want students to know, and how I
want them to know it. I use it to teach and to assess. Why can't they also use that
information (McDougall (1997)? This is yet another way to share goals and
expectations with students.
Multiple choice testing owes much of its ubiquity to the invention of the optical scanner
in 1934, and technology has continued to play a role in multiple-choice developments.
It is no surprise, then, that computing has facilitated multiple choice testing. The major
technological applications are summarized below:
Item Banking—Item banking consists of storing items and information about items
(e.g. difficulty, discrimination, content codes, etc.) in electronic format. This allows
searching based on parameters, and tracking item characteristics over multiple
administrations. There are a variety of item banking software programs available (e.g.
FastTEST, C-Quest, Examiner, Exam Manager, MicroCat). This is the least complicated
technological application in testing, but it interfaces with others.
Internet Testing—Testing on the Internet is becoming more and more common. All
of the aforementioned techniques can be done in an Internet environment. At the
moment, the main disadvantage is security. There's no good, inexpensive way yet to
make sure that the person clicking the mouse is actually the student who is supposed
to be taking the test. In Internet banking, for example, the customer is part of the
security system. They have a password or PIN number, and it's in their interest not to
share that with others. In an academic testing situation, however, the "customer" or
student cannot be considered part of the security because it is in their interest to let
someone else have access to their PIN or password. So the Internet works well for
practice tests or low stakes assessments (e.g. ones that don't lead to grades) but is
likely not yet ready for high stakes assessment (ones on which grades are based).
Multiple choice testing brings an efficiency and economy of scale that can be
indispensable. Because students can respond to dozens of questions in a class period,
it allows broad coverage of content. Multiple-choice items are flexible enough to tap
nearly any level of Bloom's taxonomy. It also prevents students from trying anything
to get some points: either they know it or they don't. Of course, these advantages only
accrue when the items are well written and the test well constructed. That takes time,
planning, creativity, and thoughtfulness.
A key downfall and disappointment with multiple choice items is that we tend to write
items that are easy to write rather than items that take hard work to craft. This bias
causes the tests to focus on the recall of facts (knowledge & comprehension level) and
to leave out analysis, synthesis and evaluation questions, something that concerns
both faculty and students (Crooks, 1988; Shifflett, Phibbs, & Sage, 1997). But this is
only a drawback if we surrender to it. Just as we want to be thoughtful and explicit
about all aspects of our assessment (Crooks, 1988), we especially want to be
thoughtful about the items on our tests. Referring to the test blueprint while writing
items is an excellent way to avoid this pitfall (Gronlund & Linn, 1990).
Through the analysis of distracters that have been carefully designed, it is possible to
gain insight into the misconceptions of students. However, simply because a student
answers correctly, it does not necessarily mean they have mastered the content
(Dufresne, Leonard, & Gerace, 2002). In other words, it is possible to answer correctly
for many reasons, including guessing, which results in “false positives”.. This issue has
led to a category of scoring models that correct for guessing (Gronlund & Linn, 1990),
which are used in some well-known testing programs like the Scholastic Assessment
Test (SAT). These techniques are not recommended for classroom use, however,
because they correct for blind guessing, not informed guessing, more common in
classroom tests (Gronlund & Linn, 1990).
A final complaint that needs careful consideration is the potential for language or
cultural bias in multiple-choice questions. Any assessment can be biased, so we should
always take care to avoid it. Following the item writing rules mitigates the potential for
bias. Eliminating extraneous verbiage and keeping the language simple and on the
students' level helps, and at least one study has shown that well-constructed items
don't preference different cognitive styles while poorly constructed ones do
(Armstrong, 1993).
Links
Frary, Robert B. (1995). More multiple choice item writing do's and don'ts.
Practical Assessment, Research & Evaluation, 4(11).
http://ericae.net/pare/getvn.asp?v=4&n=11.
Kehoe, Jerard (1995). Basic item analysis for multiple choice tests.
Practical Assessment, Research & Evaluation, 4(10).
http://ericae.net/pare/getvn.asp?v=4&n=10.
Kehoe, Jerard. Basic Item Analysis for Multiple choice Tests. -- "This Digest offers
some suggestions for the improvement of multiple choice tests using "item analysis"
statistics. These statistics are typically provided by a measurement services, where
tests are machine-scored, as well as by testing software packages."
http://www.robmccormack.com/qphelp/qphowto4.php
Carneson, J., Delpierre, G. & Masters, K. Designing and Managing Multiple Choice
Questions
http://www.le.ac.uk/cc/ltg/castle/resources/mcqman/mcqman01.php
Resources
Haladyna, T. M. (1999). Developing and validating multiple choice test items. 2nd Ed.
Mahwah, NJ: Lawrence Erlbaum Associates.
This text is primarily focused on large-scale testing, but much of it is still very
relevant to classroom testing.
Nitko, A. J. (2001). Educational assessment of students. 3rd Ed. Columbus, OH: Merrill
Prentice Hall.
http://vig.prenhall.com/catalog/academic/product/1,4096,0130137081,00.php
Though this book is for K-12 teachers, Nitko is unparalleled in his copious
examples and checklists when it comes to item writing. See especially chapter
8.
Gronlund, N. E., & Linn, R. L. (1990). Measurement and evaluation in teaching (6th
Ed.). New York: Macmillan.
Another text written for K-12 teachers, it also has excellent information about a
broad range of topics relevant to multiple choice testing.
References
Brissenden & Slater (n.d.) Assessment Primer. Retrieved June 26, 2002 from the
Internet at http://www.flaguide.org/start/start.php
Dufresne, R. J., Leonard, W. J., & Gerace, W. J. (2002). Making sense of students'
answers to multiple-choice questions. The Physics Teacher, 40, 174-180.
Gronlund, N. E., & Linn, R. L. (1990). Measurement and evaluation in teaching (6th
Ed.). New York: Macmillan.
Haladyna, T. M. (1999). Developing and validating multiple choice test items (2nd
Ed.). Mahwah, NJ: Lawrence Erlbaum Associates.
Hansen, J. D., & Dexter, L. (1997). Quality multiple-choice test questions: Item-
writing guidelines and an analysis of auditing test banks. Journal of Education for
Business, 73(2), 94-97.
Lederhouse, J. E., & Lower, J. M. (1974). Testing college professor's tests. College
Student Journal, 8(1), 68-70.
Murray, J. P. (1990). Better testing for better learning. College Teaching, 38(4), 148-
152.
Shifflett, B., Phibbs, K., & Sage, M. (1997). Attitudes toward collegiate level classroom
testing. Educational Research Quarterly, 21(1), 15-26.
Wiggins, G. (1989). A true test: Toward more authentic and equitable assessment. Phi
Delta Kappan, 70 (9) , 703-713.
Wiggins, G. (1992). Creating tests worth taking. Educational Leadership, 49(8), 26-33.
Similar to a learning log, the book response journal is a place for students to express
personal reactions and to wonder about events, themes, and ideas in a book. Children are
encouraged to react to everything they read. Teachers may use these journals to respond to
each child individually, sharing their questions, feelings, and ideas about literature and
making suggestions for further reading or related activities. Some teachers hold individual
reading conferences with their students and use these book response journals as part of the
conferences.
Comparison Charts
Comparison charts are one of a number of graphic organizers. They involve the
examinations of similarities and differences among ideas, events, characteristics, etc.
Comparison charts may take a number of forms and are an excellent way to engage
students individually or in groups as they seek to focus characters, events, or themes within
a single story or compare books, events, or properties within a given theme.
Conferences
There are many types of conferences including reading, writing, goal-setting, evaluation,
and coaching. The major purposes are to collaborate, assess, and guide.
Cooperative Learning Activities
Demonstrations
A demonstration transforms ideas into something concrete and observable through visual,
audio, art, drama, movement, and/or music. This could also include opportunities to
demonstrate and explain procedures and strategies such as a science experiment or a
solution to a non-routine math problem.
Discussion
A discussion provides a safe, open forum where children are encouraged to speak, listen,
and respond to opinions, feelings, and ideas regarding the designated topic.
Goal Setting
Setting goals with children provides the basis for monitoring student performance through
collaboration and self reflection.
Graffiti Walls
Graffiti walls are free form spaces for brainstorming or communicating words, phrases, or
ideas on a topic. These are often used as evolving records. A teacher may use them to
facilitate brainstorming on a particular theme at the beginning of a unit, as well as
encourage students to add new words or phrases relating to the theme as the unit
progresses. In addition to encouraging children to search for new and interesting words, the
graffiti wall serves as a class dictionary/thesaurus as students need novel words to enrich
their writing.
"I Learned" statements may be in either written or oral form. Their purpose is merely to
give students a chance to self-select one or more of the things they learned during a class
session, an investigation, or a series of lessons.
Interviews
Investigations
Investigations may be related to a specific subject area or may involve several areas,
integrating curriculum. The most typical form of investigation is a collection of student
writing, diagrams, graphs, tables, charts, posters, experiments and other products. When
students become involved in practical or mathematic investigations, assessment activities
and/or questions can be presented to students without their awareness of any difference
between the assessment and instruction.
KWLs
A KWL is a technique used by teachers to assess what students "know," "wish to know,"
and "have learned about a particular topic," using a sheet divided into three columns
labeled K, W, L. At the beginning of a lesson, the KWL serves as a written record of the
students prior knowledge (K) on the topic, and allows the opportunity for the student to
note what they desire (W) to know about the topic. Following the lesson, the student can
self-assess what has actually been learned (L) about the topic.
Learning Logs
A learning log is a kind of journal that enables students to write across the curriculum. The
major reason for using them is to encourage children to be in control of their own learning
and to promote thinking through writing.
Attitude surveys note in a systematic manner students' self reflections regarding group and
individual performance and affective characteristics such as effort, values, and interest.
Providing an oral survey allows students to share their ideas, learn from others, and deepen
the way they think about the topics being discussed.
Oral Presentations
Peer Evaluations
Peer evaluations consist of student analysis and assessment of peer proficiency using either
established or self-generated criteria. An activity must be very carefully structured if
students are to receive valid feedback from their peers.
Products
Response Groups
Response groups are opportunities for small numbers of children to discuss books or events
in depth with one another. Often these groups are organized and run by children themselves
because they all have read the same book or experienced the same event and want to
discuss it. Teachers participating in a response group will gain insight into their students'
thinking skills, group behaviors, and affective characteristics.
Self-Evaluations
A key concept in alternative assessment is having the student learn to recognize his/her own
progress by taking the time to reflect. Those who are able to review their own performance,
explain the reasons for choosing the processes they used, and identify the next step, develop
insight and self-involvement. Self-reflection, an important concept in any form of
assessment, is a particularly important component of a student portfolio
Introduction
Beginnings are important. Whether the class is a large introductory course for
freshmen or an advanced course in the major field, it makes good sense to start the
semester off well. Students will decide very early - some say the first day of class -
whether they will like the course, its contents, the teacher, and their fellow students.
The following list of "101 Things You Can Do..." is offered in the spirit of starting
off right. It is a catalog of suggestions for college teachers who are looking for a fresh
way of creating the best possible environment for learning. Not just the first day, but
the first three weeks of a course are especially important, studies say, in retaining
capable students. Even if the syllabus is printed and lecture notes are ready to go in
August, most college teachers can usually make adjustments in teaching methods as
the course unfolds and the characteristics of their students become known.
These suggestions have been gathered from UNL professors and from college
teachers elsewhere. The rationale for these methods is based on the following needs:
1) to help students make the transition from high school and summer or holiday
activities to learning in college; 2) to direct students' attention to the immediate
situation for learning - the hour in the classroom: 3) to spark intellectual curiosity - to
challenge students; 4) to support beginners and neophytes in the process of learning
in the discipline; S) to encourage the students' active involvement in learning; and 6)
to build a sense of community in the classroom.
Ideas For the First Three Weeks
Here, then, are some ideas for college teachers for use in their courses as they begin a
new semester.
Helping Students Make Transitions
1. Hit the ground running on the first day of class with substantial content.
2. Take attendance: roll call, clipboard, sign in, seating chart.
3. Introduce teaching assistants by slide, short presentation, or self-introduction.
4. Hand out an informative, artistic, and user-friendly syllabus.
5. Give an assignment on the first day to be collected at the next meeting.
6. Start laboratory experiments and other exercises the first time lab meets.
7. Call attention (written and oral) to what makes good lab practice: completing
work to be done, procedures, equipment, clean up, maintenance, safety,
conservation of supplies, full use of lab time.
8. Administer a learning style inventory to help students find out about
themselves.
9. Direct students to the Learning Skills Center for help on basic skills.
10. Tell students how much time they will need to study for this course.
11. Hand out supplemental study aids: library use, study tips, supplemental
readings and exercises.
12. Explain how to study for kind of tests you give.
13. Put in writing a limited number of ground rules regarding absence, late work,
testing procedures, grading, and general decorum, and maintain these.
14. Announce office hours frequently and hold them without fail.
15. Show students how to handle learning in large classes and impersonal
situations.
16. Give sample test questions.
17. Give sample test question answers.
18. Explain the difference between legitimate collaboration and academic
dishonesty; be clear when collaboration is wanted and when it is forbidden.
19. Seek out a different student each day and get to know something about him or
her.
20. Ask students to write about what important things are currently going on in
their lives.
21. Find out about students' jobs; if they are working, how many hours a week,
and what kinds of jobs they hold.
22. Greet students at the door when they enter the classroom.
23. Start the class on time.
24. Make a grand stage entrance to hush a large class and gain attention.
25. Give a pre-test on the day's topic.
26. Start the lecture with a puzzle, question, paradox, picture, or cartoon on slide
or transparency to focus on the day's topic.
27. Elicit student questions and concerns at the beginning of the class and list
these on the chalkboard to be answered during the hour.
28. Have students write down what they think the important issues or key points
of the day's lecture will be.
29. Ask the person who is reading the student newspaper what is in the news
today.
Challenging Students
30. Have students write out their expectations for the course and their own goals
for learning.
31. Use variety in methods of presentation every class meeting.
32. Stage a figurative "coffee break" about twenty minutes into the hour; tell an
anecdote, invite students to put down pens and pencils, refer to a current
event, shift media.
33. Incorporate community resources: plays, concerts, the State Fair. government
agencies. businesses, the outdoors.
34. Show a film in a novel way: stop it for discussion, show a few frames only,
anticipate ending, hand out a viewing or critique sheet, play and replay parts.
35. Share your philosophy of teaching with your students.
36. Form a student panel to present alternative views of the same concept.
37. Stage a change-your-mind debate. with students moving to different parts of
the classroom to signal change in opinion during the discussion.
38. Conduct a "living" demographic survey by having students move to different
parts of the classroom: size of high school. rural vs. urban. consumer
preferences...
39. Tell about your current research interests and how you got there from your
own beginnings in the discipline.
40. Conduct a role-play to make a point or to lay out issues.
41. Let your students assume the role of a professional in the discipline:
philosopher, literary critic, biologist. agronomist. political scientist. engineer.
42. Conduct idea-generating or brainstorming sessions to expand horizons.
43. Give students two passages of material containing alternative views to
compare and contrast.
44. Distribute a list of the unsolved problems. dilemmas. or great questions in
your discipline and invite students to claim one as their own to investigate.
45. Ask students what books they've read recently.
46. Ask what is going on in the state legislature on this subject which may affect
their future.
47. Let your students see the enthusiasm you have for your subject and your love
of learning.
48. Take students with you to hear guest speakers or special programs on campus.
49. Plan "scholar-gypsy" lesson or unit which shows students the excitement of
discovery in your discipline.
Providing Support
50. Collect students' current telephone numbers and addresses and let them know
that you may need to reach them.
51. Check out absentees. Call or write a personal note.
52. Diagnose the students' prerequisites learning by questionnaire or pre-test ant
give them the feedback as soon as possible.
53. Hand out study questions or study guides.
54. Be redundant. Students should hear, read. or see key material at least three
times.
55. Allow students to demonstrate progress in learning: summary quiz over the
day's work. a written reaction to the day's material.
56. Use non-graded feedback to let students know how they are doing: post
answers to ungraded quizzes and problem sets, exercises in class, oral
feedback.
57. Reward behavior you want: praise, stars, honor roll, personal note.
58. Use a light touch: smile, tell a good joke, break test anxiety with a
sympathetic comment.
59. Organize. Give visible structure by posting the day's "menu" on chalk- board
or overhead.
60. Use multiple media: overhead, slides, film, videotape, audio tape, models,
sample material.
61. Use multiple examples, in multiple media. to illustrate key points and .
important concepts.
62. Make appointments with all students (individually or in small groups).
63. Hand out wallet-sized telephone cards with all important telephone numbers
listed: office department, resource centers, teaching assistant, lab.
64. Print all important course dates on a card that can be handed out and taped to
a mirror.
65. Eavesdrop on students before or after class and join their conversation about
course topics.
66. Maintain an open lab gradebook. with grades kept current. during lab time so
that students can check their progress.
67. Check to see if any students are having problems with any academic or
campus matters and direct those who are to appropriate offices or resources.
68. Tell students what they need to do to receive an "A" in your course.
69. Stop the work to find out what your students are thinking feeling and doing in
their everyday lives.
Building Community
92. Learn names. Everyone makes an effort to learn at least a few names.
93. Set up a buddy system so students can contact each other about assignments
and coursework.
94. Find out about your students via questions on an index card.
95. Take pictures of students (snapshots in small groups, mug shots) and post in
classroom, office, or lab.
96. Arrange helping trios of students to assist each other in learning and growing.
97. Form small groups for getting acquainted; mix and form new groups several
times.
98. Assign a team project early in the semester and provide time to assemble the
team.
99. Help students form study groups to operate outside the classroom.
100. Solicit suggestions from students for outside resources and guest
speakers on course topics.
Feedback on Teaching
101. Gather student feedback in the first three weeks of the semester to
improve teaching and learning.
One common answer is simply to start lecturing: "This is day one, here is lecture one,
away we go." Another possibility is: "Here is the syllabus, go buy your books and we
will see you at the next scheduled class period." Neither of these two options seems
desirable. But what are some other possibilities?
Several years ago a group of professors at the University of Oklahoma visited each
other on the first day of class and then discussed what they saw each other doing. But
the discussion quickly went from what they observed, to "What might be done?"
They eventually identified nine attractive possibilities, as described below. A
teacher should not feel obliged to do all of these, but doing even one or several of
them on the first day (or during the first week) would seem to accomplish a number
of important tasks for getting a class started in the right way.
But letting students know right from the outset that they will be active
participants seems like a good approach.
Not all students come to all classes with a clear idea of why this subject is
important. The teacher may need to help them understand the significance of
the course. The sooner this is done, the sooner the students will be ready to
invest time and energy in the task of learning the subject matter.
3. Set expectations.
This can involve such things as what the teacher considers appropriate
amounts of study time and homework for the class, the importance of turning
homework in on time, expectations about in-class behavior, how the teacher
wants to relate to students, and how much interaction among students is
desired. The first day also offers an opportunity to find out what expectations
the students have of the teacher and of the class.
4. Establish rapport.
Almost any class will be more enjoyable for both the teacher and the students
if they know each other a bit. This exchange can be started with introductions,
sharing some background information, etc.
Sometimes students can relate to the teacher more productively if they can see
him or her as a human being, i.e., as something more than just an authority
figure or subject matter expert. Sharing personal stories and being able to
laugh at yourself can help this process.
This often takes the form of going through the syllabus, presuming you have a
syllabus with this information in it: what reading material the students will
need; what kind of homework will be involved; what your office hours are;
where your office is located; how the class grade will be determined; what
your policies are regarding attendance, late papers, make-up exams, etc.
o What is it?
o What are the parts of the subject?
o How is it connected to other kinds of knowledge?
Final Note:
Remember that it is imperative that you do on the first day whatever it is you want
the class to do the rest of the semester. If you want them to discuss, discuss on the
first day. If you want them to work in small groups, find something for them to do in
small groups on the first day.
Ten Things to Make the First Day (and the Rest) of the Semester
Successful
By Mary C. Clement, Berry College, GA
As published in The Teaching Professor, Volume 21, Number 7, August/September 2007
I like to arrive in the classroom well before the students. It gives me time to get
things organized. I create an entrance table (I use chairs or desks if there's no table)
that holds handouts for students to pick up. From day one the students learn the
routine: they arrive, pick up handouts on the entrance table, and read the screen for
instructions. They know what to do, and it saves time. Here's how I recommend
introducing the routine on day one.
1. Post your name and the name and section of the class on the screen, so that
when students walk in they know that they are in the right place.
2. Write: "welcome" on the screen and have directions that tell students what
they need to do immediately. Example: "As you enter, please tell me your
name. Then pick up a syllabus, a card, and a folder from the entrance table.
Fold the card so that it will stand on your desk, and write your first name on it
in BIG letters. Add your last name and major in smaller print. Write your
name on the tab of the folder, (last name first, then first name). Read the
syllabus until class starts." Note: By asking students to tell you their name as
they enter, you can hear how the name is pronounced, and avoid the
embarrassment of pronouncing it for the first time yourself.
3. When it's time for class to start - start class! Late arrivals can catch up by
reading the screen.
4. For classes of 25 or less, I have students do brief, 10-second introductions. I
tell them there will be a verbal quiz after all the introductions and that they
can win stars if they know who is who. (Have fun with this, but remember
that these are adults and college is not like junior high.)
5. For larger classes, I have students introduce themselves to three or four
people around them, and then we might do "stand-ups" - stand up if you are a
Spanish major, stand up if you are an education major, and so on. I explain
that students need to know each other for our small group work, and in case
they have a question.
6. I collect the file folders and put them alphabetically by student name into a
big plastic carrying case. When students need to turn in assignments, they
find the box on the entrance table and they put their papers in their respective
folders. When papers are graded, they can pull their graded tests or
assignments from their folders. The beauty of this system is that time is never
wasted by passing out papers. For small classes, I put handouts in the folders
of absent students.
7. After the introductions and the explanation of the folder and box system, I
turn to the "Today we will" list that I've written on the board, posted on a
large paper flip-chart, or projected on the screen. I like to actually write this
list on the board, so I can return to it even while projecting my notes. A
"today we will" list outlines my plan for the day. For example, for the first
day, my "today we will list" says:
o See screen for instruction for card and folder.
o Introductions
o Turn in folders
o Go over syllabus completely
o Mini-lecture on _____________
o Interest inventory
o Do you know what to read/do before the next class?
Note: The "today we will" list lets me walk around the room, teach from the
projection system, and then look at the list for what I should do next. I tend
not to forget things if I have the list. As the semester progresses, the "today
we will" list might contain warm-up questions that then appear as test
questions. The list helps students who arrive late or leave early see what they
have missed.]
o If your song played when you entered the room, what would that song
be?
10. Every good class has an introduction, a body, and a conclusion. I usually
teach the mini-lesson, and then save the last six to eight minutes of class for
the interest inventory and individual questions. This way, students don't have
to wait on others to finish. I instruct students to turn in their interest inventory
as they exit. As they are writing, I alphabetize their folders and put them in
the box on the table. Another good closure is to ask if they know what to
read/do before the next class, and if they know three people to ask about the
assignment if they have a question.
BREAKICER GAMES
« Fear in a Hat
Icebreaker Questions »
Candy Introductions
Candy Introductions is a get-to-know-you game that helps people learn new facts about
each other in an easy way. They select various pieces of candy from a bag, and each candy
variety is associated with a fact about themselves which they will introduce to the others.
This game also goes by other names, including the M&M game, Candy Confessions, the
Skittles Game, the Gum Drop game, among others.
Candy Introductions can work with any group size. The icebreaker works best when the
group size is limited to 12, so if you have more than 12, divide the larger group and run the
icebreaker within the smaller sized groups. This icebreaker works best indoors, and is well
suited for classrooms or meeting rooms. Materials required are: candy with about five
different variations (color or candy type), and an optional chalkboard/whiteboard.
Purchase several variety packs of candy, enough for each person to be able to have at least
five pieces. They can be any candy type, but not too many choices (limit it to around five
or six different varieties). Alternatively, you can buy gummy bears, life savers, gum drops,
skittles, m&ms, or any other candy that already has a variety of colors.
Pass around the candy and tell each participant to choose anywhere from 1 to 5 pieces of
anything that they want. Instruct them not to eat it yet, though. After they have chosen
their candy, you will tell them what each candy type/color represents.
If you don’t have the above colors, change the above to match the candy types that you
have. Each person takes turns introducing himself or herself, beginning with their name
and then saying one fact for each candy type that they have. This easy introduction game
should go relatively quickly (assuming they weren’t greedy and that they didn’t take too
many pieces of candy!)
Defend the Egg (also known as the Great Egg Drop) is a teambuilding activity that involves
collaboration, problem solving, and creative teamwork. Groups build a structure out of
ordinary materials and try to protect a raw egg from breaking when dropped from a high
elevation.
This exercise in teambuilding can be messy, so choose an appropriate setting where making
a mess is acceptable. The recommended group size is: teams of four or five people.
Several materials are needed: raw eggs, several plastic straws, masking tape, newspaper,
and other materials of your choice. This activity is for people ages 14 and up.
This game works well with teams that are not too large. A good team size is four or five
people. Pass out one egg and a limited supply of materials (e.g. four straws, a three foot
strip of tape, one section of a newspaper, etc.) This activity is more challenging with less
materials provided, so decide how challenging you wish to make it.
Explain the rules: the mission is to protect the egg from cracking using teamwork,
creativity, and a good design. You will drop each structure at least fifteen feet, and so the
goal is for each structure to be able to withstand such a fall. Each team will only be given
limited resources, and so they must be wise with what they have. They may not use any
other resources other than what is given to them. Optionally, you can have other critera for
judging including:
Decide on an appropriate amount of time (e.g. 20-25 minutes) and then instruct them to
begin! Tell them to place their egg inside their structure.
When time is up, collect all the structures. Now is dramatic finale in which the structures
are dropped (or thrown!) from at least 15 feet in elevation and then carefully inspected to
see if the eggs survived. The winners are the groups that successfully protected the egg. If
you chose to have other awards, announce those winners also.
This activity is useful to illustrate the importance of teamwork. Ask everyone to reflect on
how their group accomplished the task, what worked, what was challenging, etc.
Variation
A much more challenging variation of this activity is to provide no materials (other than the
egg) and ask the participants to find materials from outdoors.
Did You Know? Bingo (also known as the Autograph Game) is an icebreaker that helps
people learn interesting facts about each other. People walk around the room and mingle
until they find people that match the facts listed on a bingo-style sheet.
This game is a get-to-know-you style icebreaker. The recommended group size is: large or
extra large. The game works best with a group of about 25 people. It can be played
indoors or outdoors. Materials required are: printed bingo sheets and pens. Ages 12 and
up.
The objective of this game is for people to wander around the room and to obtain the
signatures of people who have the facts listed on the bingo sheet. Once a person
successfully obtains a full row (5 in a row), whether horizontally, vertically, or diagonally,
he or she shouts “BINGO!” and wins.
This game requires a little bit of setup. Prepare a 5 by 5 table, with interesting facts written
inside the boxes. These facts can include funny or bizarre things. For example:
Likes anchovies
Has been to Hawaii
Speaks more than two languages
Has never been on a plane
Has more than four brothers
Has gone without a shower for more than three days
Be creative! You can mark the center square “FREE SPACE” like traditional bingo
games. After you have prepared the table, print out enough copies for the number of
players you are expecting.
Pass out a sheet to each person, along with a pen. Explain the objective of the game and
the following rules: (1) each person you talk to may only sign your sheet once, and (2) to
win, you must get signatures to form 5 in a row horizonally, vertically, or diagonally. Say
“Go!” and ask your participants to begin.
Once someone shouts “Bingo!” everyone returns and the person must introduce the people
who signed his or her sheet. If desired, you can ask each person to explain their fact. This
icebreaker game is fun way to get to know humorous or unique facts about people. Enjoy!
Icebreaker Questions
Icebreaker Questions is simply a list of 20 great questions that you can ask people to help
them feel more part of a group or team. These questions are fun and non-threatening. You
can use them as an icebreaker for meetings or classrooms, written on notecards and adapted
for other games, or simply as a fun activity to help people get to know each other better.
1. If you could have an endless supply of any food, what would you get?
2. If you were an animal, what would you be and why?
3. What is one goal you’d like to accomplish during your lifetime?
4. When you were little, who was your favorite super hero and why?
5. Who is your hero? (a parent, a celebrity, an influential person in one’s life)
6. What’s your favorite thing to do in the summer?
7. If they made a movie of your life, what would it be about and which actor would you want
to play you?
8. If you were an ice cream flavor, which one would you be and why?
9. What’s your favorite cartoon character, and why?
10. If you could visit any place in the world, where would you choose to go and why
11. What’s the ideal dream job for you?
12. Are you a morning or night person?
13. What are your favorite hobbies?
14. What are your pet peeves or interesting things about you that you dislike?
15. What’s the weirdest thing you’ve ever eaten?
16. Name one of your favorite things about someone in your family.
17. Tell us about a unique or quirky habit of yours.
18. If you had to describe yourself using three words, it would be…
19. If someone made a movie of your life would it be a drama, a comedy, a romantic-comedy,
action film, or science fiction?
20. If I could be anybody besides myself, I would be…
21. If you were a comic strip character, who would you be and why?
22. What thought or message would you want to put in a fortune cookie?
23. If you had to give up a favorite food, which would be the most difficult to give up?
24. What is one food you’d never want to taste again?
25. If you won a lottery ticket and had a million dollars, what would you do with it?
26. You’ve been given access to a time machine. Where and when would you travel
to?
27. If you could be any superhero and have super powers, which one would you like to
have and why?
28. Mount Rushmore honors four U.S. presidents: Washington, Jefferson, Lincoln, and
Roosevelt. If yuo could add any person to Mount Rushmore, who would you add
and why?
29. What award would you love to win and for what achievement?
30. If you could transport yourself anywhere instantly, where would you go and why?
31. In your opinion, which animal is the best (or most beautiful) and why?
32. What is one item that you really should throw away, but probably never will?
33. Growing up, what were your favorite toys to play with as a child?
Superlative Game
The Superlative Game is a simple icebreaker that asks players to line up in ascending order
for various different categories (e.g. height, birthday month, etc.). It’s very easy to learn
and play, and it doesn’t require a lot of time.
The facilitator of the Superlative game needs to prepare a list of categories. These
categories can be surface-level categories such as height (in ascending order), birthday
month (in ascending order, from January to December), shoe size, number of siblings (least
to most), etc. or you can also make deeper categories, depending on your goals.
Split the group into at least three teams. The ideal team size is at least five players per
team, but preferably no more than nine. Explain the rules and consider using one of the
variations below (such as the no talking rule).
Read the first category aloud, such as “Line up by increasing order of height. Go!” Each
team scrambles to get in the proper order. When a team thinks they are done, they must all
sit down and raise their hands. At this point, the facilitator checks the team and verifies
that they are in the proper order. If they made a mistake, they get awarded zero points that
round and the facilitator checks the next group that sat down. The first team to correctly
get in order and sits down gets awarded a point.
The facilitator then reads off the next category, and this process repeats until the game is
over (no more categories).
Variations
There are many variations to this game that are worth considering. These include:
1. No talking allowed. All players must rely on body language and hand gestures to get in
the proper order.
2. Head and feet only. You can only use your head and feet to signal where to go. No
talking or use of arms/hands to communicate.
Aligning Learning Activities and Assessment Strategies in the ESL
Classroom
Robert F. Tambini
robu_sensei [at] yahoo.com
Centenary College (New Jersey, USA)
Introduction
Accurately gauging the extent to which student progress meets the goals that have been set is one
of the most difficult tasks which educators must perform. In order to do this effectively, it is
essential that instructors design their lesson plans in such a way that expectations placed upon
students regarding performance standards, instructional objectives and learning activities
identified for use in the classroom, and assessment strategies employed in evaluating student
learning are well-aligned with one another.
As a means of demonstrating the form that this alignment might take, I will discuss this
alignment within the context of a lesson plan which I have used in since 1996. The class for
which this lesson plan was developed was (and remains) part of Centenary College's spring
Intensive English Language Program (IELP), a six-week program offered each year to
students of Centenary's sister college in Tokyo, Japan. The original class for which this
lesson plan was designed consisted of twenty 19 year-old female Japanese students at
Obirin Junior College. This lesson plan has been subsequently used in both the spring and
summer Intensive English Language Programs at Centenary College.
This discussion will begin with a presentation of the standards which informed the design
of the lesson plan, and a description of the performance standards set out in the plan which
students are expected to meet. The lesson plan itself follows. Finally, I will attempt to form
a conclusion addressing the success with which the various elements of the plan have been
aligned.It is important to note, and I wish to emphasize, that the particular lesson plan under
discussion here is an actual lesson plan which has seen three years use in the classroom, not
a hypothetical one constructed solely for the purpose of serving as the object of this study.
The Standards
Any English as a Second Language exercise is founded upon a single standard against which
students will be measured: the relative character and quality of the English used by students as
compared to the character and quality of the English used by native speakers. Of course, when we
assign levels to second language (L2) students, we are judging them not only against the standard
set by native speakers, but also relative to a) the received academic standard, and b) other L2
students. These standards, English as used by native speakers and English as used by L2 students,
are built of a large number of components, including (but not limited to) the following:
1. Vocabulary - number and quality of items known; ability to use items correctly and
appropriately
2. Grammar - knowledge of morphology and syntax; ability to apply knowledge correctly and
appropriately
3. Oral production - the ability to create utterances comprehensible to native speakers
4. Listening comprehension - the ability to comprehend utterances produced by native
speakers
5. Written production - the ability to manipulate symbols (letters, characters) and produce
language structures comprehensible in a visual context, and to express feelings, concepts
and ideas in writing
6. Reading comprehension - the ability to comprehend language symbols and structures in a
visual or tactile context, and to understand feelings, concepts, and ideas expressed in
writing
Within each of these areas any number of items might be identified by the teacher as the object of
a specific lesson. The ultimate goal of ESL teaching is mastery of the English language; however,
prior to becoming fluent in their use of English, students must first acquire a clear understanding
of and proficiency in each of the areas outlined above.
After dinner, students are expected to study and complete homework for approximately three
hours. Upon completion of the IELP, students are awarded three transfer credits in English which
they may apply as an elective toward their Obirin degrees. The lesson plan which will be described
below is used in support of a unit on "apartment hunting" in the conversation class.
The Objective
The objective of this lesson is to actively engage students in the process of seeking out and making
inquiries regarding the availability and suitability of living space in as authentic a manner as
possible. In addition to this introduction to the apartment hunting process, students are required
to make actual contacts and inquiries based on their identification of available and potentially
desirable living spaces advertised in local newspapers. This requirement is intended to force
students, who otherwise tend to become overwhelmed by and shy at the prospect of interacting
with a native speaker, to use the English skills they possess in an authentic, real-life situation
which many of them are likely to encounter again.
Pursuant to this objective, students will demonstrate: a) proficiency in their knowledge and
use of English vocabulary related to housing; b) the ability to form appropriate wh-
questions intended to elicit information which will assist in determining the suitability of a
particular apartment; c) critical thinking skills, as evidenced in descriptions (written and
oral) of the reasoning which led to the ultimate acceptance or rejection of a particular
housing unit.
Similarly, student written performance is also measured against the standard set by examples
discussed in class. These examples, whether drawn from the textbook, provided to students in the
form of handouts, or presented by the teacher during class, remain available throughout the
course of the lesson as a means by which students might judge the progress and quality of their
work. In addition, students receive an in-class explanation regarding those aspects of their report
on which they will be evaluated and the teacher's expectations in these areas. The criteria for
evaluation include: spelling, punctuation, word choice, grammatical correctness, sentence
structure and variety, clarity, coherence, organization, and overall effectiveness.
Providing Feedback
Feedback is an essential component of this lesson, and is provided in any number of ways: in-class
teacher response, one-on-one student/teacher conferences, written comments, etc. Most
important, however, is the fact that during this lesson the lines of communication are not filtered
through a one-way valve. Of course, feedback is provided to the student by the teacher; but,
students are also encouraged and expected to evaluate the quality and appropriateness of the
services they receive, and to offer constructive criticisms relative thereto. For example, the one-
on-one student/teacher conferences are not intended merely as opportunities for the teacher to
"preach from the mountaintop," but are rather meant to provide students with the chance to offer
input as to how their learning might be better facilitated.
Conclusion