You are on page 1of 5

Session F4H

Comparing the Effectiveness of Evaluating


Practical Capabilities Through Hands-On On-Line
Exercises Versus Conventional Methods
Isabel Garcia, Alfonso Duran and Manuel Castro
igarcia@ing.uc3m.es, duran@ing.uc3m.es, mcastro@ieec.uned.es

Abstract - Two interrelated methodological towards placing a higher emphasis on applied capabilities.
transformations involved in the current transition of That is in turn related to the major overhaul proposed for the
European universities towards the European Higher evaluation procedures; the currently prevailing approach
Education Area (EHEA) are the role of applied based solely on written final exams is postulated to
capabilities and the evaluation process. In this context encourage learning by rote and being inappropriate for
this paper presents the results of a structured appraising applied capabilities. According to the European
comparison, throughout a five course period, of the University Association’s Trends V report to the Conference
impact of alternative evaluation methods in courses of Ministers of Education meeting in London on 17/18 May
aimed at the development of applied engineering 2007 to discuss the culmination of the Bologna process by
capabilities. The comparison perspective is twofold: how 2010, a majority of the participating institutions continue to
accurately does the evaluation method measure the rely on traditional end-of-year examinations to assess
competence level attained by the students, and how does student knowledge [2]. Progress is, however, being made, as
it affect their active learning. The experiment was shown by the comparison with the equivalent Trends III
conducted in a simulation course from the Industrial report figures.
Engineering curriculum and the aim was the evaluation
of the capability of using a simulation software. The recently approved legal framework aimed at revamping
Evaluation was traditionally based on a written final the Spanish higher education system to adapt it to the
exam and two other evaluation methods were then EHEA’s requirements highlights the focus on the
introduced: Computer exam and team project development of capabilities, as opposed to the mere
assignment. The assessment of the evaluation methods accumulation of knowledge, and the need to establish
was carried out by both faculty members and students appropriate evaluation procedures for these capabilities [4].
(through anonymous surveys). Results suggest that both
group assignments and computer exam perform far In the USA, the Accreditation Board for Engineering and
better, in this environment, than written exams. The Technology (ABET), among the criteria it applies for
comparison between group assignments and computer accrediting engineering programs during the 2007-2008
exam is less straightforward, being dependant on which accreditation cycle, requires that Engineering programs
criterion is being appraised. demonstrate that their students attain applied capabilities
such as “an ability to design and conduct experiments, as
Index Terms – Evaluating capabilities, on-line testing, well as to analyze and interpret data” and “an ability to use
evaluation methodologies, problem based learning. the techniques, skills, and modern engineering tools
necessary for engineering practice” [5]. It also requires the
INTRODUCTION implementation of an appropriate assessment process, with
The current transition of European universities towards the documented results, that demonstrates that the degree of
European Higher Education Area (EHEA) requires a move achievement of these capabilities is being measured. There
towards student-centered higher education and away from are, however, some worrying indicators, such as the
teacher driven provision, as well as a renewed emphasis on sustained “grade inflation” reported for a wide sample of US
employability and the development of transferable skills and universities [6].
capabilities [1], [2], [3]. Out of the many methodological
transformations involved, two significant and interrelated Appropriate assessment and evaluation procedures
components are the role of applied capabilities and the contribute to the effectiveness of the educational process
assessment of learning outcomes. through two complementary mechanisms. On the one hand,
student’s expectations about the evaluation system heavily
EHEA’s recommendations encourage a shift from the highly condition their chosen course of action. On the other hand,
theoretical approach widespread in most national higher the evaluation’s results will only be used in order to
education systems, such as the Spanish university system, continuously improve the educational process if the quality
978-1-4244-1970-8/08/$25.00 ©2008 IEEE October 22 – 25, 2008, Saratoga Springs, NY
38th ASEE/IEEE Frontiers in Education Conference
F4H-18
Session F4H
of the evaluation is perceived as being high. Additionally, in Two other evaluation methods were then introduced. Group
highly competitive educational environments, such as the project assignment (team development of a simulation
Spanish engineering schools, evaluation procedures also project) was used as a major evaluation element for two
determine which students do and which ones do not finally years. The other three years, the evaluation involved a
get the engineering degree; the net impact of this filtering is practical, computer based exam, whereby students were
again contingent on the appropriateness of the assessment summoned into a computer lab and assigned a practical case,
and evaluation procedures. for which they individually had to develop a model and
carry out experiments using the simulation software. The
The choice of the most appropriate assessment method(s) is resulting model was then uploaded to the instructor’s system
dependant on a number of parameters, such as the specific for grading.
educational outcome to be measured and the resources
available, since the resource requirements by the various The results have been appraised from both perspectives
assessment approaches differ widely. Proponents of mastery (measurement accuracy and impact on active learning).
exams point at options, such as applying Item Response Assessments were carried out by both faculty members and
Theory to analyze the exam results in order to assess student students (through anonymous surveys). In each case, the first
learning and the focus on the feedback loop to continuously year was considered a “warm-up” period, during which
improve the educational program, that can lead to an overall initial difficulties were ironed out, thus comparative
satisfactory result under certain circumstances [7]. However, measurements took place in the second year. Therefore,
for some educational outcomes, such as ABET’s “soft” there are three sets of data to be compared: pre-2003 data
professional skills, conventional assessment approaches are from the steady-state, final examination based alternative,
clearly not up to the task [8]. and data from 2004 and 2006 corresponding to the second
year of the alternative evaluation methods.
OBJECTIVES AND RESEARCH DESIGN
ASSESSMENT THROUGH A PRACTICAL, COMPUTER BASED
EXAM
Within this framework, the research project presented in this
paper was started in 2003 at the Engineering School of the
University Carlos III de Madrid (UC3M). Its goal was the Until 2002, grading for this course was based on a
structured comparison, in courses in which some of the conventional written final exam. Since a large percentage of
objectives are linked to acquiring practical capabilities in the the coursework was devoted to hands-on simulation work in
use of a software tool, of the impact of alternative evaluation the laboratory, 40% of the simulation part of the written
methods. The incidence of the evaluation methods was exam consisted of questions aimed at assessing the
compared from two perspectives: how accurately do they competence of the students in actually designing and
measure the actual competence level attained, and how do developing simulation models. Additionally, attendance to
they affect active learning by the students. These two basic the practical sessions was monitored, and students were
perspectives had to be complemented with an estimation of required to carry out a set of structured exercises utilizing
resource consumption, in terms of both student time and the simulation software.
instructor time, and the parameters on which this resource
usage was dependent (e.g. number of students enrolled) in To overcome the limitations of written exams in assessing
order to understand the feasibility of their implementation. this type of applied capability, the simulation evaluation was
then split into two different exams. Theoretical concepts
The course chosen, “Quantitative Methods in Management were still tested through a conventional written final exam,
II”, from the Industrial Engineering curriculum, covers accounting for 50% of the grade. For the remaining 50%, an
discrete event simulation and optimization (60% of the on-line, computer based exam was designed.
credits devoted to simulation and 40% to optimization). The
experiment was conducted over the discrete event simulation For the computer exam design there was little former
part of the course. As programming is unavoidable in experience from which to benefit. So a careful design phase
simulation, a substantial part of the student’s effort is was required before implementation. The exam takes place
devoted to developing the capability of constructing models in the same labs as the practical sessions. This has two main
and carrying out experiments using a commercial simulation advantages: the students are familiar with the context, which
software package (Witness®). Traditionally the evaluation contributes to reduce the stress of facing this new exam, and
was solely based on a written final exam. This approach fits the reliability of the computers has been evaluated before the
well for theoretical and numerical exercises, but it was exam so that the real capacity of the lab (in terms of number
considered less adequate for assessing the capabilities of computers expected to be available) is known and the
associated to the use of a software tool. corrective actions in case of a computer failure can be better
planned.

978-1-4244-1970-8/08/$25.00 ©2008 IEEE October 22 – 25, 2008, Saratoga Springs, NY


38th ASEE/IEEE Frontiers in Education Conference
F4H-19
Session F4H
Students are given approx. one and a half hours to perform order of the construction of the system in the two versions.
individually a set of simulation/ programming/ For example, if the students are asked to program a
experimenting exercises, taking as a starting point a file that manufacturing system, it could have a transportation
is copied over the network into each student’s PC directory subsystem and a processing subsystem. In version A the
when they log in. Instructions for the exercises are handed transportation could be first and the processing second, and
out in paper. Exercises of diverse complexity are included to in version B the opposite sequence. To give coherence to
facilitate the discrimination of the various levels of both systems (the one in version A and the “inverted” in
acquisition of the capabilities. At the end of the exercise the version B) the systems may be described as being different,
students are asked to upload their exercises to a server using as long as the logic of the model to be programmed remains
an ftp client application. the same. For example, in system A the transportation
subsystem could be the arrival of a material to the plant, and
A special profile was created for the exam. It gives access in system B it could be transportation of the final product to
exclusively to a network-served Witness® license, a the warehouse. To facilitate this approach several start up
predetermined directory in the PC local disk and the ftp files are copied to every PC, and in the instruction sheet the
client application. The ftp application is configured so that students are asked to work only with the one which
file downloading and overwriting are forbidden and only file corresponds to the version of the exam they receive.
uploading is allowed. Additionally, access to removable
media such as USB is disabled. This special profile along Figure 1 shows an example of a start up working file.
with the use of various exercises versions, guarantee that the
students actually work individually. The possibility of
copying among the students was one of the concerns about
this type of exam, since it opens new ways of interaction
when compared to the written exam (e.g. exchanging
solutions through the server or through e-mail). The
proximity of the computers in the lab and the vertical
position of the screens are also specific characteristics in
these exams.

The experience gained in the first year in which the system


was implemented showed how critical it was that the whole
examination process was thoroughly familiar for the students
beforehand. Thus, practical sessions had to precisely mimic
the examination environment, including downloading the
initial files and uploading the final result. Uploading to the
assigned location in the server the file containing the work
carried out during each practical session provided an
additional way to monitor progress throughout the course,
and allowed for longer exercises, that could be solved over
several consecutive practical sessions. FIGURE 1
WORKING FILE.

The type of exercises students are asked to solve reach the


same degree of complexity as the ones solved in the
practical sessions. To save time and concentrate on the ASSESSMENT THROUGH A GROUP PROJECT ASSIGNMENT
valuable part of the exercises, some of the programming is
already given in the starting up file that gets copied when
they log in. The paper instructions ask the students to As an alternative to the computer-based practical exam, a
complete the programming following a specific sequence group project assignment was used for two years.
until the simulation model of a simple production or service
system is completed (e.g. a manufacturing area of a plant). All students were asked to study through simulation a
For some questions (typically validation proofs) the students specific type of system. For example, in 2006 (when the
are asked to complement the file solution with an survey of this type of evaluation was conducted), the
explanation that they must write on the instruction sheet. students were asked to choose a gas station in their vicinity
Usually, two different versions of the exam are given to the whose “as is” and “to be” queue designs were to be
students, to prevent them from copying. The versions are simulated. The use of the same type of system for all the
carefully designed so that the complexity of the exercises is groups allowed for a highly standardized level of complexity
the same. This can be accomplished, for example, by in all phases of the project. The likelihood of one team
dividing the system in two subsystems and inverting the copying the work of another team was reduced, by forcing
each team to choose a different gas station. This design
978-1-4244-1970-8/08/$25.00 ©2008 IEEE October 22 – 25, 2008, Saratoga Springs, NY
38th ASEE/IEEE Frontiers in Education Conference
F4H-20
Session F4H
resulted from the experience gained from the first edition of project assignment was used. In computer exam as well as in
the project assignment (academic year 2005). On that project assignment, the first year was considered a “warm-
occasion, each group chose a different system. As a result, up” period and was therefore excluded. Quantitative data
some of the groups were more fortunate than others in their included average grades for the capability-oriented and for
election, in terms of feasibility of the study, interest of the the theoretical concepts-oriented part of the grade.
results…These difficulties also affected the faculty Participating students varied according to the year, between
members, leading to a higher effort in coordination and 34 and 57. Anonymous surveys, encompassing both closed
supervision. and open questions, were filled up by the students in the
second year of using the computer-based practical exam and
Another remarkable characteristic of the design of the in the second year of using the group project assignment.
project assignment is that individual members of the team Faculty members involved in the exercise were also
had the freedom to specialize in specific tasks of the project interviewed.
assignment, although they were asked to have a reasonable
knowledge of the overall project. In the report they were On a 10 point scale, average grades for the capability-
asked to make explicit the work distribution among the oriented part of the grade were 3,8 for the 2002 data
members of the group. (conventional written final exam, in which this part had a
40% weight) and 7,2 for the second year of using the
Even though the project assignment is basically an computer-based practical exam (when this part accounted for
alternative to the written conventional exam for the 50% of the grade). As for the second year of using the group
evaluation of the practical capabilities of using the project assignment, the average grade for the assignment
simulation software, it should be highlighted that it also itself, that accounted for 33,3% of the total grade, was 7,6,
helped to attain other important and difficult to fulfill while as the assignment-related individual question in the
objectives. Therefore the design of the assignment written exam, that accounted for another 16,7%, had an
incorporates a variety of objectives. Besides the evaluation average score of 8,1. Average scores for the theoretical
of the practical capabilities in the use of simulation software, concepts-oriented part of the grade were similar in the first
the most important objective stems from the opportunity of two cases (conventional written final exam and computer-
working on an integrative applied problem, which gives based practical exam), with values of 5,4 and 5,5, whereas
participants the opportunity to work in the modeling of real for the project assignment case it was higher, 6,8. This
systems, applying the theoretical contents of the course, and higher result is not surprising, as team project assignment is
developing a complete study from beginning to the end. expected to have a positive impact on the students
Other complementary objectives are team working and understanding of the theoretical concepts.
improving oral and written capabilities.
Survey questions requested students to compare the
The main disadvantage of this approach it that it is much alternative evaluation method they were using (computer
more resource consuming, for both students and faculty, based practical exam or group project assignment) with the
than the other alternatives. It requires, for example, team conventional written final exam, that they were all familiar
work, which has a value on itself, but leads to problems in with since that is the assessment method most commonly
evaluation. Even if the possibility of copying the assignment used at the UC3M. Students were not asked to compare
among groups is not an issue thanks to its design, there is the computer based practical exam with group project
risk that some students within the team act as free-riders. To assignment since they had only experienced one of the
reduce the impact of this potential risk, the group assignment approaches. This comparison encompassed, for the closed
accounted for only 33,3% of the 50% of the grade that was questions, learning outcomes, motivation, soft skills
devoted to the practical capability. The remaining 16,7% development and workload requirements. Open questions
was evaluated through a question related to the assignment enquired about the perceived strong and weak points.
but included in the individual, conventional written final
exam. Theoretical concepts, accounting for the remaining Student feedback was generally very positive regarding
50% of the grade, were tested through conventional learning outcomes, motivation and soft skills development.
questions in this same written final exam.
Thus, on a 5-level Likert item inquiring whether the
adoption of a computer based practical exam (as opposed to
RESULTS AND DISCUSSION a conventional written final exam) increased the student’s
motivation to proactively engage in the practical sessions,
As described above, three sets of data were used for the 71% of the respondents agreed (responses 4 or 5). Average
comparison: 2002 data representing the steady state while score was 3,77, standard deviation 1,19.
using only the conventional written final exam, data for the
second year in which the computer-based practical exam Similarly, 86% answered that it had led to a higher level of
was used and data for the second year in which the group knowledge of the software tool, and 84% considered that

978-1-4244-1970-8/08/$25.00 ©2008 IEEE October 22 – 25, 2008, Saratoga Springs, NY


38th ASEE/IEEE Frontiers in Education Conference
F4H-21
Session F4H
unsupervised individual, proactive work at the lab had been criteria is being appraised. While computer exam allows for
useful for their preparation. a more accurate individual evaluation of the practical
capacity of software use, group assignment adds up other
However, only 30% thought that it had led to a better grasp important formative assets related to the whole of the course
of the theoretical concepts; that result is consistent with the (not only the practical capability of software use).
negligible impact observed on the average scores for the
theoretical concepts-oriented part of the grade (5,5 vs. 5,4 However, increased workload requirements for both students
with the conventional exam). and instructors, particularly for the group assignment option,
require careful resource planning before implementation.
On the other hand, workload requirements were perceived
by 54% of the students to be higher than with traditional
methods. REFERENCES

As for the open questions, in the case of the computer based [1] Crosier, D, Purser, L, Smidt, H, "Trends V: Universities shaping the
practical exam, most students stated that this evaluation European Higher Education Area", European University Association
procedure was more appropriate for the subject matter, and report, 2007.
therefore provided a fairer and more precise assessment. A [2] Education Ministers of Bologna Process countries, "London
significant number of responses also stated that it led to a Communiqué - Towards the European Higher Education Area:
deeper learning, even though it required additional effort. responding to challenges in a globalised world", 2007.
[3] Huba, M E, Freed, J, " Learner-centered assessment on college
83% of the students were in favor of maintaining the campuses: Shifting the focus from teaching to learning", Needham
computer based practical test, while as only 10% preferred a Heights, MA: Allyn-Bacon , 2000.
conventional written final exam and 7% had mixed feelings. [4] Ministerio de Educación y Ciencia de España, "Real Decreto
1393/2007, de 29 de octubre, por el que se establece la ordenación de
las enseñanzas universitarias oficiales", Boletín Oficial del Estado,
Regarding the project assignment, student feedback was No. 260, 2007, pp. 44037-44048.
quite similar, highlighting the positive impact on the
learning outcome. However, in this case the perception that [5] ABET, "Criteria for Accrediting Engineering Programs. Effective for
evaluations during the 2007-2008 accreditation cycle ", Engineering
workload requirements were higher than with traditional Accreditation Commission, Baltimore, MD, 2007
methods was much more acute; 100% of the students
[6] Rojstaczer, S , "Grade Inflation at American Colleges and
thought so, and over 50% described the workload Universities", Accesible at www.gradeinflation.com/, 2003.
requirements as “a lot heavier”.
[7] Qualters, D M et al., "Improving Learning in First-Year Engineering
Courses Through Interdisciplinary Collaborative Assessment",
From the faculty members’ perspective, the feedback was Journal of Engineering Education, Vol 97, No 1, 2008.
very similar, with a very positive perception of the
effectiveness of the alternative evaluation methods in [8] Shuman, L J, Besterfield-Sacre, M, McGourty, J. " The ABET
“Professional Skills” – Can They Be Taught? Can They Be
promoting the active learning of the students but at the same Assessed?", Journal of Engineering Education, Vol 94, No 1, 2005,
time leading to a much heavier assessment workload, pp. 41-55.
particularly for the project assignment option. As for their
ability to precisely and fairly measuring the knowledge
acquired by the students, both methods were considered
superior to conventional exams. The project assignment and
computer based practical exam allowed the faculty to
properly assess the level acquired by the students, although
in the case of group assignment what was accurately graded
was the team as a whole, not the individuals. In an attempt to
mitigate this intra-team blurness, 16,7% of the grade was
evaluated through an individual, assignment-related question
in the final exam.

CONCLUSIONS

Results suggest that both group assignments and computer


exam perform far better, in this environment, than the
traditional written exams. The comparison between group
assignments and computer exam is less straightforward since
their relative impact is dependant on which of the chosen
978-1-4244-1970-8/08/$25.00 ©2008 IEEE October 22 – 25, 2008, Saratoga Springs, NY
38th ASEE/IEEE Frontiers in Education Conference
F4H-22

You might also like