Professional Documents
Culture Documents
Cassandra Humphrey
Dr. Bude Su
Table of Contents
I. Introduction ............................................................................................................................ 3
II. Methodology.......................................................................................................................... 3
Prototype ........................................................................................................................................... 3
Learners ............................................................................................................................................. 4
Tryout Conditions ............................................................................................................................... 5
Process ............................................................................................................................................... 5
Pre and Post Tests ......................................................................................................................................... 5
Observation .................................................................................................................................................. 6
Feedback....................................................................................................................................................... 7
IV. Summary............................................................................................................................. 11
V. Appendices........................................................................................................................... 12
A. Demographics Survey................................................................................................................... 12
I. Introduction
In the CS Online program, students are subjected to a streamlined, fast-paced course load.
Many times, their classes overlap, and students may find themselves taking a final for one class
and attending an orientation for another within the same weekend. However, a couple of courses
require more drawn-out preparation. CST 361 is the first of these courses, and students are
expected to begin preparation about two months before the course begins. Due to the workload in
their current classes, many students find it difficult to attend synchronous information sessions so
far in advance, and view counts of recorded sessions have been consistently low. To combat this,
an asynchronous module was created in the hopes that students would be able to get background
This test aims to determine whether the prototype is functional and usable, while
investigating knowledge gain, user reactions, and areas for improvement in terms of design,
II. Methodology
Prototype
informational module created with Adobe Captivate. The module, titled CST 361 Overview,
introduced the users to the information required to begin preparation for the computer science-
based Service Learning course. It reviewed background information such as why the course is
required, how it is unique, and what steps future students will need to take to prepare before the
course begins.
While the module is fairly locked down in terms of navigation, users are required to
perform interactive tasks such as clicking buttons and answering quiz questions via several
MODULE EVALUATION 4
formats such as multiple choice, multiple selection, short answer, and drag and drop. An average
learner is expected to spend around seven minutes on the module but may take as long as ten
If the module proves to be an effective instruction tool, this prototype may be used as a
supplement to face-to-face preparation information sessions for future CST 361 courses.
Learners
This module is designed for CS Online students who are about halfway through the
program. According to archival demographic data of the first three CS Online cohorts collected
by Claudia Carter, a majority of CS Online students are male, with the percentage ranging from
69% to as high as 88% depending on the cohort. Most students fall in the 26-35 age range, with
only one cohort having a majority of 18 to 25-year-old students, and 70.6% of students reported
In the hopes of providing the most realistic testing results, the thirty-three students of
Cohort 6, who are scheduled to take the course in the spring of 2019, were invited to participate
in the module usability and effectiveness testing. However, only three students volunteered, so
the breadth of test subjects expanded to include two fully online MIST students and one student
from another college for a total of five test subjects. In order to keep the subjects as close to the
CS Online demographics as possible, they needed to have taken at least one online college class,
fall between the ages of 18 and 55, and have had minimal previous exposure to information
Three respondents were male and two were female, 100% of testers had taken at least one
fully online college course in the past year, 60% were between the ages of 25-34, and 60%
Tryout Conditions
The tryout was conducted in the same format for each tester. In order to reflect CS Online
students’ environments as closely as possible, the testing was recorded via Zoom while the
volunteers completed the module in their homes on their personal computers. The participants
used a range of operating systems and browsers. Most used Windows computers, and Chrome
was the most popular browser, but Safari and Firefox were used by one student each. Users were
asked to use which browser they were most comfortable with to alleviate any stress or issues that
Although the volunteers were in different environments: at home alone, walking in the
door from work, or taking care of a young child, this is fairly accurate of CS Online student
experiences. Everyone was able to successfully complete the module, even with the varying
Each volunteer was given a link to the Google Form Demographic Survey and the
module, which was hosted on the developer’s ITCDland account. The developer created two
more Google Forms to record responses for the pretest questions and final feedback. However,
these questions were asked verbally and transcribed by the developer as users answered in an
Process
The pre and post tests were conducted in two different formats. Although initially more
difficult to quantify and compare, this was done deliberately to prevent students’ posttest scores
from being unintentionally affected. The pretest was open-ended and was conducted verbally in
an interview style. Users did not have access to the questions beforehand to discourage prior
MODULE EVALUATION 6
research. This way, knowledge of the main topics of the course were reviewed without responses
being swayed by guessing, or by preemptively tipping students off to specific areas of interest in
the model. The posttest was conducted as an end of the module quiz, where students were
required to independently answer multiple choice, multiple selection, short answer, and drag and
The pretest responses were graded and weighted based on keywords that aligned with the
posttest answers. Posttest responses were weighted based on the number of necessary responses
required to answer the question correctly. The number of responses mirrored the number of
keywords in the pretest. Care was taken to make the grading criteria as closely aligned as
possible. Each test was out of thirteen total points. All questions for a particular topic received
the same amount of points on both tests. For example, students were questioned on the hours
requirement in both tests and received one point if they were able to respond with the correct
number. See Appendices C, D, and E for a full report of questions, responses, and rubrics.
Observation
The usability and effectiveness test was conducted fully online and synchronously on
Zoom.us in meeting rooms created by the developer. Each test occurred as a one-on-one session
between May 21 to June 5, and all tests occurred on weekdays after 5pm. The total time for each
test ranged from fifteen to thirty minutes, and consisted of a pretest, demographics survey,
module test with embedded posttest quiz, and a feedback survey. Each session was recorded and
posted to the developer’s Zoom cloud account, with copies added to Google Drive as backup.
In each session, the volunteer was given a short brief on the purpose of the module and the
components of the evening’s test. The developer administered the pretest and then shared the
demographic survey link with the user for them to fill out while the developer prepared the
MODULE EVALUATION 7
module. Typically, demographics information comes first, but in this situation, the developer
knew the participants personally and knew they fit the target characteristics before beginning.
Demographics in this case were used for analysis and grouping after the test, not for screening
participants.
After the initial forms were completed, the developer shared the link to the learning
module with the user and confirmed what operating system and browser was being used. The
user shared his or her screen with the developer during this test, and the observation checklist
was filled out for each session (see Appendix B). The developer wanted to observe how students
interacted with buttons, navigation, and other components of the module, so they were not given
any specific directions on how to progress through. One student questioned how to move
forward from the introduction slide but was able to figure it out herself without intervention. The
only task that was formally requested of the tester was turning on the closed captions to ensure
that the accessibility feature worked on different browsers and operating systems.
Feedback
When the module was completed, the user was permitted to stop sharing their screen, and
the developer transcribed the users’ spoken responses to the feedback survey (see Appendix H).
This was a time for participants to give final thoughts on their experiences with completing the
module. They were able to ask additional questions or give feedback on the module’s design,
After all data was collected and the testing was finished, the developer thanked the tester
III. Results
Entry Conditions
MODULE EVALUATION 8
The predicted and observed entry conditions were consistent. All of the testers had
minimal knowledge of the course, as outlined in Appendix C. Each one had little to no
knowledge of the specifics of the course, although some were able to summarize the general idea
of helping a nonprofit or similar entity. All users completed the module within the predicted
timeframe, with one tester who had no experience with the course taking the full ten minutes, as
she chose to review before the quiz. All others took around seven or eight minutes to complete
the module.
Instruction
Overall, the course instruction worked as intended. All users were able to navigate
through the course, review previous slides if they chose to, all pages loaded properly on all
The only discrepancy between intended and observed usage was with an ungraded
“Check Your Knowledge” question. This multiple-answer question appears right after a slide
that outlines the three main requirements for a service learning project. The directions ask
students to select the three requirements and check their answer before moving on. Four of the
five students chose three options, but one student only chose one. Adding in validation that
prevents users from moving forward unless they choose three requirements should prevent this
Outcomes
Based on analyses of pre and post test scores (see Appendix F), it can be concluded that
the module did contribute to users’ knowledge. Their pre and post test scores were calculated
based on their respective rubrics (see Appendix E), and then this data was used to calculate a t-
Test: Paired Two Sample for Means summary (see Appendix G) to illustrate the effect the
MODULE EVALUATION 9
learning module had on students’ knowledge of CST 361. The absolute value of the t-Stat,
t(4)=26.5, is well above the critical value of 2.132 (.05 alpha level, df=4).
Although great care was taken to keep the pre and post test questions and rubrics as
similar as possible without unintentionally giving away answers or highlighting upcoming topics
before the users had a chance to do the module, the difference in formatting between the pre and
At such a small scale, however, it is important to note potential for bias in results before
generalizing to project CS Online cohort results. As shown in Appendix A, all users that
participated in the module rated themselves neutral, interested, or very interested in the subject
matter before testing. In the feedback survey, all respondents reported being either interested or
very interested (Appendix H). This positive attitude towards the course or content may have led
to higher scores as well. In terms of enjoyment and usability, all students found the course easy
to navigate, all reported that they felt ready to begin preparation for CST 361, and one user was
very satisfied with the overall look and feel of the module, stating, “I thought it was very clear,
not over-complicated visually but still pleasing to look at. I liked the color scheme, graphics, and
However, it is not expected that all students will enjoy this module. As observed from
past run CST 361 courses, the topics and requirements of the class can be controversial to
students based on their personal backgrounds and prior experiences. Therefore, the goal is to
keep the module objective, providing background, requirement, and preparation information in a
straightforward and unbiased manner that can be completed quickly. While it is hoped that the
module is visually pleasing and interesting to students, the main goal is for it to deliver
Recommendations
Based on developer observations and user feedback, there are three main
First, updated versions should provide clearer navigation controls. Currently, the
narration verbally (or via closed captions) instructs users on how to move forward on the first
slide. An arrow appears and points to the next button as well, but several users requested more or
clearer instruction on navigation, while others would have liked the slides to progress forward
automatically.
One student’s hidden toolbar would appear and cover the navigation buttons when they
moved their mouse down to click the next arrow, so providing alternative ways to navigate such
as integrating arrow key commands or allowing for automatic progression may make this
Second, updated versions should replace the computer-generated narration with a human
voice. One tester stated that the current voice over made the module more challenging to
complete, saying, “As I was listening to it, I was kind of deciphering what they're saying. It made
After the module, students had questions regarding course timing and appropriate projects. While
the module is intended to be sent to students with a message explaining an overview of cohort-
specific dates, more information on types of projects and partner communication in the module
may help clarify the common questions that arose during testing. Currently, users are not
exposed to a project scenario until the quiz, although they are given the requirements during the
IV. Summary
As the analysis, observations, and student scores show, the learning module proved to be
functional and effective in introducing users to CST 361’s unique requirements. Requests for
future improvements were minor and doable, and the test shows that users will be able to
successfully review the module even with minimal knowledge about the course.
It is recommended that testing is repeated with new users after the module is updated,
preferably with individuals with lower interest levels to eliminate potential bias.
However, based on the data collected here, this module could potentially be completed by
students well before they need to begin preparation for CST 361, ensuring a smooth transition
into the course. Because some students still had specific or personal questions after reviewing the
module, it is recommended that the faculty still holds a face-to-face session before the class
V. Appendices
A. Demographics Survey
361?
B. Observation Checklist
C. Pre-Test
4. Have you heard about any course requirements, specifically required documents?
• I don’t know any specifics, but my impression is that it is a big project.
• No, I haven’t.
• It’s a project that has to be done for a nonprofit. It has to be a free project. But besides
that, I haven’t heard anything.
• No.
• No.
5. Do you know of any steps you will need to take to prepare before the course starts?
• No. Potentially ideas of how to help a company locally.
MODULE EVALUATION 14
D. Post-Test
2. You must complete [] hours at your Service Learning site (Fill in The Blank)
30
Pre-Test Rubric
Question
# 1 2 3 4 5
MODULE EVALUATION 15
Post-Test Rubric
Question 1 2 3 4 5 6
Points 1 1 4 1.5 1.5 4
t Stat -26.5
P(T<=t) one-tail 6.02596E-06
t Critical one-tail 2.131846786
P(T<=t) two-tail 1.20519E-05
t Critical two-tail 2.776445105
H. Feedback Survey
• I thought the formatting was good with asking questions right after the material, that was
really nice. And then the interactivity, being able to toy around with the questions was a
nice addition.
• I mean, I don't see anything that can really be improved. It's pretty well explained in the
slides and everything. My main thing is just scheduling when this will be released in
correspondence to the course.
• Sometimes I forgot that I had to push the next button to move on. I wish it continued
automatically.
• Overall, it was perfect. At the beginning of the module, I wasn't sure how to move
forward.
• The part about the arrow was a little confusing. Maybe be more clear on how to progress.
I thought it was very clear, not over-complicated visually but still pleasing to look at. I
liked the color scheme, graphics, and animation. I like a human voice. As I was listening
to it, I was kind of deciphering what they're saying. It made it a little hard to follow in
some parts.
I. Archival Data