Education Analytics
Education Analytics
This learning analytics study looked at the various student characteristics of all
on-campus students who were enrolled in 100 and 200 level courses that were offered in
both online and face-to-face formats during a two-year period. There is a perception that
difficult for students. The results of this study show this is not the case.
The goal of this study was to complete an in-depth analysis of student profiles
related variables to reveal any patterns for student success in either online or face-to-face
courses as measured by final grade. There were large enough differences within different
population, but overwhelmingly, the most significant predictor of success was found to
Further analysis was completed on students who declared high school credit as
their primary major based on significantly different levels of success. These students
were concurrent enrollment students or those who completed college courses for both
high school and university credit. Since most of these students were new to the
university, they did not have a cumulative GPA, so other predictive factors were
explored. The study concludes with recommendations for action based on the logistic
TABLE OF CONTENTS
Section No. Section Name Page
No.
1 Requirement Statement / Problem Statement
2 Project Objectives
3 Project Scope
4 Methodology
5 Plan vs Progress
6 Detailed description of project completed till Mid-Semester Report
7 Resource Requirements and their availability
8 Risks and Mitigations
9 Issues and Resolutions
10 Conclusions and Recommendations
Annexure-1 Survey Questionnaire (if any)
Annexure-2 Computer Programs / Code
Annexure-3 Detailed workings (if any)
Annexure-n Any other details to be added A
References
Glossary
Summary of how the feedback for Project Outline have been
addressed
1. Background
Students across the India are choosing to continue their education beyond high school at an
increasing rate. In 2012, approximately 41% of the population of 18-24- year-olds were
enrolled in an institution of higher education (National Center for Education Statistics, 2014b).
Ten years earlier only 36% of 18-24-year-olds opted to enroll in college. Online learning is
growing at an even faster rate than overall enrollments. In 2014, about 28% of post-secondary
students were enrolled in at least one distance learning course (Allen & Seaman, 2016; Hart,
2012). In contrast, in 2002, less than 10% of students opted for distance learning. The
university that was the basis of this study experienced growth in the overall student population
as well as online course enrollments. The fall 2014 enrollment was approximately 29,100
students, nearly a 20% increase from just ten years earlier. Of these students, about 11,400, or
39%, were enrolled in at least one online course. Following the national trend, the university
saw a 13% decrease in the number of students enrolled in exclusively face-to-face courses over
the past two years (eCampus Center, 2015).
Despite the growth in higher education enrollments, both online and face-to-face, retention of
students until a degree is earned is a concern. Retention is defined as an institution’s ability to
retain a student from either admission to graduation, or from one term to the next (Berger &
Lyon, 2005). Retention rates are calculated by determining the 2 percentage of students who
reenroll in the university for the next term. Nationally, the retention rate of full-time students
from year to year is 71.8%, but when students are enrolled only part-time, the retention rate
drops to 42.2% (National Center for Educational Statistics, 2014a). The university that was the
focus of this study saw significant growth in year-to-year retention of full-time students, both
face-to-face and online, over the past ten years. This number grew from 58.6% in 2002 to
71.7% in 2012, which is very close to the national average (Office of Institutional Research,
2013). Persistence is a term that is often used in relation to retention. Retention is measured
from the perspective of the university, while persistence is reenrollment or the desire to reenroll
from the student’s point of view. Students make decisions about whether to persist in their
education based on a number of factors. Researchers established a number of theories on why
students persist in their education starting in the 1970s (Astin, 1975; Bean & Metnzer, 1985;
Spady, 1970; Tinto, 1975). These theoretical frameworks consider how the needs of the
individual student align with what the institution offers to students. Astin (1975) proposed that
students enter the university system with their unique set of inputs, including demographics,
high school grades, and reasons for wanting to attend college, among many others. It is the
interaction between the inputs and the higher education environment that determine the
educational outcome. Additionally, Tinto (1975) proposed an interactional theory of retention.
His theory suggested that there are multiple interrelated reasons as to why a student might not
persist in their education. The studies completed by Astin and Tinto both address the person
who enters the university system and how their personal characteristics and past experiences
can impact their education success. This framework served as the foundation for this 3 study.
In addition to the factors described in the persistence theories, academic factors can influence a
student’s decision as to whether to persist in their education.
3. Motivation for this project
This study addressed demographic, academic, and course related factors of oncampus students
and analyzed their success rates in 100 and 200 level courses taken either online or face-to-face
at a university in the west. Only enrollments in courses that were offered in both formats
between the Fall 2013 semester and the Summer 2015 semester were included in the data
analysis. These factors led to the following research objectives: 1. Which are important
predictors from student characteristic profiles that lead to successful completion of 100 and 200
level classes taken online, as measured by final grade? 2. Which are important predictors from
student characteristic profiles that lead to successful completion of 100 and 200 level classes
taken face-to-face, as measured by final grade? 3. What predictors are common or differ
between online and face-to-face settings? 4. Which academic departments or individual courses
can be identified as significant and in need of further analysis?
5. Project Scope
Poor academic achievement is second only to financial reasons for the lack of student
persistence in higher education (Bean, 2005). Academic achievement can be measured
by grade point average (GPA), test scores, class rank, or final course grades. In addition
to academic achievement, demographic, and cultural factors, the structure of the courses
a student chooses and the level of student self-regulation can influence how a student
does in school, and in turn, affect the likelihood of a student persisting until degree
completion. All of these factors contribute to a student’s set of entry characteristics.
Table 1 provides a summary compilation of several key student predictors and the study
reporting the data. Academic Factors Class status is one of the top academic predictors
of success in both face-to-face and online courses. The longer a student has been in
school, the more likely he or she is to complete a degree (Hart, 2012; Levy, 2007;
Moore & Kearsley, 2005; Wang & Newlin, 2002). Several studies found grade point
average (GPA) to be positively correlated with success in individual courses.
6. Proposed methodology
Campbell and Oblinger (2007) and Pardo (2014) described a process for learning analytics that
includes five stages. This study adopted the five stage process of capture, report, predict, act,
and refine. This process was used to address the following research questions: 1. Which are
important predictors from student characteristic profiles that lead to successful completion of
100 and 200 level classes taken online, as measured by final grade? 2. Which are important
predictors from student characteristic profiles that lead to successful completion of 100 and 200
level classes taken face-to-face, as measured by final grade? 46 3. What predictors are common
or differ between online and face-to-face settings? 4. Which academic departments or
individual courses can be identified as significant and in need of further analysis? For the
purposes of this study, completion of a course was considered successful if a student earned a
grade of a C- or better. This definition was chosen because the university requires students to
earn a C- or better in all prerequisite courses in undergraduate programs. Participants The data
collected for this study was the entire population of on-campus students who were enrolled in
the set of 100 and 200 courses that are offered in both online and face-to-face formats between
the Fall 2013 semester and the Summer 2015 semester at the university. The collection of 100
and 200 level courses was selected because the university offers multiple sections of these
courses in both formats every term. Blended courses were excluded from the study.
Additionally, these courses have higher enrollments than many upper division courses, since
they often function as service courses. Service courses are courses that are offered by one
academic department but are required for many degrees or certificates. For example, anatomy
and physiology is a course offered by the biology department but is required by degree
programs ranging from kinesiology and nursing to criminal justice and social work.
7. Plan of Work
I am looking forward to being timely in doing all needed steps. Downloaded SPSS and will try
to use it in the best possible way.
Introduction
This chapter addresses literature relevant to the study. The first section reviews
the emphasis of retention and graduation rates for both face-to-face and online as higher
education has evolved over time in the United States. Later sections address retention
theories and factors that affect persistence as well as factors that affect student
achievement. The next section discusses some of the differences between online and
face-to-face course delivery models, and the final section reviews how learning
analytics and data mining have been used to explore student success.
Institutions of higher education were established in the United States long before
the country was founded. Many of the early institutions were founded with religious
freedom in mind. Their goal was to provide religious education for future ministers
(Geiger, 2015; Snyder, 1993). At that time, the focus of the universities was to facilitate
the spread of religion as opposed to retention of students, so records of this nature were
not kept.
curriculum, shifting beyond religious studies to a focus on the classical topics such as
classical languages, ethics, philosophy, and the sciences (Berger & Lyon, 2005;
Snyder, 1993). Also during this time, American higher education began to include
normal schools, two-year institutions designed to prepare teachers for the public school
system. Enrollment in higher education during the nineteenth century was very
exclusive.
(Snyder, 1993). Because of the elite status for university level students, retention was not
perceived as an issue and therefore was not tracked (Berger & Lyon, 2005).
The first evidence of distance education was found in Europe during the same
time higher education in the United States was in its early expansion. As early as the mid-
1800s, students in Great Britain were learning shorthand through courses offered via the
postal service. Language classes were offered in both France and Germany using a
similar approach. Learning through correspondence began in the United States a few
decades later (Colorado & Eberle, 2010; Moore & Kearsley, 2005). These courses had a
which, in turn, caused an increase in demand for a more highly educated workforce
(Berger & Lyon, 2005). This need enabled universities to either grow or become more
selective in who was accepted as a student based on the institutional goals. Universities
with increased enrollments, particularly those that were less selective in who was
accepted, began to track retention of students. The first report on retention was released
in 1938 (Berger & Lyon, 2005). This report, entitled College Student Mortality,
examined dropout rates at several universities in the 1930s. It considered the time it took
students to complete a degree as well as the impact of several student factors including
gender, age, work status, living arrangements, and location of home as compared to
university location. During this time, some innovative institutions implemented distance
education courses over radio broadcasts (Colorado & Eberle, 2010; Moore & Kearsley,
2005).
Major world events during this time frame had an impact on enrollment trends at
the higher education level. World War II had a significant effect on enrollments since
societal efforts were focused on the war as opposed to getting an education. As a result,
college enrollments dropped 20% between the 1939-1940 and 1943-1944 school years
(Snyder, 1993). Male students were a much higher portion of the group that departed
college as compared to females. However, once the war was over, enrollment numbers
grew quickly. This growth is partially due to the GI Bill that was passed by congress in
1944 to provide incentives for veterans of the war to take advantage of higher education
distance learning institute around the time of the beginning of World War II. This
military based organization offered both high school and college level courses to
members of the military (Moore & Kearsley, 2005). There were opportunities for
correspondence courses, telephone based education, and courses offered via television.
These models allowed people to continue their education wherever they were located.
enrollments. This event helped to create the mindset that getting a higher education
would help strengthen the United States as a whole. Soon after the Higher Education Act
was passed, in 1965, providing grants and low-interest loans to help students pay for
their education (Bean & Metzner, 1985; Berger & Lyon, 2005; Bok, 2013). This surge
transitioned enrollment in institutions of higher education from the elite to commonplace,
leading to a more diverse student body (Berger & Lyon, 2005; Bok, 2013). This growth
also brought students to the university system who lacked the proper preparation to be
successful. Students did not know what to expect either academically or socially, and
colleges were not prepared to provide that information to students. As a result, the more
diverse student audience brought an increase in dropouts (Berger & Lyon, 2005).
enrollment, student persistence, and satisfaction with the educational experience (Berger
& Lyon, 2005). Two major studies completed in the 1970s examined college dropouts
and a variety of factors that may have contributed to students leaving the higher
education system. Spady (1970) looked at environmental factors, while Kamens (1971)
compared dropout rates to the size and prestige of the institution. These studies
determined that there were higher dropout rates at larger institutions. The large
institutional experience was less personal because students had fewer opportunities to get
to know the faculty teaching their courses (Kamens, 1971). He also found that students
who attended a university that was perceived as more prestigious regarded their
education as having more value thereby making them more employable. Studies like
those completed by Spady (1970) and Kamens (1971) led institutions to be more strategic
in their enrollment practices. Universities worked to select students with more academic
and social preparedness, specifically students with research and writing practice, which
several post-secondary institutions. By the 1980s, the Electronic University Network had
over two hundred television based courses available to learners across the United States,
most were available on public broadcasting stations (Moore & Kearsley, 2005). These
courses were some of the early attempts to provide expanded flexibility for learners.
As higher education transitioned into the twenty-first century, retention rates were
still lower than desired. Dropouts ranged from a low of 8% at private elite institutions to
a high of 50% at open enrollment colleges (Berger, & Lyon, 2005). Before this time,
most institutions were single mode institutions, offering only one mode of instruction.
models. Some expanded to operating as dual mode institutions, offering two modes of
instruction, most often face-to-face and distance learning options. Still other institutions
had individual faculty members who opted to move their courses online. Most
institutions, offering a mix of face-to-face and online course modalities, were created
with the forethought of a sustainable model, however, when a single faculty member
chooses to move their course online without institutional support, they often do not
endure (Moore & Kearsley, 2005). Many institutions of higher education found
sacrificing the existing student population. This expansion also continued to grow the
used a combination of correspondence and media including video, via live broadcasting
or video recordings, audio, printed study guides, with assignments submitted via mail
(Colorado & Eberle, 2010; Moore & Kearsley, 2005). Another multimedia course
delivery model implemented during the late twentieth century was teleconferencing.
Teleconferencing used either one-way or two-way communication using video (Moore &
Kearsley, 2005).
The next phase of distance learning was centered on the use of computers and the
Internet (Colorado & Eberle, 2010; Moore & Kearsley, 2005). Use of this technology
allowed for a multimedia experience combining the use of text, graphics, audio, and
video in the learning experience. The phrase online learning is synonymous with distance
learning via the Internet. Early iterations of online learning were not much more than
Online learning became much more feasible and more widely adopted with the
advent of the learning management system (LMS). Learning management systems and
their improvements came in three waves. Early learning management systems provided a
structured environment for sending and receiving documents. The arrival of Web 2.0
opportunities for students to interact with the content in real time. The next, and most
recent, significant change in online learning came with combining the field of data
analytics used in business and industry with the learning management systems in learning
student retention or lack thereof (Astin, 1975; Bean & Metzner, 1985; Spady, 1970;
Tinto, 1975). Many early persistence frameworks were based on a suicide theory. These
theories worked under the assumption that a combination of academic and social
integration into the environment was critical to thriving. If the student felt they did not fit
in, either academically or socially, then they were at risk of dropping out or ending their
life at the institution (Spady, 1970; Tinto, 1975). Because of the era in which these
theories were created, they were focused on face-to-face students. However, they can be
Outcome model. He theorized that students enter higher education with a number of
foundational characteristics, or inputs, that influence their ability to persist. The input
variables include demographic characteristics, high school grades, and reasons for
wanting to attend college, as well as many other factors. Astin also identified a number of
environmental variables that were likely to affect the likelihood of success for students.
Environmental factors included variables related to the institution, like size and location
of the university; factors related to the faculty, including teaching methodologies and
values; and characteristics related to the student, including the type of residence, the level
of extracurricular involvement, academic major, and peer group factors. Astin considered
the output variables the results of the environmental variables on the input variables
(Ishler & Upcraft, 2004). The outcome variables include satisfaction with the
person might not persist in their education, to propose an interactional theory of college
departure. The theory is labeled as interactional because there are often multiple
interrelated reasons why a student chooses to leave school. Astin’s and Tinto’s theories
intersect at the point that they both consider the set of characteristics that a student has
when beginning their higher education experience (Ishler & Upcraft, 2004). Tinto’s
theory includes both sociological and psychological reasons for students to drop out or
stop out of their education (Braxton & Hirschy, 2005). Bean and Metzner (1985) added
organizational reasons to the theories for lack of persistence. All of the persistence
not reenroll for reasons determined by the institution (Berger & Lyon, 2005; Ishler &
Upcraft, 2004). The institution may deny reenrollment due to serious misconduct or
consistent failing grades. Voluntary departure most often occurs when a student feels the
Sociological Factors
Sociological reasons for persistence are related to the degree to which a student
recognizes the value of their education in relation to their career goals (Habley et al.,
2012). In conflict, lack of student retention may occur when students feel like they do not
fit into a university due to differences between their culture of origin and the culture of
the university (Braxton & Hirschy, 2005). Students may be influenced by pressures for a
certain level of academic performance, and if they are unable to achieve that expectation,
they could opt to withdraw from school. This issue can be minimized if institutions and
courses emphasize building a community. This often results in higher levels of student
satisfaction, and consequently, a higher rate of retention (Lotsari, Verykios,
educational experience.
Psychological Factors
Internal factors that can influence persistence include academic success, motivation, self-
esteem issues, and study habits. Student motivation and perception of learning can also
affect their persistence in school. Some students are only looking for surface level
learning, meaning they simply want to pass the test and get a grade. These students may
get less out of their educational experience than those looking for a deeper level of
learning. These students are looking to relate new information to previous knowledge,
find patterns in the content, and gain a deep understanding of the underlying principles
External factors can also influence a student’s decision to stay in school. These
factors include family issues, time constraints like employment demands, as well as the
perceived level of support and encouragement from family, friends, and coworkers (Bean
& Metzner, 1985; Park & Choi, 2009; Tello, 2007). External factors are likely to be more
prevalent in nontraditional students, particularly those who need to balance family, work,
and school aspects of life. These are the same factors that often cause students to choose
online courses as opposed to face-to-face options (Pontes, Hasit, Pontes, Lewis, &
Siefring, 2010).
Organizational Factors
Bean and Metzner (1985) were the first to consider retention from an
interest in getting students to stay in school until a degree is earned. Persistence requires
students to conform to the organizational norms of the institution, but the institution plays
Students must have the proper academic aptitude and skill along with personality
traits that allow them to integrate themselves into the college environment (Braxton &
Hirschy, 2005; Park & Choi, 2009). If a student does not fit into the organizational
norms of the institution, it can affect their level of satisfaction with the university. Tinto
(1975) found that students needed to adapt to the routine of the institution. They need to
learn how to participate and communicate to fit into the college environment both inside
and outside of the classroom. This adaptation is dependent on the structure of the
university as well as the flexibility of the student. If this integration does not take place, a
student is much more likely to drop out of the institution. These learning communities
exist in both the face-to-face and online learning environments. Institutions can
encourage opportunities to ease student adaptation to the organization through the use of
Often orientation activities are a student’s first exposure to the higher education
well as the learning communities that they will become a part of as they move forward in
their education. Academic advising should take place in conjunction with the orientation,
setting the student down the proper path to academic success (Ishler & Upcraft, 2004).
Economic Factors
While not included in the theories established in the 1970s, current-day students
(Braxton and Hirschy, 2005). The current average cost of tuition, fees, room and board
for a full-time undergraduate student is approximately $20,000 per year. About 84% of
full-time undergraduate students rely on financial aid in the form of grants, loans, work-
study, or other sources to help cover these costs (National Center for Educational
Statistics, 2015). Many students struggle to see the return on investment of time, money,
and effort put into their education, thus select other career options that do not require
further education. The time spent working to pay back loans can also be a deterrent to
continuing in school until a degree is attained. On the other hand, financial aid can
provide opportunities for some highly motivated students who might not otherwise be
Another economic factor that can affect students is the state of the economy. A
poor economy can mean fewer jobs are available, motivating unemployed people to
contrast, when the economy is thriving, students may choose to stopout of school in favor
of a job. On the other hand, a strong economy may push students to be more successful in
their coursework, in the hopes that there are jobs waiting for them once they graduate
measured by grade point average (GPA), test scores, class rank, or final course grades. In
addition to academic achievement, demographic, and cultural factors, the structure of the
courses a student chooses and the level of student self-regulation can influence how a
student does in school, and in turn, affect the likelihood of a student persisting until
Academic Factors
Class status is one of the top academic predictors of success in both face-to-face
and online courses. The longer a student has been in school, the more likely he or she is
to complete a degree (Hart, 2012; Levy, 2007; Moore & Kearsley, 2005; Wang &
Newlin, 2002). Several studies found grade point average (GPA) to be positively
correlated with success in individual courses (Aragon & Johnson, 2008; Campbell,
DeBlois, & Oblinger, 2007; Dupin-Bryant, 2004; Harrell & Bower, 2011; Hart, 2012;
Jayaprakash, Moody, Laura, Regan, & Baron, 2014; Menager-Beeley, 2001; Morris, Wu,
& Finnegan, 2005; Muse, 2003; Osborn, 2001; Shelton, Hung, & Baughman, 2015;
Valasek, 2001). Some of these studies also found that both the verbal and mathematic
scores on the SAT are strong predictors of academic success (Campbell et al., 2007;
Cortes, 2013; Morris et al., 2005). McKenzie and Schweitzer (2001) reported academic
Table 1 Predictors of Retention for Various Student Characteristics
Relationship of Characteristic to
Student Characteristic Studies Addressing Characteristic
Academic Retention
Academic Advising and More support is positively Swail (2004)
Support correlated with persistence
Face-to-Face Only: Adelman (1999); Thayer (2000)
Online Only: Ivankova & Slick (2007)
Academic Level/ The further in school is a positive Online Only: Dupin-Bryant (2004); Levy (2007); Muse (2003);
Year in School * predictor for online course success Osborn (2001)
Academic Load/ More credits correlate to more Campbell et al.(2007)
Number of Credits * likely to be successful
Online Only: Colorado & Eberle (2010)
Academic Readiness/ More college preparation Choy (2001); Demetriou & Schmitz-Sciborski (2011);Nora &
High School Rigor correlates to more success Crisp (2012)
Face-to-Face Only: Adelman (1999)
Online Only: Aragon & Johnson (2008), Müller (2008); Muse
(2003);
Age * Younger students are more Nora & Crisp (2012)
successful
Online Only: Hung, Hsu, & Rice (2012); Menager-Beeley (2001);
Osborn (2001); Yasmin (2013)
Older students are more successful Online Only: Muse (2003); Valasek (2001)
* Variable included in this study.
22
Relationship of Characteristic to
Student Characteristic Studies Addressing Characteristic
Academic Retention
Course Subject * Students are more successful in Online Only: Hung et al. (2012); Yasmin (2013)
some subject areas. Math tends to
be more challenging.
Entrance Exam Scores * Higher test scores are a positive Campbell et al. (2007); Cortes (2013); Reason (2003)
predictor
Online Only: Morris et al. (2005)
Ethnicity * Asians and Caucasians more likely Nora & Crisp (2012); Reason, 2003; Swail (2004)
to persist
Blacks, Hispanics, Native Bowen, Chingos, & McPherson (2009); Nora & Crisp
Americans less likely to persist (2012); Reason (2003); Swail (2004)
Financial Aid Eligibility Lower socioeconomic status Campbell et al.(2007); Swail (2004)
students are less likely to persist
First Generation Student * First-generation students are less Choy (2001); Falcon (2015); Stebleton & Soria
likely to be successful
(2013) Face-to-Face Only: Thayer (2000)
Gender * Females are more successful Online Only: Aragon & Johnson (2008); Hung et al. (2012);
Yasmin (2013)
23
Relationship of Characteristic to
Student Characteristic Studies Addressing Characteristic
Academic Retention
Grade Point Average Higher GPA correlates to Bowen et al. (2009); Campbell et al.(2007); Devadoss & Foltz
(GPA) * higher success online (1996); Reason (2003); Swail (2004)
Face-to-Face Only: Adelman (1999)
Online Only: Aragon & Johnson (2008); Dupin-Bryant (2004);
Harrell & Bower (2011); Menager-Beeley (2001); Morris et al.
(2005); Muse (2003); Osborn (2001); Valasek (2001)
High School GPA * Higher GPA a positive predictor Bowen et al. (2009); Cortes (2013); Nora & Crisp (2012);
of academic success Reason (2003)
Online Only: Morris et al. (2005)
Self-Efficacy More self-efficacy a student has Cortes (2013); Demetriou & Schmitz-Sciborski (2011)
the more likely they are to be
Online Only: Holder (2007); Ivankova & Stick (2007); Kemp
successful
(2002); Müller (2008)
* Variable included in this study.
24
Relationship of Characteristic to
Student Characteristic Studies Addressing Characteristic
Academic Retention
Self-Motivation Motivated students tend to be Demetriou & Schmitz-Sciborski (2011); Devadoss & Foltz
successful (1996); Nora & Crisp (2012)
Face-to-Face Only: Adelman (1999)
Online Only: Ivankova & Stick (2007); Liu, Gomez, & Yen
(2009); Muse (2003); Valasek (2001)
Student Age Positive effect de Freitas et al. (2015)
Similar to Peers
Student Attendance Attendance in face-to-face classes Devadoss & Foltz (1996)
is a positive predictor of success
Student Engagement More social interaction with Demetriou & Schmitz-Sciborski (2011); de Freitas et al. (2015);
faculty or other students is a Nora & Crisp (2012); Swail (2004)
positive predictor of academic
Face-to-Face Only: Thayer (2000)
success
Online Only: Hung et al. (2012); Ivankova & Stick (2007); Liu et
al. (2009); Müller (2008); Valasek (2001)
Support of Family More support correlates with more Choy (2001); Swail (2004)
and Friends persistence Face-to-Face Only: Adelman (1999)
Online Only: Holder (2007); Müller (2008); Osborn (2001); Park
& Choi (2009)
Work Commitments Students who are employed are Kemp (2002); Tello (2007); Yasmin (2012)
less likely to persist to graduation
25
26
success on a more general level finding that academic performance in higher education
mirrors that of previous academic experiences. This correlation is true for both students
with good grades as well as those who were unsuccessful (Lee & Choi, 2011). Students
who enter a post-secondary institution less prepared for the academic rigor tend to
struggle academically. This causes students to take longer to graduate (Ishler & Upcraft,
2004). Additionally, the more time that has passed since a student last took a class, the
more likely they are to struggle when reenrolling (Colorado & Eberle, 2010; Moore &
Kearsley, 2005). In contrast to these weaknesses, students who enter a course knowing
how to study are more likely to be successful (Moore & Kearsley, 2005; McKenzie &
Schweitzer, 2001). In addition to studying, students who make attendance in their classes
Demographic Factors
Early attempts at online learning were promoted as if all diversity could be hidden
in an online environment (Rovai, Ponton, & Baker, 2008). While this could never happen
in a face-to-face classroom because of visual cues, this type of utopian environment may
be possible online, although it is unlikely. In this type of class, the bias would be
removed, but only until the instructor and students start interacting with each other.
Students draw on their past experiences as learning resources, and these could not be
shared without the diversity of the group being shared to some extent.
Males and females have different approaches to learning (Ewert, 2010; Rovai et
al., 2008). Historically, males dominated the higher education student audience until the
1970s, when females surpassed males in the number of both enrollments and graduates
(Ewert, 2010; Grebennikov & Skaines, 2009). Male students have a higher incidence of
27
taking a break of a term or more while working on their post-secondary education. They
are also more likely to attend school on a part-time basis (Ewert, 2010).
Rovai et al. (2008) found that, while enrolled in courses, males generally have a
more positive attitude toward technology than their female counterparts. It may be due to
this attitude toward technology that causes male students to exude more confidence in
their online participation. Male students tend to use fewer qualifiers instead opting to use
more intensifiers in their writing. When students are given the opportunity to interact
with fellow students, females are more likely to ask questions while male students tend to
answer questions more frequently. When working on low level learning tasks female
students take notes and focus on absorbing the content where male students choose to ask
questions directly to the instructor. In contrast, female students prefer interacting with
fellow students when working on higher level learning tasks where males prefer
independent processing. The same research added that female students use a “connected
voice” when contributing to discussion forums, portraying empathy and the importance
of relationships while male students use an “independent voice” which is more certain in
its tone, and sometimes is interpreted as confrontational (Rovai et al., 2008). Overall
studies show that females are more successful than males, although studies have varying
Age is another factor that is considered in the research on retention for the
university population as a whole. Individual studies have differing results. Some studies
have found younger students are more successful (Hung, Hsu, & Rice, 2012; Osborn,
2001; Yasmin, 2013), while others determined that older students do better in their
coursework (Muse, 2003; Valasek, 2001). Older students are often classified as
28
nontraditional students. The term nontraditional student refers to a student who meets one
or more of the following characteristics: they are over the age of twenty-four, married,
have children, or are financially independent (Ewert, 2010; Watt & Wagner, 2016). Any
of these factors can have a detrimental effect on a student’s attention to school work
(Braxton & Hirschy, 2005; Park & Choi, 2009). It is these same factors that may cause a
student to select online courses as opposed to face-to-face classes for the added flexibility
higher education (Morris, n.d.; Richardson, 2012). Early researchers came up with
theories based on genetics, hypothesizing that some races have more innate abilities than
others. More recently, researchers argued that differences in educational outcomes are not
due to genetics, but instead caused by the differences in economic, cultural, social, and
historical circumstances. The nature versus nurture mentality spurred a new wave of
(Morris, n.d.).
Modern research has centered on the cultural and societal factors that can have an
(2001) defined a framework that can be used to compare cultures and how the societal
factors may define how the culture views higher education. The framework uses five
Power – Distance Dimension. A measure of the disparity between those who have
distinction between what men are expected to do from what women are
expected to do.
which people from a society are looking toward the future as opposed to living in
the present.
Cultural differences can affect how students interact with the instructor in
courses, both face-to-face and online. If the students have a different cultural background
than the instructor, it has the potential to affect student achievement. The student may be
influenced by different comfort level on the power-distance dimension, and the role of
the teacher; respecting their authority to the point that it hampers their success in the
course (Rovai et al., 2008). Specifically, college level courses often incorporate the use of
discussions. Discussions are frequently in the format of a debate where the intent is to
have students debate the instructor and fellow students. The United States has a relatively
low power-distance rating, however, students from cultures with a high power-distance
rating may not feel comfortable challenging their instructor, a person in a place of
authority (Sher, 2013). This could, in turn, adversely affect their grade, and in turn their
overall academic success. Since minorities are a growing segment of the college
30
population, it is important for university faculty and staff to have an awareness of cultural
classroom to a fully online course. One range within these delivery models is the amount
of synchronous contact between instructor and student. Some classes take place in a fully
synchronous format. This can occur in a classroom, via two-way video, or using a web-
based meeting platform. Besides the level of synchronous contact, there are many
considerations that can affect both the instructor and the student in these various course
delivery models.
The roles of both the instructor and the student vary in the different course
delivery methods. In face-to-face classes, the instructor often has the role of a “sage on
the stage,” or the subject matter expert standing in the front of the classroom distributing
their knowledge to the students (King, 1993). This aligns with the traditional idea of an
instructor lecturing while students are taking notes and attempting to absorb as much
information as possible. This means the activities are often planned and led by the
In online courses, the instructor role often changes. They act more as a “guide on
the side” (King 1993). Some instructors opt to play an active role in course facilitation,
providing regular academic support for students as they work their way through the
course content. Instructors grade assignments and provide feedback to students, as well
students who ask for help. Other instructors take the initiative to contact students who
31
seem to be struggling in their course. In this model, students have more control over their
learning.
Malcolm Knowles (1984) identified a set of characteristics that are often preferred
Learner Control. Since adult learners are independent members of society, they
Life Experience. Secondly, adult learners bring a vast array of experiences to the
classroom. Knowles emphasized that these students learn best when they are
encouraged to draw on their experiences and make connections between their past
experience and the knowledge being gained through the educational experience.
and emotionally ready for the task at hand. Adults tend to choose to continue their
Value of Learning. Adult learners need a purpose for their learning. Toward this
end, students need to be informed of the outcomes of the learning experience, and
(Knowles, 1984). This final assumption about these learners is very closely
interaction among students (Stansfield et al., 2004). Some course formats, either online or
face-to-face, allow students to work through the materials at their own pace in a relatively
independent format. In this type of course, the student has opportunities to interact with
the content and the teacher, but not fellow students. Other online courses are designed for
a cohort of students. In these courses, students have the opportunity to interact with each
other as well as with the content and the teacher. Either format requires students to be
active participants. Asynchronous online courses provide the opportunity for students to
think and reflect on the content prior to participating in class. Because of the nature of the
discussions, there is the potential for more student interaction and participation than in a
personal experiences and share them with others. This approach allows students to use
The instructor is responsible for building a sense of community within the course
they teach (Rovai et al., 2008). In a face-to-face class, this can be accomplished through
discussions and classroom activities. This is a relatively easy task when students are in a
common location and time where students have all their senses gathering information in a
similar environment. However, in an online course, without audio or video, the instructor
and students do not have the visual cues of facial expressions, nor do they have the
intonation cues available when listening to a conversation. Despite the lack of face-to-
33
face contact, there can be other advantages to online learning. The increased
opportunities for reflection, as well as unlimited access to the course content, provide a
greater degree of learner control over the learning environment (Stansfield et al., 2004).
The opportunity for reflection allows for deeper discussion as compared to those that
take place in the face-to-face classroom. These discussions can be productive if students
feel the online environment is a safe place for sharing their thoughts. In doing so, all
participants, both instructors and students, need to have respect for diverse perspectives
online and face-to-face versions of the same course should be developed around the same
set of learning objectives. Both course models should have the same measurable course
outcomes, although they may be achieved in different ways. If this is truly the case, the
two course models should have similar measures of student success (Clark, 1983). When
a study finds that student outcomes differ between face-to-face and online, those
Learning Analytics
Analytics is the science of logical data analysis (Dziuban, Moskal, Cavanagh &
Watts, 2012). The use of analytics is popular in business to predict customer choices. For
example, many online shopping websites offer suggestions based on previous browsing
on their site. Similar analytics of data can be applied in the field of education to predict
student success or inform instructors on when and how to intervene with a student to
34
reduced chances of failure, effectively allowing educators to gain similar benefits for
students as businesses do for their customers through advertising (Martin & Sherin,
2013). The Society for Learning Analytics Research defines their field as “the
measurement, collection, analysis and reporting of data about learners and their contexts
for purposes of understanding and optimizing learning and the environment in which it
Learning analytics is often confused with the field of educational data mining.
While the two fields have many similarities, some argue they evolved separately with a
slightly different focus. The International Educational Data Mining Society defines
educational data mining as “an emerging discipline, concerned with developing methods
for exploring the unique types of data that come from educational settings, and using
those methods to better understand students, and the settings which they learn in”
(Siemens & Baker, 2012, p. 1). Learning analytics focuses on data from the learner and
their context that will be used to improve either the learning process or the learning
environment. In contrast, educational data mining has a slightly broader approach. These
researchers do not specify where their data originates, but they do stipulate that their goal
is to better understand students and the various learning environments. The core
difference between the two fields is that learning analytics incorporates human judgment,
while educational data mining relies on computer automation (Baker & Siemens, 2014;
Pardo, 2014; Siemens & Baker, 2012). This difference is evident in the discovery,
analysis, and application of the data. For example, educational data mining researchers
may apply their findings through having educational software automatically adapt to
35
personalize learning experiences for users. In contrast, learning analytics results are
used to inform instructors on how to assist struggling learners (Baker & Siemens, 2014).
Both learning analytics and educational data mining are emerging as new research
fields because of the ever-increasing amount of data available (Baker & Siemens, 2014;
Wagner & Ice, 2012). Stakeholders at all levels are expressing interest in access and use
of the data including educators, institutions, government, and accrediting agencies. These
groups are using the data to make decisions about instructional strategies, judgments on
the quality of learning, student attrition and graduation rates, financial aid, and policies
about online teaching and learning (Dringus, 2011). Jayaprakash et al. (2014) stated that
“the goal of learning analytics is to uncover hidden patterns in educational data and use
those patterns to attain a better understanding of the educational process, assess student
analytics should focus on providing data that support student success as opposed to other
goals such as maximizing profits for the university (Becker, 2013; Slade, & Prinsloo,
2013).
Using data to inform instruction is not new. On a small scale, teachers have used
became more formal when learning management systems first became available as
opposed to individual websites for distance courses. Learning management systems were
able to track data for users, both students and faculty (Picciano, 2012; Reyes, 2015). The
second wave of data analysis came when Web 2.0 tools were incorporated into online
36
learning situations (Brown, 2011). These tools provided additional data not available with
learning management systems alone. The final wave of development for learning
analytics and data mining came with the increased capacity to analyze large amounts of
data. Learning managements systems and student information systems were linked to
There is an ever increasing push for stakeholders to use big data in decision
making. Globalization has pushed the demand for learning analytics by creating increased
competition for online educational opportunities. Students no longer need to live in the
same town as their chosen institution of higher education. Reduced public funding and
increased government oversight have caused a need for institutions to show a return on
investment for the education they provide to students (Lockyer, Heathcote, & Dawson,
Since the field of learning analytics is relatively new, there are only a few
proposed models to provide structure for studies. Some of these models are based on
2011). Each of these models originates from the definition of learning analytics in that
they are designed to use available data to inform and improve teaching and learning.
how businesses can make knowledge actionable. He expanded on an earlier theory, which
proposed that information lies on a Knowledge Continuum based on the depth of how the
data is used (Elias, 2011). Data is at the lowest level and used to answer “what is”
questions. The next level higher is considered information. Information is used to answer
37
questions about when and where. The third level on the spectrum is labeled as
knowledge. Knowledge is used to answer questions about why and how. The high end of
the knowledge spectrum is defined as wisdom. Information has achieved the wisdom
(2009), defines a cyclical framework in which data is gathered, processed, and presented.
Gathering data involves selecting and capturing the data. Processing the data involves
aggregating and processing the data. Presenting the data includes determining how it is
displayed. If the desired detail is not displayed, then the cycle is repeated with some level
of change in what data goes through the process (Dron & Anderson, 2009; Elias, 2011).
proposed a five-stage model for learning analytics studies. The first stage is capturing the
data. Researchers need to determine what data is needed, the level of granularity of the
data, and how to retrieve that data (Campbell & Oblinger, 2007). During this stage,
researchers should employ techniques to ensure the data is stored in a secure location
(Pardo, 2014). Once the data is retrieved, the researcher must make decisions on how to
organize the data prior to moving to the next stage of the process.
The second stage of the learning analytics process involves reporting on the data.
The data needs to be processed in a manner that it can be summarized or combined for
reporting in a usable format for the end user (Campbell & Oblinger, 2007; Pardo, 2014).
During this stage, it is usually necessary to use statistics software tools that can handle
large quantities of data. The tool selected depends on the type of data that was captured
and the research questions to be considered (Greller & Drachsler, 2012). One critical
38
display the data in a meaningful way for stakeholders (Pardo, 2014). This stage includes
computation of descriptive statistics for the data, which informs end users of what has
The next stage of the process is to make predictions based on the data and
reporting completed in the previous stage. This involves answering questions that
initiated the data capture in a manner that explains what is likely to happen. An accurate
prediction depends on the use of a reliable model. This stage revolves around the
Once a prediction is made, the next phase requires stakeholders to act on that
(Pardo, 2014). These actions can be executed either manually or automatically. The
number and type of interventions are based on the nature of the prediction that was made
in the previous stage (Campbell & Oblinger, 2007). Depending on the type of reporting
and predictions created during earlier stages of the learning analytics process, actions
may be prescriptive in nature. Prescriptive actions should vary for different end users, or
The final stage of the learning analytics process is the refining stage. This is the
stage of the process that makes this model unique. The models presented by Baker (2007)
and Dron and Anderson (2009) do not define refining the data as a unique step in the
process. Calling out the refinement of the data as a requirement of the process makes this
model stronger than the other models described in the literature. Regular evaluation
should take place on results of the actions taken during the act stage. In addition to
39
evaluating the actions that take place, researchers should revisit the predictions used to
determine those actions, the reporting that was used to predict, and even how the data
was captured. Improvements could be made at any stage in the learning analytics process
(Pardo, 2014).
There are potential ethical issues within the field of learning analytics. Primarily
these are issues related to student privacy and ownership of the data (Reyes, 2015; Slade
& Prinsloo, 2013). The Family Education Rights and Privacy Act (FERPA) is a federal
law enacted to protect student privacy. This law guides institutions on how student data
can be used for research, school improvement, and accountability, and when it is
necessary to inform students (U.S. Department of Education, 2012). While some students
may want to opt out of studies that involve learning analytics, it could change the
negative manner (Brown, 2011). Since this field is in its relative infancy, students need to
be ensured that any learning analytics research used beyond the classroom and instructor
has all personally identifiable information removed from the data prior to release to
One challenge related to learning analytics is that there are few guidelines or
regulations in place to guarantee anonymity (Pardo, 2014; Reyes, 2015). Since there are
minimal guidelines, researchers should be clear in defining the purpose of their study as
well as how the sensitive data is being handled (Slade & Prinsloo, 2013).
Another ethical consideration is related to how the data are used once the analysis
is completed. Data, especially personally identifiable data, should be used for research or
40
reasons like making a profit (Slade & Prinsloo, 2013). At times, an in-depth analysis of
data may lead to conclusions that can help stakeholders increase their understanding
about student retention and academic success, but it may not be actionable data. Other
matter how the data is used, there should be a balance between the push to gain
knowledge against harming individuals, whether they are students or instructors (Slade &
Prinsloo, 2013).
Since the results of data analysis have the potential to directly affect students and
could be adverse effects. Students may become unmotivated, academic advising could be
inaccurate, faculty members could lose opportunities for advancement, or the institution
as a whole may lose enrollments. When acting on the data, stakeholders should keep in
mind that the numbers that were analyzed represent real people. These people are part of
the population, but may not have the same needs as the group (Slade & Prinsloo, 2013).
beyond what can be measured with the data alone, so it is essential to avoid profiling of
educators have an ethical obligation to act on the knowledge gained through the research
Uses of Data
The results from learning analytics studies are used by a variety of groups. How
the data is used, and what actions are taken, depends on the needs of the group, and their
41
placement in the hierarchy of the educational process (Jayaprakash et al., 2014; Shelton
et al., 2015). Learning analytics data are used in three areas: descriptive, predictive, and
prescriptive analyses (Affendey, Paris, Mustapha, Sulaiman, & Muda, 2010; Brown,
2011). Descriptive analysis helps create a portrait of past students, instructors, or other
stakeholders, while predictive analysis predicts likely trends and outcomes for students
prior to their experience (Affendey et al., 2010; Brown, 2011; Verbert, Manouselis,
Drachsler, & Duval, 2012). Prescriptive analysis dictates interventions for various
stakeholders within the educational community (Brown, 2011). Each of the user groups
may use the data in a descriptive, predictive, or prescriptive manner based on their needs.
analysis results in a variety of ways. Data are used to describe the student body as a
admissions prospects and predict the likelihood of their success (Dziuban et al, 2012).
They detect retention issues, prescribe actions, and monitor graduation rates (Reyes,
2015). Administrators may also use data to identify issues in the learning community
beyond the classroom itself that affect the success of students at the university (Pardo,
2014). Overall, the data reporting can lead to improved accountability across the
university, leading to better use of resources, and an increased reputation, both within the
designers when creating online courses (Lockyer et al., 2013). Department level staff can
use data to inform personnel decisions including teaching assignments and training needs
(Berger & Lyon, 2005; Dziuban et al., 2012; Shelton et al., 2015). University staff that
42
refine the timing and location of various services (Becker, 2013; Campbell & Oblinger,
2007).
Faculty. Both face-to-face and online faculty members can benefit from using
data to inform their teaching. Data resulting from formative assessments can be used to
identify knowledge gaps that can be addressed immediately in the classroom, positively
helping current students (Reyes, 2015). Data from other sources, including the end of
course evaluations along with LMS data, can be used in a prescriptive manner to inform
adjustments to course content or pedagogy for future course offerings, particularly for
online courses (Pardo, 2014). Learning analytics can encourage faculty members to take
encourage professional growth for faculty in the differences between face-to-face and
online teaching and learning pedagogy (Shelton et al., 2015). Faculty members have the
power to use learning analytics to guide students to success, affect practice, and
contribute to the scholarship of teaching and learning (Campbell & Oblinger, 2007).
should be able to take advantage of the large amounts of data automatically collected
both prior to enrolling and while participating in online courses. Students may benefit
from having access to predictive analysis results on given courses. This information
should not be used to limit educational options, instead, it has the potential to inform their
midcourse to improve their academic performance. Like faculty, students will benefit
from data that encourage opportunities for self-reflection (Pardo, 2014). Reflection of this
nature can affect progress in a current course, or inform decisions on future courses.
43
Government. Policy makers use data at all levels, descriptive, predictive, and
learning analytics allows for new types of data use thereby expanding the ability to
evaluate educational objectives. The new data can provide a different viewpoint for
a refined, usable format. Toward this end, researchers have a number of responsibilities.
They are responsible for the validity and reliability of the data as it goes through the
process of analysis and is shared with others (Reyes, 2015). Additionally, they are
responsible for the de-identification of student data when details are reported beyond the
classroom.
the early years of higher education. Within the last decade, a dramatic increase in the data
available has changed the way data is used in the decision-making process. Much of this
is due to “big data” that is available in student information systems, learning management
systems, and other longitudinal data systems. If this data is properly captured and
on the data. There were a number of learning analytics models presented in the literature
review, and each learning analytics study is driven by a model that allows the research to
achieve maximum results. This study used the five-step process proposed by Campbell
44
and Oblinger (2007) because it provided a framework that matched the focus of the
study.
The decisions made based on the data are supported by the persistence theories
established in the 1970s. These theories posited that the characteristics with which each
student enters college, combined with the environment of the institution, can be used to
identify reasons why a student may not succeed in their education. The review of the
literature provided a comprehensive list of characteristics that were options for data
collection points for this study. This study attempted to address as many of the variables
listed in Table 1 as possible. However, one limitation of the purely quantitative study is
that qualitative data is not available. As a result, those student characteristics included in
Table 1 that are related to information about individual students or faculty choice were
not available for this study. This included variables related to whether study participants
accessed services offered by the university. Ultimately, this study addressed 50% of the
student characteristics addressed in the literature. Those variables are indicated in Table 1
with an asterisk.
Finally, all of the literature reviewed for this study addressed the university
environments in isolation. This study addressed both face-to-face and online course
enrollments separately as well as the population as a whole. This approach makes this
study unique and allows the study to identify predictors that differ between the two
audiences.
45
Data Collection
Board (IRB), and was approved. Data was exported from the data warehouse at the
university where this study took place. The information was pulled from the PeopleSoft
adopted by the university. A detailed list of data points collected can be reviewed in
Table 2.
To initiate the data collection process, a query was run to create a comprehensive listing
of all 100 and 200 level core courses that are offered in both online and face-to-face
formats. This list was used to determine which records to extract from the data
warehouse. Courses offered in only one format or the other were excluded from this
study. A number of courses were offered in other formats including hybrid or via
teleconferencing, but those course sections were excluded from this study. Additional
queries were run to gather demographic information as well as details on residency, first
Once the data set was reduced, there were nearly 101,000 individual course
enrollments for just over 23,800 students. Due to the large quantity of data, and the
personal nature of the records, adherence to FERPA regulations was deliberate. The data
Data Organization
Student identification numbers were included with the original data set. This
information was used to join the data from multiple queries into a single merged data set.
individual students more difficult (Nelson, 2015; Slade & Prinsloo, 2013). Any
identified while minimizing the ability to identify his or her original student identification
number. Each record collected was associated with a course enrollment. So, for example,
if a single student was enrolled in three different 100 or 200 level courses, then there
were three different records associated with that student. This approach allowed the study
to account for all online course enrollments and all face-to-face course enrollments for
courses included in the study. The data was delivered in a format that was easily
imported into Microsoft Excel for data cleaning and then imported into SAS, a statistical
With a large data set, it is likely there will be invalid data (Hand, 1998). The
dataset was evaluated, field by field, for any missing data points. Based on the nature
of the data that was missing, many records were omitted from the study. For example,
this occurred when details like the final grades were listed as incomplete or audit, or if
the full-time status was not included. In some fields, a value of unknown was used
(ethnicity), and in other fields, unknown values were left blank (entrance exam scores,
49
residency information). Sorting and filtering strategies were used to identify anomalies in
the data. For example, students with a GPA above a 4.0 were omitted from the dataset.
The initial data set included all records for active enrollments as well as students who
dropped the course prior to the course drop deadline. Since none of these records had a
final grade associated with them, they were omitted from the study. Additionally,
enrollments that were for a course that used a pass/fail grading format, courses for zero
credits, enrollments that were audits of full courses and enrollments where a student
Some data required modification prior to analysis. One critical field was the
reporting of final grades. Instructors at the university are given freedom in how they
report final grades. Some issue only letter grades while others opt to use a +/- system. At
some universities, a grade of C- is considered not passing, but that is not the case at the
university where the study took place. For this study, a C- was considered passing. To
minimize confusion in this field, all grades were truncated to consider only the letter
grade. If a student withdrew from the course, their grade was considered equivalent to an
F for statistical analysis. For calculation purposes, the standard 4.0 grade scale was used
where an A was worth four points, a B was worth three points, a C was worth two points,
The year in school field was calculated based on the number of credits a student
had completed based on the definition used by the university. A student is considered a
credits have been earned, a junior when between 58 and 89 credits have been earned, a
50
senior when 90 or more credits have been earned. Students who are pursuing a second
Several of the variables were reduced for the logistic regression analysis.
Reducing variables minimizes the number of different values for the variable. The
individual course subjects were reduced from individual subjects to departments for
initial analysis and then further reduced to the college offering the course for logistic
regression analysis. A similar reduction of values was completed for the primary majors
declared by the students. The degree type was reduced from eight different types of
toward a bachelor’s degree, from those enrolled as college students while enrolled
concurrently as high school students. All other degree types were grouped into a category
labeled as other. Additionally, several variables were transformed to normalize the data
distribution prior to the logistic regression analysis. These variables include the age at
course start, the cumulative credits earned, the degree count, and the total enrolled in
course.
Report
To report on the data, the analysis must be completed. For statistical testing, the
independent, or outcome variable for this study was the course delivery model. This
variable has two possible values, face-to-face and online. Two variables were used to
measure success in each course enrollment. The final grade variable and a reduced
An initial analysis of the cleaned data was completed using descriptive statistics.
This analysis provided an overall picture of the students who enroll in either online or
face-to-face courses. The categorical variables were interpreted using percentages and
graphs to describe the distribution of the population, while numerical data was described
set of correlational tests to identify which demographic, academic, and course related
factors were related to student success in either online or face-to-face course enrollments.
The correlation analysis was followed by a logistic regression analysis to create reports
Predict
The results of the various analyses were used to create a prediction model. A
comprehensive set of correlational tests were used to identify which academic and
demographic factors were most closely associated with student success in either online or
logistic regression analyses. These results were used to create figures and tables for the
predict phase of the learning analytics process. The model highlights the likelihood of
As part of the predict step, the results were used to identify a specific area with
significantly different data. Concurrently enrolled students, those who are simultaneously
both high school and college students, were identified as this group.
Act
The act step of this study involved creating recommendations for university
personnel on student enrollment strategies, and for instructional designers working with
instructors to create both online and face-to-face courses. These recommendations relate
back to the data analyzed and current research. Additionally, the recommendations for
Refine
The refine step of this methodology includes the further analysis of the courses
taken as concurrent enrollment courses that were included in this study. Through the
refinement process, the reduced dataset was analyzed in an attempt to identify reasons for
the variations in final grades for students in courses taken for both high school and
college credit.
53
Introduction
The purpose of this study was to identify what types of students were more
successful face-to-face and what types were more successful online. Correlations were
used to identify trends for students based on a number of demographic, academic, and
course related factors. Then logistic regression tests were completed to identify
predictive models for student success. This chapter reports the findings from the
quantitative data analysis. The results presented in this chapter are organized into
sections on demographics of the study population, the courses addressed in the study, and
the enrollment details. The next section reports on the details of the various statistical
tests completed as part of this study. The individual research questions will be addressed
Demographics
The study population was determined based on the enrollment choices made by
students. It included all students who were actively enrolled in a 100 or 200 level course
that was offered in both online and face-to-face formats during all semesters between Fall
2013 and Summer 2015. An actively enrolled student is defined as one who has not
dropped the course by the drop date for the term, typically the tenth day of the semester.
Overall
20,875) opted to take a face-to-face course during the study time frame, while only
54
46.5% of the students (N = 11,076) chose to take a course online. These numbers make it
clear that many students are enrolled in a combination of face-to-face and online courses.
Of the students, 53.5% (N = 12,760) opted to enroll exclusively in the more traditional
face-to-face courses that were included in this study, although there is a possibility they
were enrolled in online courses that were excluded from the study. Additionally, 12.4%
of the students (N = 2,961) were enrolled in only online courses. The number of students
who chose to enroll in a mix of face-to-face and online courses was 34.0% (N = 8,115).
Gender
that 54% of students were female and 45% were male, with approximately 1% opting not
to disclose their gender (Office of Communications and Marketing, 2014). Students who
opted not to report their gender were omitted from this study. The students in the study
population used for this study had a slightly lower percentage of females (52.8%) and a
displayed in Table 3, the gender in the face-to-face courses has a shift from the entire
population, with fewer females (51.7%) as compared to males (48.3%). A much higher
Age
The age of the students in the study population were categorized into six groups.
students into groups by age helped identify traditional aged students (18-24 years old) as
compared to nontraditional students. The figure shows data for the entire student
population at the university as well as for students within the study population enrolled
face-to-face and online. Despite both a higher minimum (13 years old) and maximum (82
years old), face-to-face students (M = 22.28, SD == 7.17) were slightly younger than the
online students (M = 24.43, SD 7.69) who ranged between 12 and 76 years of age.
This ethnic distribution of the study population was very similar to the population
of the university as a whole. There was not a significant difference in the proportion of
56
different ethnic groups between the online and face-to-face student groups. Table 4
provides a breakdown of the ethnic groups for the study population as a whole, those
enrolled in the face-to-face courses that were part of this study, and those enrolled in the
online courses.
In fall of 2014, the university began to collect data as to whether or not students
were a first generation college student. Since the data for this study spans the semesters
between Fall 2013 and Summer 2015, this data exists for some, but not all students (N =
12,577). Of these students, 44.9% (N = 5,652) are first generation university level
students. The majority of the first generation college students, 54.7% (N = 3,089), chose
to attend exclusively face-to-face courses, while 9.5% (N = 535) selected only online
courses, and 35.9% (N = 2,028) opted for a combination of course delivery modes. Table
5 displays the distribution of the set of known first generation students by gender,
57
Gender
Female 3,15 55.9%
9
Male 2,49 44.1%
3
Ethnicity
American Indian/Alaskan Native 37 0.7%
Asian 119 2.1%
Black/African American 110 2.0%
Caucasian/White 4,04 71.5%
3
Hispanic/Latino 913 16.1%
Native Hawaiian/Pacific Islander 22 0.4%
Two or more races 277 4.9%
Not Reported 131 2.3%
Age
≤ 18 1,71 30.4%
6
19-20 1,31 23.2%
4
21-24 1,01 18.0%
9
25-34 1,02 18.2%
9
35-49 450 8.0%
50+ 124 2.2%
ethnicity, and age. There is a slightly higher percentage of females that are first
generation students as compared to the population used in this study or for the university
Hispanics in the group of first generation students. To account for this shift, there is a
lower percentage of whites in the first generation group, as well as fewer Asians. A
58
comparison of the spread of the ages of the first generation students was completed.
While there were some minor differences between the study population and this
Residency
Data on residency was available for approximately 59% of the students included
residency in the state in which the university is located, and as a result was charged the
in-state tuition rate. Students identified as nonresidents were required to pay the higher
out-of-state tuition rates. Table 6 displays the residency status of students based on their
residency status. The distribution of students opting for face-to-face as opposed to online
based on residency status. Students that are not residents of the state are much more
Majors/Minors/Certificates
The students that were part of this study (N = 23,836) declared a large number of
graduate, they need to demonstrate they have met all the requirements for that particular
degree. The university allows students to declare majors, minors, and certificates.
degree. Table 7 displays the distribution of the different types of degrees identified in the
student information system. Students who have not yet identified a major were
distinguished from students taking courses of interest based on the understanding that at
some point they would identify a major and complete a degree. Minors and certificates
must be completed in conjunction with a major, although that major may be undeclared.
Students who took university level courses while still enrolled in high school were
Table 8 displays the number of degrees declared by the students in the study
population. While most students declared a single major (66.89%), there were several
students who identified multiple degrees with the intention to complete the requirements
for each degree. The data did not allow the researcher to identify if students were
The most recently declared major was labeled as the primary major for each
student in the study population. Table 9 displays the distribution of primary major for the
students in the study population. The College of Arts and Sciences (COAS) was divided
to identify students declaring arts related majors as opposed to those in science and
mathematics (STEM) fields. There were an additional 737 students (3.09%) who had
taken a college level courses while in high school, but later declared a different major.
N Percent
Courses
The students included in this study (N = 23,836) were enrolled in courses that were
offered in both online and face-to-face modalities during the time period between Fall
2013 and Summer 2015. Enrollments from 2,811 unique course sections were included in
the study. Table 10 shows the distribution of course sections across modalities and
semesters offered. During the fall and spring semester, face-to-face course sections
outnumber the online course sections. During the fall semester, face-to-face courses were
80% of the course offerings, that number fell in the spring semester to approximately
72%. The summer semester had a different proportion of face-to-face and online course
sections. In the summer terms included in this study, online sections made up 57% of the
course sections.
The set of courses included in this study were offered by 29 of the 61 different
academic departments across the university. Many of the courses included in the study
are selected by students to meet the core graduation requirements, while others are
chosen by a more select audience as part of a specific program, to fulfill the requirements
College of Education 25 22 47
Educational Technology 17 16 33
Special Education 8 6 14
College of Engineering 13 11 24
Sections
Academic Department Online Face-to-face Online
Enrollments
The students in the study population represent 100,943 different course
enrollments throughout the two year, six semester time frame. Of the enrollments,
completely online (N = 21, 730). The students enrolled in courses ranged from freshman
status to graduate students. The academic level is determined by the number of credits
earned by a student prior to the beginning of the term. The distribution of academic level
of students enrolled in the classes included in the study can be seen in Figure 3.
The academic load of the students that carried these enrollments included in
this study varied. The vast majority (83.5%) of the enrollments were for students
enrolled full-time at the university. The remaining enrollments were students enrolled
Grades Earned
Grades earned as a result of the courses completed for the entire study population
and for both course modalities are displayed in Figure 4. The mean grade point average
(GPA) for all course enrollments in the study population was M = 2.658 (SD = 1.372).
The GPA for only face-to-face enrollments (M = 2.653, SD = 1.352) was slightly lower,
while the online GPA (M = 2.676, SD = 1.445) was somewhat higher than that of the
45.0%
40.0%
35.0%
Percent of Enrollments
30.0%
25.0%
20.0%
15.0%
10.0%
5.0%
0.0%
A B C D F
Sample 35.6% 28.0% 17.2% 5.1% 14.1%
Face-to-Face 34.5% 28.6% 18.1% 5.4% 13.4%
Online 39.7% 25.5% 14.1% 4.2% 16.5%
Final Letter
Grade
Figure 4 Final Grades Earned in Courses Included in Study
Population
65
significant difference in enrollment success based on course modality. The mean success
rate for face-to-face students (M = 0.812, SD = 0.391) was higher than the success rate
is shown in Table 12. A review of the table can be completed to identify which values for
the characteristics had higher final grade averages than their counterparts for each
variable in the population as a whole as well as for both the face-to-face and online
subsets.
Course Subjects
So far, the comparison of final grades and success have been focused on student
based factors, either demographic or academic. Another area that was found to be a
differentiating factor in the final grade and success in a course was the subject of the
course the student was enrolled in. The descriptive statistics for each of the
individual courses are listed in Table 13. Comparisons revealed a number of courses in
which students earned significantly higher grades than other courses. For the entire study
significantly higher grades than the other departments: Academic Advising and
Technology. This same list of classes differs when restricting to only face-to-face course
68
21-24 2.463 1.429 21,426 2.411 1.410 14,97 2.583 4.464 6,452
4
25-34 2.585 1.447 13,346 2.539 1.429 9,391 2.673 1.476 4,955
35-49 2.738 1.431 4,713 2.718 1.416 2,796 2.767 1.452 1,917
50 + 2.969 1.339 1,016 2.914 1.363 642 3.064 1.294 374
Study Population Face-to-Face Online
Mean St Dev N Mean St Dev N Mean St Dev N
First Generation
Reported First Generation 2.542 1.410 27,669 2.554 1.386 22,45 2.494 1.507 5,214
5
Reported Non-First Generation 2.745 1.352 35,636 2.742 1.335 29,92 2.758 1.442 5,707
9
Residency Status
Resident 2.626 1.402 44,587 2.631 1.383 36,84 2.603 1.491 7,672
1
Non-Resident 2.883 1.235 18,587 2.888 1.212 15,96 2.850 1.369 2,622
5
Academic Variables
Academic Level
Freshman 2.619 1.391 39,183 2.647 1.369 35,45 2.356 1.554 3,731
2
Sophomore 2.636 1.366 31,030 2.644 1.338 24,32 2.605 1.465 6,701
9
Junior 2.654 1.362 17,881 2.619 1.342 12,01 2.727 1.400 5,869
2
Senior 2.783 1.336 11,018 2.697 1.328 6,337 2.900 1.337 4,681
Graduate 3.171 1.258 1,831 3.209 1.220 1,083 3.118 1.311 748
Academic Load
Full-time 2.627 1.373 84,307 2.619 1.357 68,27 2.662 1.442 16,034
3
69
Part-time 2.816 1.358 16,636 2.867 1.305 10,94 2.717 1.453 5,696
0
Term of Enrollment
Fall 2.660 1.368 52,691 2.667 1.354 44,11 2.623 1.442 8,580
1
Spring 2.631 1.380 41,401 2.619 1.356 32,46 2.674 1.461 8,941
0
Summer 2.809 1.353 6,851 2.840 1.256 2,642 2.790 1.410 4,209
Study Population Face-to-Face Online
Mean St Dev N Mean St Dev N Mean St Dev N
Primary Major College
CID 2.818 1.352 44 2.906 1.304 32 2.583 1.505 12
COAS – Arts 2.598 1.384 21,566 2.618 1.355 16,51 2.532 1.472 5,047
9
COAS – Sciences 2.619 1.409 9,620 2.611 1.398 7,872 2.653 1.456 1,748
COBE 2.673 1.359 19,400 2.673 1.340 15,74 2.677 1.437 3,655
5
COED 2.731 1.371 3,285 2.729 1.336 2,382 2.735 1.460 903
COEN 2.561 1.417 10,726 2.511 1.412 8,653 2.767 1.419 2,073
COHS 2.716 1.355 25,269 2.693 1.336 2,382 2.783 1.422 6,318
SPS 2.611 1.332 6,647 2.619 1.304 5,235 2.583 1.730 1,412
Undeclared – Courses of Interest 2.157 1.526 1,699 2.054 1.508 1,306 2.499 1.537 393
Undeclared – High School Credit 3.363 0.862 2,687 3.387 0.828 2,518 3.006 1.213 169
70
Table 13 Descriptive Statistics for Grade Value by Course Subject
Study Population Face-to-Face Online
Std Std Std
Rank Mean N Rank Mean N Rank Mean N
Dev Dev Dev
Academic Advising and
1 3.496 1.012 450 2 3.588 0.911 325 2 3.256 1.211 125
Enhancement
Accounting 11 2.965 1.428 482 9 3.091 1.365 372 19 2.536 1.554 110
Anthropology 23 2.573 1.335 2,767 22 2.636 1.277 1,878 22 2.439 1.441 889
Art 14 2.826 1.268 3,215 20 2.717 1.294 2,260 5 3.085 1.167 955
Biology 29 2.383 1.287 7,724 28 2.346 1.279 6,148 20 2.527 1.309 1,576
Business Communications 13 2.946 1.186 1,600 12 2.923 1.143 1,137 7 3.000 1.286 463
Chemistry 31 2.275 1.391 4,505 33 2.248 1.378 4,172 16 2.613 1.506 333
Chinese 22 2.607 1.466 178 16 2.796 1.324 137 30 1.976 1.739 41
Communications 3 3.294 1.359 17 1 3.846 0.554 13 33 1.500 1.732 4
Criminal Justice 28 2.424 1.269 2,565 27 2.432 1.267 2,151 24 2.382 1.279 414
Economics 21 2.617 1.336 4,460 21 2.707 1.304 3,933 31 1.945 1.383 527
Educational Technology 5 3.168 1.486 708 4 3.318 1.455 358 6 3.014 1.504 350
Engineering 7 3.063 1.374 1,366 32 2.256 1.295 355 1 3.346 1.286 1,011
English 9 3.030 1.320 11,69 7 3.110 1.278 9,034 12 2.761 1.421 2,665
9
Environmental Health 6 3.079 1.131 391 8 3.104 1.099 376 21 2.467 1.685 15
71
Environmental Studies 15 2.802 1.195 822 15 2.820 1.162 656 13 2.729 1.318 166
Study Population Face-to-Face Online
Std Std Std
Rank Mean N Rank Mean N Rank Mean N
Dev Dev Dev
French 20 2.669 1.451 242 19 2.744 1.369 215 28 2.074 1.920 27
General Business 17 2.776 1.106 1,317 14 2.826 1.122 1,043 17 2.584 1.025 274
Geography 25 2.523 1.489 172 25 2.506 1.533 83 18 2.539 1.454 89
Health Studies 8 3.046 1.271 5,386 6 3.188 1.077 1,658 8 2.983 1.344 3,728
History 24 2.539 1.404 3,781 24 2.621 1.341 2,767 25 2.315 1.540 1,014
Humanities 18 2.727 1.421 714 13 2.856 1.293 285 15 2.641 1.495 429
Japanese 33 2.223 1.574 251 31 2.278 1.544 198 29 2.019 1.681 53
Kinesiology 2 3.343 0.982 1,223 3 3.364 0.961 1,085 4 3.174 1.120 138
Korean 26 2.500 1.743 52 23 2.625 1.705 40 27 2.083 1.881 12
Mathematics 32 2.265 1.423 18,16 29 2.313 1.404 16,04 32 1.899 1.507 2,120
8 8
Philosophy 27 2.429 1.420 2,777 26 2.502 1.393 2,187 26 2.159 1.487 590
Political Science 19 2.715 1.267 1,840 18 2.754 1.224 1,645 23 2.390 1.547 195
Psychology 30 2.336 1.393 5,251 30 2.310 1.392 4,890 14 2.681 1.363 361
Sociology 16 2.790 1.322 4,242 17 2.757 1.262 2,723 11 2.848 1.421 1,519
Special Education 4 3.241 1.087 502 5 3.243 1.049 272 3 3.239 1.133 230
Theater Arts 10 3.027 1.257 2,530 10 3.065 1.228 1,881 9 2.917 1.332 649
72
University Foundations 12 2.957 1.298 9,546 11 2.965 1.284 8,888 10 2.853 1.471 658
courses, the list includes the following course subjects: Engineering, Academic Advising
Doing appropriate ,timely and regular performance appraisal will be overall good for the whole
team both in qualitative way by giving best quality service students academic performance and
quantitative by not only improving student career returns and teachers also doing justified
improvement in self achievements and reward accomplishment.
CONCLUSIONS
Introduction
The purpose of this study was to identify which demographic, academic, and
course related characteristics are most closely related to successful completion of 100 and
200 level courses in both face-to-face and online formats. This chapter will discuss the
results of the analysis and make connections between the literature and the findings from
the data collected for this study. It will also include suggestions for further research, the
The following research questions will be used to provide focus for the discussion
and make connections between the various analyses described in the previous chapter:
successful completion of 100 and 200 level classes taken online, as measured
by final grade?
population of the university. Male students appeared to have a preference for face-to-face
course enrollments as the percentage of males enrolled in online courses was significantly
lower than in the face-to-face courses. There were also differences in enrollments based
on age. Based on the data, younger, traditional students appeared to prefer face-to-face
courses when given an option. In contrast, there were more students in the older age
groups enrolled in online courses. This could be related to the many other competing
priorities nontraditional students must balance, including employment and care for
dependents (Ewert, 2010; Watt & Wagner, 2016), as opposed to a genuine course
modality preference.
The courses included in this study were limited to the 100 and 200 level courses
that were offered in both online and face-to-face formats during the two-year period from
Fall 2013 through Summer 2015. As can be observed in Table 10, during the traditional
school year, which included the fall and spring semesters, approximately 76% of the
course sections included in the study were face-to-face courses. During these semesters,
the average age of students enrolled in the courses is 21.9 years. This indicates traditional
aged students are the majority during the school year. However, in the summer, the
balance of face-to-face and online courses shifted such that only 44% of the courses were
offered face-to-face. This shift appears to be associated with a common reason that
students opt for online courses. Students choose to take online courses for the flexibility
of time, location, and pace (Stansfield et al., 2004). Additionally, the average age of the
student during summer rose to 24.5 years of age. This implies that nontraditional students
work toward completing their education year-round as opposed to only during the school
year, while traditional students take the summer off to spend with family or to earn
money. This aligns with the research that found nontraditional students tend to take
courses that fit their schedule as opposed to conforming to the traditional school year
online and face-to-face environments: gender, ethnicity, age, first generation status,
residency status, academic level academic load, the term of enrollment, and primary
major college. This finding indicates that these characteristics are predictive of stronger
whether they were face-to-face or online. This finding is in agreement with the studies
completed by Aragon & Johnson (2008), Hung et al. (2012), Reason (2003), Valasek
(2001), and Yasmin (2013). In general, ethnicity was not a strong predictor. One common
finding in this study was that students of Asian descent performed slightly better than all
other ethnic groups. This is in alignment with other studies addressing ethnicity
Age was challenging to use as a predictive behavior because both older and
younger students earned higher average grades than students in the middle age ranges.
Studies reviewed in the literature had mixed results based on the use of age as a predictor,
so these results match the previous studies. Several studies found younger students were
more likely to be successful in their course enrollments (Hung et al, 2012; Osborn, 2001;
Yasmin, 2013), while other studies found older students were more likely to be
successful (Muse, 2003; Valasek, 2001). The younger student success is likely due to the
number of students enrolled in concurrent enrollment courses while the older students
often have a different level of intrinsic motivation for their learning (Stansfield et al.,
2004).
(2011), and Thayer (2000), this study found that first generation students earned lower
grades than their counterparts who are not first generation students. First generation
students tend to have lower levels of college readiness and a lack of support from family
and friends as compared to students who are not first generation (Falcon, 2015; Stebleton
& Soria, 2013). These challenges for first generation students may be real, but sometimes
commuter students, which was found to be an indicator of success in some studies, but
residents. Non-resident students are required to pay the higher out-of-state tuition rates.
While no information on a correlation between tuition rates and academic success were
found in the literature, there were studies that identified a positive relationship between
students who received educational grants and academic success (Conrood, 2008).
Another explanation for the higher grades from nonresident students is the opportunity
for nonresident scholarships. Students who meet minimum GPA (3.6 and above) and
entrance exam requirements (ACT 26 or higher, SAT 1240 or higher) from partner states
can receive scholarships to cover the difference between nonresident and resident tuition
rates (Office of Financial Aid, 2016). If a high level of achievement isnot maintained,
One academic factor from this study that contradicts the existing literature is
success based on academic load. For the student audience in the study population,
students enrolled part-time performed better in both face-to-face and online courses as
compared to those enrolled full-time. The literature from other studies consistently found
that full-time students were more likely to succeed (Adelman, 1999; Aragon & Johnson,
2008; Colorado & Eberle, 2010; Demetriou & Schmitz-Sciborski, 2011; de Freitas et al.,
2015). There are many possible explanations for this finding, but not one identified in the
existing literature. One study, completed by Ibrahim, Freeman, and Shelley (2011),
evaluated demographic and job satisfaction variables related to the academic success of
part-time students. They found that students were more successful in their courses if they
were satisfied with their employment and if their job was related to their field of study.
Another academic factor of interest was a student’s high school GPA. This data
point was available for only about 80% of the enrollments, there was a very weak
correlation between high school GPA and final grade in a course (r = 0.09202, p <
0.0001). While this result aligns with the literature, it is a very weak correlation. It is not
nearly as strong as what Bowen, Chingos, and McPherson (2009) concluded when they
stated that the high school GPA is one of the best predictors of college graduation.
Demetriou and Schmitz-Sciborski (2011) also reported on the connection between high
school GPA and success at the university level. The university that was the basis of this
study had a mix of traditional and nontraditional students. There were only two fields
used in this study that can be used to distinguish traditional from nontraditional students.
Those were the age and academic load. Using these two fields to distinguish
result, there were many students who did not begin their higher education directly after
high school. That delay is likely to change the level of motivation for students as well as
provide time for additional maturity when it comes to study skills and prioritization of
schoolwork.
References
Adelman, C. (1999). Answers in the tool box: Academic intensity, attendance
patterns, and bachelor’s degree attainment. Washington D. C.: Office of
Education Research and Improvements, U. S. Department of Education.
Affendey, L. S., Paris, I. H. M., Mustapha, N., Sulaiman, M. N., & Muda, Z. (2010).
Ranking of influencing factors in predicting students’ academic
performance. Information Technology Journal, 9(4), 832-837.
Allen, I. E., & Seaman, J. (2016). Online report card: Tracking Online Education in
the United States. Babson Park, MA. Retrieved from
http://onlinelearningsurvey.com/reports/onlinereportcard.pdf
Astin, A. W. (1975). Preventing students from dropping out. San Francisco, CA: Jossey-
Bass.
Baker, R., & Siemens, G. (2014) Educational data mining and learning analytics. In
Sawyer, K. (Ed.) Cambridge Handbook of the Learning Sciences: 2nd Edition,
pp. 253-274. Retrieved from:
http://www.columbia.edu/~rsb2162/BakerSiemensHandbook2013.pdf
Baum, S., Ma, J., & Payea, K. (2013). Education pays 2013: The benefits of higher
education for individuals and society. New York, NY: College Board. Retrieved
from: https://trends.collegeboard.org/sites/default/files/education-pays-2013-full-
report.pdf
Becker, B. (2013). Learning analytics: Insights into the natural learning behavior of our
students. Behavioral and Social Sciences Librarian 32(1) 63-67. Doi:
10.1080/01639269.2013.751804
Berger, J. B., & Lyon, S. C. (2005). Past to present: A historical look at retention. In A.
Seidman (Ed.) College Student Retention (pp. 1-29). Westport, CT: American
Council on Education and Praeger Publishers.
Bok, D. (2013). Higher education in America. Princeton, NJ: Princeton University Press.
Bowen, W. G., Chingos, M. M., & McPherson, M. S. (2009). Crossing the finish
line: Completing college at America's public universities. Princeton, NJ:
Princeton University Press.
Braxton, J. M., & Hirschy, A. S. (2005). Theoretical developments in the study of college
student departure. In A. Seidman (Ed.) College Student Retention (pp. 61-87).
Westport, CT: American Council on Education and Praeger Publishers.
Campbell, J. P., deBlois, P., & Oblinger, D. G. (2007). Academic analytics: A new tool
for a new era, EDUCAUSE Review, 42(4), 40-57.
Campbell, J. P. & Oblinger, D. G. (2007). Academic analytics. EDUCAUSE Whitepaper.
1-24. Retrieved from http://net.educause.edu/ir/library/pdf/PUB6101.pdf
Choy, S. (2001). Students whose parents did not go to college: Postsecondary access,
persistence, and attainment, (NCES Publication No. 2001–126). Retrieved from
National Center for Educational Statistics:
http://nces.ed.gov/pubs2001/2001126.pdf
Colorado, J. T., & Eberle, J. (2010). Student demographics and success in online learning
environments. Emporia State Research Studies, 46(1), 4–10.
de Freitas, S., Gibson, D., DuPlessis, C., Halloran, P., Williams, E., Ambrose, M.,
Dunwell, I., Arnab, S. (2015). Foundations of dynamic learning analytics: Using
university student data to increase retention. British Journal of Educational
Technology, 46(6), 1175-1188.
Devadoss, S., & Foltz, J. (1996). Evaluation of factors influencing student class
attendance and performance. American Journal of Agricultural Economics,
78(3), 499-507. Retrieved from http://www.jstor.org/stable/1243268
Dringus, L. P. (2011). Learning Analytics Considered Harmful. Journal of
Asynchronous Learning Networks, 16(3), 87–101.
Dron, J., & Anderson, T. (2009). On the design of collective applications. Proceedings of
the International Conference on Computational Science and Engineering, Vol 4,
pp. 368-374. Retrieved from http://hdl.handle.net/2149/2305
Dziuban, C., Moskal, P., Cavanagh, T., & Watts, A. (2012). Analytics that inform the
university: Using data you already have. Journal of Asynchronous Learning
Network, 16(3), 21–38.
eCampus Center, Extended Studies. (2015). Boise State eCampus Snapshot 2015 [Fact
sheet]. Retrieved from http://ecampus.boisestate.edu/wp-
content/uploads/2015/06/2014-2015-eCampusSnapshot.pdf
Ewert, S. (2010). Male and female pathways through four-year colleges: Disruption and
sex stratification in higher education. American Educational Research
Journal, 47(4), 744-773.
Geiger, R. L. (2015). The history of American higher education: Learning and culture
from the founding to World War II. Princeton, NJ: Princeton University Press.
Grebennikov, L., & Skaines, I. (2009). Gender and higher education experience: A case
study. Higher Education Research and Development, 29(1), 71-84. doi:
10.1080/07294360802444370
Greller, W., & Drachsler, H. (2012). Translating learning into numbers: A generic
framework for learning analytics. Educational Technology and Society, 15(3), 42-
57.
Habley, W. R., Bloom, J. L., Robbins, S., & Gore, P. A. (2012). Increasing persistence:
Research-based strategies for college student success. Hoboken, NJ: John Wiley
& Sons.
Hand, D. J. (1998). Data mining: Statistics and more? The American Statistician, 52(2),
112-118.
Harrell, I. L., & Bower, B. L. (2011). Student characteristics that predict persistence in
community college online courses. American Journal of Distance Education,
25(3), 178–191. doi: 10.1080/08923647.2011.590107
Hofmann, E. (2012), Why dual enrollment? New Directions for Higher Education,
2012(158), 1–8. doi:10.1002/he.20009
Hung, J. L., Hsu, Y. C., & Rice, K. (2012). Integrating data mining in program evaluation
of K-12 online education. Educational Technology and Society, 15(3), 27-41.
Ibrahim, N., Freeman, S. A., & Shelley, M. C. (2011). Identifying predictors of academic
success for part-tie students at polytechnic institutes in Malaysia. International
Journal of Adult Vocational Education and Technology, 2 (4), 1-16.
Ishler, J. L. C., & Upcraft, M. L. (2004). The keys to first-year student persistence. In M.
L. Upcraft, J. N. Gardner, & B. O. Barefoot (Eds.) Challenging and
Supporting the First –Year Student: A Handbook for Improving the First Year
of College (pp. 27-46). Indianapolis, IN: Jossey-Bass.
Jayaprakash, S. M., Moody, E. W., Laura, E. J., Regan, J. R., & Baron, J. D. (2014).
Early alert of academically at-risk students: An open source analytics initiative.
Journal of Learning Analytics, 1(1), 6-47.
Kalsbeek, D. H., & Zucker, B. (2013). Reframing retention strategy: A focus on profile.
New Directions for Higher Education, 2013(161), 15-25. doi: 10.1002/he.20042
Kamens, D. H. (1971). The college charter and college size: Effects on occupational
choice and college attrition. Sociology of Education, 44(3), 270-296.
Karp, M. M., & Hughes, K. L. (2008). Dual enrollment can benefit a broad range of
students. Techniques: Connecting Education and Careers, 83(7), 14-17.
King, A. (1993). From sage on the stage to guide on the side. College Teaching, 41(1),
30.
Lee, Y., & Choi, J. (2011). A review of online course dropout research: Implications for
practice and future research. Educational Technology Research and
Development, 59. 593-618.
Lockee, B. B., Burton, J. K., & Cross, L. H. (1999). No comparison: Distance education
finds a new use for ‘no significant difference’. Educational Technology,
Research and Development, 47(3), 33-42.
Lockyer, L., Heathcote, E., & Dawson, S. (2013). Informing pedagogical action:
Aligning learning analytics with learning design. American Behavioral
Scientist, 57(10), 1439-1459.
Lotsari, E., Verykios, V. S., Panagiotakopoulos, C., & Kalles, D. (2014). A learning
analytics methodology for student profiling. Lecture Notes in Computer
Science, 8445(2014), 300-312.
Martin, T., & Sherin, B. (2013). Learning analytics and computational techniques for
detecting and evaluating patterns in learning: An introduction to the special issue.
Journal of the Learning Sciences, 22(4), 511-520. doi:
10.1080/10508406.2013.840466
McKenzie, K., & Schweitzer, R. (2001). Who succeeds at university? Factors predicting
academic performance in first year Australian university students. Higher
Education Research and Development, 20(1), 21-33. doi:
10.1080/07924360120043621
Moore, M. G., & Kearsley, G. (2005). Distance education: A systems view (2nd ed.).
Belmont, CA: Thomson Wadsworth.
Nora, A., Barlow, E., Crisp, G. (2005). Student persistence and degree attainment beyond
the first year in college. In A. Seidman (Ed.) College Student Retention (pp. 129-
Oblinger, D. G. (2012). Let’s talk analytics. EDUCAUSE Review, 47(4), 10-13.
Office of the Provost, Boise State University. (2012). Focus on effectiveness: Boise
State University’s strategic plan for 2012-2017. Retrieved from
http://academics.boisestate.edu/strategic-plan/
Paechter, M., Maier, B., & Macher, D. (2010). Student’s expectations of, and their
experiences in e-learning: Their relation to learning achievements and course
satisfaction. Computers and Education, 54(1), 222-229.
Park, J.-H., & Choi, H. J. (2009). Factors influencing adult learners’ decision to drop out
or persist in online learning. Educational Technology and Society, 12(4), 207-217.
Peng, C-Y. J., Lee, K. L., & Ingersoll, G. M. (2002). An introduction to logistic
regression analysis and reporting. Journal of Educational Research, 96 (1), 3-14.
doi: 10.1080/00220670209598786
Picciano, A. G. (2012). The evolution of big data and learning analytics in American
higher education. Journal of Asynchronous Learning Networks, 16(3), 9-20.
Pontes, M. C. F., Hasit, C., Pontes, N. M. H., Lewis, P. A., & Siefring, K. T. (2010).
Variables related to undergraduate student preferences for distance education
classes. Online Journal of Distance Learning Administration, 8(2). Retrieved
from http://www.westga.edu/~distance/ojdla/summer132/pontes_pontes132.html
Reason, R. D. (2003). Student variable that predict retention: Recent research and new
developments. NASPA Journal (National Association of Student Personnel
Administrators), 40(4), 172-191.
Reyes, J. A. (2015). The skinny on big data in education: Learning analytics simplified.
Tech Trends, 59(2), 75-79.
Rovai, A. P., Ponton, M. K., & Baker, J. D. (2008). Distance learning in higher
education: A programmatic approach to planning, design, instruction, evaluation,
and accreditation. New York, NY: Teachers College Press.
Shelton, B. E., Hung, J-L., & Baughman, S. (2015). Online graduate teacher
education: Establishing an EKG for student success intervention. Technology,
Knowledge, and Learning, 20(1) doi: 10.1007/s10758-015-9254-8
Sher, S. (2013). Teaching value issues in china to chinese students enrolled in american
universities. (M. Goldman, Ed.) Teaching Philosophy, 36(4), 373–397.
doi:10.5840/teachphil20131014
Siemens, G., & Baker, R. (2012). Learning analytics and educational data mining:
Towards communication and collaboration. Proceedings of the 2nd
International Conference on Learning Analytics and Knowledge. Retrieved from
http://www.columbia.edu/~rsb2162/LAKs%20reformatting%20v2.pdf
Slade, S., & Prinsloo, P. (2013). Learning analytics: Ethical issues and dilemmas.
American Behavioral Scientist, 57(10), 1510-1529. doi:
10.1177/0002764213479366
Swail, W. S. (2004). The art of student retention: A handbook for practitioners and
administrators. Austin, TX: Educational Policy Institute.
Verbert, K. Manouselis, N., Drachsler, H., & Duval, E. (2012). Dataset-driven research to
support learning and knowledge analytics. Education Technology and Society,
15(3), 133-148.
Wagner, E., & Ice, P. (2012). Data changes everything: Delivering on the promise of
learning analytics in higher education. EDUCAUSE Review, 47(4), 32-42.
Watt, C., & Wagner, E. (2016). Improving Post-Traditional Student Success. Retrieved
from
https://www.hobsons.com/res/Whitepapers/PostTraditionalStudents_ParFramewo
rk_February2016.pdf
Glossary
This study found that course modality, either face-to-face or online, was not a
determining factor of success at the university level, nor were most demographic or
academic factors. In some cases, the course itself played a role in the likelihood of a
student’s success, but the best predictor was a student’s previous academic success, as
observed through cumulative GPA. This success was either at the high school level, in
One concern was the higher number of withdrawals in the online courses. Despite
increased enrollments in online courses, online learning is still a modality that many
students have not experienced. Because of this situation, the expectations for courses
success. This may help to equalize withdrawals in online courses and bring it closer to the
withdrawal rate of face-to-face courses, an area of concern for online course offerings at
the university. One misconception that is common among college students is that online
courses will be easier, or less rigorous than face-to-face courses. Some students who
enroll in online courses may discover this is not necessarily the case upon enrolling in a
class and a review of the syllabus and end up withdrawing from the class.
The results of this study can be used by a number of stakeholders both within the
university and beyond. The university administrators can draw from this information to
alter admissions standards that can affect the likelihood of success in course enrollments,
and in turn impact the graduation and retention rates (Dziuban et al., 2012). If the
university chooses to grow enrollments, they would lower entrance requirements. If, on
the other hand, they want to focus on increased graduation rates, they can use the results
of this study to restrict admissions in a manner that encourages success. To do so, they
could look at the factors that were indicators of success like entry level GPA. While
university cumulative GPA is the greatest predictor of success, other factors can be used
Faculty and support staff at the university can use the information to identify
success rates between the online and face-to-face modalities. The reason for these
differences may be due to the design of the courses, or the instructional techniques
employed in the course. These courses and instructors can be identified and reviewed by
instructional designers for a redesign that can narrow the performance gap (Lockyer, et
al., 2013). Some examples include courses offered by the College of Engineering, the
Faculty in both face-to-face and online courses can use information on the
demographic and academic factors of the students enrolled in their courses to perform
some preliminary student analysis. For example, if an instructor learns that most of the
students enrolled in their course has work experience and is enrolled on a part-time
status, he or she may choose to integrate some of the andragogical techniques outlined by
Knowles (1984) such as providing them with opportunities to share their life experiences
and apply them to their learning. Academic advising can apply this information in
helping students select classes and to inform which students are in need of
additional support.
In conclusion, the action that can be taken on the specific results of this study
can help universities integrate statistical modeling and other learning analytics
techniques into their decision making processes. The type of data included in this study
can be combined with learning activity data to advance the analytics to a prescriptive
level. As the field of learning analytics continues to grow, universities will find these
tools to be an invaluable resource for advising students and making informed decisions