Abstract
Online education has grown rapidly in recent years with many universities now offering fully online degree programs even in STEM disciplines. These programs have the potential to broaden access to STEM degrees for people with social identities currently underrepresented in STEM. Here, we ask to what extent is that potential realized in terms of student enrollment and grades for a fully online degree program. Our analysis of data from more than 10,000 course-enrollments compares student demographics and course grades in a fully online biology degree program to demographics and grades in an equivalent in-person biology degree program at the same university. We find that women, first-generation to college students and students eligible for federal Pell grants constitute a larger proportion of students in the online program compared to the in-person mode. However, the online mode of instruction is associated with lower course grades relative to the in-person mode. Moreover, African American/Black, Hispanic/Latinx, Native American, and Pacific Islander students as well as federal Pell grant eligible students earned lower grades than white students and non-Pell grant eligible students, respectively, but the grade disparities were similar among both in-person and online student groups. Finally, we find that grade disparities between men and women are larger online compared to in-person, but that for first-generation to college women, the online mode of instruction is associated with little to no grade gap compared to continuing generation women. Our findings indicate that although this online degree program broadens access for some student populations, inequities in the experience remain and need to be addressed in order for online education to achieve its inclusive mission.
Introduction
Online college education has been growing rapidly, not only in terms of the number of students enrolled in online courses and the number of courses offered, but also in the number of fully online degree programs. About 33% of college students take online courses and 15% of college students are enrolled in fully online degree programs, including 7.3% of students in 4-year undergraduate degree programs [1, 2].
Because fully online education removes many of the constraints imposed by a brick and mortar location, it has enormous potential to increase access to students who may otherwise not be able to pursue a college degree [3]. First, online education could increase access for higher education for students who live in education deserts. Geographical inequalities in access to resources such as high quality K-12 public education and nutritious food along the lines of race and class are well documented in the US [4–6]. Similar spatial inequalities in access to higher education have been termed “education deserts” and are defined as places with no colleges or public universities within 25 miles or with access to only one community college as the only public broad access institution within 25 miles [7, 8]. According to one estimate, about 17.6% of American adults live in an education desert [8]. Akin to other forms of structural inequalities in our society such as food deserts, these education deserts are more common in low-income communities [7, 9]. While online education cannot entirely solve these inequities in access to educational opportunities because of the availability of the technology required for online degrees in some areas, many of the students who live in education deserts do have sufficient internet service to access online education [8]. A second way that online education can increase access is by accommodating a broader range of schedules. Asynchronous online education could provide college access for students with severe time constraints that necessitate a flexible course schedule, such as students who serve as primary caregivers (which are often women with childcare responsibilities) and those students who are seeking a degree while working a full-time job. Several studies comparing the demographics of students in online degree programs to in-person degree programs show that online programs indeed increase access for such student populations. Students in fully online degree programs are more likely to be women, older than students attending a 4-year residential college, and eligible to receive federal financial aid (e.g., Pell grants), although they are as or more likely to be white [10–15].
In addition to broadening access to higher education for students, online education could help meet the demand for a larger and more diverse STEM workforce. The President’s Council of Advisors on Science and Technology (PCAST) has called for increasing the number of STEM majors by one million over the next decade [16]. Greater access to STEM degrees through online programs could contribute significantly to meeting this goal, especially because many of our brick and mortar campuses are limited in their capacity to offer physical classes and thus limited in their student enrollment [17]. Although most online degree programs currently offered are in non-STEM disciplines, fully online degree programs constitute 14% of all health professions and related associate’s or bachelor’s degree programs in the sciences (2041 out of 14578 programs) [18]. For STEM disciplines, online degree programs are more common in fields such as computer and information sciences (21%), and the numbers are considerably smaller for biological and biomedical sciences, where fully online degree programs represent only 1% of associate’s or bachelor’s degree programs (33 out of 3300 programs) [18]. We know of only three bachelor’s degree programs in biology from research intensive institutions that are fully online: University of Florida, Florida International University, and our program at Arizona State University. Further, only Arizona State University offers a Bachelor of Science degree.
It has been argued that online education’s route to increased access to higher education could also amplify the diversity of the STEM workforce [19, 20]. A diverse STEM workforce composed of individuals with different backgrounds, perspectives, and identities can help make science and technology research and their applications more just in terms of what questions are examined, who is most likely to benefit, and who could potentially be harmed. Historically, the benefits of science and technology have been distributed disproportionately to more privileged groups in our society (e.g., disparities in federal funding between diseases such as Cystic Fibrosis that more often affects white people compared to sickle cell disease that more often affects people with sub-Saharan African ancestry [21]), while harms have predominantly fallen on people of color (e.g., unethical medical experimentation on Black Americans [22]). Greater participation in the STEM workforce across gender, race/ethnicity, and social class would contribute positively to social justice in STEM, increase objectivity in science by including diverse perspectives, and overall lead to higher quality science and technology [23, 24].
Women, Black students, Latinx students, Native American students, and students with low socioeconomic resources continue to be underrepresented in most STEM undergraduate degree programs [25, 26]. Online education seems to increase the diversity of students in STEM undergraduate degree programs to some extent. Studies show that women and students who grew up in low-income households constitute a larger proportion of students in STEM online courses as well as in fully online STEM degree programs. In a comparison of 174 online and 440 in-person biology students at a large public research-intensive (R1) institution in the southwestern USA, Cooper et al. [15] found that a greater proportion of online students were first-generation college students, grew up in a low-income household, worked more than 21 hours a week, and identified as primary caregivers. Other studies using the large NCES National Postsecondary Student Aid Study 2008 dataset document that the likelihood of taking an online STEM course was higher for women and students with non-traditional characteristics (e.g. single parents, full-time workers), but lower for Black and Hispanic/Latinx students even after taking other demographic variables into account [27, 28]. Thus, the current data suggest that online degree programs offer a way for a greater number of women and low-income students to pursue science, but there is no existing evidence that it can help racial or ethnic representation. This is potentially worrisome because there are large racial/ethnic disparities in science, particularly biology, and fully online degree programs may not address these racial/ethnic inequities.
Further, online degree programs may not increase the number of STEM graduates because student enrollment in online coursework is not equivalent to completion. Studies report 10–15% higher withdrawal rates from online courses compared to in-person courses in community colleges, even after controlling for students’ prior academic achievement and other demographic variables [29, 30]. It is important to note that all of the studies referred to here have been examining courses for credit within a degree program; this does not include massive online open courses (MOOCs), which are well-known to have extremely low completion rates [31]. Another factor to consider is student performance in online courses. While several studies report comparable learning outcomes for students in online courses compared to face-to-face courses [14, 32], other studies show worse student performance in online courses [11, 12, 33, 34]. Xu & Xu [18] suggest that these differences in results might be explained by the selectivity of institutions studied. Studies done on community college systems and non-selective institutions consistently demonstrate negative student learning outcomes online, whereas studies at selective institutions tend to have mixed results [18]. Differences in the population of students who enroll in online courses at selective compared to non-selective institutions in terms of prior academic preparation and/or availability of time or resources might contribute to this observed pattern. Some studies show that Black students [35, 36], Latinx students [36, 37], and low-income students [38] do worse in online courses compared to face-to-face courses and students with these social identities constitute a larger proportion of the student body in non-selective institutions [39].
Whether student performance in online courses is comparable to in-person courses may depend on the academic field examined. Some research has found that grade differences between online and in-person courses are smaller in the natural sciences than other fields [11, 35, 36]. In contrast, a comparison of retention rates for online and in-person courses taught by the same instructor in the same semester showed lower retention rates in STEM courses, particularly elective courses, compared to non-STEM courses for online students [40]. However, numerous studies have reported gender, racial, and socioeconomic disparities in grades received by students in in-person STEM courses and shown that men, white, and middle-or upper class students tend to receive higher grades compared to women, students of color, and low income students, respectively [41–47]. Because there is a greater degree of anonymity in participation in online courses, it could be hypothesized that students could experience less discrimination and stereotype threat in online courses, so students with social identities underrepresented in STEM could do better in online courses. However, the evidence so far does not support this idea. Wladis et al. [30] showed that women are more likely to get a grade equal to or higher than a C compared to men in face-to-face STEM courses, but that both men and women have similar probability of getting a grade equal to or higher than C in online courses, even after controlling for other confounding variables such as race/ethnicity, age, and prior GPA. However, the lack of gender disparity in grades received in online courses observed in this study was a result of women doing significantly worse in the online courses compared to in-person courses [30]. While Latinx and Black students generally received lower grades in STEM courses compared to white and Asian students, this grade gap has been shown to be similar in online and face-to-face courses, so previous studies suggest that the online format does not seem to lessen these grade disparities [30].
While there have been many studies of student performance in online courses compared to in-person courses, very few studies have examined student performance in fully online vs in-person degree programs. Fully online asynchronous degree programs offer much greater geographic and temporal flexibility than programs where students can take some online courses within the purview of an in-person degree program. Student experiences in such fully online programs likely differ significantly from student experiences in online courses nested within in-person degree programs, potentially because students enrolled in online programs tend to be older and have more work experience. Some of us published a study on students that completed a fully online astrobiology course that showed that students in the fully online degree program began the course with a higher sense of personal value of science compared to students in the in-person degree program [48]. We also conducted a study demonstrating that students in a fully online biology degree program who wanted to become doctors were more confident about their career choice, but knew less about the criteria that medical schools consider in their admissions process, compared to students in an equivalent in-person biology program [15]. Thus, beyond just learning outcomes and course grades, there could be downstream effects for students in online degree programs that influence their career trajectory. However, in order to begin to understand these effects, we argue that we first need to establish how student representation and student grades differ between fully online and in-person degree programs.
In this study, we examined student representation and grades in a fully online biology degree program compared to an equivalent in-person biology degree program offered at a large non-selective public university in the southwestern United States. Systematic disparities in grades and other academic outcomes along the lines of gender, race/ethnicity and class reflect problems with institutional structures such as course and assessment design as well as broader social inequities, i.e. an “opportunity gap” [49]. In addition to looking at disparities in grades between students with different social identities, we also took a quantitative intersectional approach based on Black Feminist Theory developed by Kimberlé Crenshaw, Angela Davis, Combahee River Collective and Patricia Hill Collins among others [50–53] and asked how the interlocking systems of oppression of racism, classism and sexism could affect disparities in the grades that students receive. Several studies have shown that race/ethnicity, social class, and gender interact in a variety of ways to influence differences in grades and graduation rates among students with different social identities in college education [45, 54–57]. For example, Harackiewicz et al. found that the average grade disparity in an undergraduate biology course between first-generation BLNP (Black, Latinx, Native American, and Pacific Islander) students and continuing generation white and Asian students was about twice that between continuing generation BLNP students and white and Asian students [45]. Compounding effects of this nature or other forms of interactions might be observed in online courses as well.
Research questions
Research question 1
Has the online biology degree program increased representation of students with social identities that are underrepresented in STEM such as women, Black students, Latinx students, first-generation students, and low-income students compared to an in-person degree program?
Research question 2
(a) Within the biology program’s core required courses for all majors, do students with social identities that are historically underrepresented in STEM receive lower grades (women, Black, Latinx, first-generation, low-income) compared to students with social identities that are historically overrepresented in STEM (men, white, continuing generation, middle- to high-income)? (b) Do these grade gaps differ between the online science courses and the in-person courses? We define grade gaps here as the difference in grades received by students with social identities historically overrepresented in STEM compared to the grades received by students with social identities historically underrepresented in STEM. While women are no longer underrepresented at the undergraduate biology level, they still are underrepresented in the professoriate.
Research question 3
Are grade gaps exacerbated for students holding multiple social identities that are underrepresented in STEM?
Research question 4
Do grade gaps between students holding multiple identities that are underrepresented in STEM and students with social identities that are overrepresented in STEM differ between online and in-person courses?
Methods
Data
This study uses student course grades, student GPAs, and demographic data from students enrolled in in-person and fully online biology degree programs from a single non-selective university. All data were collected from the university registrar. The data were anonymized, and our analysis was performed in accordance with an institutional review board-approved protocol. We included data from students enrolled in the Biological Sciences major and from the set of core science courses required for that major. These included introductory biology courses, an evolution course, and a genetics course (six biology courses total); two general chemistry courses and two organic chemistry courses (four chemistry courses total); and two general physics courses (two physics courses total). Course grades were taken from the Fall 2014–Spring 2018 semesters. This research was conducted under a protocol approved by the Arizona State University institutional review board (STUDY #9105). No consent was required because all data was analyzed anonymously.
The Biological Sciences major was chosen because it is a high enrollment program that is represented in both the in-person and fully online settings. The degree requirements for the in-person and online degrees are equivalent, and online and in-person courses have been backward designed so that they have similar learning goals/outcomes and course content. Many of the same instructors of in-person courses have designed the online version of the course. Although this unified design makes a comparison between courses in these two degree programs possible, it is important to acknowledge that while the courses are intended to be equivalent, there are likely to be some differences between them. A notable difference is that in-person courses are only offered over 15-week, whereas online courses are only offered over 7.5-weeks. Students are advised to take half as many courses in the 7.5 week session than the 15-week session, so even though the course pace of one course is accelerated, the overall course load should be the same. Given our study design, it was not possible to control for the assessment formats, grading schemes, or other factors that are left at the discretion of individual course instructors. Although we know that there are differences among instructors in their assessment decisions even in the same in-person course offering [58], we do not know of any systematic differences between online and in-person courses as far as assessments.
Course grades are on a 4.33 scale (4.33 = A+, 4.0 = A, 3.66 = A-,…, 0 = Fail). Students who received “withdraw” grades were converted to failing (0.0) grades. We made this choice because all of the courses in question are required for graduation, thus, withdrawing and needing to later retake the course has a similar negative effect on a student as failing the course even though withdraws do not impact student GPA. It is important to note that there are systematic differences in withdraw rates by demographic groups and between in-person and online courses (Table 1). More students withdraw from online sections (13.9% compared to 9.0%, Χ2(1) = 53.211, p < .001), and, considering there is a shorter window of time to drop courses in the online program, any decision regarding treatment of withdraws will impact our overall conclusions. For interested readers, we provide a complete alternative analysis in the supplemental materials in which withdraws were excluded (S1–S6 Tables in S2 File).
Table 1. Withdraw percentages by student group and mode of instruction.
Demographic Category | Withdraw Percentage | |
---|---|---|
Gender | Women | Men |
In-person | 8.2 | 8.9 |
Online | 14.2 | 12.5 |
Race/Ethnicity | BLNP | White |
In-person | 10.1 | 8.5 |
Online | 16.5 | 12.5 |
College generation status | First-generation | Continuing generation |
In-person | 10.4 | 7.6 |
Online | 13.1 | 14.2 |
Pell eligibility | Pell eligible | Non-Pell eligible |
In-person | 10.2 | 7.2 |
Online | 14.0 | 13.5 |
We used grade point average in other courses (GPAO) to indicate each student’s prior academic success [41, 59]. For any student-course enrollment in our dataset, we calculated GPAO by removing the contribution of the course in question from the end-of-term GPA record. We preferred this method over using start-of-course GPA because it allows us to have an academic success indicator that is valid for first-semester students. Because this institution does not require incoming test scores, significant missing data prevented us from relying on high school GPA or standardized test scores as a proxy for academic achievement.
Demographic data were collected from the registrar. For race/ethnicity, we included only students identified as white, Black, Latinx, Native American, or Pacific Islander. We then operationalized this category to the majority group (white) and groups that are underrepresented in STEM, which include Black, Latinx, Native American, and Pacific Islander (which we abbreviate in the figures as BLNP). We prefer this abbreviation to the commonly used term URM (underrepresented minority), because it retains the names that people in these groups use to identify themselves [60, 61]. We purposely do not capitalize white because “white” does not represent a shared culture and history in the way Black does and has long been capitalized by hate groups [62]. Our analysis excludes Asian students, students reported as belonging to two or more races, and any students for whom these data are unreported or missing. We chose to exclude Asian students because our focus in this study is on groups that are historically underrepresented in science in the US and Asian students have not been included in the National Science Foundation’s definition [26]. We acknowledge that there are many subsets of Asian students, some of whom are underrepresented in STEM in the US [63]. Unfortunately, we were not able to disaggregate Asian students because of the way that these data were collected by the institution. We also excluded the two or more races category due to uncertainty about group membership. For gender, we only included women and men in our analyses. This excluded two students who had non-binary genders recorded. We acknowledge that gender exists as a continuum, but we were not able to statistically analyze the two non-binary students. For income level, we included federal Pell grant eligibility. Students who received any Pell grant money in any academic year studied were classified as Pell eligible and considered low-income. All other students were considered non-Pell eligible. For college generation status, we used first-generation college student data collected from students’ college application and financial aid documents. We categorized students as first-generation to college if neither parent had earned a college degree and as continuing generation otherwise. This resulted in a final data set of 10,249 student-course-enrollments across 24 courses (treating in-person and online versions of the same course as distinct courses).
Data analysis
To assess if student representation in the online biology program was significantly different from the in-person program, we compared the proportions of students using a chi-square test. We used mixed effects linear modeling to determine the effects of four demographic variables (gender, race/ethnicity, college generation status, and federal Pell grant eligibility) as well as online modality in predicting course grades. We used GPAO to adjust for each student’s general academic success. GPAO was centered separately among the in-person and online students, using each cluster’s mean GPAO [64]. This centering choice follows from the fact that the two populations have no enrollment overlap—online course sections are open to only online students, and vice versa—and that there are systematic differences in how students are graded between the online and in-person modes. We also included random effects components on the model intercept. We used separate random effect terms to account for the data dependency (i.e., clustering) caused by (1) multiple student-enrollment records coming from the same students, and (2) individual records nested within course sections, respectively. We calculated intraclass correlation coefficients for our base model (model A below) with random effects for both students and sections and compared it to models with only one of these random effects. The intraclass correlation coefficient (ICC, Eq 1) [65] is a ratio of the variance explained by random effects to the total variance explained by random and fixed effects plus residual variance. In model A (below) ICC was 0.356. Variants of model A including only one of these random effects show ICC values for course sections of 0.100 and for individual students of 0.236. These ICC values show that the random effects are appreciable and should be included in our modeling.
Eq 1 |
We began with a regression that included the GPAO term and the four demographic terms. We examined each possible addition to the model in a stepwise manner and made model selections based on comparison of AIC values and statistical comparisons using likelihood ratio tests.
Statistical analyses were performed in R [66] and made use of the lme4 [67], lmerTest [68], MuMIn [69], performance [70], and sjPlot [71] packages for the modeling. The anonymized data used in this study as well as the R code used in our analyses are provided in S2 File.
Results
Finding 1: Compared to in-person program, online biology degree program increases representation of women, first-generation, and Pell eligible students, but not BLNP students
Women, white students, and continuing generation students are in the majority in both the in-person and online populations we studied. Non-Pell eligible students represent a slight majority of in-person students, whereas Pell eligible students are a slight majority online (Fig 1). A summary of racial/ethnic data within the BLNP category and between in-person and online populations is provided in S7 Table in S2 File. Statistical analysis of the differences between in-person and online populations shows that the proportion of BLNP students between the in-person and online biology program is not statistically different (X2(1) = 0.00, p = .95), while the proportions of women (X2(1) = 233.76, p < .001), first-generation students (X2(1) = 62.21, p < .001), and Pell eligible students (X2(1) = 75.27, p < .001) are each significantly higher online compared to in-person (Fig 1).
Finding 2: Women, first-generation, Pell eligible, and BLNP students receive lower grades regardless of modality. For first-generation students, this grade gap is eliminated in online courses
To assess whether students with historically underrepresented social identities in STEM receive lower grades in science courses than students with overrepresented social identities, we first ran a model with course grade as the outcome and demographic variables as predictors while controlling for student grades in other courses (see Table 2 for full model specification). This initial model (A), shows that women, first-generation, Pell eligible, and BLNP students are each associated with lower course grades (Tables 2 and 3). The size of these effects in model A ranges from around -0.08 grade units for women and first-generation students, to -0.11 for Pell eligible students, to -0.22 for BLNP students.
Table 2. Models compared for research question 2.
A: Course Grade ~ GPAO + Gender + College Generation + Pell Eligibility + Race/Ethnicity + (1 | STUDENT) + (1 | SECTION) |
B: ~ {Model A} + Online |
C: ~ {Model B} + Gender: Online |
D: ~ {Model B} + College Generation: Online |
E: ~ {Model B} + Pell Eligibility: Online |
F: ~ {Model B} + Race/Ethnicity: Online |
Table 3. Regression results for research questions 2a and 2b.
Model A | Model B | Model D | |
---|---|---|---|
Intercept | 2.664*** (0.038) | 2.721*** (0.039) | 2.740*** (0.040) |
GPAO | 0.643*** (0.015) | 0.644*** (0.015) | 0.643*** (0.015) |
Gender (W) | -0.078* (0.032) | -0.066* (0.032) | -0.064* (0.032) |
College Generation (FG) | -0.085* (0.033) | -0.081* (0.033) | -0.151*** (0.043) |
Pell Eligibility (Y) | -0.109*** (0.032) | -0.104** (0.032) | -0.095** (0.032) |
Race/Ethnicity (BLNP) | -0.215*** (0.034) | -0.215*** (0.032) | -0.212*** (0.033) |
Online (Y) | -0.500*** (0.078) | -0.571*** (0.082) | |
Online (Y): College Generation (FG) | 0.170** (0.065) | ||
AIC | 30058 | 30019 | 30014 |
R-squaredmarginal | 0.215 | 0.241 | 0.242 |
No. Observations | 10249 | 10249 | 10249 |
Standard errors are reported in parentheses.
*, **, *** indicate significance at the 95%, 99%, and 99.9% levels, respectively.
Next, in model B, we added the mode of instruction term (online or in-person). This significantly improved model fit (likelihood ratio test, p < .001) and showed that the online mode of instruction was associated with a large, negative grade difference. This means that on average, students earn lower grades online. Finally, to consider whether this online effect is differential by student demographics, we added interactions between mode of instruction and each demographic variable, in turn, in models C–F. Model D, which includes an interaction between first-generation status and mode of instruction, was the best-fitting model based on AIC and likelihood ratio tests (likelihood ratio test compared to model B, p = .009). According to this model, while first-generation students earn lower grades in-person than continuing generation students, there is almost no grade gap between first-generation and continuing generation students in online courses. Effects plots of model D are shown in Fig 2 and fit statistics are shown for all models in S8 Table in S2 File.
Although the inclusion of the first-generation status and mode of instruction interaction in model D was the only change among models C–F to significantly improve upon the no-interactions model (B), model F that includes an interaction between race/ethnicity and mode of instruction does at least marginally improve model B as well. Model F shows a positive interaction between BLNP and online instruction, analogous to the interaction effect in model D, i.e. the grade disparity between white and BLNP students is lower online than in-person. Although it was not significant in our dataset, it is possible that this is a real effect and may merit further examination, especially with disaggregated race/ethnicity data. Interestingly, in alternative modeling with withdraw grades excluded, model F is the best fitting of models C–F (S2 File).
For readers interested in viewing the full dataset, S1–S4 Figs present simple linear regressions for each demographic subpopulation. Model fit statistics and results for all models are shown in S2 File. The full data and R code for analyses are also included in S2 File.
Finding 3: Grade disparities are generally not exacerbated for students holding multiple identities underrepresented in STEM
Beginning from the initial, no-interaction model from the previous section (model A), we tested additional pairwise interactions between the demographic terms (G–L, Table 4). None of these represent a significant improvement over model A. Of these, model G (Pell Eligibility: Race/Ethnicity) has the most support from AIC, offering some evidence that BLNP students who are Pell eligible receive lower grades than BLNP students who are not Pell eligible (S10 Table in S2 File). Other than BLNP and Pell eligibility, grade gaps were not exacerbated for students holding any combination of identities examined here.
Table 4. Models compared for research question 3.
A: Course Grade ~ GPAO + Gender + College Generation + Pell Eligibility + Race/Ethnicity + (1 | STUDENT) + (1 | SECTION) |
G: ~ {Model A} + Pell Eligibility: Race/Ethnicity |
H: ~ {Model A} + College Generation: Race/Ethnicity |
I: ~ {Model A} + Gender: Race/Ethnicity |
J: ~ {Model A} + College Generation: Pell Eligibility |
K: ~ {Model A} + Gender: College Generation |
L: ~ {Model A} + Gender: Pell Eligibility |
Finding 4: Continuing generation women students receive better grades than first-generation women in-person, but this difference is not observed online. Men receive higher grades than women online, but not as much in-person
Using model D (College Generation: Online) as a point of comparison, we again tested the inclusion of all possible pairwise interactions for the demographic categories, but here we also included an interaction between these pairwise combinations and instruction mode (Table 5). As with the findings from RQ3, we find that most of the models tested are similar or worse in terms of AIC (S8 Table in S2 File). The best fitting of these new models was Model Q, which includes a three-way “gender: first-generation: instruction mode” interaction (Table 6). This model indicates that the interaction between first-generation status and mode of instruction in model D is predominantly driven by the additional interaction with gender. There is little grade gap between first-generation and continuing generation men, both online and in-person (Fig 3C). However, continuing generation women receive higher grades than first generation women in in-person courses, but not in online courses (Fig 3C). This model also indicates that there is an interaction between gender and mode of instruction. While all students receive better grades in in-person courses than online courses, this effect is particularly large among continuing generation women, meaning that continuing generation women receive lower grades than continuing generation men online (Fig 3D).
Table 5. Models compared for research question 4.
B: Course Grade ~ GPAO + Online + Gender + College Generation + Pell Eligibility + Race/Ethnicity + (1 | STUDENT) + (1 | SECTION) |
D: ~ {Model B} + College Generation: Online |
M: ~ {Model B} + Pell Eligibility: Race/Ethnicity: Online |
N: ~ {Model B} + College Generation: Race/Ethnicity: Online |
O: ~ {Model B} + Gender: Race/Ethnicity: Online |
P: ~ {Model B} + College Generation: Pell Eligibility: Online |
Q: ~ {Model B} + Gender: College Generation: Online |
R: ~ {Model B} + Gender: Pell Eligibility: Online |
Table 6. Regression results for models examining whether online modality exacerbates or reduces any grade disparities for students holding multiple social identities underrepresented in STEM.
Model D | Model Q | |
---|---|---|
Intercept | 2.740*** (0.040) | 2.697*** (0.045) |
GPAO | 0.643*** (0.015) | 0.641*** (0.015) |
Gender (W) | -0.064* (0.032) | 0.011 (0.049) |
College Generation (FG) | -0.151*** (0.043) | -0.074 (0.065) |
Pell Eligibility (Y) | -0.095** (0.032) | -0.093** (0.032) |
Race/Ethnicity (BLNP) | -0.212*** (0.033) | -0.210*** (0.033) |
Online (Y) | -0.571*** (0.082) | -0.454*** (0.101) |
Online (Y): College Generation (FG) | 0.170** (0.065) | 0.018 (0.118) |
Online (Y): Gender (W) | -0.178* (0.089) | |
College Generation (FG): Gender (W) | -0.133 (0.083) | |
Online (Y): College Generation (FG): Gender (W) | 0.235† (0.141) | |
AIC | 30014 | 30015 |
R-squaredmarginal | 0.242 | 0.243 |
No. Observations | 10249 | 10249 |
Standard errors are reported in parentheses.
†, *, **, *** indicate significance at the 90%, 95%, 99%, and 99.9% levels, respectively.
Discussion
Among the students with social identities that are historically underrepresented in STEM, we found that all except BLNP students are enrolled in higher proportions in the online biology degree program compared to the equivalent in-person biology degree program. This speaks to one of the commonly discussed benefits of online degree programs: increased access. However, all of our regression models of course grade show that the online instruction mode is associated with lower course grades for all students after controlling for the other factors. This is also consistent with prior literature showing lower student grades in online courses, particularly at nonselective institutions. When employers and advanced degree programs are evaluating these students, particularly their GPAs, these lower course grades in core courses may limit their ability to maximally leverage their biology undergraduate degree. It is important to note that we only examined grade differences. Prior studies have indicated that there are other important disparities between fully online degree program students and students in an in-person degree program, including sources of career advice and access to high-impact practices such as undergraduate research [15]. Thus, increasing access to a degree is only the first step as these other inequities need to be considered.
The cause of the overall lower online grades is difficult to identify definitively within this study. Our finding is consistent with some prior work [11, 12, 33, 34], but not all [14, 32]. There are several factors that could be influencing the lower online grades in our study. Online instructors may be less sensitive to the specific concepts that students are struggling with, there may be fewer opportunities for formative feedback because of the modality, or students may struggle with building relationships with the instructor. In addition, in our study specifically, the shortened 7.5 week terms for the online courses mean that a student who suffers an illness, experiences a major life event, or is otherwise prevented from doing coursework for several days or a week will fall disproportionately further behind than a student who is enrolled in a 15 week term. Finally, although our models do account for the students’ college performance, they do not fully control for academic preparation, thus the online effect could be partially explained by lesser college preparedness. This would be consistent with Cavanaugh & Jacquemin (2015) who, while finding a negligible difference in online vs. face-to-face courses, also found that online courses tend to magnify the effect of differences in GPA, with low GPA students performing even worse and high GPA students performing even better [14, 32].
Regarding the specific demographic effects on course grades and considering only main effects, we found that women, first-generation students, low-income students eligible for federal Pell grants, and BLNP students are more likely to receive lower course grades than men, continuing generation students, middle- to high-income students, and white students both online and in-person. So, the online format does not seem to eliminate grade gaps for these populations historically underrepresented in STEM undergraduate degree programs. However, by considering differential effects of online instruction on demographic groups, we found that the grade gap between first-generation and continuing generation students differs between the online and in-person degree programs. While continuing generation students receive better grades than first-generation students in in-person courses, grades are lower for both groups online such that there is almost no grade difference between these two groups in online courses.
By also considering interactions of social identities, we were able to point to a more specific explanation of this interaction between first-generation status and online instruction. When we added an additional interaction with gender (model Q), we observe the significant first-generation and online interaction only among women. Thus, first-generation women earn similar grades compared to continuing generation women in the online degree program, whereas first-generation women in-person earn lower grades than their continuing generation peers. This model also shows an interaction effect between gender and online instruction, suggesting that, overall, women receive lower grades online compared to men.
Explanations
Our results show considerable similarity between in-person and online degree program courses in that BLNP students and students eligible for federal Pell grants received lower grades compared to white students and students that are not eligible for Pell grants in both in-person and online degree programs. This is consistent with prior literature comparing online and in-person STEM courses such as Wladis et al. [30] whose modeling found neither race/ethnicity nor Pell grant eligibility to significantly explain online course grades. Therefore, it is quite possible that the same institutional and societal mechanisms that are hypothesized to drive grade gaps in-person have a similar effect online. One of the most common suggestions for lower exam scores of BLNP students in college science courses is stereotype threat—the phenomenon wherein individuals from negatively stereotyped groups feel pressured to perform well lest they be seen as fulfilling those stereotypes [30, 72, 73]. Although our results do not indicate, as Wladis and coauthors suggest, that stereotype threat is more detrimental online, we still consider it to be a possible factor in explaining the observed grade differences. One reason for similar levels of stereotype threat in online and in-person science courses could be similar prevalence of high-stakes assessments such as timed exams which are known to exacerbate stereotype threat [73, 74]. Reducing the contribution of exams toward final grades and providing frequent low-stakes assessments have been shown to reduce grade gaps in STEM courses [46, 75]. Lastly, given the larger societal context in the United States, structural racism and large opportunity gaps by race and social class are additional possible reasons for persistent grade gaps in online and in-person science courses by race/ethnicity and Pell grant eligibility. In other words, our finding of similar demographic grade differences surely derives in part from the fact that the online and in-person degree programs both operate within a society that is highly unequal (particularly with regard to academic preparation and opportunities) and in part due to institutional structures and traditions such as high-stakes timed exams that exacerbate these inequalities.
Although our results suggest in-person and online equivalence in grade gaps with respect to race/ethnicity and Pell grant eligibility, our models show that women receive lower grades than men in online courses but this grade gap is smaller in-person courses. Previous studies focused on the performance of women in in-person science courses may help explain why women experience larger grade gaps online. Eddy & Brownell [42] reviewed the literature of gender disparities in undergraduate STEM education and reported widespread gender gaps in academic performance, engagement, and affective measures such as self-efficacy, sense of belonging, and science identity. They describe four sociocultural factors underlying the observed gender gaps: stereotype threat, implicit or explicit biases held by peers and/or instructors, conflict between personal goals and stereotypes about STEM professionals, and implicit theories of intelligence. As we suggested above, stereotype threat is a plausible driver of gaps in both instructional modes. However, classroom interactions may also be a factor. Eddy et al. [47] observed substantially lower classroom participation in-person among women compared to men. Subsequent studies in the context of in-person courses have documented gender differences in student comfort participating in front of the whole class [76], lower likelihood of women participating in larger classes [77], student perception of their own intelligence when comparing it to others [78], and preference for working with a friend in small group work [76]. Such differences not only represent missed learning opportunities for women to participate but can also perpetuate stereotypes about men being smarter or more capable in science. The effect of class participation is difficult to predict. Given that class participation in online courses often involves posting on a discussion board that is visible to the full class, many of the same biases and social pressures discussed by Eddy et al. and Ballen et al. as far as whole class discussions may be present in online discussion boards too. On the other hand, some studies show that women participate disproportionately more in web-based classroom discussions [79, 80]. Finally, student age data show that in our dataset, men enrolled in online degrees are on average about two years older than women. This may imply other differences in life circumstances between men and women in online programs. Thus, whereas for traditional-aged students, women typically come to college with better preparation [81], it is possible that within the population interested in pursuing online degrees, men are more prepared. Lastly, in explaining their finding that women performed worse in online courses, Wladis et al. [30] suggest that women taking online courses are more likely to be primary caregivers of small children and thus have less time available for coursework, which is consistent with a previous study of a fully online biology degree program [15].
The second difference we find with respect to mode of instruction is a reduction of the grade gap for first-generation students in the online setting, particularly among women. Prior work studying the success of first-generation to college students in the context of in-person courses suggests a range of factors that may drive lower grades compared to continuing generation students, including general academic preparation, poverty, and feeling unconnected to or threatened by the college environment [82]. Model D which includes an interaction between online modality and first-generation status shows that the first-generation grade gap is eliminated in the online instruction mode, while model Q which includes a three-way interaction between gender, online modality and first-generation status shows this reduction being limited to first-generation women online. This could reflect that the online college experience is in fact equivalent for first-generation and continuing generation students. Because online college is a relatively novel experience, one possibility is that there is no advantage from having a family history in college. Alternatively, because first-generation students make up a larger percentage of online students, these students may perceive the online program to be more welcoming and may not feel out of place in the way that some first-generation students feel in in-person settings when they interact with other students. A different possibility is that students are not interacting with each other as much online, so there are fewer opportunities for them to compare themselves to each other and not feel as though they belong. Finally, instructors may be more explicit with the norms and expectations of online courses than they are with in-person courses, so information about the course may be more equitably distributed to students through a syllabus or online course materials. However, it is important to note that the lack of grade gap in the online modality is due to poorer performance of continuing generation students online, not because of better performance of first-generation students online.
Having discussed possible explanations for these effects, it is important to explain that these estimates represent a conservative lower bound on the in-practice demographic effects on college course grades. There are two reasons why this is the case, which we will discuss in turn along with alternative models that provide possible estimates of the true effects (S11 Table in S2 File).
Alternative subgroup-centered GPAO model
The first, and most significant, reason that our estimates are conservative is that there are systematic differences in GPAO between the demographic groups studied and the demographic compositions are systematically different across instruction mode. Because of these systematic differences and because our models used GPAO centered within each of the in-person and online instruction modes, the estimated demographic effects are smaller than those that would be found by simply calculating average grades by group. To estimate the maximum variance that can be ascribed to demographics, we can run alternative models in which the GPAO term is centered within both instruction modes and demographic subgroups. In effect, this forces all of the between-group variation onto the demographic terms. In this alternative model, the explanation for the demographic effects includes both the pre-college academic preparation differences and any university policy, culture, or classroom causes.
We ran two subgroup-centered models (S and T, S12 Table in S2 File), based on models D (College Generation: Online) and Q (Gender: College Generation: Online), respectively. As expected, these yield larger estimates of the demographic effects and interaction effects than our main models. It is also notable that for these subgroup-centered models, the more complex model with a three-way interaction between gender, first-generation status, and online instruction mode (T) is a significantly better fit than model S, which only has a two-way interaction between first-generation status and online instruction mode (likelihood ratio test, p < .001). This adds credibility to the effects seen in the main model Q, which was a non-significant improvement over model D. That is, it provides further evidence that the grade gap between first-generation and continuing generation is lower among women in the online program compared to in-person program, but there is little grade gap between first-generation and continuing generation men in both the online and in-person programs.
Alternative prior GPA model
The second reason our estimates are conservative stems from our use of students’ college GPAs to account for overall academic performance. We chose to use college GPA because measures of pre-college academic performance, such as high school GPA or college transfer GPA, suffer from significant missing data for our student population in ways that may skew the data. However, when we run our models on the subset of data for which high school GPA or college transfer GPA are available, we find slightly larger effects that are otherwise consistent with our main model results (S13 Table in S2 File). Detailed results from these models are reported in S2 File.
The most likely explanation for this difference in magnitude is that students from these demographic groups receive lower grades in all of their courses, thus by using GPAO we are somewhat muting the true demographic effects. That said, the general concordance of results from this alternative model lends support to our overall conclusion that there are systematic demographic grade gaps in these science courses. We chose to be conservative in our approach presented in the study, but the real impact on students is likely larger than our estimates.
Limitations
All of our main models use GPAO to adjust for a student’s general academic success. However, because in-person degree students cannot enroll in the same courses as online degree students, and vice versa, we cannot test the consistency or similarity of grading between these two modes. This somewhat increases the uncertainty of our model estimates for the effect of instruction mode. However, the equivalent in-person and online courses are offered by the same academic units, so we assume that the grading policies and student expectations are similar in all offerings.
Previous studies have included student age as a predictor of online course grades. We did examine student ages within our dataset and found the expected pattern of online students being substantially older than in-person students (Monline = 27.3 yrs; Min-person = 23.4 yrs; t.test = 32.425, p < .001). There were also smaller age differences (±1 year) between demographic groups within each instruction mode population. The largest of these showed that men in the online degree program were 2.1 years older, on average, than women. Because the age difference is so strongly correlated with instruction mode, it was not feasible with the available data to separately estimate predictive effects due to student age and instruction mode.
As described in the methods section, we treated withdraw grades as fails, but we also noted that withdraws are more common in the online courses (Table 1), a pattern also seen in prior research [29, 30]. One possible cause for this difference within our particular dataset stems from differences in the academic schedule. All of the studied online courses are taught in abbreviated 7.5-week terms. As a result, students have a very short window during which to drop a course and avoid having it appear on their transcript. Beyond that time, students must either complete the course for a conventional letter grade or withdraw. Other possible causes include differences in academic advising or differences in academic preparation. To explore the degree to which the treatment of withdraws alters our findings, we ran alternative models corresponding to models A–Q on a dataset with withdraws excluded instead of counted as failures (see S2 File). These models show reduced, but still significant and large, negative online effects. The models also differ somewhat in the magnitude of the demographic effects. These differences can be explained by the overall greater withdraw percentage in the online population and how withdraw percentages vary among the demographic groups studied (Table 1). Thus, the differences in withdrawal rates imparts some uncertainty to our results, but we do not feel it detracts from our overall conclusions. This analysis does highlight the complexity inherent in studying academic outcomes and the importance of exploring multiple outcome measures, where appropriate.
Another limitation of our study is that we did not disaggregate BLNP students due to low sample sizes, even though people with these racial/ethnic identities have vastly different histories and experiences in the US. However, these groups share the experience of being underrepresented in STEM in the US, which is the specific context we examined in our study and was our rationale for aggregating them and comparing them to white students. Finally, we acknowledge that Pell grant eligibility is a crude measure of socioeconomic status and does not allow us to gain a detailed understanding of the effects of social class on student experiences in STEM courses. However, it was the only indicator of socioeconomic status we could access from the university registrar. We encourage future studies to survey and interview low-income students to further probe their experiences and factors that may influence their experience in online degree programs.
Future directions
There are a number of points within this discussion that would benefit from additional, targeted research. Because this study used only administrative data, it was not possible to speak to or control for the many ways that a student’s life circumstances, resources they could access, academic opportunities, or personal motivations could impact course grades. Similarly, institutional structures and traditions that shape individual instructors’ course and assessment design could contribute to the patterns we have observed here. Our previous work in one in-person biology course showed that cognitive difficulty and format of exams were associated with gender and socioeconomic grade disparities [43]. Follow-up research should be undertaken that examines the effects of course design and assessment practices on grade gaps systematically or even experimentally while maintaining the focus on fully online degree programs. Whereas a number of studies exist that examine online learning broadly, fully online degree programs represent a new and important area of growth—one with potential to reduce historical inequities in STEM and higher education more broadly, but only if the institutional deficits presented here are addressed.
Conclusions
Our results show that a fully online biology degree program increases access to college biology degree programs for women, first-generation students, and students eligible to receive federal Pell grants. However, it does not increase access for BLNP students, i.e. students that identify as Black/African American, Latinx/Hispanic, Native American/Native Alaskan or Pacific Islander/Native Hawaiian. Our results also show that gender, racial and socioeconomic disparities in grades received by students in science courses persist in both in-person and online courses. More specifically in our main analyses, BLNP students, first generation students, and students eligible for Pell grants, receive lower grades than white, continuing generation, and students not eligible for Pell grants, in both online and in-person modalities. Women receive lower grades than men in online courses, but not as much in in-person courses. However, for first-generation college students, particularly first-generation women, online instruction is associated with a complete reduction in the grade gap seen in the context of in-person instruction.
We hope that these results will encourage a continued effort to address the outcome inequalities that clearly exist for students with different social identities in both in-person and online programs. All colleges and universities should strive to provide inclusive excellence and access is only the first step. While online instruction can mitigate some of these access issues, there are still inequities online. Further, our alternative modelling shows how adjusting for prior academic performance may minimize the true size of outcome inequalities. In particular, we observe a much larger grade gap for women in online programs in this alternative model. Although this gap is probably driven by pre-college differences, it is important to consider what colleges and universities could do to help reduce the impact of pre-college differences themselves.
As more degree programs are moved online and as these programs continue to attract women, first-generation students, and students from low-income households, any disadvantage felt by online students will disproportionately disadvantage these students. Worse yet, this is potentially a hidden problem to these students if the programs are advertised as being truly equivalent or if they are being presented as ways to close opportunity gaps. To fully reach their potential in making college degrees attainable by all qualified students, online programs should carefully monitor such disparities in academic outcomes and be proactive in working to eliminate them.
Supporting information
Data Availability
All relevant data are within the manuscript and its Supporting Information files.
Funding Statement
This work was supported by grant #GT11046 from the Howard Hughes Medical Institute (www.hhmi.org), awarded to JPC, SEB, PL, and ADA and grant #1711272 from the National Science Foundation (www.nsf.gov), awarded to SEB. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Protopsaltis S, Baum S. Does online education live up to its promise? A look at the evidence and implications for federal policy. 2019. Available: https://mason.gmu.edu/~sprotops/OnlineEd.pdf [Google Scholar]
- 2.Ginder SA, Kelly-Reid JE, Mann FB. Enrollment and employees in Postsecondary institutions, Fall 2017; and financial statistics and academic libraries, Fiscal year 2017: First look (Provisional data). National Center for Education Statistics; 2019. January Report No.: NCES 2019-021Rev. Available: https://nces.ed.gov/pubsearch/pubsinfo.asp?pubid=2019021REV [Google Scholar]
- 3.Varty AK. Options for online undergraduate courses in biology at American colleges and universities. LSE. 2016;15: ar58 10.1187/cbe.16-01-0075 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Tieken MC, Auldridge-Reveles TR. Rethinking the School Closure Research: School Closure as Spatial Injustice. Review of Educational Research. 2019;89: 917–953. 10.3102/0034654319877151 [DOI] [Google Scholar]
- 5.Kolak M, Bradley M, Block DR, Pool L, Garg G, Toman CK, et al. Urban foodscape trends: Disparities in healthy food access in Chicago, 2007–2014. Health & Place. 2018;52: 231–239. 10.1016/j.healthplace.2018.06.003 [DOI] [PubMed] [Google Scholar]
- 6.Soja EW. Seeking Spatial Justice. U of Minnesota Press; 2013. [Google Scholar]
- 7.Hillman N, Weichman T. Education deserts: The continued significance of “place” in the twenty-first century. Washington, DC: American Council on Education; 2016. Available: http://hdl.voced.edu.au/10707/408013 [Google Scholar]
- 8.Rosenboom V, Blagg K. Disconnected from higher education: how geography and internet speed limit access to higher education. Urban Institute; 2018. Jan. Available: https://vtechworks.lib.vt.edu/handle/10919/89125 [Google Scholar]
- 9.Hillman NW. Geography of college opportunity: the case of education deserts. American Educational Research Journal. 2016;53: 987–1021. 10.3102/0002831216653204 [DOI] [Google Scholar]
- 10.Bailey A, Vaduganathan N, Henry T, Laverdiere R, Pugliese L. Making digital learning work—success strategies from six leading universities and community colleges. Boston, MA: Boston Consulting Group; 2018. Available: https://www.voced.edu.au/content/ngv:80989 [Google Scholar]
- 11.Hart CMD, Friedmann E, Hill M. Online course-taking and student outcomes in California community colleges. Education Finance and Policy. 2018;13: 42–71. 10.1162/edfp_a_00218 [DOI] [Google Scholar]
- 12.Bettinger EP, Fox L, Loeb S, Taylor ES. Virtual classrooms: How online college courses affect student success. American Economic Review. 2017;107: 2855–2875. 10.1257/aer.20151193 [DOI] [Google Scholar]
- 13.Ashby J, Sadera WA, McNary SW. Comparing student success between developmental math courses offered online, blended, and face-to-face. Journal of Interactive Online Learning. 2011;10: 128–140. [Google Scholar]
- 14.Cavanaugh JK, Jacquemin SJ. A large sample comparison of grade based student learning outcomes in online vs. face-to-face courses. Online Learning. 2015;19 Available: https://eric.ed.gov/?id=EJ1062940 [Google Scholar]
- 15.Cooper KM, Gin LE, Brownell SE. Diagnosing differences in what Introductory Biology students in a fully online and an in-person biology degree program know and do regarding medical school admission. Advances in Physiology Education. 2019;43: 221–232. 10.1152/advan.00028.2019 [DOI] [PubMed] [Google Scholar]
- 16.PCAST. Engage to excel: producing one million additional college graduates with degrees in Science, Technology, Engineering, and Mathematics. Report to the President. Executive Office of the President. Executive Office of the President; 2012. Available: https://eric.ed.gov/?id=ED541511 [Google Scholar]
- 17.Morrison JL. US higher education in transition. On the Horizon. 2003;11: 6–10. 10.1108/10748120310698307 [DOI] [Google Scholar]
- 18.Xu D, Xu Y. The promises and limits of online higher education: understanding how distance education affects access, cost, and quality. American Enterprise Institute; 2019. Available: https://eric.ed.gov/?id=ED596296 [Google Scholar]
- 19.Flowers L. Online courses can be used to boost minority numbers in STEM fields. Diverse Issues in Higher Education; Fairfax. 8 December 201128: 19. [Google Scholar]
- 20.Drew JC, Galindo-Gonzalez S, Ardissone AN, Triplett EW. Broadening participation of women and underrepresented minorities in STEM through a hybrid online transfer program. LSE. 2016;15: ar50 10.1187/cbe.16-01-0065 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Farooq F, Mogayzel PJ, Lanzkron S, Haywood C, Strouse JJ. Comparison of US federal and foundation funding of research for Sickle cell disease and cystic fibrosis and factors associated with research productivity. JAMA Netw Open. 2020;3: e201737–e201737. 10.1001/jamanetworkopen.2020.1737 [DOI] [PubMed] [Google Scholar]
- 22.Washington HA. Medical apartheid: the dark history of medical experimentation on Black Americans from colonial times to the present. Doubleday; 2006. [Google Scholar]
- 23.Intemann K. Why diversity matters: understanding and applying the diversity component of the National Science Foundation’s broader impacts criterion. Social Epistemology. 2009;23: 249–266. 10.1080/02691720903364134 [DOI] [Google Scholar]
- 24.Harding S. “Strong Objectivity”: A response to the new objectivity question. Synthese. 1995;104: 331–349. [Google Scholar]
- 25.Niu L. Family socioeconomic status and choice of stem major in college: An analysis of a national sample. College Student Journal. 2017;51: 298–312. [Google Scholar]
- 26.National Science Foundation, National Center for Science and Engineering Statistics. Women, minorities, and persons with disabilities in science and engineering: 2019. Alexandria, VA: National Science Foundation; 2019. Report No.: Special Report NSF 19–304. Available: https://www.nsf.gov/statistics/wmpd [Google Scholar]
- 27.Wladis C, Hachey AC, Conway KM. The representation of minority, female, and non-traditional STEM majors in the online environment at community colleges: A nationally representative study. Community College Review. 2015;43: 89–114. 10.1177/0091552114555904 [DOI] [Google Scholar]
- 28.Wladis C, Hachey AC, Conway K. Which STEM majors enroll in online courses, and why should we care? The impact of ethnicity, gender, and non-traditional student characteristics. Computers & Education. 2015;87: 285–308. 10.1016/j.compedu.2015.06.010 [DOI] [Google Scholar]
- 29.Jaggars SS. Online learning: Does it help low-income and underprepared students? (Assessment of evidence series). 2011. [cited 16 Dec 2019]. Available: https://ccrc.tc.columbia.edu/publications/online-learning-low-income-underprepared.html [Google Scholar]
- 30.Wladis C, Conway KM, Hachey AC. The online STEM classroom—who succeeds? An exploration of the impact of ethnicity, gender, and non-traditional student characteristics in the community college context. Community College Review. 2015;43: 142–164. 10.1177/0091552115571729 [DOI] [Google Scholar]
- 31.Jordan K. Massive open online course completion rates revisited: assessment, length and attrition. irrodl. 2015;16: 341–358. 10.19173/irrodl.v16i3.2112 [DOI] [Google Scholar]
- 32.Paul J, Jefferson F. A comparative analysis of student performance in an online vs. face-to-face environmental science course from 2009 to 2016. Front Comput Sci. 2019;1 10.3389/fcomp.2019.00011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Figlio D, Rush M, Yin L. Is it live or is it internet? Experimental estimates of the effects of online instruction on student learning. Journal of Labor Economics. 2013;31: 763–784. 10.1086/669930 [DOI] [Google Scholar]
- 34.Alpert WT, Couch KA, Harmon OR. A randomized assessment of online learning. American Economic Review. 2016;106: 378–382. 10.1257/aer.p20161057 [DOI] [Google Scholar]
- 35.Xu D, Jaggars SS. Performance gaps between online and face-to-face courses: differences across types of students and academic subject areas. The Journal of Higher Education. 2014;85: 633–659. 10.1353/jhe.2014.0028 [DOI] [Google Scholar]
- 36.Johnson H, Mejia MC. Online learning and student outcomes in community colleges. 2014. Available: https://www.ppic.org/publication/online-learning-and-student-outcomes-in-community-colleges/ [Google Scholar]
- 37.Kaupp R. Online penalty: the impact of online instruction on the Latino-white achievement gap. Journal of Applied Research in the Community College. 2012;19: 8–16. [Google Scholar]
- 38.Xu D, Jaggars SS. The impact of online learning on students’ course outcomes: Evidence from a large community and technical college system. Economics of Education Review. 2013;37: 46–57. 10.1016/j.econedurev.2013.08.001 [DOI] [Google Scholar]
- 39.Bailey T, Jenkins D, Leinbach T. What we know about community college low-income and minority student outcomes: descriptive statistics from national surveys. Community College Research Center. 2005. Available: https://eric.ed.gov/?id=ED484354 [Google Scholar]
- 40.Wladis C, Hachey AC, Conway K. Are online students in STEM (science, Technology, Engineering and Mathematics) courses at greater risk of non-success? American Journal of Educational Studies; Cary. 2013;6: 65–84. [Google Scholar]
- 41.Matz RL, Koester BP, Fiorini S, Grom G, Shepard L, Stangor CG, et al. Patterns of gendered performance differences in large introductory courses at five research universities. AERA Open. 2017;3: 2332858417743754 10.1177/2332858417743754 [DOI] [Google Scholar]
- 42.Eddy SL, Brownell SE. Beneath the numbers: A review of gender disparities in undergraduate education across science, technology, engineering, and math disciplines. Phys Rev Phys Educ Res. 2016;12: 020106 10.1103/PhysRevPhysEducRes.12.020106 [DOI] [Google Scholar]
- 43.Wright CD, Eddy SL, Wenderoth MP, Abshire E, Blankenbiller M, Brownell SE. Cognitive Difficulty and Format of Exams Predicts Gender and Socioeconomic Gaps in Exam Performance of Students in Introductory Biology Courses. LSE. 2016;15: ar23 10.1187/cbe.15-12-0246 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Theobald EJ, Hill MJ, Tran E, Agrawal S, Arroyo EN, Behling S, et al. Active learning narrows achievement gaps for underrepresented students in undergraduate science, technology, engineering, and math. PNAS. 2020;117: 6476–6483. 10.1073/pnas.1916903117 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Harackiewicz JM, Canning EA, Tibbetts Y, Priniski SJ, Hyde JS. Closing achievement gaps with a utility-value intervention: Disentangling race and social class. Journal of Personality and Social Psychology. 2016;111: 745–765. 10.1037/pspp0000075 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Eddy SL, Hogan KA. Getting Under the Hood: How and for Whom Does Increasing Course Structure Work? CBE Life Sci Educ. 2014;13: 453–468. 10.1187/cbe.14-03-0050 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Eddy SL, Brownell SE, Wenderoth MP. Gender gaps in achievement and participation in multiple Introductory Biology classrooms. LSE. 2014;13: 478–492. 10.1187/cbe.13-10-0204 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Perera V, Mead C, Buxner S, Lopatto D, Horodyskyj L, Semken S, et al. Students in fully online programs report more positive attitudes toward science than students in traditional, in-person programs. LSE. 2017;16: ar60 10.1187/cbe.16-11-0316 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Carey RL. A Cultural Analysis of the Achievement Gap Discourse: Challenging the Language and Labels Used in the Work of School Reform. Urban Education. 2014;49: 440–468. 10.1177/0042085913507459 [DOI] [Google Scholar]
- 50.Crenshaw K. Demarginalizing the Intersection of Race and Sex: A Black Feminist Critique of Antidiscrimination Doctrine, Feminist Theory and Antiracist Politics. U Chi Legal F. 1989;1989: 139. [Google Scholar]
- 51.Combahee River Collective. A Black Feminist Statement. 1977. [Google Scholar]
- 52.Collins PH. Learning from the Outsider Within: The Sociological Significance of Black Feminist Thought. Soc Probl. 1986;33: s14–s32. 10.2307/800672 [DOI] [Google Scholar]
- 53.Davis AY. Women, Race, & Class. New York: Random House; 1981. [Google Scholar]
- 54.Nix S, Perez-Felkner L. Difficulty orientations, gender, and race/ethnicity: an intersectional analysis of pathways to STEM degrees. Social Sciences. 2019;8: 1–29. [Google Scholar]
- 55.López N, Erwin C, Binder M, Chavez MJ. Making the invisible visible: advancing quantitative methods in higher education using critical race theory and intersectionality. Race Ethnicity and Education. 2018;21: 180–207. 10.1080/13613324.2017.1375185 [DOI] [Google Scholar]
- 56.Alon S. Overlapping disadvantages and the racial/ethnic graduation gap among students attending selective institutions. Social Science Research. 2007;36: 1475–1499. 10.1016/j.ssresearch.2007.01.006 [DOI] [Google Scholar]
- 57.Dusen BV, Nissen J. Equity in college physics student learning: A critical quantitative intersectionality investigation. Journal of Research in Science Teaching. 2020;57: 33–57. 10.1002/tea.21584 [DOI] [Google Scholar]
- 58.Wright C, Huang A, Cooper K, Brownell S. Exploring Differences In Decisions About Exams Among Instructors Of The Same Introductory Biology Course. International Journal for the Scholarship of Teaching and Learning. 2018;12 10.20429/ijsotl.2018.120214 [DOI] [Google Scholar]
- 59.Huberth M, Chen P, Tritz J, McKay TA. Computer-tailored student support in Introductory Physics. PLOS ONE. 2015;10: e0137001 10.1371/journal.pone.0137001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Bensimon EM. The misbegotten URM as a data point. Los Angeles, CA: Center for Urban Education, Rossier School of Education, University of Southern California; 2016. [Google Scholar]
- 61.Williams TL. “Underrepresented minority” considered harmful, racist language. 2020. [cited 15 Jul 2020]. Available: https://cacm.acm.org/blogs/blog-cacm/245710-underrepresented-minority-considered-harmful-racist-language/fulltext [Google Scholar]
- 62.The New York Times. Uppercasing ‘Black.’ In: The New York Times Company [Internet]. 30 Jun 2020 [cited 15 Jul 2020]. Available: https://www.nytco.com/press/uppercasing-black/
- 63.Teranishi RT. Asian Pacific Americans and critical race theory: an examination of school racial climate. Equity & Excellence in Education. 2002;35: 144–154. 10.1080/713845281 [DOI] [Google Scholar]
- 64.Enders CK, Tofighi D. Centering predictor variables in cross-sectional multilevel models: a new look at an old issue. Psychological methods. 2007;12: 121–138. 10.1037/1082-989X.12.2.121 [DOI] [PubMed] [Google Scholar]
- 65.Nakagawa S, Johnson PCD, Schielzeth H. The coefficient of determination R2 and intra-class correlation coefficient from generalized linear mixed-effects models revisited and expanded. Journal of The Royal Society Interface. 2017;14: 20170213 10.1098/rsif.2017.0213 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.R Core Team. R: A language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing; 2019. Available: http://www.R-project.org [Google Scholar]
- 67.Bates D, Maechler M, Bolker BM, Walker S. Fitting linear mixed-effects models using lme4. Journal of Statistical Software. 2015;67: 1–48. 10.18637/jss.v067.i01 [DOI] [Google Scholar]
- 68.Kuznetsova A, Brockhoff PB, Christensen RHB. lmerTest package: tests in linear mixed effects models. Journal of Statistical Software. 2017;82: 1–26. 10.18637/jss.v082.i13 [DOI] [Google Scholar]
- 69.Bartoń K. MuMIn: multi-model inference. 2019. [Google Scholar]
- 70.Lüdecke D, Makowski D, Waggoner P, Patil I. Performance: assessment of regression models performance. 2020. Available: https://easystats.github.io/performance/ [Google Scholar]
- 71.Lüdecke D. sjPlot: Data visualization for statistics in Social Science. 2020. Available: https://CRAN.R-project.org/package=sjPlot [Google Scholar]
- 72.Spencer SJ, Steele CM, Quinn DM. Stereotype threat and women’s math performance. Journal of Experimental Social Psychology. 1999;35: 4–28. 10.1006/jesp.1998.1373 [DOI] [Google Scholar]
- 73.Steele CM, Aronson J. Stereotype threat and the intellectual test performance of African Americans. Journal of personality and social psychology. 1995;69: 797–811. 10.1037//0022-3514.69.5.797 [DOI] [PubMed] [Google Scholar]
- 74.Nguyen H-HD, Ryan AM. Does stereotype threat affect test performance of minorities and women? A meta-analysis of experimental evidence. Journal of Applied Psychology. 2008;93: 1314–1334. 10.1037/a0012702 [DOI] [PubMed] [Google Scholar]
- 75.Cotner S, Ballen CJ. Can mixed assessment methods make biology classes more equitable? PLOS ONE. 2017;12: e0189610 10.1371/journal.pone.0189610 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Eddy SL, Brownell SE, Thummaphan P, Lan M-C, Wenderoth MP. Caution, Student Experience May Vary: Social Identities Impact a Student’s Experience in Peer Discussions. LSE. 2015;14: ar45 10.1187/cbe.15-05-0108 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Ballen CJ, Aguillon SM, Awwad A, Bjune AE, Challou D, Drake AG, et al. Smaller classes promote equitable student participation in STEM. BioScience. 2019;69: 669–680. 10.1093/biosci/biz069 [DOI] [Google Scholar]
- 78.Cooper KM, Krieg A, Brownell SE. Who perceives they are smarter? Exploring the influence of student characteristics on student academic self-concept in physiology. Advances in Physiology Education. 2018;42: 200–208. 10.1152/advan.00085.2017 [DOI] [PubMed] [Google Scholar]
- 79.Caspi A, Chajut E, Saporta K. Participation in class and in online discussions: Gender differences. Computers & Education. 2008;50: 718–724. 10.1016/j.compedu.2006.08.003 [DOI] [Google Scholar]
- 80.Morante A, Djenidi V, Clark H, West S. Gender differences in online participation: examining a History and a Mathematics open foundation online course. Australian Journal of Adult Learning. 2017;57: 266–293. [Google Scholar]
- 81.Conger D, Long MC. Why are men falling behind? Gender gaps in college performance and persistence. The ANNALS of the American Academy of Political and Social Science. 2010;627: 184–214. 10.1177/0002716209348751 [DOI] [Google Scholar]
- 82.Harackiewicz JM, Canning EA, Tibbetts Y, Giffen CJ, Blair SS, Rouse DI, et al. Closing the social class achievement gap for first-generation students in undergraduate biology. Journal of Educational Psychology. 2014;106: 375–389. 10.1037/a0034679 [DOI] [PMC free article] [PubMed] [Google Scholar]