Abstract
Assessing learning across a biology major can help departments monitor achievement of broader program-level goals and identify opportunities for curricular improvement. However, biology departments have lacked suitable tools to measure learning at the program scale. To address this need, we developed four freely available assessments—called Biology-Measuring Achievement and Progression in Science or Bio-MAPS—for general biology, molecular biology, ecology/evolution, and physiology programs. When administered at multiple time points in a curriculum, these instruments can provide departments with information on how student conceptual understanding changes across a major and help guide curricular modifications to enhance learning.
INTRODUCTION
Program-level assessments can be important tools to measure student learning during undergraduate majors. Assessing students at multiple time points throughout the curriculum can motivate faculty to discuss and come to agreement on the essential learning outcomes of their program and consider how students will achieve these outcomes regardless of the specific courses they take. By providing information on student understanding at key time points, program-level assessments can also assist departments in determining the cumulative impact of their courses and pinpointing areas for improvement.
At a broader level, program assessment has been propelled by a national interest in ensuring that college participation leads to measurable learning outcomes. Accreditation agencies are increasingly requiring that programs collect and respond to assessment data (1). A survey of chief academic officers revealed that regional and program accreditation represent the primary drivers for program assessment, even outranking departmental and institutional commitment to improvement (2). In light of rising tuition costs, colleges and universities are also facing increasing pressure to document impacts on students, while accounting for incoming attributes and abilities (3, 4). Currently, programs use a variety of approaches to gauge student outcomes and respond to questions regarding the added value of a college degree. These approaches include capstone projects, learning portfolios, and performance assessments. To ensure that the approaches and assessment results have meaning, it is important for departments to involve faculty throughout the assessment process, including in decisions on what to assess, which approaches would provide adequate evidence of student learning, and how student performance aligns with broader disciplinary standards (5).
To conduct program assessment within a discipline, departments must first identify the central goals of their program. Program-level learning goals specify what students should be able to do at the end of a major. Although crafting and reaching consensus on these goals takes time, biology departments can capitalize on a growing pool of available guides. For example, the Vision and Change report articulates core concepts and competencies for undergraduate biology programs (6), and the BioCore Guide (7) and Conceptual Elements Framework (8) further delineate how the core concepts apply within biology. CourseSource also houses extensive lists of learning goals for biology subdisciplines developed by multiple life science professional societies, including the American Society for Microbiology (click on the relevant course at https://www.coursesource.org/courses).
Assessing student achievement requires that departments have access to instruments that align with their learning goals. Since departments typically have limited financial and personnel resources for program assessment, these instruments ideally would be freely available and easily administered and scored. Until this point, biology departments have had few options for valid and reliable tools to measure advancement toward program-level conceptual learning goals. Most tools have measured student understanding of a specific topic, and few instruments gauge student understanding across the breadth of a major. Although general biology assessments have been developed commercially (e.g., the Biology Majors Field Test), these tests are usually administered only at the end of a degree program and have costs that must be covered by either the student or institution.
An illustrative example: Using program assessment to promote transformation in microbiology education
Within the literature, one prime example stands out in which instructors used program assessment as a mechanism to build and refine their undergraduate program. In this case, biologists who taught related microbiology courses at the same institution formed a teaching group to coordinate their curriculum and improve undergraduate learning. Working together, this group identified content linkages and overarching learning goals, chose specific microorganisms to serve as recurring model systems, and implemented active learning methods across their courses (9). Building on their learning goals, they developed a response-validated assessment focused on host–pathogen interactions and administered this assessment within introductory and advanced courses (10). Students demonstrated growth within their introductory course, maintained this performance over time, and showed further improvements within some advanced courses. The authors proposed using the varied student performance in the advanced courses to make data-driven changes to content and instructional practices. Importantly, assessment data enabled this group to refine their expectations for student learning, reflect on how curriculum coverage compared with actual performance, and understand how the curriculum served particular student groups (11).
New tools for program assessment: Bio-MAPS
While this example provides a roadmap for departments to follow, several steps in the process, such as generating an assessment, can be time consuming and require specialized expertise. To help alleviate issues associated with designing program assessments, we developed a suite of freely available instruments called Biology-Measuring Achievement and Progression in Science or Bio-MAPS. These assessments align with the core concepts from Vision and Change (6, 7), with separate instruments for general biology (12), molecular biology (13), physiology (14), and ecology/evolution (15). We followed established methods of assessment development (16), including optimizing response validity through student interviews, ensuring technical accuracy through expert reviews, pilot testing at a variety of institutions across the country, and addressing any potential item biases favoring particular demographic groups. Each question stem presents a scenario, followed by a series of statements that students evaluate as “true/false” or “likely/unlikely to be true” (Fig. 1). These formats allow the rapid collection of fine-grained information on student understanding of many concepts, while retaining the convenience of automatic grading (17, 18).
FIGURE 1.
Example question from the GenBio-MAPS assessment (12). The question stem presents drawings of chromosomes, gene loci, and alleles. Students answer true/false statements related to the question stem. Correct answers are highlighted in green. Note that any number of statements can be true or false for different question stems.
To support faculty in using these assessments for large-scale evaluations, we also developed a web portal to facilitate Bio-MAPS administration (http://cperl.lassp.cornell.edu/biomaps), building on existing portals in physics education research (19). Through this portal, instructors or assessment coordinators fill out a brief survey and then receive a web link that they share with their students. Students complete the assessment online and outside of class, eliminating the need to devote class time to administering the assessment. This approach has been found to produce scores similar to in-class administrations (20). Because the instruments are typically administered within the context of individual courses, faculty can offer a few points of participation credit, which helps motivate student participation. After students complete the assessment and access has closed, the portal returns summary reports to the instructor or assessment coordinator, providing aggregated information on overall student performance and performance for each of the Vision and Change core concepts (Fig. 2).
FIGURE 2.
Example score report from students at one institution who took the GenBio-MAPS assessment just before graduation (advanced time point, n=127 students). Central bars represent median test scores, boxes represent inner quartiles, and whiskers represent minimum/maximum scores. Each dot represents one student’s average percent correct for all true/false items aligned with the indicated Vision and Change core concept.
Bio-MAPS assessments are specifically designed to measure student conceptual understanding at key time points across a major (Fig. 3); they are not intended to be given as pre/post tests around a single course. These time points typically include beginning of introductory biology: the start of the program (e.g., prior to classes or at the beginning of the first course in the major), end of introductory biology: after the completion of an introductory course series, and advanced: at the end of courses typically taken by seniors just prior to graduation. Within our multi-institution data sets, we find overall increases in student performance over time for each Bio-MAPS instrument (12, 14, 15). Students demonstrated large gains for some concepts, while other concepts remained challenging, even for students about to graduate. Importantly, we showed that student performance trajectories differ across institutions, indicating that these assessments can detect program-specific strengths and limitations.
FIGURE 3.
Example assessment administration timeline based on two-semester academic calendar. Prior to adopting an assessment, departments should discuss their program goals and come to agreement on the use of a particular instrument. Bio-MAPS instruments can be administered through the web portal (A) at the beginning of the fall semester to students who are starting the intro series and incoming transfer students, (B) at the end of the fall semester to students who are finishing the intro series and graduating seniors, (C) at the beginning of the spring semester to students who are starting the intro course series and incoming transfer students, and (D) at the end of the spring semester to students who are finishing the intro course series and graduating seniors. After administration, the score reports automatically generated through the web portal can be reviewed. Throughout the academic year, faculty can meet to discuss data-driven adjustments to their program.
Using program assessment data to facilitate change
After collecting and processing assessment data, departments can compare performance on individual items over time to diagnose student understanding of specific concepts. By comparing longitudinal or cross-sectional data across time points (example shown in Fig. 4), departments can identify informative trends in student performance, such as when a particular concept shows high incoming performance, little growth during a time period, or high achievement at the end of advanced courses. In addition, departments can administer a single instrument across all time points, or they can administer different instruments at different time points to gauge mastery of advanced subdisciplinary concepts. For example, in programs where students take an introductory course series and then specialize into subdisciplinary tracks, faculty may choose to administer the general biology instrument at the first two time points and the subdiscipline-focused instruments at the last time point to students in the corresponding tracks. Furthermore, departments can incorporate more focused assessments within specific courses as needed to achieve fine-scale information on student learning and progress (21).
FIGURE 4.
Student performance at different time points on two of the GenBio-MAPS true/false statements shown in Figure 1 (n=137 beginning of intro, 176 end of intro, and 127 advanced students from one institution). Students show increased performance on Statement a, which assesses the concept that genes located close together on a chromosome are linked and will tend to be inherited together more frequently than genes located far apart on chromosomes or on different chromosomes. In contrast, students show little growth on Statement d, which assesses understanding of the differences between sex chromosomes and the location of genes involved in the development of sex organs.
Program-level assessments support a data-driven approach to helping departments advance their educational missions, rather than relying on the perceptions of individual faculty or anecdotal feedback from a small group of students. Capturing data at the program level over multiple time points encourages faculty to engage in a broad dialogue about overall goals and solutions rather than limiting discussions to the single-course level. For concepts showing little improvement over time, faculty can discuss whether and how current courses address these concepts and consider structural, curricular, or pedagogical changes to improve learning. For example, information about student growth on evolution concepts inspired one department to start requiring that all students take an evolution course at the sophomore level (15). For particularly challenging concepts, faculty can consult the literature to find additional studies and classroom lessons to incorporate into individual courses. In cases where students in a particular curricular track (e.g., students with sub-specializations or taking particular courses) demonstrate low performance, it is helpful for departments to keep the conversation at the curriculum level and focus on collective program goals rather than individual courses.
CONCLUSIONS
The promise of program-level assessment lies in its ability to help departments identify central goals, monitor student learning, and chart a course for program improvement. These assessments also enable departments to determine the impact of subsequent interventions, such as changing the course sequence, introducing clicker questions with peer discussion, or developing online courses. When combined with demographic information, these data allow departments to determine whether their programs adequately serve students from particular groups, including community college transfer students, students from historically underrepresented groups, or students participating in summer bridge programs. Access to assessment data may also prompt departments to explore new questions, such as how background preparation influences student performance and what additional resources faculty and students need to achieve success. Program assessment ultimately has the potential to improve student retention by helping departments understand how students with different backgrounds progress through their major, determine how conceptual understanding at each time point relates to persistence, and make targeted changes to help boost student achievement. Bio-MAPS results disaggregated by demographic data can also inspire departmental discussions about the importance of ensuring that curricular decisions promote inclusive excellence and help provide opportunities for all students to maximize their learning regardless of background (22, 23). While the impact of college cannot be fully captured by a single instrument, Bio-MAPS assessments enable departments to measure student learning of core concepts across key time points and make data-driven decisions about their undergraduate curriculum.
ACKNOWLEDGMENTS
This work was supported by a collaborative grant from the National Science Foundation (NSF) Transforming Undergraduate Education in Science, Technology, Engineering and Mathematics (TUES) program to JKK and BAC (DUE-1322364), MKS (DUE-1322556), and SEB and AJC (DUE-1323010). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. Example data collection was approved by the Institutional Review Board at the University of Nebraska-Lincoln (protocol 14618). Web portal development was classified as exempt by the Cornell Institutional Review Board (protocol 1808008199). The authors declare that there are no conflicts of interest.
REFERENCES
- 1.Beno BA. The role of student learning outcomes in accreditation quality review. New Dir Community Coll. 2004;126:65–72. [Google Scholar]
- 2.Gannon-Slater N, Ikenberry S, Jankowski N, Kuh G. Institutional assessment practices across accreditation regions. National Institute for Learning Outcomes Assessment; Urbana, IL: 2014. [Google Scholar]
- 3.Arum R, Roksa J. Academically adrift: limited learning on college campuses. University of Chicago Press; Chicago, IL: 2011. [Google Scholar]
- 4.Cunha JM, Miller T. Measuring value-added in higher education: possibilities and limitations in the use of administrative data. Econ Educ Rev. 2014;42:64–77. doi: 10.1016/j.econedurev.2014.06.001. [DOI] [Google Scholar]
- 5.Bok D. Higher education in America, Revised Edition. Princeton University Press; Princeton, NJ: 2018. [Google Scholar]
- 6.American Association for the Advancement of Science (AAAS) Vision and change in undergraduate biology education: a call to action: a summary of recommendations made at a national conference organized by the American Association for the Advancement of Science; July 15–17, 2009; Washington, DC. 2011. [Google Scholar]
- 7.Brownell SE, Freeman S, Wenderoth MP, Crowe AJ. BioCore guide: a tool for interpreting the core concepts of Vision and change for biology majors. CBE Life Sci Educ. 2014;13:200–211. doi: 10.1187/cbe.13-12-0233. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Cary T, Branchaw J. Conceptual elements: a detailed framework to support and assess student learning of biology core concepts. CBE Life Sci Educ. 2017;16(2):ar24. doi: 10.1187/cbe.16-10-0300. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Marbach-Ad G, Briken V, Frauwirth K, Gao L-Y, Hutcheson SW, Joseph SW, Mosser D, Parent B, Shields P, Song W, Stein DC, Swanson K, Thompson KV, Yuan R, Smith AC. A faculty team works to create content linkages among various courses to increase meaningful learning of targeted concepts of microbiology. CBE Life Sci Educ. 2007;6:155–162. doi: 10.1187/cbe.06-12-0212. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Marbach-Ad G, Briken V, El-Sayed NM, Frauwirth K, Fredericksen B, Hutcheson S, Gao LY, Joseph SW, Lee V, McIver KS, Mosser D, Booth Quimby B, Shields P, Song W, Stein DC, Yuan RT, Smith AC. Assessing student understanding of host pathogen interactions using a concept inventory. J Microbiol Biol Educ. 2009;10(1):43–50. doi: 10.1128/jmbe.v10.98. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Marbach-Ad G, McAdams KC, Benson S, Briken V, Cathcart L, Chase M, El-Sayed NM, Frauwirth K, Fredericksen B, Joseph SW, Lee V, McIver KS, Mosser D, Quimby BB, Shields P, Song W, Stein DC, Stewart R, Thompson KV, Smith AC. A model for using a concept inventory as a tool for students’ assessment and faculty professional development. CBE Life Sci Educ. 2010;9:408–416. doi: 10.1187/cbe.10-05-0069. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Couch BA, Wright CD, Freeman S, Knight JK, Semsar K, Smith MK, Summers MM, Zheng Y, Crowe AJ, Brownell SE. GenBio-MAPS: a programmatic assessment to measure student understanding of Vision and change core concepts across general biology programs. CBE Life Sci Educ. 2019;18:ar1. doi: 10.1187/cbe.18-07-0117. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Couch BA, Wood WB, Knight JK. The molecular biology capstone assessment: a concept assessment for upper-division molecular biology students. CBE Life Sci Educ. 2015;14:ar10. doi: 10.1187/cbe.14-04-0071. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Semsar K, Brownell S, Couch BA, Crowe AJ, Smith MK, Summers MM, Wright CD, Knight JK. Phys-MAPS: A programmatic physiology assessment for introductory and advanced undergraduates. Adv Physiol Educ. 2019;43:15–27. doi: 10.1152/advan.00128.2018. [DOI] [PubMed] [Google Scholar]
- 15.Summers MM, Couch BA, Knight JK, Brownell SE, Crowe AJ, Semsar K, Wright CD, Smith MK. EcoEvo-MAPS: an ecology and evolution assessment for introductory through advanced undergraduates. CBE Life Sci Educ. 2018;17:ar18. doi: 10.1187/cbe.17-02-0037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Adams WK, Wieman CE. Development and validation of instruments to measure learning of expert-like thinking. Int J Sci Educ. 2011;33:1289–1312. doi: 10.1080/09500693.2010.512369. [DOI] [Google Scholar]
- 17.Hubbard JK, Potts MA, Couch BA. How question types reveal student thinking: an experimental comparison of multiple-true-false and free-response formats. CBE Life Sci Educ. 2017;16:ar26. doi: 10.1187/cbe.16-12-0339. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Couch BA, Hubbard JK, Brassil CE. Multiple–true–false questions reveal the limits of the multiple–choice format for detecting students with incomplete understandings. BioScience. 2018;68:455–463. doi: 10.1093/biosci/biy037. [DOI] [Google Scholar]
- 19.Wilcox BR, Zwickl BM, Hobbs RD, Aiken JM, Welch NM, Lewandowski HJ. Alternative model for administration and analysis of research-based assessments. Phys Rev Phys Educ Res. 2016;12 doi: 10.1103/PhysRevPhysEducRes.12.010139. [DOI] [Google Scholar]
- 20.Couch BA, Knight JK. A comparison of two low-stakes methods for administering a program-level biology concept assessment. J Microbiol Biol Educ. 2015;16:178–185. doi: 10.1128/jmbe.v16i2.953. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Knight JK. Biology concept assessment tools: design and use. Microbiol Aust. 2010;31:5–8. [Google Scholar]
- 22.Asai DJ, Bauerle C. From HHMI: doubling down on diversity. CBE Life Sci Educ. 2016;15(3) doi: 10.1187/cbe.16-01-0018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Springer JT, Hatcher JA, Powell A. High-impact practices: the call for a commitment to quality educational experiences and inclusive excellence. Assess Update. 2018;30:6–11. doi: 10.1002/au.30141. [DOI] [Google Scholar]




