Skip to main content
Educational and Psychological Measurement logoLink to Educational and Psychological Measurement
. 2021 Feb 8;81(6):1221–1233. doi: 10.1177/0013164421992112

A Short Note on Optimizing Cost-Generalizability via a Machine-Learning Approach

Zhehan Jiang 1,, Dexin Shi 2, Christine Distefano 2
PMCID: PMC8451023  PMID: 34565822

Abstract

The costs of an objective structured clinical examination (OSCE) are of concern to health profession educators globally. As OSCEs are usually designed under generalizability theory (G-theory) framework, this article proposes a machine-learning-based approach to optimize the costs, while maintaining the minimum required generalizability coefficient, a reliability-like index in G-theory. The authors adopted G-theory parameters yielded from an OSCE hosted by a medical school, reproduced the generalizability coefficients to prepare for optimizing manipulations, applied simulated annealing algorithm to calculate the number of facet levels minimizing the associated costs, and conducted the analysis in various conditions via computer simulation. With a given generalizability coefficient, the proposed approach, virtually an instrument of decision-making supports, found the optimal solution for the OSCE such that the associated costs were minimized. The computer simulation results showed how the cost reductions varied with different levels of required generalizability coefficients. Machine learning–based approaches can be used in conjunction with psychometric modeling to help planning assessment tasks more scientifically. The proposed approach is easy to adopt into practice and customize in alignment with specific testing designs. While these results are encouraging, the possible pitfalls such as algorithmic convergences’ failure and inadequate cost assumptions should also be avoided.

Keywords: generalizability theory, cost, reliability, OSCE, optimization

Introduction

Assessment of students’ performance across the curriculum involves a variety of methods that address the assessment of knowledge, skills, and behaviors. Compared with other methods, an observation measure often yields more information. However, it possesses multiple sources of measurement errors, for example, rater effects, occasion effects, and item effects. Accommodating this type of assessment, generalizability theory (G-theory; Cronbach et al., 1972) provides a statistical solution for both decomposing the sources of errors on observation measures and indicating the reliability changes of the design manipulations; these two functions are called G-study and D-study in the literature (Brennan, 1992). To be specific, G-study yields parameter estimates of variances of all defined error sources and provides reliability-like indexes, such as for measurement quality evaluation. On the other hand, if a source of error (called facet) yielded by a G-study is small/large, the number of observations of the facet can be reduced/increased, where the consequences are calculated in a D-study (Brennan, 2010).

Approaches have been proposed to maximize reliability within a budget constraint. Woodward and Joe (1973) derived equations for their constrained optimization; Saunders et al. (1989) deployed discrete optimization, which was further updated by Saunders (1992) with the Cauchy–Schwartz inequity approach; Marcoulides (1993, 1995) as well as Marcoulides and Goldstein (1990, 1991, 1992) had developed the LaGrange multiplier approach and other related variants to handle the optimization within both univariate and multivariate G-theory. Meyer et al. (2014) extended the LaGrange multiplier approaches to G-theory with nested designs. Devising these approaches to a particular G-theory design requires mathematical deriving procedures, which may be a challenge to many applied researchers.

Different from the aforementioned studies, this article addresses the optimization problem from a different perspective: Given a minimal reliability value (namely a reliability threshold), what is the optimal combination of facets’ sample sizes for achieving the lowest costs? The inquiry’s context stems from practical needs in the field of medical education, which is demonstrated in the coming section. Instead of using the existing methods such as the LaGrange multiplier one, this article proposes a machine-learning approach called simulated annealing (SA) to handle the task of interest.

Objective Structured Clinical Examination

In the field of medical education, objective structured clinical examination (OSCE) is adapted to assess clinical skills so that the process can be more consistent and objective. Harden et al. (1975) proposed the original version of the OSCE—a checklist to assess students’ clinical skills using direct observation of their interactions with patients at multiple stations. Studies addressing the use of OSCEs are found to be tremendous in health professional education (Hastie et al., 2014; Hodges et al., 2014; Patrício et al., 2013; Setyonugroho et al., 2015; Smith et al., 2012; Walsh et al., 2009), and they together verify the necessity of using multiple stations in OSCEs, as opposed to the traditional long-case examination, which is not adequate to predict students’ performance in a different clinical situation (Hubbard, 1971). Accordingly, a typical OSCE design nowadays is likely to be based on a large sample of clinical cases and longer testing time, ensuring a satisfactory reliability level.

In addition to reliability, the costs of organizing OSCEs make up another concern. Compared with other assessment forms, an OSCE tends to be more resource-intensive: It involves examinees rotating round multiple standardized stations responding to test items or prompts, with or without the standardized patient situation, and raters scoring examinee performance against some predefined rubrics (Cusimano et al., 1994). Correspondingly, OSCE providers are required to make a budget for equipment, accommodations, subsistence, standardized patients, staff payrolls, the fee for content experts/consultants, and others (Brown et al., 2015). As a result, vast investments in OSCEs across countries are seen each year (Walsh & Jaye, 2013).

The trade-off between the costs and the reliability of OSCEs can be regarded as a reflection of cost-effectiveness in medical education assessment. Although numerous studies investigate either of the topics, addressing the two components together is rarely found in the literature. The fiscal constraints due to COVID-19 and the merging calls for raising transparency in medical education strengthen the practice of ensuring that the funding spent on medical education should be as cost-reliable as possible (Dacre & Walsh, 2013).

Generalizability Theory

A trustable examination demands certain statistical frameworks to provide a scientific understanding of the measurement quality, as it is believed that there are errors (i.e., variance off the interest) occurring simultaneously during the measurement process, which results in a discrepancy between the raw scores and the examinees’ true skill levels. The errors are called measurement errors, mathematically introducing construct-irrelevant turbulence into the analysis, posing a threat to the interpretation of the scores (Downing & Haladyna, 2004; Messick, 1998).

To address the measurement errors, classical test theory (CTT) is often adopted to yield the reliability, which is defined as a correlation between the true skill levels and the test scores. That said, the estimation of reliability, referring to the reproducibility of assessment data over time or occasions, is a pathway to ensure appropriate validity (Downing, 2003). Cronbach’s α is known as a CTT reliability estimate of a given test: It indicates the proportion of variance in test scores that can be attributed to true score variance (Cronbach, 1951). The measurement errors in CTT, however, oversimplify research designs in many situations as it only assumes two components for a test score (i.e., true score and errors). To overcome the shortcoming, CTT and α were further extended to G-theory (Brennan, 1992, 2010). There are ways of estimating G-theory, for example, the mean square method (e.g., Cornfield & Tukey, 1956; Henderson, 1953; Rao, 1970), the mixed model method (e.g., Huebner & Lucht, 2019; Jiang, 2018; Jiang et al., 2020), and the Bayesian method (e.g., Jiang & Skorupski, 2018; LoPilato et al., 2015)

Different from traditional tests such as multiple-choice questions and essays, OSCEs contain more sources of variance contributing to the measurement process. The variance can therefore be decomposed to more specific facets, for example, the effects of raters, items/questions, stations, and other sources of variation over which generalization is desired (Boulet et al., 2003; Myford & Wolfe, 2003). Therefore, a critical element in OSCE quality control is using a reliability index that takes all relevant measurement errors into the analysis (Haladyna & Downing, 2004). Naturally compatible with multiple-facet situations, G-theory provides an appropriate solution to OSCE studies. Articles addressing both summative and formative OSCEs through G-theory are plentiful in the literature (Baig & Violato, 2012; Donnon & Paolucci, 2008; Newble & Swanson, 1988), and the span of the reported reliability level ranges from .12 to .85.

An OSCE Within G-Theory

The illustration in this section is based on a published report: A G-theory study conducted in National Autonomous University of Mexico Faculty of Medicine in Mexico City, which organized a multiple-station OSCE in summative end-of-career final examination and recruited 278 examinees (Trejo-Mejía et al., 2016). Specifically, there were four exam versions, each of which was delivered at 18 equivalent stations from six areas (pediatrics, obstetrics and gynecology, surgery, internal medicine, emergency medicine, and family medicine). In total, there were 72 stations of which the workflows were navigated via the same blueprint and measured the same skills but in different forms of cases. Finally, the exam was applied in six testing sites and examinees were randomly assigned to only on site.

Accordingly, four facets (i.e., sources of measurement errors) were considered: the examinees, the stations, the versions, and the sites. The design, however, is not fully crossed as examinees were nested within sites. To construct a G-theory model, one should treat the final scores of the examinees X as the dependent variable, where the facets θ s are the independent variables. Mathematical expressions for the G-theory model are

Xpilj=μ+θp+θi+θl+θj+θij+θlj+θli+θilj+ϵ, (1)
σ(X)pilj2=σp2+σi2+σl2+σj2+σij2+σlj2+σli2+σilj2+σϵ2, (2)

Equation 1 shows that an observed score, X, for examinee p on version j of station l at site i is made of the grand mean μ, examinee effect θp (i.e., true skill level of examinee p), site effect θi, version effect θj, station effect θl, both nested- and unnested-interaction effects, and residual error ϵ. Correspondingly, the relevant variance components are outlined in Equation 2. The variance estimates of the G-theory model were σp2 = 17.652, σi2 = 0.000, σl2 = 42.157, σj2 = 0.737, σij2 = 0.867, σlj2 = 38.968, σli2 = 46.631, σilj2 = 34.692, and σϵ2 = 187.374.

The relative- and absolute-error-based generalizability coefficients (i.e., reliability-like index within G-theory framework) can be calculated as follow: Eρδ2=σp2σp2+σδ2 and EρΔ2=σp2σp2+σΔ2, where σδ2=σilj2ni*nj+σϵ2ni*nl*nj and σΔ2=σi2ni+σj2nj+σij2ni*nj+σl2nl+σlj2nl*nj+σil2ni*nl+σilj2ni*nl*nj+σϵ2ni*nl*nj are relative and absolute error variances respectively. As a result, Eρδ2 and EρΔ2 were 0.94 and 0.81, respectively. Both Eρδ2 and EρΔ2 were substantially higher than the acceptable range of G-theory being applied to OSCE research (0.51 to 0.78) according to published reports (Trejo-Mejía et al., 2016). This space provides the possibility of reducing the costs of the OSCE, while maintaining the generalizability coefficients to certain acceptable levels.

Method

Simulated Annealing

As an essential component of machine-learning toolkits, optimization can be defined as a problem where one maximizes or minimizes a target function by systematically choosing input values from an allowed numerical space and computing the value of the function (i.e., outcome). That said, optimization is essentially about finding the best solution. Unlike traditional optimization techniques that heavily rely on deriving mathematical procedures, machine-learning approaches, also known as metaheuristics, are related with general purpose solvers based on computational methods that use few domain knowledge, iteratively improving an initial solution (or population of solutions) to an optimal one at the end (Boussaïd et al., 2013). Well-known optimization techniques include SA, tabu search, genetic algorithms, genetic programming, differential evolution, and particle swarm optimization (Goffe et al., 1994). These techniques are particularly useful for solving:

  1. Complex problems where no specialized optimization algorithm has been developed,

  2. Dynamic questions in which a change in the model requires rederiving efforts to accommodate the new functional needs,

  3. Irregular conditions are imposed so that traditional optimization techniques become inappropriate.

This article adopts SA as it has been widely used in practice and incorporated in most statistical software programs, in which a user only needs to specify a target expression (i.e., the function that the algorithm aims at minimizing or maximizing). SA mimics physical annealing in real life that contains the process of heating up a material until it reaches an annealing temperature and then it will be cooled down slowly in order to change the material to a desired structure. The SA application for optimal design problems with associated costs has a successful history, especially in the construction sector where maximization of profit of a project is of interest (Bettemir, 2010; Jaśkowski & Sobotka, 2006; Li & Coster, 2014, Reeves, 1995).

Similar to other machine-learning approaches, SA requires a set of parameters to drive the algorithm. To be specific, SA requires initial values, starting temperatures, perturbation determination, temperature changing modes, and others. These parameters are proved to be critical in a task as inappropriate selections of them could result in nonoptimal solutions (Feurer & Hutter, 2019). In this article, the R package optimization was used to execute SA (Husmann et al., 2017), where the algorithmic parameters listed in Figure 1 were adopted as they functioned well in our pilot tests. Proposed by Husmann et al. (2017, p. 9), the vf function in Figure 1 defines that the solutions yielded by SA should be integers. Alternatively, one can simply specify the requirement by rounding or ceiling approaches in the target function.

Figure 1.

Figure 1.

Parameters setting for simulated annealing algorithm.

OSCE Cost Specification

The present optimization inquiry falls into the venue of G-theory’s D-study, which is about manipulating each facet’s level to obtain new generalizability coefficients. However, D-study does not accommodate the tasks of minimizing the associated costs of OSCEs. Therefore, the proposed solution, a machine-learning approach, is used to optimize the cost-generalizability.

The costs were summarized from Brown et al. (2015): £15,896 per site, £6,677 per version, and £4,843 per station. With the original levels of the facets, the total cost was as high as £209,240. It is informative to outline the levels of minimal generalizability coefficients acceptable to decision makers to reduce the costs. Without loss of generalizability, EρΔ2 is used to represent the generalizability coefficient as the optimization requires more parameters to estimate so that the inquiry would not be oversimplified.

Monte Carlo Simulation

This section consists of two parts. The first part is verifying the utility of the proposed approach by comparing with the LaGrange multiplier approach in the scenarios illustrated in Marcoulides and Goldstein (1990). Specifically, they provided a closed-form solution for a two-facet fully crossed case (i.e., item and occasion effects represented by σi2 and σo2, respectively) with the information about cost per item per occasion and total budget available represented by c and c¯, respectively:

ni=σpi2σpo2*(c¯c)andno=σpo2σpi2*(c¯c), (3)

while σpi2 and σpo2 are the variance components of interaction effects between examinee (i.e., p) and item as well as occasion facets. With the provided solution, the budgetary constraint can be satisfied such that cninoc¯, while the generalizability coefficient is also maximized. Note that simply rounding the LaGrange multiplier solutions would result in violating the budgetary constraints. Goldstein and Marcoulides (1991) proposed a simple table approach that converts optimal noninteger solutions to acceptable candidate integer solutions while satisfying the cost constraint can be also met; this approach was used in the simulation. In the example of Marcoulides and Goldstein (1990), that σp2 = 6.30, σpi2 = 1.60, σpo2 = 0.30, σϵ2 = 1.95, c = $5 and c¯ = $100 yielded ni = 2 and no = 10 via Equation 3, which resulted in the maximum generalizability coefficient Eρδ2 = 0.94; the simulation adopted the same two-facet fully crossed design as well as the c and c¯ values but randomly generated variance parameters by uniformly sampling from [σ2σ22,σ2+σ22]. For example, in an arbitrary replication, σp2 is uniformly sampled from [6.303.15,6.30+3.15] and other variance components are also sampled in the same fashion. It was replicated 200 times where the optimal solutions yielded by SA approach and the LaGrange multiplier approach were compared.

The second part of the simulation section is based on present OSCE planning inquires: minimizing costs when a generalizability coefficient threshold is given. Let [0.50, 0.55, 0.60, 0.65, 0.70, 0.75, 0.80] be the levels of minimal EρΔ2. The task of interest then becomes finding the optimal combination of the facet levels without producing a EρΔ2 lower than the selected threshold. The evaluation was essentially conducted a Monte Carlo simulation design, meaning we used computers to simulate different conditions based on a real setting and examine the performance of the proposed approach within these conditions. Assuming that β is the generalizability threshold, the target function f that SA attempts to minimize in the present context is

f=£15,896*ni+£6,677*nj+£4,843*nl,

while

σp2σp2+σi2ni+σj2nj+σij2ni*nj+σl2nl+σlj2nl*nj+σil2ni*nl+σilj2ni*nl*nj+σϵ2ni*nl*njβ.

Results

In the first part of the simulation study, SA results had a perfect match with the closed-form solutions (i.e., 200/200 replications that two approaches produced equal solutions), meaning that the proposed approach has great utility for conditional optimization works in the present context. In addition to the replications, the results yield by SA approach and the LaGrange multiplier approach were identical in the example of Marcoulides and Goldstein (1990).

For the second part of the simulation study, Table 1 lists the combinations yielded by SA optimization. Unsparingly, the OSCE cost reduces as the generalizability coefficient threshold falls, as the facets’ levels are lessened. Meanwhile, the actual EρΔ2 s are quite close the thresholds showing that the optimization converges well to a desirable degree. It can be interpreted as, for example, if EρΔ20.65 is an acceptable condition, an OSCE testing setting with three sites, two versions, and 13 stations will result in £114,769 as the associated cost. Compared with the original amount, the cost decreases by £94,471. Other rows of the results in Table 1 can be explained similarly and therefore will not be repeated here.

Table 1.

Optimization Results via Simulated Annealing Approach.

Threshold Actual EρΔ2 Costs(£) Site Version Station
0.50 0.51 85,717 3 2 7
0.55 0.57 95,401 3 2 9
0.60 0.60 102,078 4 2 9
0.65 0.66 114,769 3 2 13
0.70 0.71 131,130 4 2 15
0.75 0.75 147,491 5 2 17
0.80 0.80 176,078 4 3 21

Discussion and Conclusion

This article provides an interdisciplinary study for OSCE planning, fiscal elements, psychometric properties, and data science optimization. Although the optimization can be achieved by exhaustive search, trying all possible combinations of the solution set is not viable (sometimes even wholly impossible) when the problem of interest is complex. The growth of the facet size can result in an exponential increment of the exhaustive search’s trial-and-error attempts. For example, many OSCE providers would like to consider the rater effect, the occasion effect, and their interactions with other facets.

The example above may be not generalizable as OSCE testing settings, and the associated costs can be quite different; however, the proposed approach can be easily adopted and customized to meet the practical requirements. OSCEs are both money- and time-consuming; scientific tools and methods should be applied to the validity of the exam per se and also the planning from the fiscal standpoint. The definitions of “cost-effectiveness” are not consistent in different contexts. This article treats OSCE generalizability as the “effectiveness,” where other studies may consider the term as the proportion of examinees that become qualified physicians/doctors.

Using optimization approaches, like any techniques, is not risk-free. Algorithmic failures and local convergences should always be cautioned. That said, if the optimization stops unexpectedly or yields a nonoptimal solution, the results can be misleading to decision makers. A recent trend in the machine-learning literature is calling for an ensemble paradigm, which operates various optimization approaches simultaneously and combines the results via a certain weighting schema.

Footnotes

Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Funding: The author(s) received no financial support for the research, authorship, and/or publication of this article.

References

  1. Baig L. A., Violato C. (2012). Temporal stability of objective structured clinical exams: A longitudinal study employing item response theory. BMC Medical Education, 12(1), 1-6. 10.1186/1472-6920-12-121 [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Bettemir Ö. H. (2010). Experimental design for genetic algorithm simulated annealing for time cost trade-off problems. International Journal of Engineering and Applied Sciences, 3(1), 15-26. [Google Scholar]
  3. Boulet J. R., McKinley D. W., Whelan G. P., Hambleton R. K. (2003). Quality assurance methods for performance-based assessments. Advances in Health Sciences Education, 8(1), 27-47. 10.1023/A:1022639521218 [DOI] [PubMed] [Google Scholar]
  4. Boussaïd I., Lepagnot J., Siarry P. (2013). A survey on optimization metaheuristics. Information Sciences, 237(July), 82-117. 10.1016/j.ins.2013.02.041 [DOI] [Google Scholar]
  5. Brennan R. L. (1992). Generalizability theory. Educational Measurement: Issues and Practice, 11(4), 27-34. 10.1111/j.1745-3992.1992.tb00260.x [DOI] [Google Scholar]
  6. Brennan R. L. (2010). Generalizability theory and classical test theory. Applied Measurement in Education, 24(1), 1-21. 10.1080/08957347.2011.532417 [DOI] [Google Scholar]
  7. Brown C., Ross S., Cleland J., Walsh K. (2015). Money makes the (medical assessment) world go round: The cost of components of a summative final year Objective Structured Clinical Examination (OSCE). Medical Teacher, 37(7), 653-659. 10.3109/0142159X.2015.1033389 [DOI] [PubMed] [Google Scholar]
  8. Cornfield J., Tukey J. W. (1956). Average values of mean squares in factorials. Annals of Mathematical Statistics, 27(40), 907-949. 10.1214/aoms/1177728067 [DOI] [Google Scholar]
  9. Cronbach L. J. (1951). Coefficient alpha and the internal structure of tests. Psychometrika, 16(3), 297-334. 10.1007/BF02310555 [DOI] [Google Scholar]
  10. Cronbach L. J., Gleser G. C., Nanda H., Rajaratnam N. (1972). The dependability of behavioral measurements: Theory of generalizability scores and profiles. Wiley. [Google Scholar]
  11. Cusimano M. D., Cohen R., Tucker W., Murnaghan J., Kodama R., Reznick R. (1994). A comparative analysis of the costs of administration of an OSCE (objective structured clinical examination). Academic Medicine, 69(7), 571-576. 10.1097/00001888-199407000-00014 [DOI] [PubMed] [Google Scholar]
  12. Dacre J., Walsh K. (2013). Funding of medical education: The need for transparency. Clinical Medicine, 13(6), 573-575. 10.7861/clinmedicine.13-6-573 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Donnon T., Paolucci E. O. (2008). A generalizability study of the medical judgment vignettes interview to assess students’ noncognitive attributes for medical school. BMC Medical Education, 8(1), 58. 10.1186/1472-6920-8-58 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Downing S. M. (2003). Validity: On the meaningful interpretation of assessment data. Medical education, 37(9), 830-837. 10.1046/j.1365-2923.2003.01594.x [DOI] [PubMed] [Google Scholar]
  15. Downing S. M., Haladyna T. M. (2004). Validity threats: Overcoming interference with proposed interpretations of assessment data. Medical education, 38(3), 327-333. 10.1046/j.1365-2923.2004.01777.x [DOI] [PubMed] [Google Scholar]
  16. Feurer M., Hutter F. (2019). Hyperparameter optimization. In Automated machine learning (pp. 3-33). Springer. [Google Scholar]
  17. Goldstein Z., Marcoulides G. (1991). Maximizing the coefficient of generalizability in decision studies. Educational and Psychological Measurement, 51(1), 79-88. 10.1177/0013164491511006 [DOI] [Google Scholar]
  18. Goffe W. L., Ferrier G. D., Rogers J. (1994). Global optimization of statistical functions with simulated annealing. Journal of Econometrics, 60(1-2), 65-99. 10.1016/0304-4076(94)90038-8 [DOI] [Google Scholar]
  19. Haladyna T. M., Downing S. M. (2004). Construct-irrelevant variance in high-stakes testing. Educational Measurement: Issues and Practice, 23(1), 17-27. 10.1111/j.1745-3992.2004.tb00149.x [DOI] [Google Scholar]
  20. Harden R. M., Stevenson M., Downie W. W., Wilson G. M. (1975). Assessment of clinical competence using objective structured examination. British Medical Journal, 1(5955), 447-451. 10.1136/bmj.1.5955.447 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Hastie M. J., Spellman J. L., Pagano P. P., Hastie J., Egan B. J. (2014). Designing and implementing the objective structured clinical examination in anesthesiology. Anesthesiology: Journal of the American Society of Anesthesiologists, 120(1), 196-203. 10.1097/ALN.0000000000000068 [DOI] [PubMed] [Google Scholar]
  22. Henderson C. R. (1953). Estimation of variance and covariance components. Biometrics, 9(2), 226-252. 10.2307/3001853 [DOI] [Google Scholar]
  23. Hodges B. D., Hollenberg E., McNaughton N., Hanson M. D., Regehr G. (2014). The psychiatry OSCE: A 20-year retrospective. Academic Psychiatry, 38(1), 26-34. 10.1007/s40596-013-0012-8 [DOI] [PubMed] [Google Scholar]
  24. Hubbard J. P. (1971). Measuring medical education, the tests and test procedures of the National Board of Medical Examinations. Lea & Febiger. [Google Scholar]
  25. Huebner A., Lucht M. (2019). Generalizability theory in R. Practical Assessment, Research & Evaluation, 24(5), 2. 10.7275/5065-gc10 [DOI] [Google Scholar]
  26. Husmann K., Lange A., Spiegel E. (2017). The R package optimization: Flexible global optimization with simulated-annealing. r-project.org. https://cran.r-project.org/web/packages/optimization/vignettes/vignette_master.pdf
  27. Jaśkowski P., Sobotka A. (2006). Scheduling construction projects using evolutionary algorithm. Journal of Construction Engineering and Management, 132(8), 861-870. 10.1061/(ASCE)0733-9364(2006)132:8(861) [DOI] [Google Scholar]
  28. Jiang Z. (2018). Using linear mixed-effect model framework to estimate generalizability variance component in R: A lme4 package application. Methodology, 14(3), 133-142. 10.1027/1614-2241/a000149 [DOI] [Google Scholar]
  29. Jiang Z., Raymond M., Shi D., DiStefano C. (2020). Using a linear mixed-effect model framework to estimate multivariate generalizability theory parameters in R. Behavior Research Methods, 52(6), 2383-2393. 10.3758/s13428-020-01399-z [DOI] [PubMed] [Google Scholar]
  30. Jiang Z., Skorupski W. (2018). A Bayesian approach to estimating variance components within a multivariate generalizability theory framework. Behavior Research Methods, 50(6), 2193-2214. 10.3758/s13428-017-0986-3 [DOI] [PubMed] [Google Scholar]
  31. Li C., Coster D. C. (2014). A simulated annealing algorithm for D-optimal design for 2-way and 3-way polynomial regression with correlated observations. Journal of Applied Mathematics, 2014, Article 746914. 10.1155/2014/746914 [DOI]
  32. LoPilato A. C., Carter N. T., Wang M. (2015). Updating generalizability theory in management research: Bayesian estimation of variance components. Journal of Management, 41(2), 692-717. 10.1177/0149206314554215 [DOI] [Google Scholar]
  33. Marcoulides G. A. (1993). Maximizing power in generalizability studies under budget constraints. Journal of Educational Statistics, 18(2), 197-206. 10.3102/10769986018002197 [DOI] [Google Scholar]
  34. Marcoulides G. A. (1995). Designing measurement studies under budget constraints: Controlling error of measurement and power. Educational and Psychological Measurement, 55(3), 423-428. 10.1177/0013164495055003005 [DOI] [Google Scholar]
  35. Marcoulides G. A., Goldstein Z. (1990). The optimization of generalizability studies with resource constraints. Educational and Psychological Measurement, 50(4), 761-768. 10.1177/0013164490504004 [DOI] [Google Scholar]
  36. Marcoulides G. A., Goldstein Z. (1991). Selecting the number of observations in multivariate measurement studies under budget constraints. Educational and Psychological Measurement, 51(3), 573-584. 10.1177/0013164491513005 [DOI] [Google Scholar]
  37. Marcoulides G. A., Goldstein Z. (1992). The optimization of multivariate generalizability studies with budget constraints. Educational and Psychological Measurement, 52(2), 301-308. 10.1177/0013164492052002005 [DOI] [Google Scholar]
  38. Messick S. (1998). Test validity: A matter of consequence. Social Indicators Research, 45(1-3), 35-44. 10.1023/A:1006964925094 [DOI] [Google Scholar]
  39. Meyer J. P., Liu X., Mashburn A. J. (2014). A practical solution to optimizing the reliability of teaching observation measures under budget constraints. Educational and Psychological Measurement, 74(2), 280-291. [Google Scholar]
  40. Myford C. M., Wolfe E. W. (2003). Detecting and measuring rater effects using many-facet Rasch measurement: Part I. Journal of Applied Measurement, 4(4), 386-422. [PubMed] [Google Scholar]
  41. Newble D. I., Swanson D. B. (1988). Psychometric characteristics of the objective structured clinical examination. Medical Education, 22(4), 325-334. 10.1111/j.1365-2923.1988.tb00761.x [DOI] [PubMed] [Google Scholar]
  42. Patrício M. F., Julião M., Fareleira F., Carneiro A. V. (2013). Is the OSCE a feasible tool to assess competencies in undergraduate medical education? Medical Teacher, 35(6), 503-514. 10.3109/0142159X.2013.774330 [DOI] [PubMed] [Google Scholar]
  43. Rao C. R. (1970). Estimation of heteroscedastic variances in linear models. Journal of the American Statistical Association, 65(329), 161-172. 10.1080/01621459.1970.10481070 [DOI] [Google Scholar]
  44. Reeves C. R. (1995). A genetic algorithm for flowshop sequencing. Computers & Operations Research, 22(1), 5-13. 10.1016/0305-0548(93)E0014-K [DOI] [Google Scholar]
  45. Saunders P. F. (1992). Alternative solutions for optimization problems in generalizability theory. Psychometrika, 57, 351-356. 10.1007/BF02295423 [DOI] [Google Scholar]
  46. Saunders P. F., Theunissen T. J., Baas S. M. (1989). Minimizing the number of observations: A generalization of the Spearman-Brown formula. Psychometrika, 54, 587-589. 10.1007/BF02296398 [DOI] [Google Scholar]
  47. Setyonugroho W., Kennedy K. M., Kropmans T. J. (2015). Reliability and validity of OSCE checklists used to assess the communication skills of undergraduate medical students: A systematic review. Patient Education and Counseling, 98(12), 1482-1491. 10.1016/j.pec.2015.06.004 [DOI] [PubMed] [Google Scholar]
  48. Smith V., Muldoon K., Biesty L. (2012). The Objective Structured Clinical Examination (OSCE) as a strategy for assessing clinical competence in midwifery education in Ireland: A critical review. Nurse Education in Practice, 12(5), 242-247. 10.1016/j.nepr.2012.04.012 [DOI] [PubMed] [Google Scholar]
  49. Trejo-Mejía J. A., Sánchez-Mendiola M., Méndez-Ramírez I., Martínez-González A. (2016). Reliability analysis of the objective structured clinical examination using generalizability theory. Medical Education Online, 21, 31650. 10.3402/meo.v21.31650 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Walsh K., Jaye P. (2013). Cost and value in medical education. Education for Primary Care, 24(6), 391-393. 10.1080/14739879.2013.11494206 [DOI] [PubMed] [Google Scholar]
  51. Walsh M., Bailey P. H., Koren I. (2009). Objective structured clinical evaluation of clinical competence: An integrative review. Journal of Advanced Nursing, 65(8), 1584-1595. 10.1111/j.1365-2648.2009.05054.x [DOI] [PubMed] [Google Scholar]
  52. Woodward J. A., Joe G. W. (1973). Maximizing the coefficient of generalizability in multifacet decision studies. Psychometrika, 38(2), 173-181. 10.1007/BF02291112 [DOI] [Google Scholar]

Articles from Educational and Psychological Measurement are provided here courtesy of SAGE Publications

RESOURCES