Skip to main content
BMC Musculoskeletal Disorders logoLink to BMC Musculoskeletal Disorders
. 2009 Oct 1;10:120. doi: 10.1186/1471-2474-10-120

Are classifications of proximal radius fractures reproducible?

Fabio T Matsunaga 1,, Marcel JS Tamaoki 1, Eduardo F Cordeiro 1, Anderson Uehara 1, Marcos H Ikawa 2, Marcelo H Matsumoto 1, João BG dos Santos 1, João C Belloti 1
PMCID: PMC2761854  PMID: 19793401

Abstract

Background

Fractures of the proximal radius need to be classified in an appropriate and reproducible manner. The aim of this study was to assess the reliability of the three most widely used classification systems.

Methods

Elbow radiographs images of patients with proximal radius fractures were classified according to Mason, Morrey, and Arbeitsgemeinschaft für osteosynthesefragen/Association for the Study of Internal Fixation (AO/ASIF) classifications by four observers with different experience with this subject to assess their intra- and inter-observer agreement. Each observer analyzed the images on three different occasions on a computer with numerical sequence randomly altered.

Results

We found that intra-observer agreement of Mason and Morrey classifications were satisfactory (κ = 0.582 and 0.554, respectively), while the AO/ASIF classification had poor intra-observer agreement (κ = 0.483). Inter-observer agreement was higher in the Mason (κ = 0.429-0.560) and Morrey (κ = 0.319-0.487) classifications than in the AO/ASIF classification (κ = 0.250-0.478), which showed poor reliability.

Conclusion

Inter- and intra-observer agreement of the Mason and Morey classifications showed overall satisfactory reliability when compared to the AO/ASIF system. The Mason classification is the most reliable system.

Background

Fractures of the proximal radius are relatively common injuries, accounting for approximately 1/3 of all elbow fractures and about 1.7 to 5.4% of all fractures in adults [1,2]. Most of these (85%) occur in adults between 20 and 60 years of age (average of 30 to 40 years) with a male:female ratio of about 2:3 [1,2]. One third of the lesions are associated with others upper limb injuries, such as carpal bone fractures [3], distal radioulnar joint injuries [4,5], interosseous membrane injuries [6,7], capitellar fractures [8], and damage of the medial collateral ligament.

The most common mechanism of injury in these fractures is a fall onto an outstretched hand with an axial load on the radius. In such cases, the radial head and/or neck fracture when they collide with the capitellum, usually with forearm in pronation and the elbow in partial flexion [9-11].

Fractures of the proximal radius play an important role in injuries of the elbow, not so much by how frequently they occur, but mainly by the potential difficulties in treatment and the complications that can arise, sometimes with serious impairment of function due to pain and loss of mobility. Therefore, proper classification is essential in order to render the proper treatment.

Systems have been developed to help surgeons in classifying fractures into different and clinically useful groups for treatment definition.

In 1954, Mason described the first classification system dividing proximal radius fractures into three types. Type I fractures were nondisplaced or minimally displaced fractures of the head or neck; type II were displaced fractures (more than 2 mm) of the head or neck; and type III were severely comminuted fractures of the proximal radius [12]. In 1962, Johnston expanded the classification of Mason adding type IV, a fracture associated with dislocation of the elbow [13].

In 2008, van Riet and Morrey published a revision of the Mason classification, distinguishing between injuries associated with coronoid fractures, the olecranon fractures, and ligamentous injuries [14].

The AO/ASIF classification was created in 1986 and revised in 2007 [15]. It considers the seriousness of the bone injury and serves as a basis for treatment and prognosis.

The AO/ASIF system specifies three basic types: extra-articular, articular of the radius or ulna, and articular of the radius and ulna. With each group, the fractures are organized in increasing order of severity with regard to morphological complexity, difficulty in treatment, and prognosis. While this system is the most comprehensive, its intra- and inter-observer reliability has shown to be limited [16,17].

The purpose of a classification system is to name and describe the fractures according to their characteristics providing a hierarchy of those characteristics. A classification system should also guide action or intervention and assist in predicting outcomes of an intervention or treatment [18]. A good system needs to be valid, reliable, and reproducible. The perfect classification system should also standardize the language used to describe the fractures, offer guidelines for treatment, indicate the possibility of complications, and help determine the prognosis. The ideal system should also provide a mechanism to evaluate and compare the results with treatment of similar fractures treated at various centres and reported at different times in the literature [19].

Considering the need to classify the fractures of the proximal radius in an appropriate and reproducible manner, we sought to assess the reproducibility of the three most widely used classification systems. Thus we evaluated the intra- and inter-observer agreement of the Mason modified by Hotchkiss, Morrey and AO/ASIF classifications of proximal radius fractures.

Methods

We analyzed 65 consecutive elbow radiographs performed on patients with fractures of the proximal radius. The patients were treated in the same hospital. Each of the 65 radiographs consisted of two views, anterior-posterior and lateral, and they were numbered, with the patients' names and ages concealed. Radiographs were excluded if the patient had incomplete skeletal development, pathologic fractures or previous elbow surgery.

The image quality was determined by two orthopaedic surgeons. The radiography was accepted only when both of these surgeons considered the radiographs acceptable.

Four observers familiar with the classification systems were selected for analysis. These observers were a second-year resident of orthopaedics (R2), a general orthopaedist (GO), a shoulder and elbow surgery specialist (SES), and a radiologist (RD).

To standardize the information for all observers, each were given self-explanatory diagrams with the classification systems. Each observer classified the 65 images at three different times, according to the three systems (Mason modified by Hotchkiss, Morrey and AO/ASIF). In the first evaluation (T1), the 65 digitized images of radiographs of each patient were viewed on a computer in numerical sequence. Three weeks later, in the second assessment (T2), the sequence of radiographs was randomly altered, as it was in the third assessment (T3), three weeks after T2. This sequence of randomization was known only by a person uninvolved in the assessment of the images.

The data were collected in spreadsheets and kappa coefficients were calculated for analysis according to the method proposed by Fleiss et al [20,21]. This method not only calculates the agreement expected by chance, as described earlier in the method of Scott and Cohen [22,23], but also the correlation among more than two observers in the evaluation of nominal variables. The kappa coefficient of agreement indicates the proportion of agreement among observers. The kappa values range from -1 to +1: values between -1 and 0 indicate that observed agreement was lower than that expected by chance, 0 indicates a level of agreement equal to that expected by chance, and +1 indicates total agreement. Overall, kappa values below 0.5 are considered unsatisfactory; values between 0.5 and 0.75 are considered satisfactory and appropriate, and values over 0.75 are considered excellent [24,25].

This project was approved by the Research Ethics Committee under No. 363/08, on April 4, 2008.

Results

From the 65 initial radiographs six were excluded because of poor technical quality, leaving a sample size to 59 radiographs.

Table 1 summarizes the kappa values for intra-observer comparisons for each observer, at the three time points. The concordance was higher among the different with the Mason and Morrey classifications (mean κ = 0.582 and 0.554, respectively) than with the AO/ASIF classification (mean κ = 0.483).

Table 1.

Kappa coefficients for intra-observer comparisons between the three time points (T1, T2 and T3).

Observer Classification
AO/ASIF MORREY MASON

R2 0.676 0.666 0.710

GO 0.439 0.548 0.523

SES 0.305 0.356 0.409

RD 0.513 0.645 0.685

Mean 0.483 0.554 0.582

Observers: second-year resident of orthopaedics (R2), general orthopaedist (GO), shoulder and elbow surgery specialist (SES), radiologist (RD).

Table 2 shows the inter-observer kappa coefficients at each of the times of assessment. Inter-observer agreement was higher in the Mason (κ = 0.429-0.560) and Morrey (κ = 0.319-0.487) classifications than in the AO/ASIF classification (κ = 0.250-0.478).

Table 2.

Kappa coefficients for inter-observer comparisons at each time point. (T1, T2, T3)

Time Point Classification
AO/ASIF MORREY MASON

T1 0.250 0.319 0.429

T2 0.386 0.443 0.560

T3 0.478 0.487 0.551

Discussion

The classification systems for this study were selected because they are the most commonly used and studied for fractures of the proximal radius [14]. Classification systems are of great importance in orthopaedic practice because they are used to describe fractures, guide treatment, and compare treatment outcome within and between studies in the literature. As a result, intra- and inter-observer concordances are essential for any classification system.

In the analysis of intra-observer agreement between three time points, the average kappa coefficient for the AO/ASIF classification was unsatisfactory (κ = 0.483), ranging from 0.305 (for the SES) to 0.676 (for the R2). The broad variability in these results is probably due to the complexity of the classification, and accord with the results of other studies that evaluated the classification for fractures of other bones [17,26]. Professional experience had no effect on intra-observer agreement with this system, as indicated by the highest kappa coefficient for the R2 and the lowest for the SES. For the Mason and Morrey classifications, intra-observer concordances were satisfactory and similar (κ of 0.692 and 0.644, respectively). The similarity of these coefficients was expected because the Morrey classification is derived from and more complex than the of Mason classification [14].

Regarding inter-observer concordances, the AO/ASIF rating was unsatisfactory at all times, although it improved with time, probably due to the learning curve. A similar pattern occurred with the Morrey classification, but the values were slightly higher but still unsatisfactory. The Mason classification had an unsatisfactory mean kappa coefficient at T1 (0.429). However, the means at T2 and T3 were satisfactory (0.560 and 0.551, respectively). These values may also be explained by the complexity of the classifications.

The AO/ASIF is a more complex system that involves the proximal radius as well as associated ulnar injuries and ligamentous injuries. These added variables will presumably have a negative impact on its reliability.

The results of this study were similar to those reported by Sheps et al. [27], which indicated that the correlation was unsatisfactory for the AO/ASIF classification and better for the system adapted by Hotchkiss from the Mason system. However, even this adapted Mason system was unsatisfactory when considering the lower limit of 95% of the confidence interval.

It is important to mention that the present study was limited to evaluating the agreement between the observers' opinions. This study was unable to measure the accuracy of each observer's opinion. To clarify the accuracy issue, studies in which clinical-radiographic diagnoses made by each observer compared with an examination result or a gold-standard procedure (i.e. one with high sensitivity and specificity) would be needed in order to prove the proposed diagnosis.

We have made the first step to the development of the ideal classification once we studied the reliability of the most widely used classification systems. In the continuing search for this ideal classification there is a need to perform new prospective studies with large number of patients to determine which variables (displacement of the fragments, intra-articular/metaphyseal comminution, associated injuries, patient's age) can guide the ideal treatment and predict the prognosis of such fractures through radiographic examination.

Conclusion

Inter- and intra-observer agreement of the Mason and Morey classifications showed overall satisfactory reliability when compared to the AO/ASIF system.

The Mason classification is the most reliable system.

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

FTM participated in the design of the study, collected, digitalized and randomized the radiographs, helped to draft the manuscript, revised the bibliography for references and performed the statistical analysis; MJST conceived the study design, analyzed the radiographs and classified the fractures, helped to draft the manuscript and revised the bibliography for references; EFC participated in the design of the study, analyzed the radiographs and classified the fractures and translated the manuscript to English; AU: participated in the design of the study, analyzed the radiographs and classified the fractures; MHI participated in the design of the study, analyzed the radiographs and classified the fractures; MHM conceived the study design, revised the article and gave the final approval of the version to be published; JBGS participated in the design of the study, revised the article and gave the final approval of the version to be published; JCB participated in the design of the study, revised the article; FF revised the article and gave the final approval of the version to be published. All authors read and approved the final manuscript.

About the authors

Fabio T Matsunaga1, Marcel J S Tamaoki2, Eduardo F Cordeiro3, Anderson Uehara4, Marcos H Ikawa5, Marcelo H Matsumoto6, João B G dos Santos7, João C Belloti8.

1Second-year resident in the Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

2Attending physician in Shoulder and Elbow Surgery Sector in the Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

3Second-year resident in the Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

4Attending physician in Shoulder and Elbow Surgery Sector in the Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

5Attending physician in the Department Radiology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

6Head of the Shoulder and Elbow Surgery Sector, Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

7Adjunct professor and head of Hand Surgery Clinic, Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

8Attending physician in Hand Surgery Clinic, Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

Pre-publication history

The pre-publication history for this paper can be accessed here:

http://www.biomedcentral.com/1471-2474/10/120/prepub

Acknowledgments

Acknowledgements

The authors thank Dr Walter Manna Albertoni and Dr Flavio Faloppa for their on-going support and suggestions in the preparation of this manuscript.

Contributor Information

Fabio T Matsunaga, Email: fteruo@gmail.com.

Marcel JS Tamaoki, Email: marceltamaoki@terra.com.br.

Eduardo F Cordeiro, Email: eduardo68@gmail.com.

Anderson Uehara, Email: andersonuehara@hotmail.com.

Marcos H Ikawa, Email: mhikawa@gmail.com.

Marcelo H Matsumoto, Email: marcelohmatsumoto@terra.com.br.

João BG dos Santos, Email: jbgsantos@ig.com.br.

João C Belloti, Email: jbelloti@terra.com.br.

References

  1. Van Riet R, Morrey BF, O'Driscoll SW, Van Glabbeek F. Associated injuries complicating radial head fractures: a demographic study. Clin Orthop Relat Res. 2005;441:351–355. doi: 10.1097/01.blo.0000180606.30981.78. [DOI] [PubMed] [Google Scholar]
  2. Kaas L, Van Riet R, Vroemen JPAM, Eygendaal D. The incidence of associated fractures of the upper limb in fractures of the radial head. Strategies Trauma Limb Reconstr. 2008;3:71–4. doi: 10.1007/s11751-008-0038-8. Epub 2008 Jul 10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Jupiter JB. The management of fractures in one upper extremity. J Hand Surg. 1986;11A:279–82. doi: 10.1016/s0363-5023(86)80071-5. [DOI] [PubMed] [Google Scholar]
  4. Curr J, Coe W. Dislocation if the inferior radio-ulnar joint. Br J Surg. 1946;34:74–77. doi: 10.1002/bjs.18003413312. [DOI] [PubMed] [Google Scholar]
  5. Edwards G, Jupiter JB. The Essex-Lopresti lesion revised. Clin Orthop. 1988;234:61–69. [PubMed] [Google Scholar]
  6. McDougall A, White J. Subluxation of the inferior radio-ulnar joint complicating fracture of the radial head. J Bone Joint Surg. 1957;39[B]:278–87. doi: 10.1302/0301-620X.39B2.278. [DOI] [PubMed] [Google Scholar]
  7. Taylor TKF, O'Connor BT. The effect upon the inferior radio-ulnar joint and excision of the head of the radius in adults. J Bone Joint Surg. 1964;46[B]:83–84. [PubMed] [Google Scholar]
  8. Osborne G, Cotterill P. Recurrent dislocation of the elbow. J Bone Joint Surg. 1966;48[B]:340–46. [PubMed] [Google Scholar]
  9. Ring D, Jupiter JB. Fracture-dislocation of the elbow. J Bone Joint Surg [Am] 1998;80-A:566–580. [PubMed] [Google Scholar]
  10. Ring D. Rockwood and Green's Fractures in Adults. Sixth. Vol. 1. Philadelphia: Lippincot Williams & Wilkins; 2006. Fractures and Dislocations of the Elbow; pp. 1011–12. [Google Scholar]
  11. Amis AA, Miller JH. Mechanisms of elbow fractures: investigation using impact tests in vitro. Injury. 1995;26:163–168. doi: 10.1016/0020-1383(95)93494-3. [DOI] [PubMed] [Google Scholar]
  12. Mason ML. Some observations on fractures of the head of the radius with a review of one hundred cases. Br J Surg. 1954;42:123–132. doi: 10.1002/bjs.18004217203. [DOI] [PubMed] [Google Scholar]
  13. Johnston GW. A follow-up of 100 cases of fracture of the head of the radius with review of the literature. Ulster Med J. 1962;31:51–56. [PMC free article] [PubMed] [Google Scholar]
  14. Van Riet RP, Morrey BF. Documentation of associated injuries ocurring with radial head fractures. Clin Orthop Relat Res. 2008;466:130–134. doi: 10.1007/s11999-007-0064-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Marsh JL, et al. Fracture and dislocation compendium - 2007: Orthopaedic Trauma Association classifications, database and outcomes comittee. J Orthop Trauma. 2007;21:S1–S133. doi: 10.1097/00005131-200711101-00001. [DOI] [PubMed] [Google Scholar]
  16. Burstein AH. Fracture classification systems: do they work and are they useful? J Bone Joint Surg [Am] 1993;75-A:1743–4. [PubMed] [Google Scholar]
  17. Kreder HJ, Hanel DP, Mckee M, Jupiter J, Mcgillivary G, Swiontkowski MF. Consistency of AO fracture classification for the distal radius. J Bone Joint Surg [Br] 1996;78-B:726–31. [PubMed] [Google Scholar]
  18. Dirschl DR, Cannada LK. Rockwood and Green's Fractures in Adults. Sixth. Vol. 1. Philadelphia: Lippincot Williams & Wilkins; 2006. Classification of Fractures; pp. 43–4. [Google Scholar]
  19. Martin JS, Marsh JL. Current classification of fractures. Rationale and utility. Radiol Clin North Am. 1997;35:491–506. [PubMed] [Google Scholar]
  20. Fleiss JL. Measuring nominal scale agreement among many raters. Psychological Bulletin. 1971;76:378–82. doi: 10.1037/h0031619. [DOI] [Google Scholar]
  21. Fleiss JL, Slakter MJ, Fischman SL, Park MH, Chilton NW. Inter-examiner reliability in caries trials. J Dent Res. 1979;58:604–609. doi: 10.1177/00220345790580021101. [DOI] [PubMed] [Google Scholar]
  22. Scott WA. Reliability of content analysis: the case of nominal scale coding. Public Opinion Quarterly. 1995;19:321–5. doi: 10.1086/266577. [DOI] [Google Scholar]
  23. Cohen J. A coefficient of agreement for nominal scales. Educ Psychol Meas. 1960;20:37–46. doi: 10.1177/001316446002000104. [DOI] [Google Scholar]
  24. Martin JS, Marsh JL, Bonar SK, DeCoster TA, Found EM, Brandser EA. Assessment of the AO/ASIF Fracture Classification for distal tibia. J Orthop Trauma. 1997;11:477–483. doi: 10.1097/00005131-199710000-00004. [DOI] [PubMed] [Google Scholar]
  25. Svanholm H, Starklint H, Gunderson HJ, Fabricius J, Barlebo H, Olsen S. Reproducibility of histomorphological diagnosis with special reference to the kappa statistic. APMIS. 1989;97:689–698. doi: 10.1111/j.1699-0463.1989.tb00464.x. [DOI] [PubMed] [Google Scholar]
  26. Flinkkila T, Nikkola-Sihto A, Kaarela O, Paakko E, Raatikainen T. Poor interobserver reliability of AO classification of fractures of distal radius. J Bone Joint Surg [Br] 1998;80-B:670–2. doi: 10.1302/0301-620X.80B4.8511. [DOI] [PubMed] [Google Scholar]
  27. Sheps D, Hildebrand K, Keifer K. The inter-observer reliability of classifications systems for radial head fractures. J Bone Joint Surg [Br] 2008;90-B:140. [Google Scholar]

Articles from BMC Musculoskeletal Disorders are provided here courtesy of BMC

RESOURCES