Skip to main content
Journal of Medical Internet Research logoLink to Journal of Medical Internet Research
. 2003 Jun 13;5(2):e11. doi: 10.2196/jmir.5.2.e11

Development and Evaluation of the Virtual Pathology Slide: A New Tool in Telepathology

Sean SP Costello 1, Daniel J Johnston 1, Peter A Dervan 2, Daniel G O'Shea 1,
Reviewed by: V Della Mea
PMCID: PMC1550558  PMID: 12857667

Abstract

Background

The Virtual Pathology Slide is an interactive microscope emulator that presents, via the Internet or CD-ROM, a complete 15.53 mm x 11.61 mm digitalized tissue section. The Virtual Pathology Slide mimics the use of a microscope in both the stepwise increase in magnification (from 16x up to 2000x) and in lateral motion in the X and Y Cartesian directions. This permits a pathologist to navigate to any area on a slide, at any magnification, similar to a conventional microscope.

Objective

The aim of this study was to assess the diagnostic accuracy and acceptability of the Virtual Pathology Slide.

Methods

Ten breast needle core biopsies were randomly selected and presented to 17 pathologists or trainee pathologists with at least 2 years experience in pathology practice. Participants were required to examine each case online and provide a diagnostic classification using online feedback forms. The recorded data permitted examination of interobserver variability and user satisfaction.

Results

Agreement between original glass-slide diagnosis and consensus diagnosis using the Virtual Pathology Slide was reached in 9 out of 10 slides. Percentage concordance for slides lay in the range of 35.3% to 100% with an average percentage concordance between slides of 66.5%. The average Kappa statistics for interobserver agreement was 0.75 while average percentage concordance amongst participants was 66.5%. Participants looked at an average of 22 fields of view while examining each slide. Confidence: 81.25% of the participants indicated confidence using the Virtual Pathology Slide to make a diagnostic decision, with 56.25% describing themselves as "reasonably confident," 18.75% as "confident," and 6.25% as "very confident." Ease of use: 68.75% reported the system as "easy" or "very easy" to use. Satisfaction: 87.5% of participants expressed satisfaction with image quality, with 43.75% describing the image quality as "adequate," 25% describing it as "good," and 18.75% describing the image quality as "excellent." Pathologists with a working bandwidth greater than 20 kilobits per second found the download speed of the Virtual Pathology Slide "adequate" or better.

Conclusions

Results from this study show that the Virtual Pathology Slide can be used to make a correct diagnostic decision, and that the system is a realistic alternative to dynamic telepathology.

Keywords: Telepathology, Internet, telemicroscopy, remote diagnosis, virtual slide, pathology, imaging

Introduction

Definition of Telepathology

Telepathology is the practice of diagnostic pathology by a remote pathologist utilizing images of tissue specimens transmitted over a telecommunications network [1,2]. Traditionally telepathology systems are defined as either dynamic or static. Dynamic systems allow a telepathologist to view images transmitted in real time from a remote robotic microscope that permits complete control of the field of view and magnification [3]. Static (or store-and-forward) telepathology involves the capture and storage of images followed by transmission over the Internet via e-mail attachment, file transfer protocol, or a Web page, or distribution via CD-ROM. Dynamic hybrids also exist, which incorporate aspects of both technologies [3].

Applications of Telepathology

The diversity in telepathology systems reflects growing technological expertise in this area and the increasing importance of telepathology in education, training, quality assurance, and teleconsultation [4- 9]. Numerous pathology archives abound on the Internet providing links to both educational and commercial telepathology websites. These offer access to either static or dynamic image delivery systems [7- 19].

Limitations of Telepathology

Image quality and the ability to make diagnostic decisions from electronically-compressed images is a contentious issue [3,19- 20]. In order for telepathology to be of clinical use, studies have attempted to access the diagnostic accuracy of store-and-forward telepathology, and have shown accuracy in the range of 77% to 100% [3,20- 28]. The diverse nature of this technology makes it difficult to draw comparisons between studies, or to form a consensus on a method of best practice. There is no universally-accepted standardization in hardware, software, image resolution, color-depth, or image compression and storage [3]. However, studies have shown that the use of images with as low a resolution as 1024 pixels x 768 pixels resolution x 24-bit color does not impair diagnostic performance [3,20,27- 29]. To contend with such nonstandardization, guidelines have been formulated for the capture and treatment of diagnostic images and for the practice of telepathology [30- 31].

Recent improvements in Internet-browser technology have facilitated the development of interactive store-and-forward Web pages. These feature the ability to show the spatial relationship between individual images in low-power and high-power views. This technology is commonly visualized using a small image gallery constructed from one or two microscopic fields out of a possibility of thousands, displaying images of the same fields at higher magnifications [3,20,21]. Field selection and interpretation are thought to be the primary reasons specific to store-and-forward telepathology that account for its discordance with diagnosis in a conventional pathology setting [19- 20]. Studies involving multiple pathologists provide the most robust and accurate method of assessing a telepathology technique [20- 27]. However it is difficult to distinguish the performance of the technology from the skill of the pathologist and the degree of difficulty of the cases being presented [21]. Until recently, the development of a tool for routine diagnosis and teleconsultation was the driving goal for the evolution of telepathology systems. Initial expense, lack of broadband Internet connections, potential liabilities, and a lack of knowledge transfer from expert to potential user have all contributed to preventing the incorporation of telepathology into everyday practice [32- 35]. The emerging role of telepathology in the area of education and quality assurance is not encumbered by the same difficulties. It has been demonstrated that the application of telepathology in such roles has the advantage of lower cost, less logistical effort, and a positive response to its use by the end user [36- 39]. Coupled with the growing presence of ultra-fast slide scanners, this should ensure an increasing role for telepathology in this area [40- 42].

The Virtual Pathology Slide (VPS)

To overcome problems attributable to sampling bias and interpretation resulting from limited field selection, telepathologists must be able to navigate to any field of view, at magnifications comparable to that of a conventional microscope, using images of sufficient resolution to render a correct diagnosis [21,22,35]. To meet such criteria we have developed the Virtual Pathology slide (VPS) [43,44]. This is a microscope emulator that displays digitized representations of tissue slides, allowing inspection of numerous fields of view, over a wide range of magnifications. Similar applications, commonly referred to as Virtual Slides, have been developed by other commercial and academic bodies [16,18,20,34,35,40- 42]. A screenshot of our VPS is shown in Figure 1; further screenshots are in Multimedia Appendix 1.

Figure 1.

Figure 1

Screenshot of the Virtual Pathology Slide Web browser (for further screenshots, see Multimedia Appendix 1)

In an important new departure, the VPS can also record and quantify the diagnostic trace of a pathologist, as a discrete data set on a central server. This allows the decoupling of a pathologist's field selection from the technical functionality of the telepathology system. In this paper, we report on the development of this system, its acceptability among a group of evaluating pathologists, the level of diagnostic agreement among this group, and the potential future applications of the VPS in telepathology.

Methods

A comprehensive document detailing the scanning algorithm and system architecture of the VPS is in Multimedia Appendix 2.

Construction of the VPS

Development of VPS Imaging Workstation

To create VPS slides, an imaging workstation was developed in-house. An Olympus BX-40 microscope (Olympus, Melville, NY, USA) incorporating a 40x plan apochromat lens with a 0.95 numerical aperture was used. The microscope was fitted with a robotic stage (Prior Scientific Inc, Rockland, Mass, USA) and a JVC 3-CCD (3-chip charge-coupled device) video camera.

Development of VPS Slide Scanning Algorithm

Using Optimas 6.5 imaging software (Media Cybernetics, Inc, Silver Spring, Md, USA), an algorithm was written in ALI (Analytical Language for Images) to perform a raster scan of 15.53 mm x 11.61 mm (180 mm 2) of tissue at 40x objective magnification. The VPS raster scan acquires 128 x 128 images in the X and Y Cartesian directions, one row at a time. Each acquired image represents 0.011 mm 2 at a resolution of 768 pixels by 574 pixels. Images were saved using a JPEG (Joint Photographic Experts Group) format at 10% compression, resulting in image-file sizes in the range of 100 to 150 KB (kilobytes). To build layers of lesser magnification, a second algorithm was developed, which tiles and resizes multiple images from the raster scan into composite images [16]. Images were subsequently uploaded onto the VPS Web server.

Development of VPS Web Interface

To view images via the Internet a graphical user interface was constructed [45]. This is a Web page powered by server-side scripting in PHP (PHP = Hypertext Preprocessor). The interface emulates the experience of using a conventional microscope by allowing a user to increase or decrease magnification or move laterally while examining a tissue section.

A customized browser was developed to control the user's access to the VPS during dedicated studies, to optimize the integrity of recorded data, and to provide a uniform experience for users who would otherwise experience subtle differences due to variation in currently-existing versions of Web browsers.

The VPS browser is a Microsoft Foundation Class (MFC) application written in Visual C++, which utilizes Internet Explorer file libraries to behave as a customized browser. The VPS customized browser opens up prescribed Web pages on the VPS server. The VPS browser is optimized for PC users with Microsoft Internet Explorer 5 or greater.

Development of VPS Database

When a user examines a VPS slide, data describing the user's interaction with the VPS is transmitted from the user's workstation to the VPS server and stored in an Oracle database. The VPS examination database is structured to contain the following data types:

System Configuration Data

This consists of data automatically recorded on the VPS server and includes parameters such as user's browser version, operating system, screen resolution, screen color depth, and IP (Internet Protocol) address.

User Tracking Data

This data records a user's "diagnostic pattern" as the user examines a slide. Information recorded includes image file name, image magnification, and the time spent viewing each image.

User-submitted Data

Diagnostic and descriptive data is submitted to the VPS server by participants, using HTML (Hypertext Markup Language) forms. Information recorded includes the report submitted by the user at the end of each slide examination and a final questionnaire. The observer also has the option to record or annotate every field of view examined.

VPS Deployment

The user has two choices on how he or she wishes to use the system. Users with high-speed Internet access can download the VPS browser from the VPS homepage and view images downloaded directly from the VPS server. To accommodate users with slow Internet connections, users may launch the VPS browser from a VPS CD and view images stored on the CD. However, an Internet connection is still required to record data on the VPS database, and to provide essential data for statistical analysis and playback facilities.

Validation of the VPS

Slide Selection

Ten needle core biopsies were obtained from the Department of Pathology, Mater Misericordiae Hospital, Dublin, Ireland. The slides were randomly selected by a pathologist (P.A.D.) with a special interest in breast pathology. The slides represented a range of diagnostic classifications. Two of the slides are presented in Figure 2. All 10 slides can be viewed in Multimedia Appendix 3.

Figure 2.

Figure 2

Two examples of the 10 breast needle core biopsies presented to 17 pathologists or trainee pathologists using the VPS (for all 10 images, see Multimedia Appendix 3)

Participants

Fifty-four pathologists with at least 2 years experience in pathology practice registered for the study. Of the 54 pathologists, 17 examined all 10 slides and 8 initiated the study but did not complete it. Of the 17 participants who completed the study, 8 were members of the European Working Group of Breast Screening Pathology. Of the 17 participants who examined all 10 slides, 13 subsequently completed a questionnaire on user perception of the VPS. Of the 8 participants that initiated the study but did not complete it, 3 completed the questionnaire.

Examination Procedure

Upon launching the VPS browser, participants were prompted to log in using the username and password they received at registration. This made them identifiable to the system. On successful log-in, the VPS needle core examination guidelines [46] were displayed.

After stating they read the guidelines, users were permitted to browse the slides available for examination and select one from a slide gallery. The slide gallery displayed a thumbnail image of each slide and indicated the patient's age and sex, and a brief case description.

Upon selecting a slide for examination, participants were presented with the VPS user interface. While examining a slide, participants could if desired annotate the fields of view using the text area provided. Upon completing a slide examination, participants submitted an online report that provided a diagnostic classification for the case, using an adaptation of the Core Biopsy Reporting Guidelines for Non-operative Diagnostic Procedures and Reporting in Breast Cancer screening [47] as used by the British National Co-ordinating Committee for Breast Screening Pathology. Users were requested to classify the slides as one of the following:

B1: Unsatisfactory/normal tissue only.

B2: Benign.

B3: Benign but of uncertain malignant potential.

B4: Suspicious of malignancy.

B5: Malignant.

For slides categorized as B5, participants were required to subclassify their decision as malignant, in-situ, or invasive. Upon making a classification, participants were returned to the slide gallery from which another slide could be selected for examination.

Utilization of this data allowed the following to be determined:

  • Percentage concordance for a user, calculated as the number of slides (expressed as a percentage) for which the user's diagnosis is in agreement with the consensus VPS diagnosis.

  • Percentage concordance of a slide, calculated as the percentage of users who concur as to the correct diagnosis of a slide.

  • Cohen's Kappa [48- 49], a measure of agreement between observers taking into account agreement that could occur by chance. Kappa values range from 0 to 1 with a score greater than 0.7 indicating "substantial agreement."

Participants who completed examination of the 10 slides were subsequently requested to complete an online questionnaire describing their experience using the VPS. Participants were asked to give a subjective evaluation of diagnostic confidence in using the VPS, reasons for uncertainty, an evaluation of image quality, and perceived download speed.

Results

User Performance Using the VPS

Table 1 shows strong diagnostic agreement between original glass-slide diagnosis and the most-common diagnosis offered by users of the VPS, with agreement being reached in 9 out of the 10 slides. Disagreement by 1 diagnostic degree occurred with slide 8 (glass slide diagnosis was B3; most- common VPS diagnosis was B4). The diagnostic classification of slide 8 had the lowest level of agreement between participants at 38.5%. The second most popular choice for slide 8 was split between B3 and B2, 6 participants (35.3% of users) classified it as B4 while 4 participants (23.5% of users) classified it as B3 and 4 participants (23.5% of users) classified it as B2). Participants with the 4 highest Kappa scores (23.5% of users) classified slide 8 as B4.

Table 1.

Comparison of glass slide needle core surgical biopsy diagnosis and most-common Virtual Pathology Slide (VPS) diagnosis, in order of level of agreement (concordance) for each slide

Virtual Pathology Slide
S6 S2 S3 S4 S7 S9 S10 S1 S5 S8
Diagnosis Glass B5 B5 B5 B2 B2 B2 B2 B5 B5 B3
Diagnosis VPS B5 B5 B5 B2 B2 B2 B2 B5 B5 B4
Concordance, % 100 94.1 82.4 76.5 64.7 58.8 52.9 52.9 47.1 35.3
Fields of view* 243 208 464 410 463 309 591 431 487 299

A more-detailed analysis of the diagnostic classifications made by participants is described in Table 2. The average percentage concordance between participants on all cases was 66.5%. Of the 17 participants, 14 attained a percentage concordance of between 90% and 60%.

Table 2.

For each participant: years of experience in pathology practice, diagnostic classification of slides, level of agreement with each other (% concordance and Kappa index), and number of fields of view examined

ID* EXP Virtual Pathology Slide Concordance,% Kappa§ Fields of View
S6 S2 S3 S4 S7 S9 S10 S1 S5 S8
5 5 B5 B5 B5 B2 B2 B2 B2 B4 B5 B4 90 0.97 321
62 5 B5 B5 B5 B2 B2 B3 B2 B5 B4 B4 80 0.94 326
35 5 B5 B5 B5 B2 B2 B1 B1 B4 B5 B4 70 0.94 122
10 5 B5 B5 B5 B2 B3 B3 B2 B5 B4 B4 70 0.91 157
39 5 B5 B5 B5 B1 B3 B3 B2 B5 B5 B5 60 0.90 343
55 5 B5 B5 B5 B2 B2 B2 B2 B5 B3 B3 80 0.87 289
87 5 B5 B5 B5 B1 B2 B2 B2 B3 B5 B3 70 0.86 130
18 3 B5 B5 B4 B1 B3 B2 B1 B5 B4 B3 40 0.86 252
68 5 B5 B5 B5 B2 B3 B2 B2 B5 B4 B2 70 0.85 228
65 5 B5 B5 B5 B2 B2 B3 B1 B3 B4 B4 60 0.80 234
22 5 B5 B5 B5 B2 B2 B2 B5 B5 B5 B5 80 0.75 204
41 5 B5 B5 B5 B1 B2 B2 B2 B2 B4 B4 70 0.75 216
7 5 B5 B5 B5 B2 B2 B2 B1 B4 B2 B3 60 0.73 121
1 5 B5 B5 B4 B2 B4 B2 B2 B5 B5 B1 70 0.67 223
75 5 B5 B5 B5 B2 B2 B2 B5 B4 B5 B2 70 0.65 120
36 3 B5 B5 B2 B2 B3 B3 B4 B5 B2 B2 40 0.26 201
6 5 B5 B2 B5 B2 B2 B4 B5 B3 B5 B2 50 0.23 418
Average 66.5 0.76 23

* ID = identification number of participant.

EXP = years of experience in pathology practice.

Concordance = number of slides (expressed as a percentage) for which the user's diagnosis is in agreement with the consensus Virtual Pathology Slide diagnosis.

§ Kappa = Cohen's Kappa, a measure of agreement between observers, taking into account agreement that could occur by chance. Kappa greater than 0.7 indicates "substantial agreement."

The average Kappa value achieved by participants was 0.76. Participants 36 and 6 achieved a Kappa of 0.26 and 0.23 respectively indicating "fair agreement" [31- 32] with other participants while the remaining 15 participants achieved a Kappa of between 0.97 and 0.65.

The average percentage concordance for slides was 66.5% with a minimum concordance of 35.3% for slide 8 and a maximum concordance of 100% for slide 6. The percentage concordance for slide 5 was 47%. For all remaining slides there was greater than 50% agreement between participants.

The average number of fields of view examined by each participant was 23 per slide. Participant number 5, who achieved the highest Kappa, examined 321 views, while participant number 6, who had the lowest Kappa, examined 418 fields of view.

The highest number of number of fields of view examined for a particular slide was 118 by participant number 6 while examining slide 10. This slide had a percentage concordance between participants of 52.9%. The lowest number of views examined while examining a slide was 3; this was by participant 10 who achieved a Kappa score of 0.91 and agreed with the group consensus for slide 2. Diagnosis for slide 2 had a percentage concurrence amongst participants of 94%.

The average time taken for participants to examine a slide was recorded as 6 minutes 11 seconds. The maximum time taken to examine a slide was recorded as 12 minutes 49 seconds by participant number 36 with an average bandwidth of 20 kilobits per second while examining slide 7. The minimum examination time was recorded as 43 seconds by participant number 1 with an average bandwidth of 64 kilobits per second while examining slide 2.

User Perception of the VPS

Participants were asked to assess their own computer competency and the frequency with which they use a telepathology system. Participants described themselves as "advanced" (18.75%), "competent" (18.75%), or "adequately competent" (62.5%) with computers, while 44% of participants indicated they had never used a telepathology system prior to the study.

Figure 3 illustrates that 68.75% of participants rated the VPS "easy" (62.5%) to use or "very easy" to use (6.25%). Participants were requested to rate their degree of confidence in making a diagnostic decision using the VPS.

Figure 3.

Figure 3

Ease of use of the Virtual Pathology Slide (VPS). Yrs = years of pathology experience. Percentages = percentage of 16 participants for that rating

Figure 4 illustrates that 80.25% of participants expressed confidence in using the VPS with 56.25% indicating they were "reasonably confident," while 18.75% were "confident," and 6.25% were "very confident" in making a diagnosis.

Figure 4.

Figure 4

Degree of confidence in using the Virtual Pathology Slide (VPS) to make a diagnostic decision. Yrs = years of pathology experience. Percentages = percentage of 16 participants for that rating

Figure 5 illustrates that 87.5% of participants expressed satisfaction with the image quality with 43.75% indicating the quality as "adequate," 25% as "good," and 18.75% of participants indicating the image quality as "excellent."

Figure 5.

Figure 5

Perceived image quality of the Virtual Pathology Slide (VPS). Yrs = years of pathology experience. Percentages = percentage of 16 participants for that rating

Discussion

The VPS system is a realistic alternative to dynamic telepathology, in terms of its ability to mimic a conventional microscope, its accessibility via the Internet, and its simplicity of operation. Of the 17 participants, 15 achieved a Kappa of between 0.97 and 0.65 and 14 attained a percentage concordance of between 90% and 60%. This demonstrates "substantial" agreement between users when using the VPS [31- 32]. The calculation of Kappa was weighted to reflect the degree of variation of a participant's diagnostic decision from the most popular choice. For example, participant 18 achieved a high Kappa of 0.86 despite being in agreement with other participants for 4 out of the 10 slides. This is because for each of the other 6 slides, participant 18 was inconsistent with the popular choice by one degree. Participant 36 achieved the same percentage concordance as participant 18 but only achieved a Kappa of 0.26. This is because the diagnostic categories selected by participant 36 deviated to a greater degree from the popular choice than those selected by participant 18 [48- 49].

Participant 36 and participant 6 attained the lowest Kappa scores of 0.26 and 0.23 respectively. This reduced the overall average Kappa value considerably. Confidence in using the VPS was described as "reasonably confident" by participant 36, who had 3 years experience in pathology and examined 201 fields of view while examining the entire set of slides. Further analysis of the images viewed is necessary to elucidate reasons for the diagnostic decisions made by participant 36; however, inexperience with breast pathology coupled with insufficient examination of the slides may have contributed to poor performance. Use of telpathology was described as "infrequently" by participant 6 who was "confident" in making a diagnostic classification using the VPS and described the use of the VPS as "easy." However, participant 6 attributed some diagnostic uncertainty to "problems with assessing significance of small subtle lesions without having the whole slide to look at." Participant 6 examined 418 fields of view, the highest number examined by any participant.

The average percentage concordance for the entire set of slides was 66.5%. Full agreement between participants was achieved for slide 6, which demonstrates that full agreement can be achieved using the VPS.

The average number of views examined by participants while examining the entire set of slides was 230. The percentage concordance for a particular slide decreases as the average number of fields examined for that slide increases. For example, the average number of fields examined for slide 6 (100% concordance amongst participants) was 14.3, while the average number of fields examined for slide 10 (52.9% concordance amongst participants) was 34.8. Conversely, participants with a high Kappa score tend to view a greater number of fields of view than participants with a low Kappa score, suggesting that the greater the amount of tissue viewed by a pathologist, the more likely they are to make a correct diagnosis.

Slide 8 had the lowest level of concordance at 35.3%. This reduced the average percentage concordance for the set of slides by 3.46%. Table 2 shows there is a broad distribution of diagnostic categorization for slide 8 by participants. As shown in Table 1, for slide 8 the number of fields of view examined by participants is low (299) given the apparent complexity of the case. It is apparent that users are rapidly coming to a conclusion that usually does not concur with the original glass slide diagnosis. Further study of the examination traces from this slide will be required to evaluate the reasons for the diagnostic spread.

Participants with 3 years or less experience did not have access to broadband Internet connection and recorded bandwidth speeds of less than 15 kilobits per second. These participants expressed least satisfaction with the VPS in terms of ease of use, image quality, and diagnostic confidence. All 3 participants, who indicated they were "not confident," attributed difficulty in using the VPS to poor download speed, with comments such as "Poor download speed was extremely slow and made the viewing experience disjointed and basically unworkable." Of these 3 participants, 2 had a working bandwidth of 12.6 kilobits per second and 31.5 kilobits per second respectively. A bandwidth could not be determined for the third, however the third did offer such comments as "too long to download images" and "problem was on my end, slow connection." High-speed broadband Internet connectivity is still unavailable to many pathologists. This is a major limiting factor for acceptability of Web-driven telepathology due to the time taken to download large image files over the Internet [2,4,39]. We have attempted to overcome this with the development and deployment of a CD-ROM VPS system to selected participants. This facilitates rapid retrieval of images from a CD while data pertaining to the examination is transmitted and stored on the VPS web server.

Participants were asked to comment on improvements to the VPS that they would like implemented. A number of participants suggested they would like additional magnification ranges. For example, "Navigation within the slide was disjointed and it was difficult to maintain perspective whilst moving from field to field. The range of magnifications was too limited, especially in the intermediate magnification range."

There are a growing number of interactive pathology sites available via the Internet [7- 19]. The diversity in their principle of operation, their application in telepathology, and their degree of sophistication promises an encouraging future in telepathology. The contribution of the VPS to the field of telepathology is notable in that it records the diagnostic pathway of a pathologists slide examination. We now have the diagnostic traces of 17 pathologists examining 10 cases. We intend to utilize this data to elucidate the cognitive and decision-making process of pathologists as they render a diagnosis when using a microscope. This will provide valuable insight into interobserver variability and the subjective process of microscopic diagnosis.

Acknowledgments

The authors would like to thank the Irish Health Research Board (HRB), for their partial funding of this work. The authors would also like to thank the pathologists who took time to partake in this study and provided such useful and constructive feedback.

Abbreviations

VPS

Virtual Pathology Slide

Multimedia Appendix 1

Screenshots of the Virtual Pathology Web browser.

jmir_v5i2e11_app1.ppt (525KB, ppt)

Multimedia Appendix 2

Technical Documentation.

jmir_v5i2e11_app2.pdf (552KB, pdf)

Multimedia Appendix 3

Ten breast needle core biopsies presented to 17 pathologists or trainee pathologists using the VPS.

jmir_v5i2e11_app3.ppt (74KB, ppt)

Footnotes

None declared.

References

  • 1.Weinstein R S. Prospects for telepathology. Hum Pathol. 1986 May;17(5):433–4. doi: 10.1016/s0046-8177(86)80028-4. [DOI] [PubMed] [Google Scholar]
  • 2.Weinstein R S, Bloom K J, Rozek L S. Telepathology. Long-distance diagnosis. Am J Clin Pathol. 1989 Apr;91(4 Suppl 1):S39–42. [PubMed] [Google Scholar]
  • 3.Leong FJW-M, Graham AK, Gahm G, McGee JO'D. Telepathology clinical utility and methodology. In: Lowe David G, Underwood James C E., editors. Recent advances in Histopathology. 18. London: Churchill Livingstone; 1999. pp. 217–240. [Google Scholar]
  • 4.Brauchli K, Christen H, Haroske G, Meyer W, Kunze K D, Oberholzer M. Telemicroscopy by the Internet revisited. J Pathol. 2002 Feb;196(2):238–43. doi: 10.1002/path.1036.10.1002/path.1036 [DOI] [PubMed] [Google Scholar]
  • 5.Wells C A, Sowter C. Telepathology: a diagnostic tool for the millennium? J Pathol. 2000 May;191(1):1–7. doi: 10.1002/(SICI)1096-9896(200005)191:1&#x0003c;1::AID-PATH624&#x0003e;3.0.CO;2-S.10.1002/(SICI)1096-9896(200005)191:1<1::AID-PATH624>3.0.CO;2-S [DOI] [PubMed] [Google Scholar]
  • 6.Petersen I, Wolf G, Roth K, Schlüns K. Telepathology by the Internet. J Pathol. 2000 May;191(1):8–14. doi: 10.1002/(SICI)1096-9896(200005)191:1<8::AID-PATH558>3.0.CO;2-9.10.1002/(SICI)1096-9896(200005)191:1<8::AID-PATH558>3.0.CO;2-9 [DOI] [PubMed] [Google Scholar]
  • 7.Afework A, Beynon MD, Bustamante F, Cho S, Demarzo A, Ferreira R, et al. Proceedings AMIA Symposium 1998. Bethesda, MD: American Medical Informatics Association; 1998. Digital dynamic telepathology--the virtual microscope; pp. 912–916. http://www.cc.gatech.edu/systems/papers/schwan/Bustamante98DDT.pdf. [PMC free article] [PubMed] [Google Scholar]
  • 8.Taylor R N, Gagnon M, Lange J, Lee T, Draut R, Kujawski E. CytoView. A prototype computer image-based Papanicolaou smear proficiency test. Acta Cytol. 1999;43(6):1045–51. doi: 10.1159/000331352. [DOI] [PubMed] [Google Scholar]
  • 9.Barbareschi M, Demichelis F, Forti S, Dalla Palma P. Digital Pathology: Science Fiction? Int J Surg Pathol. 2000 Oct;8(4):261–263. doi: 10.1177/106689690000800401. [DOI] [PubMed] [Google Scholar]
  • 10.Telepathology City, authors. Commercial solutions. [2003 Apr 17]. http://www.telepathologycity.com/systems.htm.
  • 11.Department of Pathology, University of Pittsburgh School of Medicine, authors. On line case database. [2003 Apr 17]. http://path.upmc.edu/cases.html.
  • 12.The Johns Hopkins Medical Institutions, Department of Pathology, Division of Hematopathology, authors. Flow cytometry tutorial. [2003 Apr 17]. http://162.129.103.34/leuk/toc.htm.
  • 13.University of Medicine and Dentistry of New Jersey, Department of Pathology, authors. Bone tumors: tutorial for residents. [2003 Apr 17]. http://www.umdnj.edu/tutorweb/index.htm.
  • 14.Committee on Quality Assurance Training and Education of the European Federation of Cytology Societies, authors. Home page. [2003 Apr 17]. http://crsg.ubc.kun.nl/quate/EFCS/EFCSindex.htm.
  • 15.Armed Forces Institute of Pathology, authors. Home page. [2003 Apr 17]. http://www.afip.org/
  • 16.Bacus Laboratories Inc, authors. Microscope imaging and virtual microscopy products for pathology, research and medical education. [2003 Apr 17]. http://www.bacuslabs.com/
  • 17.Histokom, authors. Telemedicine, telepathology, teleconsultation, telemicroscopy, teleendoscopie, telesonography, teledermatology. [2003 Apr 17]. http://www.ipe.uni-stuttgart.de/res/ip/histkome.html.
  • 18.Fairfield Imaging, authors. Software for digital pathology. [2003 Apr 17]. http://www.fairfield-imaging.co.uk/
  • 19.Walter-Reed Army Medical Center, authors. Pathology and area laboratory sciences: pathology to go! [2003 Apr 17]. http://www.wramc.amedd.army.mil/departments/pathology/links.htm.
  • 20.Cross S S, Dennis T, Start R D. Telepathology: current status and future prospects in diagnostic histopathology. Histopathology. 2002 Aug;41(2):91–109. doi: 10.1046/j.1365-2559.2002.01423.x.1423 [DOI] [PubMed] [Google Scholar]
  • 21.Halliday B E, Bhattacharyya A K, Graham A R, Davis J R, Leavitt S A, Nagle R B, Mclaughlin W J, Rivas R A, Martinez R, Krupinski E A, Weinstein R S. Diagnostic accuracy of an international static-imaging telepathology consultation service. Hum Pathol. 1997 Jan;28(1):17–21. doi: 10.1016/S0046-8177(97)90273-2. [DOI] [PubMed] [Google Scholar]
  • 22.Weinberg DS, Allaert FA, Dusserre P, Drouot F, Retailliau B, Welch W R, Longtine J, Brodsky G, Folkerth R, Doolittle M. Telepathology diagnosis by means of still digital images: an international validation study. Hum Pathol. 1996 Feb;27(2):111–118. doi: 10.1016/s0046-8177(96)90363-9. [DOI] [PubMed] [Google Scholar]
  • 23.Weinstein M H, Epstein J I. Telepathology diagnosis of prostrate needle biopsies. Hum Pathol. 1997 Jan;28(1):22–9. doi: 10.1016/S0046-8177(97)90274-4. [DOI] [PubMed] [Google Scholar]
  • 24.Wold LE, Weiland LH. Telepathology at the Mayo. Clin Lab Manage Rev. 1992;(1):174–175. [Google Scholar]
  • 25.Shimosato Y, Yagi Y, Yamagishi K, Mukai K, Hirohashi S, Matsumoto T, Kodama T. Experience and present status of telepathology in the National Cancer Center Hospital, Tokyo. Zentralbl Pathol. 1992 Dec;138(6):413–7. [PubMed] [Google Scholar]
  • 26.Eide T J, Nordrum I. Frozen section service via the telenetwork in northern Norway. Zentralbl Pathol. 1992 Dec;138(6):409–12. [PubMed] [Google Scholar]
  • 27.O'brien M J, Takahashi M, Brugal G, Christen H, Gahm T, Goodell R M, Karakitsos P, Knesel E A, Kobler T, Kyrkou K A, Labbe S, Long E L, Mango L J, Mcgoogan E, Oberholzer M, Reith A, Winkler C. Digital imagery/telecytology. International Academy of Cytology Task Force summary. Diagnostic Cytology Towards the 21st Century: An International Expert Conference and Tutorial. Acta Cytol. 1998;42(1):148–64. doi: 10.1159/000331540. [DOI] [PubMed] [Google Scholar]
  • 28.Gombás Péter, Skepper Jeremy N, Hegyi László. The image pyramid system--an unbiased, inexpensive and broadly accessible method of telepathology. Pathol Oncol Res. 2002;8(1):68–73. doi: 10.1007/BF03033705. http://www.webio.hu/por/2002/8/1/0068/0068a.pdf.PAOR.2002.8.1.0068 [DOI] [PubMed] [Google Scholar]
  • 29.Foran D J, Meer P P, Papathomas T, Marsic I. Compression guidelines for diagnostic telepathology. IEEE Trans Inf Technol Biomed. 1997 Mar;1(1):55–60. doi: 10.1109/4233.594046. [DOI] [PubMed] [Google Scholar]
  • 30.Loane M, Wootton R. A review of guidelines and standards for telemedicine. J Telemed Telecare. 2002;8(2):63–71. doi: 10.1258/1357633021937479. [DOI] [PubMed] [Google Scholar]
  • 31.The American Telemedicine Association Special Interest Group for Telepathology, authors. Clinical guidelines for telepathology: a draft version. [2003 May 20]. http://telepathology.upmc.edu/ata/guideline.htm.
  • 32.Mairinger T, Netzer T T, Schoner W, Gschwendtner A. Pathologists' attitudes to implementing telepathology. J Telemed Telecare. 1998;4(1):41–6. doi: 10.1258/1357633981931263. [DOI] [PubMed] [Google Scholar]
  • 33.Della Mea V, Cortolezzis D, Beltrami C A. The economics of telepathology--a case study. J Telemed Telecare. 2000;6 Suppl 1(6 Suppl 1):S168–9. doi: 10.1258/1357633001934555. [DOI] [PubMed] [Google Scholar]
  • 34.Steinberg D M, Ali S Z. Application of virtual microscopy in clinical cytopathology. Diagn Cytopathol. 2001 Dec;25(6):389–96. doi: 10.1002/dc.10021.10.1002/dc.10021 [DOI] [PubMed] [Google Scholar]
  • 35.Demichelis F, Barbareschi M, Dalla Palma P, Forti S. The virtual case: a new method to completely digitize cytological and histological slides. Virchows Arch. 2002 Aug;441(2):159–64. doi: 10.1007/s00428-001-0561-1. [DOI] [PubMed] [Google Scholar]
  • 36.Demichelis Francesca, Della Mea Vincenzo, Forti Stefano, Dalla Palma Paolo, Beltrami Carlo Alberto. Digital storage of glass slides for quality assurance in histopathology and cytopathology. J Telemed Telecare. 2002;8(3):138–42. doi: 10.1258/135763302320118979. [DOI] [PubMed] [Google Scholar]
  • 37.Leong F J, Graham A K, Schwarzmann P, Mcgee J O. Clinical trial of telepathology as an alternative modality in breast histopathology quality assurance. Telemed J E Health. 2000;6(4):373–7. doi: 10.1089/15305620050503834. http://screening.iarc.fr/planningmanual.php. [DOI] [PubMed] [Google Scholar]
  • 38.Heidger Paul M, Dee Fred, Consoer Daniel, Leaven Timothy, Duncan James, Kreiter Clarence. Integrated approach to teaching and testing in histology with real and virtual imaging. Anat Rec. 2002 Apr 15;269(2):107–12. doi: 10.1002/ar.10078. [DOI] [PubMed] [Google Scholar]
  • 39.Mackinnon M. Broadband multimedia for education. J Telemed Telecare. 1998;4 Suppl 1(4 Suppl 1):48–9. doi: 10.1258/1357633981931434. [DOI] [PubMed] [Google Scholar]
  • 40.Microbrightfield, Inc, authors. Home page. [2003 Apr 17]. http://www.microbrightfield.com/index.html.
  • 41.Aperio Technologies, authors. Virtual microscopy solutions for analysis & discovery. [2003 Apr 17]. http://www.aperio.com/home.asp.
  • 42.Interscope Technologies, authors. The anatomy of a better way to work. [2003 Apr 17]. http://www.interscopetech.com.
  • 43.Costello SP, Johnston DJ, Dervan PA, O'Shea DG. The virtual pathology slide: a new internet telemicroscopy tool for tracing the process of microscopic diagnosis and evaluating pathologist behaviour [abstract] Archives of Pathology and Laboratory Medicine. 2002;126(7):781–802. http://arpa.allenpress.com/arpaonline/?request=get-document&doi=10.1043/0003-9985(2002)126%3c0781:APIIAT%3e2.0.CO;2. [Google Scholar]
  • 44.Costello SP, Johnston DJ, Dervan PA, O'Shea DG. Evaluation of the virtual pathology slide: using breast needle core biopsy specimens. Br J Cancer Suppl. 2002;86:S34. [Google Scholar]
  • 45.Virtual Pathology Slide, authors. Please select the slide you wish to view. [2003 May 21]. http://www.telepathology.dcu.ie/index16.php3?route=guest.
  • 46.Virtual Pathology Slide, authors. Virtual pathology slide document archive. [2003 May 21]. http://www.telepathology.dcu.ie/document.htm.
  • 47.Non-operative Diagnosis Subgroup of the National Coordinating Group for Breast Screening Pathology, authors. Guidelines for non-operative diagnostic procedures and reporting in breast cancer screening. [2002 Nov 13]. http://www.cancerscreening.nhs.uk/breastscreen/publications/nhsbsp50.pdf.
  • 48.Cohen J. A coefficient of agreement for nominal scales. Educ Psychol Meas. 1960;(20):37–46. [Google Scholar]
  • 49.Landis J R, Koch G G. The measurement of observer agreement for categorical data. Biometrics. 1977 Mar;33(1):159–74. [PubMed] [Google Scholar]

Articles from Journal of Medical Internet Research are provided here courtesy of JMIR Publications Inc.

RESOURCES