Abstract
Background
Validation of the autoverification function is one of the critical steps to confirm its effectiveness before use. It is crucial to verify whether the programmed algorithm follows the expected logic and produces the expected results. This process has always relied on the assessment of human–machine consistency and is mostly a manually recorded and time-consuming activity with inherent subjectivity and arbitrariness that cannot guarantee a comprehensive, timely and continuous effectiveness evaluation of the autoverification function. To overcome these inherent limitations, we independently developed and implemented a laboratory information system (LIS)-based validation system for autoverification.
Methods
We developed a correctness verification and integrity validation method (hereinafter referred to as the "new method") in the form of a human–machine dialog. The system records personnel review steps and determines whether the human–machine review results are consistent. Laboratory personnel then analyze the reasons for any inconsistency according to system prompts, add to or modify rules, reverify, and finally improve the accuracy of autoverification.
Results
The validation system was successfully established and implemented. For a dataset consisting of 833 rules for 30 assays, 782 rules (93.87%) were successfully verified in the correctness verification phase, and 51 rules were deleted due to execution errors. In the integrity validation phase, 24 projects were easily verified, while the other 6 projects still required the additional rules or changes to the rule settings. Taking the Hepatitis B virus test as an example, from the setting of 65 rules to the automated releasing of 3000 reports, the validation time was reduced from 452 (manual verification) to 275 h (new method), a reduction in validation time of 177 h. Furthermore, 94.6% (168/182) of laboratory users believed the new method greatly reduced the workload, effectively controlled the report risk and felt satisfied. Since 2019, over 3.5 million reports have been automatically reviewed and issued without a single clinical complaint.
Conclusion
To the best of our knowledge, this is the first report to realize autoverification validation as a human–machine interaction. The new method effectively controls the risks of autoverification, shortens time consumption, and improves the efficiency of laboratory verification.
Keywords: Autoverification, Correctness verification, Integrity validation, Human–computer interaction, Risk control, Laboratory information system
Background
Autoverification is a powerful tool for the batch processing of test results and has been widely used in recent years. It has obvious advantages in reducing reporting errors, shortening turnaround time and improving audit efficiency [1–5].
Current status and challenges
Our self-developed autoverification system has been used for 6 years in many disciplines, such as biochemistry, immunology, hematology, microbiology, molecular analysis and pathology. To date, 25,487 rules have been set. The system judges test results 1.1 million times a day and provides audit recommendations for 250,000 report forms, accounting for 87% of the total number of report forms. Approximately 80,000 reports are automatically generated every day. To ensure the effectiveness and safety of the autoverification system, its validation process is very important. The College of American Pathologists clauses GEN43875 [6] and ISO 15189:2012 [7] 5.9.2b both require that autoverification systems undergo functional verification before use.
According to published studies, in laboratories that use autoverification, the majority of laboratories have performed personnel-based and automatic system audits with the same results, manually recorded consistency, and reached a conclusion after a statistical analysis of the results [2, 4, 8, 9]. The manual verification method is less difficult to operate but has the following limitations:
Massive validation workload. Based on the requirements of WS/T 616-2018 (China Health Organization recommended standard) [10] for validation of the autoverification of quantitative clinical laboratory test results, every test and every sample type involved in the autoverification procedure should be tested; the validation time should be no less than 3 months and/or the number of reports released should be no less than 50,000; and periodic verification should be performed every year for no less than 10 working days and/or for no less than 5000 reports. The validation workload is large, and it is difficult to rely on manual comparison and recording, which greatly increases the postanalytical workload.
Reporting risk [2]. During manual verification, personnel are prone to inertia or judgment errors. The lack of a system control mechanism for this kind of validation can generate reporting risks and directly affect clinical diagnosis and treatment.
Therefore, there is an urgent need to design a verification method that minimizes the workload and systematically controls risks. We report a rule verification system with a small workload and ease of operation that can be used as a reference for self-built and automatic test auditing for laboratories and manufacturers.
Methods
System design
Based on the American Clinical and Laboratory Standards Institute (CLSI) AUTO-10 [11] standards and current review processes, we established an autoverification system including 11 rule categories. Technicians set the rules according to audit requirements and rule categories. Each item can set multiple rules, including limited range check, combined mode judgment, Delta check, sampling time validity judgment, sample abnormality (hemolysis, lipemia) judgment and quality control check. The autoverification system determines whether the report is abnormal according to the rules. Tests that do not trigger contradiction mode are displayed in green, while failed tests (triggering rules, contradictory modes set by the rules) are displayed in red, and the cause of the contradiction is indicated. If all the tests in the report are green, the barcode of the report is also green. If any test in the report is red, the report shows a red barcode, which signals a warning in the system.
According to the above steps, the autoverification system displays colors and abnormal prompts after judging the rules in a process called automatic early warning. The automatic warning is only for judgment and is not involved in the decision to issue a report. Based on this, the system automatically sends out a report with a green barcode in a process called automated reporting. Automatic early warning and automatic reporting comprise autoverification. This system is especially useful in the review of complex diagnostic projects (e.g., molecular testing, pathological testing). These projects prompt absurd values from personnel. For some moderately complex projects (e.g., biochemical, blood), the combination of report reviewing, automatic warning and automated reporting is equivalent to the autoverification system in a large number of literature reports and laboratory information system (LIS) automatic reports. The autoverification process used by our laboratory is shown in Fig. 1.
Validation scheme
On the premise that automatic audits are divided into automatic warnings and automatic reports, we divide the verification system into two stages. The first stage is called correctness verification, which verifies that the operation of the rules is consistent with the expectations set by the personnel. If there is a problem, the responsible party may be the program development department. The second stage is called integrity validation. Based on the results from the first stage, this stage verifies whether the set rules include all the elements from the personnel’s audit report. The functional design of the two-stage system is shown in Table 1.
Table 1.
Phase | Object | Validation method | Explanation | Inconsistent solutions |
---|---|---|---|---|
Automatic warning | Warning rules | Correctness verification | To verify that the warning rules behave as expected and produce the expected outcome | If the warning rule setting is wrong, delete and reset the rules |
Automated reporting | Laboratory tests | Integrity validation | To confirm that the laboratory test results that pass the automatic warning can be reported automatically | Add more warning rules according to the laboratory report criteria |
Correctness verification
The correctness verification phase confirms whether the execution of a single rule is correct. It is implemented as follows: (1) For newly added rules, the system adds the label "Pending Verification". (2) When the report is reviewed, the system displays the rule judgment result, and a purple color block is displayed to remind the staff to judge whether the execution result of the "Pending Verification" rule is correct. (3) The staff input the judgment result. (4) The system changes the rule status according to the staff input. If it is consistent, the rule label is set to "verified", prompting the personnel to continue to the next stage of verification. If it is inconsistent, the staff are prompted to delete the rule. Figure 2 is a schematic diagram of the correctness verification using the example of C-reactive protein (CRP).
Integrity validation
Integrity validation can be started only after the correctness verification of all rules of a project is completed. It is implemented as follows. (1) After the report shows the result of the automatic warning, if the system detects that the report has been changed, a dialog box will pop up and ask the reviewer to select the reason for the modification. These reasons include (a) a rule execution error, (b) a rule setting value that is inappropriate, (c) the required addition of new rules, (d) the lack of involvement of other issues related to automatic review, and (e) automatic warning and prompt modification. The LIS records the modified content and the reasons for personnel analysis. (2) If the laboratory wants to implement automated reporting, a validation number, such as 5000, can be set according to the complexity of the project review. (3) If the automatic warning result of the report is green (approved), the personnel will issue the report directly, and the validation number of the report will automatically increase by one. (4) If the validation number of all items on the report exceeds the set number, the report will be automatically released. (5) If the automatic warning result of the report is green (approved), but the result is modified, with the reason for the modification specified as any of a, b or c, then the LIS will clear the validation number for the related items and stop automated reporting. Figure 4 shows the integrity validation process. The validation goals and validation amount for six projects are shown in Fig. 5.
Accuracy guarantee
The accuracy of the autoverification includes whether the rules can be identified and whether there are omissions (completeness) in the report review. Therefore, our method confirms the accuracy of the new method from these two aspects. We perform function correction and system improvement through correctness verification and integrity validation. In the validation system, we design the following logic to ensure the accuracy of the function:
The new rule is automatically deleted if it fails the correctness verification within 10 days;
The rule is not allowed to be modified;
If the rule fails the correctness verification, it is forbidden to be converted;
If the autoverification of a single project fails the integrity validation, the historical validation amount is cleared.
Data collection
The validation data of 30 assays from October 2019 to January 2020 were collected for analysis, and in total, 833 early warning rules were obtained. A total of 926,195 reports was used to evaluate the accuracy of the new method.
Time consumption statistics
We used HBV as an example to introduce the comparison of the validation time before and after the new method was used. In the measurement of the validation time, we divided the complete autoverification into 10 stages. Time statistics were collected for manual verification and new method verification for each step. We used systematic records and estimates to develop time statistics for different stages.
Satisfaction survey
We used questionnaires to evaluate the effectiveness of new methods used by laboratory technicians. The survey was launched using the online tool WJX (www.wjx.cn) which feeds back the percentage of responses and the total numbers.
Results
Correctness verification results
Among the 833 rules, 782 (93.88%) were successfully verified for correctness, with a total of 3814 validations, including 2230 (58.47%) released tests and 1584 (41.53%) intercepted tests. The inconsistencies were verified, and 51 (0.06%) error rules were deleted. The reasons for verification failure are shown in Table 2.
Table 2.
Error type | Proportion (%) | Sample | Solution |
---|---|---|---|
Human error | 63.3 | Incorrect English letter case in the text of the rules, resulting in no warning | Reset the rules |
Specific warning target | 24.9 | Early warning of diagnostic results and microscopy results in a special report interface for pathology | Add a supplementary algorithm code |
Algorithm code error | 8.4 | HPV typing results could not be verified with the Delta Check; the results of the microbial project identification could not be correlated with a variety of drug sensitivity combinations | Fix the algorithm code |
Software compatibility problem | 3.4 | Problem with the precision of the number comparison script | Fix the algorithm code |
Integrity validation results
We collected integrity validation data through system export and department feedback. The reasons provided for rule modification were automatic warning prompt and rule modification (5, 10.6%), rule execution error (0, 0%), improper setting values (15, 31.9%), new rule added (18, 38.3%), and no automatic warning involving other questions (9, 19.2%). The integrity of all projects was verified within 1 month, and the problems found are shown in Table 3.
Table 3.
Test | Reason for not passing | Solution |
---|---|---|
HPV genotyping | There was no comprehensive analysis of the combined thin-layer cytology results | Analyze the results associated with thin-layer cytology |
Urea | The limit range was too wide | Reduce the limit range |
Albumin | Review of the detection system produces an error | Specify the detection system |
CBC | Test results were checked only on the same day as the barcode | Extend the backdating of the historical results |
HBsAg HBsAb HBeAg HBeAb HBcAb | Not all composite mode scenarios were covered | Add a joint audit of the portfolio project results |
Cortisol | There was no warning of abnormal rhythms | Add a rule about checking sampling time |
Comparison of the two methods
The comparison of manual record analysis and the new method for different steps is shown in Table 4. The new method performs 4 automation steps, reduces the personnel workload, and automatically controls the enabling and disabling of automatic report release through system monitoring report modification. The increased accuracy verification can quickly eliminate rule setting exceptions and development loopholes while reducing the time needed for personnel analysis. The manual record analysis and the new method took 452 h and 275 h to complete, respectively.
Table 4.
Steps | Manual validation (h) | New method (h) |
---|---|---|
1. Set 65 rules | 1.5 | 1.5 |
2. Perform Rule 130 test | 2.5 | 2.5 |
3. Correctness verification | 0 | 0.25a |
4. Personnel comparison report and results review | 240 | 240 |
5. Record comparison result | 100 | 0b |
6. Analysis of the verification number | 10 | 0b |
7. Determine whether to activate automatic approval | 5 | 0b |
8. Personnel analysis of the reasons for inconsistent audit results | 90 | 30 |
9. Add and modify rules | 1 | 1 |
10. Determine whether to turn off autoverification | 1 | 0c |
Total | 452 | 275 |
In the measurement of the validation time, we divided the complete autoverification into 10 stages. The statistics of manual verification and the new method for each step are shown in Table 4. In steps 4–6, in total, 3000 reports are used for statistics. The time consumption of the consistent work content in the new and old methods is subject to the following: the manual timing of the old method, such as steps 1, 2, 4, and 9; the inconsistent steps in the two methods; the new steps that are recorded in the system, such as step 3; the saving step time clearing, such as steps 5, 6, 7, and 10; and the remaining steps that are estimated, such as step 8
For automatic implementation, the time is calculated as zero
aReasons for invalid locking rules
bReduced workload
cControlled risks
Satisfaction survey
After using the new method for 1 year, we conducted a satisfaction survey of laboratory personnel who used the function, distributed 182 questionnaires, and recovered 168 copies, with a response rate of 92.3%. The survey results showed that 94.6% of laboratory users believed that the new method could greatly reduce the workload, effectively control the report risk, and produce satisfactory or very satisfactory assessments of the new method.
Discussion
The core of the use of autoverification lies in the validation of system functions and rules. Due to the complexity of these rules, it is impossible to find all the functional defects by relying solely on function validation before the system goes online, and even human input errors cannot be carried out in the validation [12]. Such functional defects must be found in actual application scenarios with multiple different rule settings, such as the incorrect input of full-width symbols, that is, correctness verification. Furthermore, the premise of rule verification is to include a review of the logic of all reviewers in the system, which can be discovered only in actual application scenarios. Additionally, integrity validation can be performed in actual application scenarios to truly find problems [13, 14].
We initially designed the system in two parts, automatic warning and automated reporting, to allow complex detection items (molecular and pathological examinations, final human reports, and system prompt errors) to be included in the automatic review. Laboratory technicians can then choose to address the needs of different measurements. These two parts correspond to two verification steps: the automatic warning part performs correctness verification, and the automated reporting part performs integrity validation.
Compared with other systems reported in the literature [4], the advantages of the new method are mainly simplifying the verification process, reducing the verification workload and ensuring the accuracy of the verification results. As shown in Fig. 6, in the manual verification scheme, the junior staff will review the results, and then the intermediate staff will repeat reviewing the results and combine the autoverification results to determine the human–machine consistency to complete the autoverification function validation. The whole process focuses on human leadership. The new method uses system monitoring to judge the accuracy of the autoverification based on the operational trajectories of different personnel. By interacting with personnel, the system collects validation data and controls the operation of autoverification.
We delineate the advantages of the new method compared to manual validation in Table 5.
Table 5.
Advantages | Difference | Manual validation | New method | Explanation |
---|---|---|---|---|
Efficiency improvement | Whether to add extra workload? | YES | NO | No additional personnel are required to manually record the reason for the inconsistency. The new method is that the system completes judgment and records while personnel review the reports normally. The system will control the operation of the autoverification program based on the consistency results |
Can the cause of inconsistency be quickly determined? | NO | YES | The main reasons for the inconsistency are abnormal rule settings and lack of necessary rules. The new method correspondingly sets up correctness verification and integrity validation for these two main reasons. In different verification stages, only the main reason for that stage can be traced back | |
Risk control | Is it possible to skip the validation process? | YES | NO | Starting from setting the rules, the system will pull the validation process, and no validation link can be skipped |
Whether to ensure sufficient amount of validation data? | NO | YES | In the process of normal personnel issuance, the system will truthfully record the validation data. Before the set data volume is reached, the automated reporting function is prohibited | |
Can autoverification be used in the case of failed validation? | YES | NO | When the system confirms that the validation fails due to a defect in the autoverification, it will prohibit the rule conversion or the automated reporting from being enabled |
Compared with the traditional method, the true positives and false positives of the "personal and machine-based audit results" are easy to understand, but if the indicators are abnormal, it can be difficult to find the cause of this abnormality, especially after all the reasons are verified after thousands of reports are released [15]. Consequently, the audit scenario has become blurred in the auditor's memory, and it becomes inefficient to check the problems one by one. The process-based validation scheme that we developed is more practical and advantageous: (1) It can be easily operated and quickly initialized; (2) its self-traction and control of online functions can ensure that every rule is fully verified; and (3) the amount of manual work is small, allowing technicians to complete the verification steps during their daily work.
We divided the entire validation into two modules, correctness verification and integrity validation, based on the concept of process management. Rules are the basic unit of the entire autoverification system. If basic rule verification is not performed at the beginning of the entire process, when the human–machine judgment is inconsistent, it is difficult to confirm whether the problem is caused by algorithm error, execution error or another reason, inevitably increasing the analysis workload. In contrast, if correctness verification is completed when the rules are established, the only reason for an inconsistency between man and machine during the release of the report issuance would be "rule omission", requiring the technician to add only the corresponding rules.
During the entire verification process, we implemented human–computer interaction, which includes the following:
An "expected sense of play": Before the laboratory personnel view the results, they already possess a logical expectation, and in the process, they establish a comparison of the rules and effects;
The use of visual stimulation methods (red, green, and purple backgrounds) that can be quickly identified and relax the laboratory personnel; and
System pull—once the verification succeeds or fails, it is automatically counted with the click of a button, which automatically opens the automatic report function. All the functions ensure that laboratory personnel, particularly those of the new generation, can derive enjoyment from completing the verification process, thus increasing its core value [16].
According to the experience of this research, the logic of the autoverification validation process is not difficult, but if it is applied to other laboratories on a large scale, the intermediate software supplier needs to develop the original autoverification system. The validation system is based on the autoverification system developed by our laboratory, so it is more compatible in adding new functions. However, as a supplementary function, it is difficult to graft to existing systems. We suggest that peers can refer to the program logic provided in this study. On the basis of the current functions, we will further strengthen the learning ability of the validation system and convert validation records into learning cases that can serve as a guide for laboratory technicians to use the autoverification function more efficiently.
Conclusions
In the 2 years that the online validation has been in use, there have never been any defects or reporting risks due to autoverification. We believe that for both intermediate and self-built autoverification systems, online validation is a useful tool for controlling the risks of autoverification and improving the quality of reports. The detailed process for this method can serve as reference for the development and implementation of LIS-based autoverification systems.
Acknowledgements
We would like to gratefully acknowledge the technicians working in the laboratory for their helpful collaboration. In addition, we thank Xinyu Li and Jiazhen Ren for their technical support.
Abbreviations
- LIS
Laboratory information system
- TAT
Turnaround time
- CLSI
American Institute of Clinical and Laboratory Standardization
- AUTO 10-A
Autoverification of Clinical Laboratory Test Result 10-A; Approved Guideline
- CRP
C-reactive protein
- CBC
Complete blood cell count
- HBsAg
Hepatitis B virus surface antigen
- HBsAb
Hepatitis B virus surface antibody
- HBeAg
Hepatitis B virus e antigen
- HBeAb
Hepatitis B virus e antibody
- HBcAb
Hepatitis B virus core antibody
- HBV
Hepatitis B virus
- HPV
Human papilloma virus
Authors' contributions
All of the authors had full access to all of the data in the study and taking responsibility for the content of the manuscript. RT conceived and designed the study. DJ, QW, BJL, DZP and JJW performed the case and sample collection, analysis, and interpretation of the data. YTC, XYD and NXM performed the analysis with constructive discussions.DJ wrote the first draft of the paper. RT reviewed and approved the final manuscript. All authors have read and approved the final manuscript.
Funding
This research was supported by Guangdong Medical Science and Technology Research Fund (Program Grant A2020597). The funding bodies were not involved in the design of the study and collection, analysis, and interpretation of data and in writing the manuscript.
Availability of data and materials
All data generated or analyzed during this study are included in this published article. The data underlying this study are available and researchers may submit data requests to the corresponding author on reasonable request.
Declarations
Ethics approval and consent to participate
This study was approved by the ethics review board of KingMed Diagnostics. The study adhered to relevant guidelines and regulations. The patient consent was waived by the approving ethics review board, as utilization of anonymized history data does not require patient consent.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Li J, Cheng B, Ouyang H, Xiao T, Hu J, Cai Y. Designing and evaluating autoverification rules for thyroid function profiles and sex hormone tests. Ann Clin Biochem. 2018;55(2):254–263. doi: 10.1177/0004563217712291. [DOI] [PubMed] [Google Scholar]
- 2.Wang Z, Peng C, Kang H, Fan X, Mu R, Zhou L, He M, Qu B. Design and evaluation of a LIS-based autoverification system for coagulation assays in a core clinical laboratory. BMC Med Inform Decis Mak. 2019;19(1):123. doi: 10.1186/s12911-019-0848-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Wu J, Pan M, Ouyang H, Yang Z, Zhang Q, Cai Y. Establishing and evaluating autoverification rules with intelligent guidelines for arterial blood gas analysis in a clinical laboratory. SLAS Technol. 2018;23(6):631–640. doi: 10.1177/2472630318775311. [DOI] [PubMed] [Google Scholar]
- 4.Randell EW, Short G, Lee N, Beresford A, Spencer M, Kennell M, Moores Z, Parry D. Strategy for 90% autoverification of clinical chemistry and immunoassay test results using six sigma process improvement. Data Brief. 2018;18:1740–1749. doi: 10.1016/j.dib.2018.04.080. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Randell EW, Short G, Lee N, Beresford A, Spencer M, Kennell M, Moores Z, Parry D. Autoverification process improvement by Six Sigma approach: clinical chemistry & immunoassay. Clin Biochem. 2018;55:42–48. doi: 10.1016/j.clinbiochem.2018.03.002. [DOI] [PubMed] [Google Scholar]
- 6.College of American Pathologists. Laboratory general checklist. Commission on Laboratory Accreditation. Laboratory Accreditation Program; 2013.
- 7.Medical laboratories—requirements for quality and competence (ISO 15189:2012, corrected version 2014-08-15); German version EN ISO 15189:2012; 2014.
- 8.Palmieri R, Falbo R, Cappellini F, Soldi C, Limonta G, Brambilla P. The development of autoverification rules applied to urinalysis performed on the AutionMAX-SediMAX platform. Clin Chim Acta. 2018;485:275–281. doi: 10.1016/j.cca.2018.07.001. [DOI] [PubMed] [Google Scholar]
- 9.Sediq AM, Abdel-Azeez AG. Designing an autoverification system in Zagazig University Hospitals Laboratories: preliminary evaluation on thyroid function profile. Ann Saudi Med. 2014;34(5):427–432. doi: 10.5144/0256-4947.2014.427. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Autoverification of Clinical Laboratory Quantitative Test Results. In: National Health Commission of the People's Republic of China; 2018.
- 11.CLSI: Autoverification of Clinical Laboratory Test Results; Approved Guideline, AUTO10AE; 2006.
- 12.van Rossum HH. An approach to selecting auto-verification limits and validating their error detection performance independently for pre-analytical and analytical errors. Clin Chim Acta. 2020;508:130–136. doi: 10.1016/j.cca.2020.05.026. [DOI] [PubMed] [Google Scholar]
- 13.Krasowski MD, Davis SR, Drees D, Morris C, Kulhavy J, Crone C, Bebber T, Clark I, Nelson DL, Teul S, et al. Autoverification in a core clinical chemistry laboratory at an academic medical center. J Pathol Inform. 2014;5(1):13. doi: 10.4103/2153-3539.129450. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Jones JB. A strategic informatics approach to autoverification. Clin Lab Med. 2013;33(1):161–181. doi: 10.1016/j.cll.2012.11.004. [DOI] [PubMed] [Google Scholar]
- 15.Fu Q, Ye C, Han B, Zhan X, Chen K, Huang F, Miao L, Yang S, Wang W. Designing and validating autoverification rules for hematology analysis in Sysmex XN-9000 hematology system. Clin Lab. 2020;66(4):549–556. doi: 10.7754/Clin.Lab.2019.190726. [DOI] [PubMed] [Google Scholar]
- 16.Guidi GC, Poli G, Bassi A, Giobelli L, Benetollo PP, Lippi G. Development and implementation of an automatic system for verification, validation and delivery of laboratory test results. Clin Chem Lab Med. 2009;47(11):1355–1360. doi: 10.1515/CCLM.2009.316. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
All data generated or analyzed during this study are included in this published article. The data underlying this study are available and researchers may submit data requests to the corresponding author on reasonable request.