Table 4.
“Best Estimate” Event Prevalence and System Performance
Metric* | Derivation | Value (95% CI) | |
---|---|---|---|
Prevalence | |||
Case rate: proportion of cases with one or more true events | 53 ÷ 1,000 | 0.053 (0.040–0.069) | |
Event rate: true events per case |
65 ÷ 1,000 |
0.065 (0.051–0.082) |
|
System performance for detecting cases with events | |||
Sensitivity: proportion of cases with true events that had apparent events | 15 ÷ 53 | 0.28 (0.17–0.42) | |
Specificity: proportion of cases with no true events that had no apparent events |
|
0.985 (0.984–0.986)† | |
Positive predictive value: proportion of cases with apparent events that had true events | 652 ÷ 1,461 | 0.45 (0.42–0.47) | |
Negative predictive value: proportion of cases with no apparent events that had no true events |
930 ÷ 968 |
0.96 (0.95–0.97) |
|
System performance for detecting individual events | |||
Sensitivity: proportion of true events that were identified by the system | 16 ÷ 65 | 0.25 (0.15–0.37) | |
Specificity: proportion of cases without true events of a given type that the system did not identify |
|
0.9996 (0.9996–0.9997)† | |
Positive predictive value: proportion of apparent events that were true | 704 ÷ 1,590 | 0.44 (0.42–0.47) | |
Negative predictive value: proportion of cases without true events of a given type that had no true event |
|
0.9989 (0.9986–0.9992) |
CI = confidence interval.
A true event was detected by manual review; an apparent event was identified by the system.
See the text for an explanation of the difference between case specificity and event specificity.