Abstract
In this paper, we present an efficient statistical method (denoted as ‘Adaptive Resources Allocation CUSUM’) to robustly and efficiently detect the hotspot with limited sampling resources. Our main idea is to combine the multi-arm bandit (MAB) and change-point detection methods to balance the exploration and exploitation of resource allocation for hotspot detection. Further, a Bayesian weighted update is used to update the posterior distribution of the infection rate. Then, the upper confidence bound (UCB) is used for resource allocation and planning. Finally, CUSUM monitoring statistics to detect the change point as well as the change location. For performance evaluation, we compare the performance of the proposed method with several benchmark methods in the literature and showed the proposed algorithm is able to achieve a lower detection delay and higher detection precision. Finally, this method is applied to hotspot detection in a real case study of county-level daily positive COVID-19 cases in Washington State WA) and demonstrates the effectiveness with very limited distributed samples.
Keywords: Multi-arm bandit, change point detection, adaptive resources allocation, count data, CUSUM statistics
Mathematics subject classifications: 62P10, 62L15
1. Introduction
Nowadays, streaming data from multiple data sources or streams have become more and more common in public health surveillance. Rapid hotspot detection for the streaming data across different spatial regions over time is important and can provide valuable information to the decision-makers to mitigate the risk as soon as possible. In this paper, we define the hotspot as the structured outliers that are persistent after a certain time point. Besides, we also assume that full observation is not possible due to the limited sampling resources.
A motivating example for this research is to monitor the number of COVID-19 confirmed cases for different spatial regions in the United States. Monitoring the entire population is not possible due to the limited testing resources or labels. Here, we would like to focus on adaptively distributing the testing resources such that the hotspot can be detected as soon as possible. Please see Section 2 for a more detailed description of the COVID-19 monitoring problem.
Generally speaking, hotspot detection in spatio-temporal data can be treated as a change-point detection problem. Take the COVID-19 monitoring as an example. There are three types of challenges for the hotspot detection problem: (1) High-dimensionality: the number of sites or locations to be monitored is often quite large. (2) Spatial sparsity of the hotspot: the affected counties are often quite sparse in the high-dimensional space. It implies that the counties undergoing significant changes in the number of confirmed cases should be small at certain time points. (3) Temporal consistency: the hotspots should last for a reasonably long period of time.
Besides the aforementioned challenges, there are also additional challenges in sequential change detection related to the limited sensing resource as follows: (1) Lack of access to the fully observational data: existing research for high dimensional sequential sparse change detection focuses on a fully observable process, i.e. at each sampling time point, all the variables can be observed for analysis. However, in reality, it is infeasible to acquire full measurements for the entire population. (2) Partial sampling with count data observation: existing research often focuses on the continuous observational data and often assumes that a location is either observed or not observed. However, in many examples, such as the COVID-19 monitoring case, the decision variables are not binary, which implies that you can distribute more test kits to a particular region to obtain a more accurate estimation of the percentage of the infected person. In such cases, the distribution of the sampling resources is much more complicated than simply deciding whether a region should be observed or not. (3) Adaptive decision-making to balance the exploration and exploitation: On one hand, we would distribute more tests in the particular region that has been detected with a high infection rate. On the other hand, other unobserved regions may be potentially at risk as well, and some testing resources should be distributed in these regions as well. Overall, the trade-off between exploration and exploitation is a fundamental challenge in multi-arm bandit (MAB) and reinforcement learning. Recently, Such a balance is also an important aspect to be considered in adaptive resource allocation problems for change point detection, e.g. in [13,36,38,41]. Take Figure 1 as an example, infection rates in Florida and Louisiana were higher than in other states on Sep 13, 2020. However, the resource allocation algorithm should distribute enough samples randomly to all states to discover such a fact (i.e. exploration) before distributing more tests to these two states specifically (i.e. exploitation) for final confirmation and change point detection.
Figure 1.
Infection rate on Sep 13, 2020 in the United States.
This paper aims to develop an adaptive resource allocation method to automatically distribute the number of testing resources into multiple regions for the quickest change point and hotspot detection. More specifically, we will focus on the case where the data follows a binomial distribution, where the total number of tests for each region i (i.e. ) depends on the number of testing resources distributed to region i. Overall, we propose to borrow the idea from the Upper Confidence Bound (UCB) method in solving the MAB problem to automatically distribute the testing resources to balance the exploration and exploitation. Bayesian statistics are used to update the estimation of the infection rate online with the consideration of uncertainty. A ‘reward’ function is proposed based on the CUSUM statistics such that the change point can be identified as soon as possible.
Intuitively, the algorithm should work adaptively. Typically, the testing resources should be distributed equally to different regions in the early stages (i.e. exploration). If some regions with potentially high infection rates are observed, more testing resources should be distributed in these regions for confirmation (i.e. exploitation). Furthermore, for resource allocation, we will borrow the idea from MAB. MAB aims to sequentially allocate a limited set of resources between competing ‘arms’ to maximize their expected gain, where each arm's reward function is unknown. MAB provides a principled way to balance exploration and exploitation. To apply MAB for change point detection under the sampling constraint, we propose to use MAB to combine the CUSUM statistics into the reward function in the MAB problem. However, different from [13,41], the ‘arm’ does not refer to whether a single sensor should be observed or not, but the combinations of the number of testing resources (e.g. testing kits) should be distributed to each particular region. Finally, an integer programing algorithm is proposed to optimize such reward function for resource allocation.
The rest of the paper is organized into the following parts. In Section 2, we describe the motivation example of the problem in detail. In Section 3, we review some related methods in change point detection. In Section 4, we introduce the formulation and algorithm of the proposed method in detail. In Section 5, we use the simulation to study the behavior and some properties of the algorithm. A threshold is also generated to be applied to the case study. In Section 6, we study the case of county-level daily positive cases in Washington State (WA). Section 7 summarizes the result and contributions of this study.
2. Motivation example
Since the initial outbreak of the novel coronavirus in early January 2020, the COVID-19 pandemic has rapidly spread across the world and brought enormous disruption to the economy and society. Various emergency measures, such as social distancing, school closures, and economic shutdowns, have been taken by different countries to control the first wave of the pandemic. To balance saving lives and the economy, a condition-based phased approach has since been adopted in the USA, where the strictness of public health measures is set to adapt according to the present epidemic condition. The success of such a phased approach critically depends on the accurate assessment of the current and near-future status of the pandemic.
Much recent research has been focusing on modeling and prediction of the spread of the COVID-19 case report data. For example, Jiang et al. [16] studied the pandemic from a statistical point of view. The piece-wise linear trend model has been applied to model the positive case growth and the self-normalization-based method was applied to detect the change points and predict cases in the future. Tariq et al. [31] monitored pandemic data in Singapore and estimated the effective reproduction number . They also took advantage of different local clusters and some international transmission. Despite various studies to forecast the positive cases and deaths, Luo [21] pointed out that it would be hard to predict the future of the pandemic. The direction of the mutant of a virus is random and, thus, hard to predict whether it will be more contagious or not. For example, the Delta and Omicron variant of COVID-19 caused a remarkable increase of positive cases in the US starting in June 2021 and Jan 2022.
Despite the prediction and modeling of the COVID-19 spread, many recent works focus on developing a monitoring framework to detect the pandemic outbreak while taking advantage of a more scientific resource allocation method to distribute the test kits and detect the outbreak as soon as possible. Scobie et al. [29] monitored and demonstrated the importance of the vaccine in terms of the reduction of the incidence rate ratios (IRRs). Astley et al. [1] used the data from The University of Maryland Global COVID Trends and Impact Survey (UMD-CTIS) in cooperation with Facebook to monitor the pandemic. In general, this algorithm can monitor the COVID-19-related daily spatio-temporal cases with limited testing resources, and in this paper, we call this problem the adaptive partially observed spatio-temporal hotspot detection problem. However, most of these works assume that the data is passively collected but do not focus on how to adaptively distribute the testing resources. Chatzimanolakis et al. [4] discussed the optimal resources allocation problem in a different stage to achieve maximum information gain.
To better understand the COVID-19 status, different types of testing resource is typically distributed in different regions. Centers for Disease Control and Prevention (CDC) has classified the testing for COVID-19 into the following two categories: (1) Diagnostic testing is intended to identify current infection in individuals and is performed when a person has signs or symptoms consistent with COVID-19, or is asymptomatic, but has recently known or suspected exposure to COVID-19.) (2) Screening tests are recommended for unvaccinated people to identify those who are asymptomatic and do not have known, suspected, or reported exposure to COVID-19. Screening helps to identify unknown cases so that measures can be taken to prevent further transmission [25].
Overall, the screening test is very useful in testing employees in a workplace setting or universities, testing a person before or after travel, or randomly distributed test in some underdeveloped areas to identify unknown cases so that measures can be taken to prevent further transmission. In this paper, we will focus on hotspot detection for the case report data with limited screening test resources. For simplicity, we assume that the screening test is the only available source of information, and limited testing resource is available.
In this paper, we will use the real COVID-19 test report data from Johns Hopkins University Center for Systems Science and Engineering (JHU CSSE) (Dong et al. [8]). The dataset is available on https://github.com/CSSEGISandData/COVID-19. More specifically, we will use the confirmed COVID-19 cases from all 39 counties in WA. The source of daily positive cases in WA is the Department of Health (https://www.doh.wa.gov/Emergencies/COVID19). The time-series data is updated daily around 23:59 (UTC). We will use the data from Jan 23, 2020, to Sep 13, 2020, a total of 235 days, as an illustration. On each day, the confirmed cases are recorded in all counties in the United States. We aim to identify hotspots in such a region and see if we can discover the hotspot with a much smaller number of screening test resources.
In this study, we will focus on the distribution of the number of screening testing resources for efficient hotspot detection. With this technique, we are able to conduct a real-time online monitoring process and actively emphasize the regions potentially at risk. First, we showed the infection rates in some example counties in Figure 2. From this figure, we can see that the positive rates of Yakima County started to rise around day 70, and increased dramatically. Furthermore, we have added a figure of the proportion of infected cases in some of the time points during the spread of the COVID-19 data in Figure 3. We can see that the infection rates on day 100 are not very high for all counties. The infection rates in some of the counties started to rise.
Figure 2.
Example of infection rates in WA.
Figure 3.
Comparison of county-wise infection rate in WA on different dates. (a) Infection proportion on day 100: May 1 2020 and (b) Infection proportion on day 200: Aug 9 2020.
3. Literature review
In this Section, we will present a literature review on the change point detection methodology that can be potentially applied to the proposed spatio-temporal problem, including change point detection on the high dimensional data stream, change point detection with sampling control, and change-point detection on count data.
3.1. Change point detection for multiple data streams
Originally, the change point detection is first proposed to detect some sudden changes in the univariate data stream with the shortest detection delay. The CUSUM statistics is proposed by Page [26,27] to minimize the supremum average detection delay [24]. Then, the Shiryaev–Roberts procedure was introduced by Shiryaev [30] in a Bayesian decision framework. Please see [32] for a review paper of the existing change point detection methods. However, as mentioned, these methods cannot be applied to problems with multiple data streams.
To extend the change-point detection method for multiple data streams, Mei [22] extended the CUSUM statistics to multiple data streams by the summation of the CUSUM statistics for each data stream. Xie and Siegmund [34] compared the mixed procedure method they developed with some generalized log-likelihood ratio-based procedures to detect the change point in multiple but unknown data streams under the assumption that the data are normally distributed with known variance, and the change is a mean shift with unknown post-change mean. There are some works to extend these methods to multiple change-point detections with abrupt changes, such as [3,40], which considered the functional subspace learning and exhaustively search of the change point. Enikeeva and Harchaoui [11] studied the change point detection when the number of dimensions and~the length of each data stream goes to infinity based on the maximum of scan statistic and linear test statistic. However, these methods typically assume that the data follow a normal distribution or the data is fully observed.
3.2. Change point detection for partially observed data with sampling control
Another line of research focuses on detecting anomalies for the spatio-temporal data. For example, Das et al. [7] applied some tests to study the change point of winter maximum, minimum, and average temperature from 6 meteorological stations in 102 year time period. Chen et al. [5] proposed a new statistic to detect change points in spatio-temporal data with mean shift or covariance structure change. Zhao et al. [42] developed a likelihood-based algorithm to solve the multiple change detection in non-stationary spatio-temporal processes by splitting the data into piecewise stationary processes. Knoblauch and Damoulas [18] developed a Bayesian online change-point detection algorithm to model the process between change points in spatio-temporal data. Zhang et al. [39] studied the change point detection problem when there are system-inherent variations and sensor feature differences and is able to identify the local change point in each cluster. Again, these methods focus on the fully observed data and do not provide a method to iteratively update the sampling resources.
To generalize the change point detection for partially observed data, some works focus on the change point detection by assuming that only partial data streams at each time point can be observed due to limited sampling resources. Such setting has been studied in social network monitoring [12], biological network [28], and manufacturing systems [19]. To deal with partially observed data, Xie et al. [33] developed a change point detection method by combining subspace manifold learning and multiscale analysis to solve the missing data problem. Dubey et al. [9] studied the online change point detection problem in the network where the pattern of the missing data of the network is heterogeneous. Corradin et al. [6] studied multiple change-point detections in multivariate time series with missing values. The conditional distribution of each data point given the rest data is derived to handle the missing value problem. However, these methods do not actively control which subsets of data to measure, which hinders the detection performance.
Recently, many existing works have focused on the change-point detection problem with sampling control, where the sensor(s) to be observed in each epoch are determined adaptively based on the current and historical observation. For example, Liu et al. [19] proposed to combine CUSUM statistics and a top-r strategy to select the data streams to be observed for online change detection. Xu and Mei [35] studied the problem when there is only one sensor to be observed at each time and an adaptive sampling strategy based on the CUSUM statistic and showed that it is asymptotically optimal in the sense of minimizing the detection delay. Zhang and Hoi [38] developed an online learning algorithm for anomaly detection for independent data streams using the combinatorial MAB approach. Zhang and Mei [41] studied the change point detection under sampling control by the combination of Thompson Sampling and a top-r local Shiryaev-Robert statistic to raise a global alarm. Guo et al. [13] proposed a Bayesian spike-and-slab prior distribution to decompose the data into the smooth background and sparse anomaly and focus on identifying the anomaly by selecting the sensors to observe using Thompson sampling. However, all aforementioned methods focus on change detection under the continuous observation data and assume each sensor is either fully observed or not, which cannot be used for more complex resource allocation problems for anomaly detection.
3.3. Change point detection on counting data
Much existing research has been focusing on extending the change point detection method for count data or binary data. For example, Hinkley and Hinkley [15] studied the change point detection problem for binary variables based on the asymptotic distribution of the likelihood ratio test statistics. Yu et al. [37] developed the CUSUM statistics to study the~change point detection problem in the binomial thinning process and applied it to the 2001-2012 influenza data. Jiang et al. [17] proposed a weighted CUSUM control chart for monitoring Poisson processes with varying sample sizes. Gut and Steinebach [14] studied the change point detection problem in renewal counting data with the discrete-time frame, where the term ‘partially observed’ meant the discrete-time frame. Ellenberger et al. [10] developed a binary segmentation method to solve the change point detection problem when the sample size of the binomial distribution is small. However, these methods typically assume that the sample size of the binomial distribution is pre-determined, which cannot be used to solve the resource allocation problem.
4. Proposed methodology
In this Section, we will explore the adaptive sampling method on discrete data with the binomial distribution. In Section 4.1, we will review the overall problem formulation. In Section 4.2, we will introduce the procedure of the proposed algorithm. In Section 4.3, we will introduce the tuning parameter selection procedure for the proposed algorithm.
4.1. Problem formulation
We assume that the data is , where is the time index and is the region index. In the COVID-19 monitoring example, can be the number of newly observed confirmed cases in region k and time t.
We further assume that all prior-change follows a binomial distribution as , with the number of test and the probability p. Furthermore, we assume that after the change , only one hotspot region (i.e. ) is impacted by the change. In contrast, we assume that the regions affected by the changes follow another post-change distribution where is another binomial distribution as , with the number of tests as with the probability q, as shown in Equation (1). The superscript 0 means in control, and the superscript 1 means out of control.
| (1) | 
Finally, we assume that there are limited sampling resources. The constraint we assume is that limited testing kits C are distributed to all the regions at each time t with the constraint . For simplicity, we assume that the actively distributed test kits are the only source of observation data. The objective of this paper is to sequentially distribute the testing resources for region k at time t for faster change point detection.
4.2. Proposed adaptive resources allocation CUSUM for binomial count data (ARA-CUSUM) method
In this Section, we first introduce the overall step of the algorithm and then discuss three major components of the proposed methods as follows: monitoring statistics updates, planning for resource allocation, and change-point detection decision. The overall framework is shown in Figure 4. The detailed steps are discussed as follows.
Monitoring statistics update: We propose to update the monitoring statistics at each time t for the quickest change-point detection. This procedure will be discussed in detail in Section 4.2.1.
Planning for adaptive sampling: We would like to distribute the testing resource for the next time point based on previous results by the UCB bandit algorithm. The detailed planning formulation and the optimization algorithm are discussed in Sections 4.2.2 and 4.2.4.
Change point detection decision: Finally, the alarm is raised when the largest CUSUM statistics for all regions exceeds some threshold. Finally, the corresponding region that triggers such change can also be used for hotspot identification.
Figure 4.
Flow chart for the algorithm.
4.2.1. Monitoring procedure
In this Section, we will first introduce the definition of CUSUM statistics and then derive the monitoring procedure for online change detection.
Definition 4.1
Suppose in a sequential data, is the score of nth observation. The CUSUM statistics Page [26] is defined by
(2) 
Moustakides [23] suggested that the CUSUM statistic is mini-max optimal in the sense of detection delay [20].
| (3) | 
with the constraint of in-control average run length . In change point detection, the score for each observation can be the log-likelihood ratio, which is , where is the nth observed data, is the pdf of before change and is the pdf of after change.
In the proposed formulation, is defined as the observed positive cases in time t and region k. Inspired by the CUSUM statistics definition, the CUSUM statistics is formulated as:
| (4) | 
where and represent the distribution of null and alternative hypothesis, respectively. The initial value is set to be 0 for all k. Following the CUSUM statistics definition, we define the CUSUM statistics in the proposed formulation for each county k and time t in Proposition 4.2.
Proposition 4.2
Suppose that at time t, the number of tests distributed at region k is , the number of positive cases collected is . For each region k, for , the CUSUM statistics can be computed for each time t and region k can be derived as
(5) 
The proof is detailed in Appendix 1.
Here, we will raise the alarm when the largest of the CUSUM statistics exceeds some threshold c at some time T. Finally, we will discuss the procedure of selecting the threshold c in Section 4.3.
4.2.2. Planning objective function for resource allocation
In this Section, we present an efficient algorithm for resource allocation at the next time point for hotspot detection. Overall, a good resource allocation algorithm for change detection should achieve a good balance between exploration and exploitation. Here, exploration in the problem formulation implies the test kits should be distributed to the regions that were not measured before, and exploitation means that when the change occurs, the planning will focus more on the region with the change for the quickest change detection. To better tackle the exploration-exploitation trade-off problem, we apply the UCB as the target function to be optimized. Auer [2] suggested that the UCB criteria for the MAB problem maximize the final reward. This method takes advantage not only of the mean reward μ but also the standard deviation of the reward σ. In this case, we should not only focus on the region leading to the larger reward but also the action with larger variability, which has the potential to lead to a larger reward. Specifically, in this hotspot detection application, for the true hotspot region with a larger infection rate, we may want to continue to distribute more test kits to the particular region the next day to encourage further exploitation. However, when a region has not been observed enough in previous days, it typically has a larger variance, and some testing resources will be distributed in the regions with large variability the next time to encourage further exploration.
Given that the log-likelihood ratio for change point detection is often a good quantification of how likely the change may happen in a particular region k, here we propose to use the sum of UCB of the likelihood ratio in the CUSUM statistics as the reward function. For the convenience of notation, we denote the log-likelihood ratio in the CUSUM statistics in each region k by
| (6) | 
Therefore, the UCB function of the summation of the likelihood ratio statistics can be expressed as
| (7) | 
Here, we consider a special form of UCB function with equal weights between mean and standard deviation rewards, and the ideas can be easily extended to other UCB functions with different weights. The goal is to find the best resource allocation planning method to optimize the resource distribution at time t. However, one of the challenges is that the is the number of positive testing results at time t, which is a random variable at time t. In Section 4.2.3, we will derive the Bayesian updating procedure to update the posterior distribution of , which can be used estimate the and .
4.2.3. Bayesian updating of
To estimate the distribution of , we assume that it follows a binomial distribution of , where is the true infection rate to be estimated. Therefore, an online estimation of the rate is needed for better planning and resource allocation. Here, we propose a weighted Bayesian update method to estimate the infection rate in Proposition 4.3. The reason to consider such a weighted updating procedure is that the samples in recent time points are typically more likely to represent the current state of the system and should be used to better detect the change. Therefore, more weight should be put into the recent time points. For simplicity, we propose to use a exponential decayed weight for each sample .
Finally, the posterior distribution can be derived as
| (8) | 
where is the prior distribution of the infection rate, which is assumed to follow a Beta distribution as . is the posterior distribution of , which can be derived in Proposition 4.3.
Proposition 4.3
Suppose that the data of the kth region until time T is collected, where is the number of total tests distributed to region k and time t and is the daily positive cases. We further assume the prior distribution of is , the weighted posterior distribution of can be derived as
(9) 
The proof is given in Appendix 2. Here, the updated posterior distribution of can be used to calculate the uncertainty in to derive the UCB function in the planning method.
4.2.4. Optimization algorithm for planning
By plugging in the weighted posterior distribution of into the UCB reward function in (7), we will derive the objective function to be optimized and propose an integer programing algorithm to optimize the number of test kits in each region k at time T + 1.
Proposition 4.4
Suppose and the posterior distribution . The target function can be derived as
(10) 
The proof is shown in Appendix 3. Therefore, Algorithm 1 is proposed to determine the best resource allocation at time T + 1 as follows:
| (11) | 
Here, we would like to give a brief discussion on the reward function defined in (10). An example of the function with different infection rates are given in Figure 5. The expectation part and standard deviation part is larger for regions with a higher estimated infection rate. However, there is a diminishing return effect on this reward function, which implies that as more testing resources are distributed in the highly infected regions, the increase of the reward function will become smaller. At a certain point, it may be less rewarding to distribute additional tests to the highly infected region compared to the low infected region. To demonstrate, Figure 5 shows that for k = 1, 2, 3, 4 and , but for some k = 1, 2, 3, 4 and .
Figure 5.
Example of reward functions with different estimated infection rate.
To optimize problem (11), a greedy algorithm is proposed. The idea of the greedy algorithm is to distribute the testing kits one by one to the region that can result in the largest target increment until all the tests are properly distributed. Given the concavity of , such a greedy approach can guarantee the global optimum. This property will be discussed in Section 4.6. The greedy algorithm is designed as Algorithm 1.
Finally, the entire process can be shown in Algorithm 2. Overall, the steps of the algorithm are described here: (1) On day one, we have an initial test distribution. If the prior distribution is set the same for all regions, uniform distribution can be used. The confirmed number of will be collected on Day 1. (2) The collected dataset can be used to update the posterior distribution of the rate by (9). (3) Such distribution information can be plugged into (11) to optimize the test distribution of the next day. Such a procedure can be estimated recursively until the current time T.
4.2.5. Property of the optimization algorithm
In this subsection, we would like to prove that the greedy approach proposed in Section 4.2.4 can obtain the global optimum.
Lemma 4.5
The target function (10) is the sum of K concave functions and is therefore concave.
The target function in each region can be expressed as . One can easily verify that the second order derivative of is negative when b, c>0.
Proposition 4.6
The Algorithm 1 can obtain the global optimum for the optimization problem (11).
Proof.
From Lemma 4.5, is an increasing and strictly concave function. For any fixed time t, we denote be the increment of the reward function to distribute one more tests on region k. will be a positive decreasing function for any k. Then the total reward function given the test distribution is
Now consider the increment parameter matrix
In Algorithm 1, the first iteration will choose the largest value in the first column. Since each row in the matrix is decreasing, the first iteration will also choose the largest value in the matrix. Each of the remaining iterations will choose the largest value in the matrix that has not been chosen. Therefore, Algorithm 1 chooses the top C largest values in the matrix. And because the target function can be written by the sum of C values in the matrix, Algorithm 1 gives the optimal solution to the optimization problem (11)
4.3. Tuning parameter selection
There are several tuning parameters that need to be determined for the algorithm. The key parameters are prior distribution parameter a, b, weight w and threshold c. In this Section, we will discuss the effect of these tuning parameters and how to decide them.
Prior Distribution: There are two parameters a, b in the prior distribution. The prior distribution reflects the prior knowledge about the infection rate. Given that the infection rate for the null hypothesis is p, we can also set the ratio to be p, i.e. , assuming no change happens. Since p is a preset constant, we only need to determine a. When a is large, the influence of prior distribution is large, and the difference in the infection rate estimation from the posterior distribution will be smaller. Thus, the difference in test distribution will not be very large. In the simulation, we will choose .
Weight w: The weight is used in Bayesian update to get the posterior distribution of . Weight allows the algorithm to take not only the information from one day before but also the previous days. A large weight will make the test distribution in each region smooth across different days. Therefore, when occasionally the observed infection rate is high, the algorithm will only distribute a reasonably larger amount of testing kits on the next day rather than some extreme testing kit distribution. In the simulation, we choose w = 0.3.
Total Resources C: This parameter determines the total number of tests to be distributed in all regions. Typically, C is pre-determined by the resource constraint.
Threshold c: This parameter determines whether an alarm should be raised in each region every day. Higher threshold c leads to high in-control average run length ( ) and higher out-of-control . Typically, we would like to maintain a preset , which determines the threshold c. Overall, for a preset C, p, q, a simulation study can be used to determine the best C with certain p and q. We will determine this parameter by the simulation work and use a preset average run length.
5. Simulation
In this Section, we will start with the simulation setup in Section 5.1 and then discuss the result evaluation in Section 5.2. The balance of exploration and exploitation is discussed in Section 5.3.
5.1. Simulation setup
In this Section, we will evaluate the proposed methodology using a simulation study. To generate the simulation dataset, we will simulate a true infection rate for each county in the WA as an example.
There are 39 counties in WA, i.e. K = 39. Before the distribution change, the true infection proportion is set to be p = 0.01 for all counties. We will compare the detection delay for different infection proportions where the out-of-control infection rates are q = 0.025, 0.03, 0.04, 0.05 in the first county after the distribution change. We further assume that the total number of tests is 3900, where each county takes an average of 100 tests per day. The weighted update parameter w is set to be 0.3 and the prior distribution parameter a, b are set as the of the total tests to make the prior distribution as , .
The process of the simulation is described as follows:
The test distribution in each county determined from the previous iteration can be used.
Before the change point , we will sample the observed positive cases by .
After the change point , we will sample the observed positive cases in the first county by the out-of-control infection rate q and . We will sample the rest counties still by
Finally, we will follow Algorithm 1 for the proposed algorithm. The CUSUM statistics are updated, and the test distribution of the next day is determined by the proposed method or benchmark methods. When studying the average run length and detection delay, we will stop the process when the alarm is raised.
In this simulation study, we will compare the proposed algorithm with the following two benchmark methods.
Even Distribution: The first benchmark (denoted as Even) is a simple approach, which always evenly distributes the tests to all the counties. This even distribution will focus on exploration of the testing resources but do not have the power to exploit for a specific region.
Top-R Distribution: The second benchmark (denoted as top-R') is inspired by the Top-R strategy proposed in [19]. We first evenly distribute all tests evenly into 20 batches. For in total 3900 testing kits, it averages out 195 testing kits per batch. We then select the top 20 regions with the largest CUSUM statistics in the previous day. Here, we select the top 20 regions, given it is a little more than half of the whole regions.
5.2. Evaluation of the detection delay and precition
To compare this algorithm with the benchmarks, we will first need to determine the thresholds for the proposed method and benchmark methods. Here, binary search can be used to decide the optimal threshold for each benchmark method to maintain the in-control average run length (ARL) as 200 for all benchmark methods with 1000 replications.
To evaluate the proposed methods, we use the following three metrics: (1) the out-of-control average run length ( ): is defined as the average detection delay after the change has happened. Overall, we aim to reduce the ARL when keeping the same ; (2) Detection precision (DP): DP is evaluated as the proportion in the 1000 iterations that the alarm is raised in the correct county. Overall, we would aim to increase the DP so that we can identify the county that is accountable for the change more accurately. (3) the standard deviation of run-length (SDRL): SDRL is evaluated by computing the standard deviation with 1000 iterations. Overall, a smaller SDRL would imply a more robust detection with smaller variability for detection delay. These three metrics for the proposed method and the benchmark methods (namely the evenly distributed and the Top-R distributed) are evaluated in Table 1.
Table 1.
The result of detection delay and detection precision when .
| Metric | Methods | q = 0.025 | q = 0.03 | q = 0.04 | q = 0.05 | 
|---|---|---|---|---|---|
| Proposed | |||||
| Even | 14.885 | 7.93 | 4.408 | 3.299 | |
| Top-R | 10.385 | 5.688 | 3.443 | 2.891 | |
| DP | Proposed | 0.938 | 0.939 | ||
| Even | 0.892 | 0.926 | |||
| Top-R | 0.904 | 0.918 | 0.936 | 0.937 | |
| SDRL | Proposed | ||||
| Even | 12.81 | 5.48 | 2.28 | 1.45 | |
| Top-R | 10.556 | 3.72 | 1.63 | 1.12 | 
From Table 1, we can conclude that ARL is the smallest for the different q values. Especially, when q is small (i.e. 0.025 and 0.03), the detection delay can even be greatly reduced (i.e. 7.893 for ‘proposed’, 14.885 for ‘Even’, and 10.385 for ‘Top-R’ for q = 0.025). For larger q, the differences between the proposed methods and the second best method, such as ‘Top-R’ are similar, given it becomes easy to detect all the methods. However, the proposed method still outperforms others in this case. Finally, we find that the SDRL of the proposed method is also the smallest, especially when q is small. This implies that the proposed methods can also make the detection much more robust by reducing the standard deviation of the detection delay.
5.3. Exploration and exploitation
As formerly mentioned, a good change point detection with a sampling control algorithm should balance both exploration and exploitation behavior. In this simulation, exploration means that we will need to explore all counties, given that any county can undergo large changes at certain times. Exploitation means that we will need to distribute more resources to the counties that have been detected or have the potential to be the hotspot. Figures 6 and 7 give a more intuitive understanding of the exploration and exploitation behavior of this proposed algorithm from the following simulation study setup. In the first 500 days, all the counties are in control with the same infection rate p = 0.01. From day 501 to day 1000, only the first county, Adams County is out of control with the infection rate q = 0.05, while the rest counties stay the same with p = 0.01. Figure 6 shows the density and the boxplot of the number of tests in control. Figure 7 shows the median number of test kits in each county under different conditions, in control and out of control.
Figure 6.
Visualization of test distributions in control. (a) Test distributions density in control and (b) Boxplot of number of tests in control.
Figure 7.
Median number of tests. (a) Median number of tests in control and (b) Median number of tests out of control.
Here are some behaviors of the algorithm. When there is no distribution change, the distributions of the number of tests in each county are similar, see Figures 6 and 7(a). This shows that the proposed method is able to explore all the counties undergoing potential changes. However, if a county has a higher infection rate on the current day, it is likely to have higher tests the next day. If the sampled positive proportion is small, the increment of CUSUM statistics will be likely to be 0. If the distribution change happens in that county, the CUSUM test statistic will be likely to grow dramatically and we will continue to focus on that county further. See Figure 7(b), when Adams County is out of control, the median number of the distributed tests of Adams County is significantly larger than the number of other counties. In this case, the CUSUM test statistic of that Adams County is more likely to increase dramatically, which is a good behavior of exploitation.
6. Case study
In this case study, we would like to evaluate the proposed methods to the real case study. The data we are working on is the county-level daily positive cases, and the parameters are K = 39, p = 0.01, q = 0.05, a = 19.5, b = 1930.5 and w = 0.3. The average number of tests in each county is 100.
6.1. Detection of the out-of-control county
The alarms were raised in Yakima county on Jun 19, 2020. The comparison figure of the test statistics is in Figure 8. Starting from day 135, the infection rate of Yakima county started to show a dominating trend over the other counties. Therefore, the algorithm emphasizes Yakima county to seek potential risk. But the true infection rate is not large enough to produce a positive increment in the CUSUM statistics. Thus, we will see this oscillating trend. When the infection rate of Yakima county increases consistently, the CUSUM statistics will increase dramatically. Finally, on June 21, 2021, the CUSUM statistics in Yakima County exceeded the threshold of 6.5, and the alarm was raised.
Figure 8.
Comparison of the test statistics.
6.2. Behavior of sampling resource distribution
The plots in Figure 9 are examples of the test distribution of this case study before and after the alarm is raised. This shows similar behavior as the simulation study. The test distribution before the alarm is close to an even distribution to explore all the counties for the risks. The test distribution after the alarm focuses on Yakima county and consistently monitors this county to show good behavior of exploitation.
Figure 9.
Example of test distribution in WA. (a) Example of test distribution in control and (b) Example of test distribution out of control.
7. Conclusion
In conclusion, the proposed method is able to handle the challenge of exploration and exploitation trade-offs and limited resources. In the simulation work, we show that the adaptive sampling method can beat the two benchmarks either in the sense of detection delay or detection precision. In the case study in WA, the alarm was raised in Yakima county. This algorithm works well when there is only one out-of-control region. When there are multiple out-of-control regions, this algorithm will focus on the first region with a higher out-of-control risk. If there are multiple potential regions at risk, one can start the algorithm from the beginning again when an alarm is raised and remove the alerted region. In the future, we can address this concern and detect multiple out-of-control regions in a time series manner. Some other future work includes taking geometric distance and graphing into consideration and solving the detection problem when there are multiple out-of-control regions.
Acknowledgments
The authors are grateful to the Editor, Dr. Jie Chen, the Associate Editor, and anonymous reviewers for their constructive comments that greatly improved the quality of this article.
Appendices.
Appendix 1. Proof of Proposition 4.2
Proof.
The distribution of is a binomial distribution with total counts . In the null hypothesis and alternative hypothesis, the probability is p and q, respectively. Thus
Plugging in the formula above to Equation (4), we have
And we have the result of Equation (5).
Appendix 2. Proof of Proposition 4.3
Proof.
For binomial distribution, a reasonable conjugate prior distribution is the beta distribution. The posterior distribution from a weighted update is
where is a series of time decayed weights with increasing order. Here we set it to be . Recall that , we have
This is the kernel of the distribution
Appendix 3. Proof of Proposition 4.4
Proof.
We refer to Equation (6) to get of the increment. Recall that , and are the decision variables, which can be regarded as constant. We have
The first term is
The second term is
Then, we can calculate the variance of as
Recall that the target function is formulated to be the sum of the mean and standard deviation. We also have a constrain that . Then when we sum up , the first term of the expectation part will be a constant and can be left out from the target function. We also notice that when taking the squared root of , the coefficient will be the same as the coefficient of the second term in . Thus, this coefficient can also be removed. The simplified reward function can now be formulated as a formula (10) and the tests distribution of next day t + 1 is to solve the optimization problem (11)
Appendix 4. Sensitivity analysis
In this section, we will study the effect of hyperparameters w and a, b on the average run length by performing an additional simulation study. Here, we will start with the effect of w on the average run length. When , , the average run length is around 200, when w = 0.01, 0.3, 0.6. The detection delay and detection precision are in Table A1.
From Table A1, we can conclude that when q is small, a smaller w results a larger or higher detection delay. However, w = 0.3 and w = 0.6 yield similar . When studying one specific iteration, we found that the hyperparameter w affects the distribution of the testing kits. Figure A1 is the box-plot of a region for different weights as w = 0.01, 0.3, 0.6. We can see that the median of test distributions is similar for different weights. However, when w is large, it results in more days with extremely large numbers of testing kits. This is because when w is large, it puts more weight on the previous days to update the posterior distribution of the infection rate. Therefore, it is harder to reduce the number of testing kits the next day compared to when w is small.
Table A1.
Comparison of detection delay ( ) and Detection precision (DP) for different weight w.
| w | q = 0.025 | q = 0.03 | q = 0.04 | q = 0.05 | |
|---|---|---|---|---|---|
| 0.01 | 9.362 | 5.23 | 3.391 | ||
| 0.3 | 7.893 | 4.958 | 2.863 | ||
| 0.6 | 3.438 | 2.88 | |||
| DP | 0.01 | 0.907 | 0.919 | 0.931 | |
| 0.3 | 0.918 | ||||
| 0.6 | 0.929 | 0.939 | 
Figure A1.
Boxplot of the testing kit distribution in one county before the change for different w.
We also studied the impact of different a in the prior distribution, given that the ratio of a and b are kept the same in the simulation study. The testing kit distribution of one county in the in-control state is shown in Figure A2. Intuitively, a higher a makes it harder to distribute more tests in a region. For example, if a is 19.5, the algorithm will allocate fewer tests in the region with 10 tests and 2 positive cases than when a is 1. The simulated result of detection delay and detection precision when the average run length in control is around 200 is as Table A2. We can see that when a increases from 0.1 to 19.5, the detection delay decreases, and the detection precision increases. Therefore, the choice of a and b depends on how much weight we would like to rely on the prior knowledge.
Figure A2.
Boxplot of the testing kit distribution in one county before the change for different a.
Table A2.
Comparison of detection delay ( ) and Detection precision (DP) for different a.
| a | q = 0.025 | q = 0.03 | q = 0.04 | q = 0.05 | |
|---|---|---|---|---|---|
| 0.1 | 17.463 | 10.229 | 8.097 | 7.247 | |
| 1 | 12.453 | 8.064 | 6.214 | 5.373 | |
| 19.5 | 7.893 | 4.958 | 3.358 | 2.863 | |
| DP | 0.1 | 0.88 | 0.91 | 0.925 | 0.93 | 
| 1 | 0.901 | 0.923 | 0.93 | 0.939 | |
| 19.5 | 0.918 | 0.932 | 0.938 | 0.939 | 
Funding Statement
H. Yan was supported in part by NSF grant [DMS-1830363]. Y. Mei was supported in part by NSF grant [DMS-2015405]. S. E. Holte's research was supported in part by NSF grant [DMS-1830372] and an NIH grant [AI027757]. In addition, this research was supported in part by an NIH grant [R21 AI157618]. The content is solely the responsibility of the authors and does not necessarily represent.
Disclosure statement
No potential conflict of interest was reported by the author(s).
References
- 1.Astley C.M., Tuli G., Mc Cord K.A., Cohn E.L., Rader B., Varrelman T.J., Chiu S.L., Deng X., Stewart K., Farag T.H., Barkume K.M., LaRocca S., Morris K.A., Kreuter F., and Brownstein J.S., Global monitoring of the impact of the COVID-19 pandemic through online surveys sampled from the facebook user base, Proc. Natl. Acad. Sci. 118 (2021), Article ID e2111455118. [DOI] [PMC free article] [PubMed] [Google Scholar]
 - 2.Auer P., Using confidence bounds for exploitation-exploration trade-offs, J. Mach. Learn. Res. 3 (2002), pp. 397–422. [Google Scholar]
 - 3.Avanesov V. and Buzun N., Change-point detection in high-dimensional covariance structure, Electron. J. Stat. 12 (2018), pp. 3254–3294. [Google Scholar]
 - 4.Chatzimanolakis M., Weber P., Arampatzis G., Wälchli D., Kičić I., Karnakov P., Papadimitriou C., and Koumoutsakos P., Optimal allocation of limited test resources for the quantification of COVID-19 infections, Swiss Med. Wkly. 150 (2020), p. W20445. [DOI] [PubMed] [Google Scholar]
 - 5.Chen J., Kim S.H., and Xie Y., S 3 t: A score statistic for spatiotemporal change point detection, Seq. Anal. 39 (2020), pp. 563–592. [Google Scholar]
 - 6.Corradin R., Danese L., and Ongaro A., Bayesian nonparametric change point detection for multivariate time series with missing observations, Int. J. Approx. Reason. 143 (2022), pp. 26–43. [Google Scholar]
 - 7.Das J., Mandal T., and Saha P., Spatio-temporal trend and change point detection of winter temperature of North Bengal, India, Spat. Inf. Res. 27 (2019), pp. 411–424. [Google Scholar]
 - 8.Dong E., Du H., and Gardner L., An interactive web-based dashboard to track COVID-19 in real time, Lancet Infect. Dis. 20 (2020), pp. 533–534. [DOI] [PMC free article] [PubMed] [Google Scholar]
 - 9.Dubey P., Xu H., and Yu Y., Online network change point detection with missing values, preprint (2021). Available at arXiv:2110.06450.
 - 10.Ellenberger D., Lausen B., and Friede T., Exact change point detection with improved power in small-sample binomial sequences, Biom. J. 63 (2021), pp. 558–574. [DOI] [PubMed] [Google Scholar]
 - 11.Enikeeva F. and Harchaoui Z., High-dimensional change-point detection under sparse alternatives, Ann. Statist. 47 (2019), pp. 2051–2079. [Google Scholar]
 - 12.Farajtabar M., Rodriguez M.G., Zamani M., Du N., Zha H., and Song L., Back to the past: Source identification in diffusion networks from partially observed cascades, in Artificial Intelligence and Statistics, PMLR, 2015, pp. 232–240.
 - 13.Guo J., Yan H., Zhang C., and Hoi S., Partially observable online change detection via smooth-sparse decomposition, preprint (2020). Available at arXiv:2009.10645.
 - 14.Gut A. and Steinebach J., Truncated sequential change-point detection based on renewal counting processes, Scand. J. Statist. 29 (2002), pp. 693–719. [Google Scholar]
 - 15.Hinkley D.V. and Hinkley E.A., Inference about the change-point in a sequence of binomial variables, Biometrika 57 (1970), pp. 477–488. [Google Scholar]
 - 16.Jiang F., Zhao Z., and Shao X., Time series analysis of COVID-19 infection curve: A change-point perspective, J. Econom. (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
 - 17.Jiang W., Shu L., and Tsui K.L., Weighted cusum control charts for monitoring poisson processes with varying sample sizes, J. Qual. Technol. 43 (2011), pp. 346–362. [Google Scholar]
 - 18.Knoblauch J. and Damoulas T., Spatio-temporal Bayesian on-line changepoint detection with model selection, in International Conference on Machine Learning, PMLR, 2018, pp. 2718–2727.
 - 19.Liu K., Mei Y., and Shi J., An adaptive sampling strategy for online high-dimensional process monitoring, Technometrics 57 (2015), pp. 305–319. [Google Scholar]
 - 20.Lorden G., Procedures for reacting to a change in distribution, Ann. Math. Statist. 42 (1971), pp. 1897–1908. [Google Scholar]
 - 21.Luo J., Predictive Monitoring of COVID-19, Vol. 446, SUTD Data-Driven Innovation Lab, Singapore, 2020. [Google Scholar]
 - 22.Mei Y., Efficient scalable schemes for monitoring a large number of data streams, Biometrika 97 (2010), pp. 419–433. [Google Scholar]
 - 23.Moustakides G.V., Optimal stopping times for detecting changes in distributions, Ann. Statist. 14 (1986), pp. 1379–1387. [Google Scholar]
 - 24.Moustakides G.V., Polunchenko A.S., and Tartakovsky A.G., Numerical comparison of cusum and shiryaev–roberts procedures for detecting changes in distributions, Comm. Statist. Theory Methods 38 (2009), pp. 3225–3239. [Google Scholar]
 - 25.Overview of testing for sars-cov-2 (COVID-19) (2022). Available at https://www.cdc.gov/coronavirus/2019-ncov/hcp/testing-overview.html.
 - 26.Page E.S., Continuous inspection schemes, Biometrika 41 (1954), pp. 100–115. [Google Scholar]
 - 27.Page E.S., A test for a change in a parameter occurring at an unknown point, Biometrika 42 (1955), pp. 523–527. [Google Scholar]
 - 28.Raue A., Kreutz C., Maiwald T., Bachmann J., Schilling M., Klingmüller U., and Timmer J., Structural and practical identifiability analysis of partially observed dynamical models by exploiting the profile likelihood, Bioinformatics 25 (2009), pp. 1923–1929. [DOI] [PubMed] [Google Scholar]
 - 29.Scobie H.M., Johnson A.G., Suthar A.B., Severson R., Alden N.B., Balter S., Bertolino D., Blythe D., Brady S., Cadwell B., Cheng I., Davidson S., Delgadillo J., Devinney K., Duchin J., Duwell M., Fisher R., Fleischauer A., Grant A., Griffin J., Haddix M., Hand J., Hanson M., Hawkins E., Herlihy R.K., Hicks L., Holtzman C., Hoskins M., Hyun J., Kaur R., Kay M., Kidrowski H., Kim C., Komatsu K., Kugeler K., Lewis M., Lyons B.C., Lyons S., Lynfield R., McCaffrey K., McMullen C., Milroy L., Meyer S., Nolen L., Patel M.R., Pogosjans S., Reese H.E., Saupe A., Sell J., Sokol T., Sosin D., Stanislawski E., Stevens K., Vest H., White K., Wilson E., MacNeil A., Ritchey M.D., and Silk B.J., Monitoring incidence of COVID-19 cases, hospitalizations, and deaths, by vaccination status-13 us jurisdictions, April 4–July 17, 2021, Morb. Mortal. Wkly. Rep. 70 (2021), p. 1284. [DOI] [PMC free article] [PubMed] [Google Scholar]
 - 30.Shiryaev A.N., Problem of most rapid detection of a disturbance in sationary processes, Dokl. Akad. Nauk SSSR 138 (1961), p. 1039. [Google Scholar]
 - 31.Tariq A., Lee Y., Roosa K., Blumberg S., Yan P., Ma S., and Chowell G., Real-time monitoring the transmission potential of COVID-19 in Singapore, March 2020, BMC Med. 18 (2020), pp. 1–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
 - 32.Truong C., Oudre L., and Vayatis N., Selective review of offline change point detection methods, Signal Processing 167 (2020), Article ID 107299. [Google Scholar]
 - 33.Xie Y., Huang J., and Willett R., Change-point detection for high-dimensional time series with missing data, IEEE J. Sel. Top. Signal Process. 7 (2012), pp. 12–27. [Google Scholar]
 - 34.Xie Y. and Siegmund D., Sequential multi-sensor change-point detection, in 2013 Information Theory and Applications Workshop (ITA), IEEE, 2013, pp. 1–20.
 - 35.Xu Q. and Mei Y., Multi-stream quickest detection with unknown post-change parameters under sampling control, in 2021 IEEE International Symposium on Information Theory (ISIT), IEEE, 2021, pp. 112–117.
 - 36.Xu Q., Mei Y., and Moustakides G.V., Optimum multi-stream sequential change-point detection with sampling control, IEEE Trans. Inform. Theory 67 (2021), pp. 7627–7636. [Google Scholar]
 - 37.Yu X., Baron M., and Choudhary P.K., Change-point detection in binomial thinning processes, with applications in epidemiology, Seq. Anal. 32 (2013), pp. 350–367. [Google Scholar]
 - 38.Zhang C. and Hoi S.C., Partially observable multi-sensor sequential change detection: A combinatorial multi-armed bandit approach, in Proceedings of the AAAI Conference on Artificial Intelligence, Hawaii, Vol. 33, 2019, pp. 5733–5740.
 - 39.Zhang C., Yan H., Lee S., and Shi J., Multiple profiles sensor-based monitoring and anomaly detection, J. Qual. Technol. 50 (2018), pp. 344–362. [Google Scholar]
 - 40.Zhang C., Yan H., Lee S., and Shi J., Dynamic multivariate functional data modeling via sparse subspace learning, Technometrics 63 (2021), pp. 370–383. [Google Scholar]
 - 41.Zhang W. and Mei Y., Bandit change-point detection for real-time monitoring high-dimensional data under sampling control, preprint (2020). Available at arXiv:2009.11891. [DOI] [PMC free article] [PubMed]
 - 42.Zhao Z., Ma T.F., Ng W.L., and Yau C.Y., A composite likelihood-based approach for change-point detection in spatio-temporal process, preprint (2019). Available at arXiv:1904.06340.
 











