Abstract
Estimating the size of hard-to-reach populations is an important problem for many fields. The Network Scale-up Method (NSUM) is a relatively new approach to estimate the size of these hard-to-reach populations by asking respondents the question, “How many X’s do you know,” where X is the population of interest (e.g. “How many female sex workers do you know?”). The answers to these questions form Aggregated Relational Data (ARD). The NSUM has been used to estimate the size of a variety of subpopulations, including female sex workers, drug users, and even children who have been hospitalized for choking. Within the Network Scale-up methodology, there are a multitude of estimators for the size of the hidden population, including direct estimators, maximum likelihood estimators, and Bayesian estimators. In this article, we first provide an in-depth analysis of ARD properties and the techniques to collect the data. Then, we comprehensively review different estimation methods in terms of the assumptions behind each model, the relationships between the estimators, and the practical considerations of implementing the methods. We apply many of the models discussed in the review to one canonical data set and compare their performance and unique features, presented in the supplementary materials. Finally, we provide a summary of the dominant methods and an extensive list of the applications, and discuss the open problems and potential research directions in this area.
Keywords: Size estimation, small area estimation, key populations, aggregated relational data
1. Introduction
Estimating the size of hard-to-reach populations is an important problem in a variety of contexts. Governments and humanitarian organizations which aim to eradicate infectious diseases and improve the lives of citizens through treatment programs are interested in population sizes because the treatment target needs to be clear and funds need to be allocated correctly. The Joint United Nations Programme on HIV and AIDS (UNAIDS) aims to limit the spread of HIV by locating large HIV populations. Groups that are particularly vulnerable are known as target- or key-populations. For example, female sex workers (FSW) are among the highest subpopulation living with HIV. It is difficult to estimate the size of FSW directly because of social stigma around sex work and because FSW comprise a relatively small percentage of the general population. Existing approaches to estimate these hard-to-reach populations include mark-recapture, mapping, and venue-based surveys. See Bernard et al. (2010) for a detailed list of population size estimation methods and Sabin et al. (2016) for a comparison of the availability and quality of different data types when estimating certain key subpopulations.
A relatively new method for estimating the size of key populations is the Network Scale-up Method (NSUM), based on the basic scale-up model (Bernard et al., 1989). The authors were in Mexico soon after an earthquake and were interested in estimating the number of people who had died in the earthquake. In this case, where the target population is people who have died from an earthquake, many existing methods were impossible to implement. One author asked people around Mexico City how many people they knew who had died in the earthquake. By leveraging only the responses about how many people each respondent knows, they were able to estimate the number of people who died from the earthquake. The NSUM uses questions from “How many X’s do you know?” surveys to estimate both average network size and subpopulation sizes.
The method provides a relatively cheap, easy, and powerful tool for researchers and can still be applied when it is impossible or difficult to reach the target population directly. Furthermore, the NSUM respects the privacy of the respondents since respondents are not asked about their own characteristics or identifying information of their alters. UNAIDS and WHO published guidelines for the NSUM and noted that the NSUM is advantageous because respondents do not need to reveal their own status, questions can be added to existing household surveys, and sizes of multiple subpopulations can be estimated from one survey (UNAIDS and WHO, 2010). Furthermore, obtaining data through NSUM can be 70 – 80% cheaper than collecting traditional network data (Breza et al., 2017). However, due to biases from violating method assumptions, the UNAIDS and WHO designated the NSUM as a method “under-development” for estimating hard-to-reach populations (UNAIDS and WHO, 2010), prompting the need for further development.
Here we standardize the notation we will use for the rest of the paper and provide a brief introduction to relevant network and NSUM terminology and notation to streamline reading of later sections. Aggregated relational data (ARD) refers to the data collected using “How many X’s do you know?” questions, while NSUM refers to the process of estimating network size or subpopulation size using ARD. Let be the size of the general population, be the size of subpopulation , and be the number of respondents. We denote the ARD responses by , the number of people that respondent reports knowing in subpopulation . The network size, or degree, for person is and for ARD, typically represents the total number of individuals that the respondent recognizes by sight or name, has contacted in the last several years, and can still contact. The degree is typically equal to the number of edges connected to the respondent node in an undirected graph since the connections are generally assumed to be symmetric, although more precisely is equal to the number of out-going edges from the respondent node in a directed graph (e.g. respondents can still know people who have died in an earthquake). Let be the number of subpopulations with known sizes. For simplicity, we assume only one subpopulation size is unknown, denoted by , but stress that any number of subpopulations can be unknown in practice. The popular McCarty ARD dataset (McCarty et al., 2001), contained 29 known subpopulations which included 12 names (Michael, James, Kimberly), Native Americans, commercial pilots, and homicide victims. The three unknown subpopulations were HIV positive individuals, women raped in the past 12 months, and homeless individuals. Thus, the responses from person are denoted . The respondents to ARD surveys are also known as the ego and the people to whom the ego can form ties are called alters (Salganik et al., 2011b).
The basic idea behind NSUM relies on the assumption that the proportion of the subpopulation to the general population is equal to the proportion of the person’s network that belongs to the subpopulation, i.e.
| (1) |
If the degree was known, then it would be easy to solve for using Equation (1). However, estimating the degree is a difficult problem in and of itself (Rogerson, 1997; Dasgupta et al., 2014). A related topic is that of the “small-world problem,” which states that only a small number of connections can connect any two people (Rogerson, 1997). Estimating the average network size is also difficult because it is nearly impossible for someone to recount their entire social network without substantial effort and network size varies dramatically between individuals.
The model in Equation (1) works well under three strong conditions: 1. everyone in the population is equally likely to know someone in subpopulation , 2. for every person in a respondent’s network, the respondent knows every subpopulation the person belongs to, and 3. respondents are able to fully recall everyone in their social network in the allotted time. These assumptions are commonly violated in practice. Barrier effects cause individuals to be more or less likely to know individuals in certain subpopulations, violating condition 1. Transmission errors block individuals from knowing everything about people in their social networks, violating condition 2. Recall errors result when people are unable to quickly count or remember everyone in their social network that belongs to a certain group, violating condition 3 (Bernard et al., 1989; Johnsen et al., 1995; Killworth et al., 1998b; McCarty et al., 2001). We further explore each of these violations in later sections.
There are currently two software packages to analyze ARD using the NSUM, both implemented in R (R Core Team, 2019). The first, NSUM (Maltiel and Baraff, 2015), implements the models proposed by Maltiel et al. (2015). The second, networkreporting (Feehan and Salganik, 2014), fits the generalized network scale-up model proposed by Feehan and Salganik (2016).
During the writing of this review, another review of the network scale-up method was published in The Oxford Handbook of Social Networks (McCormick, 2020). Like the review presented here, the author introduces the NSUM in the context of ARD, delineates the assumptions and pitfalls of the method, and introduces the proposed models to improve the method. While McCormick (2020) walks through the history of the NSUM primarily through the lens of NSUM applications, we instead reflect on NSUM through the modeling perspective. To this end, our review allocates more space to model properties while McCormick (2020) spends more time discussing designing and analyzing an ARD survey. In addition, we apply many of the models discussed in this review to one canonical data set and compare their performance and unique features. Read together, these two reviews provide a near-complete picture of the current NSUM and ARD literature.
The rest of this manuscript is organized as follows. First, we offer a background to ARD and explore the features of the data and problems that arise when collecting ARD. Then, in Section 3, we provide introductions to all significant NSU models. This section is divided into three subsections. In Section 3.1, we discuss the frequentist network scale-up estimators. These basic estimators are the most frequently used in practice due to their ease of use, but also include more complex models with increased flexibility. In Section 3.2, we introduce the Bayesian estimators. These models are reported to have improved the basic methods by better accounting for the sources of bias, but are also more difficult to use since they rely on Bayesian sampling algorithms. In Section 3.3, we review recent estimators that estimate complete network properties using only ARD. After discussing specific models, Section 4 introduces modifications to model estimates. These approaches recognize limitations of the modeling procedures and further calibrate estimates using empirical studies. Final discussion is found in Section 5.
2. Properties of Aggregated Relational Data
Before discussing the various NSUM models, it is important to look at the properties of ARD in more detail. Specifically, there are four key biases that plague ARD: 1) Transmission errors, 2) Barrier effects, 3) Recall errors, and 4) Response biases. As with any survey, ARD can suffer from poor sampling behavior. Recall errors and response biases depend on the survey implementation, while transmission errors and barrier effects depend on the subpopulations and respondents.
2.1. Transmission Error
A response suffers from transmission error when the respondent is unaware that someone in their network belongs to a subpopulation (Killworth et al., 2006). The transmission error violates the assumption that respondents have perfect knowledge of which subpopulations their alters belong to and varies widely between different subpopulations (Killworth et al., 2006; Zheng et al., 2006; Maltiel et al., 2015).
Shelley et al. (1995) studied transmission error by interviewing respondents in key subpopulations. HIV-positive respondents reported that only 49% of their relatives were aware of their HIV-status. However, note that transmission error also exists for easy-to-reach populations. A large percentage of diabetics, twins, Native Americans, and widows and widowers all reported not revealing their status to some members of their social network (Killworth et al., 2006). Focus has primarily been on accounting for the transmission rate of the hard-to-reach subpopulations, although ignoring the transmission rate of the known subpopulation may also significantly influence estimates.
In some cases, researchers can estimate the transmission rate, , of a subpopulation. Also frequently called the visibility factor, represents the fraction of a respondent’s network that is aware the respondent is in the hidden subpopulation. The methods to estimate the visibility factor include expert opinion, comparison of NSU with proxy respondent method, social respect, coming-out ratio, and game of contacts (Haghdoost et al., 2018). Thus, in addition to collecting the ARD, researchers can also collect additional data to estimate the visibility factor. To the best of our knowledge, there has been no study that compares the accuracy of the different methods to estimate the visibility factor.
2.2. Barrier Effect
Barrier effects violate the constant proportion assumption because respondents can be more or less likely to know someone in a subpopulation due to their own characteristics, and are not limited to hard-to-reach populations. (Shelley et al., 1995; Killworth et al., 2006; Shelley et al., 2006; Salganik et al., 2011a). For example, respondents were more likely to know people with whom they shared their race.
The reasons for barrier effects can be both geographical as well as social. For example, the number of Native Americans that a respondent knows is highly correlated to the state in which they reside. On the other hand, doctors are much more likely to know other doctors, regardless of where they live (Killworth et al., 2006). The influence that barrier effects can have on model estimates can be reduced by obtaining a representative sample of the population of interest. Killworth et al. (2006) cited a study that estimated an unusually high HIV-positive prevalence when compared to national surveys because the original study interviewed only Florida residents, increasing the influence of barrier effects. Unlike transmission error, there does not appear to be any feasible way of estimating barrier effects directly without studying every characteristic of each respondent, since barrier effects depend on both the respondent and the subpopulation considered.
2.3. Recall Error
Recall error occurs when respondents inaccurately recall the number of alters they know in a subpopulation (Killworth et al., 2003, 2006; McCormick et al., 2007; Maltiel et al., 2015). In many studies, respondents have only about 30 seconds to recall everyone in their social network that belongs to a subpopulation (Killworth et al., 2003). Thus, respondents will undercount or overestimate the true number of alters. Little if any research has been done on how many people respondents can recall (Killworth et al., 2003). Furthermore, given that a respondent has recalled of possible alters, the probability that they recall another member decreases as increases. Recall bias may also increase as the survey progressing, meaning later questions will suffer from larger recall error (McCarty et al., 2001).
Researchers found that respondents typically overcounted the number of alters in small subpopulations and undercounted the number of alters in large subpopulations (Killworth et al., 2003; Zheng et al., 2006; McCormick et al., 2007). Killworth et al. (2003) proposed a formula to study the relationship between the estimated degree and the rates at which respondents over and undercount. For large subpopulations, respondents have also been observed to round their responses, typically answering in multiples of 5, but enumerate the individuals in their social network for small subpopulations (McCarty et al., 2001; Killworth et al., 2003). For the large subpopulations, respondents relied on “feel” (McCarty et al., 2001). This means that even for subpopulations with low transmission error, like those based on names, the recall error may be larger than for other subpopulations. Little work has been done to reduce these recall errors when collecting ARD.
2.4. Response Bias
Response bias refers to respondents deliberately misreporting the number of individuals they know in the subpopulation. Respondents may be hesitant to report members of stigmatized subpopulations are in their social network. For example, respondents were hesitant to admit knowing FSW (female sex workers) in household settings (Jing et al., 2018). Thus, response bias can reasonably be reduced at the data-collection stage by making respondents feel comfortable enough to truthfully answer the survey.
Snidero et al. (2009) studied how the survey design and implementation affected response bias. The authors showed that question order likely did not seem to be an important factor in producing reliable estimates of degree and subpopulation size. However, individual interviewers did have a significant effect on whether interviews were interrupted, refused, or contained missing fields. Therefore, in order to reduce variability of the ARD and provide consistent estimates, it is most important to provide the interviewers with sufficient training.
NSUM was also combined with the randomized response technique (RRT) to reduce response bias (Jing et al. (2018)). RRT aims to increase the likelihood that a respondent answers sensitive questions by splitting respondents into two groups and randomly asking the respondents either a sensitive or unrelated survey question and ensuring the respondents that only the respondent knows which question they are answering. Since only the respondents knew which question was asked, the researcher must use only the proportion of the sensitive and unrelated questions to calculate the mean response to the sensitive question. Respondents were much more likely to answer the sensitive question truthfully under RRT, leading to reliable NSUM estimates. (Jing et al., 2018). Note, however, that the RRT provided only the average number of FSW that the respondents knew. Thus, the RRT can only be combined with the scale-up estimates that rely on the average number of individuals known in a subpopulation, rather than the number known for each respondent. Based on the current trend of NSU estimators in Section 3, the RRT is useful only for the most basic estimators.
3. Models
In this section, we will discuss the different ARD models. These models can be loosely categorized into three groups: 1) frequentist models which provide subpopulation size estimates, 2) Bayesian models which handle the biases through the prior distributions, and 3) complete network models which focus on estimating network properties using only ARD. Within each subsection, models will be introduced chronologically. Key theoretical properties of the models will also be discussed, although for brevity, full model properties are left to the original publications.
3.1. Frequentist Models
3.1.1. First NSUM Model
Bernard et al. (1989) first proposed the Network Scale-up Model to study the number of people who had died in the 1985 Mexico City earthquake. The authors derived bounds for the average network size and point estimates for the unknown subpopulation size and Bernard et al. (1991) provided additional empirical results from a larger survey. We focus here on the subpopulation size estimates. While the estimator for subpopulation size is limited in application, it provided a powerful stepping stone for future estimators. The first probability estimator makes no assumption for the distribution of the ARD responses, but notes that the probability of the event that a random respondent knows no one in subpopulation (denote this event ) is given by
| (2) |
where the degrees can vary over the integers from to . The authors then assume that either for a random respondent not in the subpopulation , the respondent’s social network is equally likely to have been any subset of size from the general population, or that all subsets of the general population of size were equally likely to be . Using either of these assumptions and several steps of algebra, the authors show that there exists a real number , , such that
| (3) |
where the approximation holds since is small with respect to . Now consider (3) for subpopulations 1 to , with corresponding values to . Since is an increasing function with respect to , if , then . Therefore, if for the unknown subpopulation is such that , the subpopulation size is also bounded, where
| (4) |
Thus, this procedure provides an upper and lower bound for the size of the unknown subpopulation.
3.1.2. Maximum Likelihood Estimator Models
In order to derive more precise size estimates, Killworth et al. (1998a) proposed a Binomial likelihood based estimator. Overall, six different estimators were proposed to model both personal network size and subpopulation size, but the maximum likelihood based estimator proved the most useful for subpopulation size estimation and was later extended into the most frequently used NSU estimator. The estimator, which we call the plug-in MLE (PIMLE), works by first maximizing the following likelihood with respect to :
| (5) |
When are small relative to and are small relative to , the maximum likelihood estimate of is given by
| (6) |
Plugging in these into (1) yields respondent-level subpopulation estimates . The results are then averaged into a single estimate for ,
| (7) |
where each term inside the summation is the estimated unknown subpopulation size from respondent . Note that the degree estimate for each respondent depends only on the responses from that respondent and the known subpopulation sizes. Then, each respondent is weighted equally in the final summation. It was shown that is unbiased. Furthermore, Monte Carlo simulations showed is essentially unbiased for and the back-estimates for were essentially unbiased when more than 20 subpopulations with known sizes were used for verification (Killworth et al., 1998a). Note that no statement about the bias or standard error of can be made since the final plug-in does not make any distributional assumptions.
Killworth et al. (1998b) proposed a modified version of the estimate in (7), which has become the most frequently used NSU estimator for unknown subpopulation size. Studies typically refer to this as the maximum likelihood estimator (MLE). Instead of back-estimating using (1), the MLE method instead maximizes the binomial likelihood
| (8) |
with respect to , where the are fixed at the estimated from (6). Thus, the final estimate is still uses the known subpopulation to estimate the degrees, but estimates by maximizing a likelihood instead of solving the equality of two ratios. The new MLE for is then easily found to be
| (9) |
The critical difference in the unknown subpopulation estimate is that the PIMLE averages estimates from each respondent while the MLE maximizes a likelihood using all respondent data simultaneously. The estimate is unbiased and assuming the is sufficiently small, the authors show the standard error is given by
| (10) |
which decreases as the network sizes increases. When the degrees are small or prevalence is relatively large, the standard error in (10) will be inaccurate.
3.1.3. Weighted Estimators
The MLE estimate of Killworth et al. (1998b) implicitly values information from large known subpopulations more than small known subpopulations. In order to weight the subpopulations equally, the mean of sums (MoS) estimator first estimates the based on each subpopulation, averages those , and then back-estimates based on the (Habecker et al., 2015). Thus, the estimate for network size is given by which then yields the back-estimate for ,
| (11) |
Note that has the same form as , but relies on a different method to calculate the . The MoS estimator was first proposed in Killworth et al. (1998a), but the authors noted that the variance of the estimate is extremely large when one or more of the known subpopulations are small. Habecker et al. (2015) proposed controlling this variance by choosing subpopulations of similar size.
Habecker et al. (2015) also proposed improving the the MLE and the MoS estimator by incorporating weights to adjust for survey characteristics, like probability of selection. Thus, the weighted MLE (9) and the weighted MoS estimator (11) are given by
| (12) |
and
| (13) |
respectively.
3.1.4. Generalized Scale-up Estimators
Feehan and Salganik (2016) developed the generalized scale-up estimator (GNSUM) to estimate the size of a hard-to-reach subpopulation by using the network property that the total number of in-reports equals the total number of out-reports (i.e. if person reports they know trait about person , then person reports that person knows trait about them). The new estimator also requires additional data collected from the hard-to-reach subpopulations called enriched ARD. Enriched ARD is collected by asking members of the unknown population “How many X’s do you know?” and then “How many of these X’s are aware that you belong to population ?” Clearly one limitation of the generalized network scale-up estimator is that it requires directly sampling from the hard-to-reach subpopulation. However, by leveraging the additional information from the enriched ARD, the generalized scale-up estimator can provide more accurate network and subpopulation size estimates.
First, let the hard-to-reach population of size be a subset of the entire population . In practice, researchers sample from , which is a separate subset of called the frame population. Then, let be the number of out-reports from respondent to subpopulation , i.e. how many people respondent knows in subpopulation . Furthermore, let be the number of in-reports to respondent from the entire population, also known as the visibility of person to people in . Thus, if we define to be the total number of in-reports and to be the total number of out-reports, then . Multiplying both sides by , we can write . This must also be the case for the frame population, i.e. . Then, the estimate is found by estimating the numerator and denominator separately, i.e.
| (14) |
The numerator is found using the ARD, where , is the sample population, and is the probability that respondent is included in the sample from the frame population . The enriched ARD is used to find , but the exact details of the estimator are complicated, and for brevity we leave the description of the estimator to the original manuscript.
Feehan and Salganik (2016) make several important connections between their model and the MLE estimator. First, the authors show that the GNSUM is equal to the MLE times several adjustment factors, and the MLE is only correct when all adjustments factors are equal to 1 while the generalized scale-up is correct regardless. Procedures for estimating the adjustments factors are provided in the original manuscript. This decomposition also leads an expression for the bias of the MLE. This bias expression can be used to adjust MLE estimates if the adjustment factors are known or estimated.
Recently, Verdery et al. (2019) extended the generalized network scale-up estimator to allow venue-based sampling designs instead of the link-tracing samples from the original approach, producing the venue-based generalized scale-up estimator (VB-GNSUM). The generalized NSU estimator is difficult to use with venue-based sampling because estimating requires the probability that an individual was included in the sample and because venue-based sampling produces only one sample, rather than the two samples needed for the traditional estimator. Verdery et al. (2019) use the same estimator in Equation 14, but develop new ways to estimate the numerator and denominator.
3.2. Bayesian Models
While the original NSUM literature acknowledged that network sizes varied greatly between individuals and the biases influenced the results, the basic models made it difficult to account for these factors and thus relied on averaging estimates. After the initial development of the NSUM, a wave of Bayesian models dominated the topic, allowing the parameters to vary between individuals and subpopulations and relying on posterior estimates. The Bayesian approach inherently allows us to also consider the joint distribution between and the other model parameters.
3.2.1. The Overdispered Model
Zheng et al. (2006) proposed the first Bayesian model for ARD. Noting that there was large overdispersion in the data, likely a result from barrier effects and varying network sizes, Zheng et al. (2006) proposed the overdispersed model, given by
| (15) |
The parameter represents the expected degree of respondent , represents the proportion of total links that involve subpopulation , and allows for extra variability in the model not accounted for by and . Thus, if the are constant, the number of people that respondent knows in subpopulation is dependent only on the number of people that respondent knows and the relative prevalence of subpopulation . If, however, varies widely across the , then this suggests the presence of one or more of the ARD biases. The flexibility and hierarchical formulation differentiates the Bayesian models from the frequentist models by allowing for more variation in the responses. It is difficult for the frequentist models to accommodate large changes in how likely each respondent is to know someone in each subpopulation. The addition of an overdispersion parameter allows two respondents with equal degrees to have very different responses, which is often seen in the data.
The authors let follow a gamma distribution with mean 1 and shape parameter , which integrated to the following negative binomial model:
| (16) |
where and .
Note that the and are nonidentifiable.Left untouched, this means that an increase in the expected degree of person is equivalent to a decrease in the proportion of total links that involve group . The authors chose to leave the nonidentifiability in the model and instead renormalize the using the rare names (those believed to have the least bias) after running the MCMC chain. Full details of the renormalization process can be found in Zheng et al. (2006).
The main motivation and utility of this paper is the relationship between the subpopulations and the overdispersion, . Figure 4 of the original manuscript provides a nice visual summary of the overdispersion estimates. For the considered data-set, the model showed that the “homeless” and “member of Jaycees” populations had some of the highest overdispersions while the names had some of the lowest. This means that the number of homeless and Jaycees that a respondent knew varied highly between respondents, while the propensity for respondents to know someone named “Stephanie,” for example, was roughly equal. The authors pointed out that homeless populations are both “geographically and socially localized,” explaining the largest range of propensities between respondents.
3.2.2. The Latent Profile Models
McCormick et al. (2010) note that the normalization procedure of Zheng et al. (2006) does not ensure the degrees are estimated accurately since the transmission errors and the barrier effects can still bias the degree estimates. Thus, the authors propose introducing latent nonrandom mixing to account for the biases. Unlike previous models, the McCormick et al. (2010) model estimates the propensity for respondents in ego group to know members of alter group . In their case-study, the authors chose ego and alter groups by crossing gender and age (e.g. males aged 25–64 comprised one ego group), but the ego and alter groups need not match. The initial model is given by
| (17) |
where is the mean, is the proportion of subpopulation within alter group , and is the total number of alter groups. is the mixing coefficient between group and alter group , where
| (18) |
and is the number of respondent alters that belong to subpopulation . Note that the ego groups and the alter are both exhaustive and mutually exclusive. So for any ego group , we have . The proposed model in Equations (17)–(18) accounts for the barrier effects but still suffers from recall bias. To account for recall bias, the authors add a calibration curve derived in McCormick et al. (2007) to the mean of their negative binomial model, i.e. replace the previous mean with
| (19) |
where the full details of the calibration curve can be found in the original manuscript.
The authors also provide suggestions for designing future ARD surveys. The Killworth et al. (1998a) degree estimates are equivalent to the degree estimates from the above model in expectation if either (1) there is random mixing; or (2) the known subpopulations represent a scaled-down population, i.e.
| (20) |
In words, McCormick et al. (2010) explain that “if 20% of the general population is females under age 30, then 20% of the people with the [subpopulations] used also must be females under age 30.” The utility of this requirement is that if the ARD survey is well-designed, then simple models can have the same accuracy as more complicated models. Strategies for designing such a survey and guidelines for understanding the standard error of the estimates are also provided in the original manuscript.
Note, however, that the McCormick et al. (2010) model does not estimate the size of unknown subpopulations, focusing instead on estimating the individual degree and the distribution of network sizes. In a follow-up paper, McCormick and Zheng (2012) extended the McCormick et al. (2010) model to estimate the unknown subpopulation sizes via MCMC. Via a two-stage estimation procedure, the authors first use the subpopulations where size is known to fit the model in (17) and (19) and then estimate the latent profiles for unknown subpopulation conditional on the estimated values.
3.2.3. Maltiel et al. (2015)
Maltiel et al. (2015) introduced five additional models of increasing complexity to estimate unknown subpopulation sizes from NSUM data and implemented the models in the NSUM R package (Maltiel and Baraff, 2015; R Core Team, 2019). Figure 1 in Maltiel et al. (2015) contains a helpful flowchart detailing the four basic proposed models. We discuss only the most complex version of the model and direct the reader to the original manuscript for the other models.
In order to address barrier effects, transmission effects, and recall error, Maltiel et al. (2015) proposed the recall adjustment model, given by
| (21) |
where , , , and . The parameters handle the recall error, handle the barrier effects, (fixed at 1 for known subpopulations) handle the transmission effects, and represents the varying degree between respondents. Furthermore, the hyperparameters are set to . Full prior settings for the hyperparameters are omitted here for clarity. The advantage of this model is that the parameters are clearly separated into the degree and the bias terms. Thus, the response is influenced both by the number of people that respondent knows as well as the subpopulation size and the biases. Estimating the transmission effect for the unknown subpopulation requires some estimate of the visibility factor from additional data sources, for example via the game of contacts.
Ultimately, the authors noted that this complex model was difficult to estimate and proposed removing the parameters and accounting for the recall error post hoc. As others have shown (Killworth et al., 2003; Zheng et al., 2006; McCormick et al., 2007, 2010), respondents seemed to over-report the number of people they knew in small subpopulations and under-report for large subpopulations. The post hoc adjustments are similar to approaches used by others, so details of the adjustments are provided in Section 4.2.
3.2.4. Teo et al. (2019)
In order to account for transmission error and barrier effects, Teo et al. (2019) proposed two new models that include respondent demographics as regression coefficients. Some of the covariates capture overall respondent characteristics, like age and sex. These covariates can be used to capture additional trends in the response. Other covariates measure the perception that each respondent has for each subpopulation. The main idea is if a respondent views a subpopulation poorly, then they are less likely to know individuals from that subpopulation. The transmission error model assumes a Poisson distribution for the ARD given by
| (22) |
while transmission error and barrier effect model is then given by
| (23) |
where U represents the upper bound of the Likert scale, is respondent response to the question, and represents the column-centered covariate matrix. Like McCormick et al. (2010); McCormick and Zheng (2012), the model relies on respondent characteristics, but treats them more as predictors in a regression framework rather than discretizing the predictors jointly into groups.
3.3. Complete Network Models
Full network data is common in a variety of disciplines. For example, economists can use full social networks to determine whether someone is more likely to save more money when another individual monitors their saving progress (Breza and Chandrasekhar, 2019). Recently, researchers have focused on using ARD as a substitute for full social network studies since ARD is significantly cheaper and easier to collect. ARD has been estimated to be 70–80% cheaper to collect that full network surveys (Breza et al., 2017). The question that arises is if and when ARD can be used in place of full social network data. While this line of research deviates significantly from the predominately population size estimation focus of previous methods, we want to make the reader aware of this new research area. Furthermore, to the best of our knowledge, ARD has not been studied in connection with exponential random graph models.
3.3.1. Latent Surface Model
ARD are ultimately partially observed or sampled network data. Thus, McCormick and Zheng (2015) developed a latent surface model, a popular model for complete network data, to analyze incomplete networks like ARD. The proposed model is based on the “projection model” of Hoff et al. (2002), relaxing the complete graph model to handle the incomplete nature of the ARD. In this framework, the propensity for person and to know each other is proportional to the distance between person and in the latent geometry. The latent geometry, in their case, is a -dimensional hypersphere. Furthermore, for subpopulation , the collected ARD represents , where equals 1 if person and know each other and 0 otherwise. Thus, if we denote the latent positions of and as and , then the distribution of is approximately Poisson distributed with rate . In the complete network case, are observed and known, while for ARD they are unobserved, making it impossible to calculate directly. Instead, the rate is approximated by
| (24) |
where is the size of . After computing the expectation of the observed data and reparameterizing the model in terms of , the likelihood of the latent surface model for ARD is found to be
| (25) |
where is the normalizing constant of the von-Mises Fisher distribution and is the angular distance between respondent and the center of subpopulation . The authors propose a Metropolis MCMC algorithm to sample draws from the posterior distribution.
The latent surface model is closely related to the overdispersed model in Zheng et al. (2006), and McCormick and Zheng (2015) offer a more detailed comparison of the two models. As opposed to the overdispersed model, the latent surface model can put two subpopulations on opposite sides of a latent sphere despite the two subpopulations having similar sizes and dispersion. McCormick and Zheng (2015) observed from one data set that individuals who reported knowing more people with AIDS also reported knowing more religious individuals. This ability to view the relationship between subpopulations is the main benefit of the latent surface model, but comes at the cost of increased computation.
3.3.2. Network Statistics
Breza et al. (2017) extended the latent surface models and showed that ARD can be used to estimate node- or graph-level statistics under certain situations and provided insight into when ARD is sufficient. Examples of these statistics are individual centrality and average path length of the graph. The authors showed that they could reproduce the findings of complete network studies using only ARD. As the focus of this paper is remarkably different from our previous discussions, we refer the reader to the original manuscript for model details.
Breza et al. (2019) further developed the theory behind why and when ARD is sufficient to estimate model parameters for complete networks. One key result from their manuscript is that under certain graphs and given a sufficiently large graph, certain parameter estimates from ARD are consistent. The authors further develop a system to identify when ARD is sufficient to recover graph statistics. While there are too many results to include here, ARD proves to be an extremely useful tool to estimate graph statistics, especially considering the cost savings and ease of implementation. If the interest is statistical inference of various population parameters of networks, network sampling methods (Bhattacharyya et al., 2015; Green and Shalizi, 2017; Levin and Levina, 2019; Lin et al., 2020) might be an alternative to ARD and NSUM.
4. Model Calibration
A significant portion of the NSU literature focuses on calibrating the crude estimates from the NSU models in Section 3 through post hoc adjustments. The calibrations scale the model estimates to correct for ARD biases. Calibrations exist for transmission errors and recall errors, but no adjustments exist for barrier effects because of the aforementioned difficulties in estimating the barrier effects. Note, the original adjustments discussed in this section were typically developed for specific models, but the ideas can be often applied to others, so we discuss them in a general setting.
4.1. Transmission Calibration
The two general approaches to account for transmission error are to use known subpopulations that have low transmission error and to correct crude estimates using the visibility factors. Using only low transmission error subpopulations for estimation does not require any additional data sets and can be applied to most NSU studies. The visibility factor, however, requires an additional sample to estimate and can generally only be used for the hidden subpopulation of interest.
Transmission error can be reduced by using only known subpopulations that are unlikely to have transmission error, like names (McCormick et al., 2010). It is unlikely that a respondent reports knowing someone without also knowing their name, thereby removing bias from the estimates of degree size. However, the approach does not eliminate the transmission error present in the hard-to-reach subpopulation, where the unknown subpopulation size estimates can still be biased.
More sophisticated methods for selecting the known subpopulations have also been proposed. One approach is to back-estimate each known subpopulation using a leave-one-out procedure and remove all subpopulations which are poorly estimated based on the ratio between the back-estimate and the known size (Guo et al., 2013). Similarly, Habecker et al. (2015) proposed trimming one subpopulation at a time by removing only the worst performing subpopulation ratio in the leave-one-out back-estimates. This stepwise trimming addresses the fact that all model estimates will change after removing any subpopulation, so subpopulations that originally had poor performing ratios might actually perform well after removing one other subpopulation.
The most common method to account for transmission bias is to scale the crude subpopulation size estimates by some scaling factor estimated from an additional data set. The visibility factors are used to directly scale the subpopulation size estimates from the NSU procedure by dividing the crude NSU estimate by the visibility factor. For example, if only 50% of a female sex worker’s social network is aware that they are a female sex worker, than the FSW subpopulation estimate is divided by 0.5 to account for the transmission bias. Given a estimate of the visibility factor, this approach can be easily applied to any estimator. However, only the unknown subpopulation size estimates are scaled and not the degree estimates. Combining the visibility factor method with the only names approach in McCormick et al. (2010) would likely lead to better degree and subpopulation size estimates.
4.2. Calibration Curve
While Shelley et al. (1995), Killworth et al. (2003), and Shelley et al. (2006) investigated recall bias in ARD, accounting for the recall bias in the models proved difficult. However, several approaches have been proposed, none of which require any additional data sets. McCormick et al. (2007) were the first to propose a method to account for recall bias. In order to account for models overestimating small subpopulations and underestimating large subpopulations, the authors constructed a “calibration curve” which the authors believed to match the relationship between subpopulation size and recall bias. The calibration curve attempts to scale the recalled number of alters to be equal to the true number of alters in subpopulation . As defined in McCormick et al. (2007), let be the proportion of ties in the social network for subpopulation , be for the recalled social network, and then define to be the calibration curve. The calibration curve was first defined as
| (26) |
where controls how fast the derivative of this curve approaches and controls at what value of the curve changes from correcting for over-reporting to under-reporting. The parameters and are then estimated using the subpopulation size estimates without any bias correction.
As mentioned in Section 3.2.3, Maltiel et al. (2015) proposed a similar calibration curve. Their calibration works by first treating each known subpopulation as unknown one at a time and estimating the size of that population, . Then, the errors-in-variables model is fit via maximum likelihood to estimate , , and the variances of and , where and . Estimates of from the posterior are then transformed using the estimated , , and an additional random noise. The transformation is given by
| (27) |
where is the -th MCMC sample from the posterior distribution of and .
5. Discussion
In this manuscript, we discussed the properties of ARD, explored the wide range of models for ARD, and summarized common modifications to these model estimates. For a succinct summary, Table 1 lists these models, the primary objectives, and additional modeling properties. Many of the models have multiple objectives and estimate several properties, so we record what appears to be the primary objective of the original manuscript. ARD is an increasingly popular survey type for estimating the size of unknown subpopulations due to its relatively cheap method of collecting network data and that individuals from hard-to-reach populations do not need to be surveyed. There are several biases that make modeling ARD difficult, but extensive research has been performed to improve the accuracy and precision of ARD. Recently, ARD has been used in place of full social networks to estimate network properties, and this is a promising area of research that deserved more attention. We include an extensive list of implemented and some proposed NSU studies in the Supplementary Material. The additional abbreviations we use for the target subpopulations are MCFSW (male client of a female sex worker), MSM (men who have sex with men), RWOS (relationship with opposite sex), EPMS (extra/pre-marital sex), and PED (performance enhancing drugs).
Table 1:
Brief summary of NSUM models.
| Model | Primary Objective | Requires External Data | Requires Respondent Covariates | Distributional Assumptions | Bayesian |
|---|---|---|---|---|---|
| Bernard et al. (1989) and Bernard et al. (1991) | Degree Estimation | N/A | |||
| Killworth et al. (1998a) and Killworth et al. (1998b) | Size Estimation | Binomial | |||
| Zheng et al. (2006) | Overdispersion/social structure | Negative Binomial | |||
| McCormick et al. (2010) | Degree Estimation | ✓ | Negative Binomial | ✓ | |
| McCormick and Zheng (2012) | Size Estimation | ✓ | Negative Binomial | ✓ | |
| Habecker et al. (2015) | Size Estimation | Binomial | |||
| Maltiel et al. (2015) | Size Estimation | ✓ | Binomial | ✓ | |
| Feehan and Salganik (2016) | Size Estimation | ✓ | N/A | ||
| Verdery et al. (2019) | Size Estimation | ✓ | N/A | ||
| Teo et al. (2019) | Size Estimation | ✓ | Poisson | ✓ | |
| McCormick and Zheng (2015) | Complete Network Statistics | Poisson | ✓ | ||
| Breza et al. (2017) | Complete Network Statistics | Poisson | ✓ | ||
| Breza et al. (2019) | Complete Network Statistics | N/A |
There are a variety of popular methods for estimating key populations, including census/enumeration, capture-recapture, multiplier, and respondent-driven sampling. Population size estimates from these methods and NSUM can vary widely, and in a study of heavy drug users in Curitiba, Brazil, Salganik et al. (2011a) found the NSUM and the generalized NSUM estimates were multiple times larger than the size estimates from comparable direct and multiplier methods. While several reasons for the discrepancy were discussed, there was no clear conclusion.
Respondent-driven sampling and NSUM are related through their reliance on respondents’ social networks, albeit it this connection takes a different form in each method. Respondent-driven sampling relies on the social networks of members of the target population and specifically how connections the respondent knows that are also in the target population. Thus, future responses depend highly on each respondent. This procedure makes it easy to sample directly from the target population, but may come at a cost. For respondent-driven sampling, the standard error of the population size estimate may converge at a much slower rate than expected as the number of participants increases Rohe et al. (2019). Meanwhile, the NSUM relies on social networks of individuals from the general population and each respondent is assumed to be independent of the others. These factors suggest that the NSUM does not suffer from the same problems of convergence.
There are many desirable features of the more traditional estimation techniques, but like NSUM, they all rely heavily on data assumptions. While multiple studies have compared estimates from the more traditional estimation techniques and NSUM, there is still not a clear consensus that any model out-performs the other. The main advantage of the NSUM is respondents do not need to be from the target population. This is beneficial both because it is often difficult simply to survey members from the target population, but also because ARD surveys are cheap and can be implemented quickly. Furthermore, the NSUM can easily estimate the size of impossible-to-reach population like those who have died in an earthquake and estimate the size of multiple unknown subpopulations from the same ARD survey with little extra cost. However, these features do not necessarily lead to more accurate results, and estimates depend significantly on the design of the ARD survey as well as the details of the implementation, like question order. Furthermore, NSUM relies more heavily both on the truthfulness and accuracy of the respondent answers than some of the other methods. Despite these limitations, NSUM models have developed to handle many of the biases present in the data and additional information can be included in the models, making NSUM an attractive and trustworthy method.
The models for ARD range from very simple MLE estimators to more complex latent surface approaches. Each method has different purposes and assumptions in place, so we refrain from recommending any specific model. Furthermore, the performance of each model has been shown to depend heavily on the sample and place of study. The practical challenge of implementing these methods is also of importance. The frequentist estimators are much easier to implement and have remained the most popular approaches. However, the more complex models offer potential gains in accuracy as well as potentially additional insight into the subpopulations. For example, at the cost of additional computation, the latent surface model proposed in McCormick and Zheng (2015) can place two subpopulations with similar dispersion parameters on opposite sides of the latent surface, indicating that the subpopulations are socially distinct, while the Zheng et al. (2006) model estimates only the dispersion parameters. This is not to say that the latent surface model outperforms the overdispersed model, but rather that each model has a different niche. We implement many of the models on a canonical ARD data set and present the results in the Supplementary Material. In there, we compare the performance of various methods and comment on their unique features.
ARD, and the extension of NSUM, is a promising area of research, and the models can be extended beyond size estimation. One interesting approach already considered was to invert the problem and estimate the ARD from known subpopulation sizes and degrees. Moody (2005) estimated how many people know someone affected by the United States war on terror. Furthermore, there are many hard-to-reach populations that can benefit from the NSUM. Shelton (2015) proposed using the NSUM to estimate the prevalence of sex trafficked individuals, noting how difficult it is to obtain a direct estimate. Another promising extension of ARD is to consider other forms of “How many X’s do you know?” questions. The generalized NSUM estimator requires enriched ARD. Asking additional questions to the respondents may yield helpful information that can be used in a new estimator. Finally, Breza et al. (2017) suggest that ARD should move beyond social networks. For example, the authors consider the question “How many links does the firm have to firms with trait ?” This question takes a step away from social networks and generalizes ARD to weighted and directed graphs.
The data and models still have limitations, of course. Estimates for known subpopulation sizes are often significantly incorrect. Approaches for estimating the transmission error are often expensive or impossible for certain subpopulations. Many approaches have reduced response bias, but the success of these methods depends heavily on the unknown subpopulation as well as where the study is implemented. Some papers report eliminating the recall error in the data, but it is difficult to validate these approaches since little research has been done on just recall error. Adjustments are based on leave-one-out estimates of known subpopulations, but the discrepancy between these estimates and the truth are not shown to be due to recall error.
With enough improvement, the NSUM will hopefully drop its “under-development” label and be a useful and accurate method for estimating hard-to-reach populations. The models have been used in a large number of real-world studies and have offered promising results in the field of size estimation. The simplicity of the method is attractive and we hope this review inspires new and exciting developments in the field.
Supplementary Material
Acknowledgment
The authors would like to thank Dr Tian Zheng for the helpful discussions, and Dr Tyler McCormick and Dr Alex Cook for sharing their code.
This work was supported by the National Institute of Allergy and Infectious Diseases of the National Institutes of Health under award number R01AI136664.
References
- Bernard HR, Hallett T, Iovita A, Johnsen EC, Lyerla R, McCarty C, Mahy M, Salganik MJ, Saliuk T, Scutelniciuc O, et al. (2010). Counting hard-to-count populations: the network scale-up method for public health. Sexually transmitted infections, 86(Suppl 2):ii11–ii15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bernard HR, Johnsen EC, Killworth PD, and Robinson S (1989). Estimating the size of an average personal network and of an event subpopulation. In The small world, pages 159–175. Ablex Press. [Google Scholar]
- Bernard HR, Johnsen EC, Killworth PD, and Robinson S (1991). Estimating the size of an average personal network and of an event subpopulation: Some empirical results. Social science research, 20(2):109–121. [Google Scholar]
- Bhattacharyya S, Bickel PJ, et al. (2015). Subsampling bootstrap of count features of networks. Annals of Statistics, 43(6):2384–2411. [Google Scholar]
- Breza E and Chandrasekhar AG (2019). Social networks, reputation, and commitment: evidence from a savings monitors experiment. Econometrica, 87(1):175–216. [Google Scholar]
- Breza E, Chandrasekhar AG, McCormick TH, and Pan M (2017). Using aggregated relational data to feasibly identify network structure without network data. Technical report, National Bureau of Economic Research [DOI] [PMC free article] [PubMed] [Google Scholar]
- Breza E, Chandrasekhar AG, McCormick TH, and Pan M (2019). Consistently estimating graph statistics using aggregated relational data. arXiv preprint arXiv:1908.09881 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dasgupta A, Kumar R, and Sarlos T (2014). On estimating the average degree In Proceedings of the 23rd international conference on World wide web, pages 795–806. [Google Scholar]
- Feehan D and Salganik M (2014). The networkreporting package
- Feehan DM and Salganik MJ (2016). Generalizing the network scale-up method: a new estimator for the size of hidden populations. Sociological methodology, 46(1):153–186. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Green A and Shalizi CR (2017). Bootstrapping exchangeable random graphs. arXiv preprint arXiv:1711.00813 [Google Scholar]
- Guo W, Bao S, Lin W, Wu G, Zhang W, Hladik W, Abdul-Quader A, Bulterys M, Fuller S, and Wang L (2013). Estimating the size of hiv key affected populations in chongqing, china, using the network scale-up method. PloS one, 8(8):e71796. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Habecker P, Dombrowski K, and Khan B (2015). Improving the network scale-up estimator: Incorporating means of sums, recursive back estimation, and sampling weights. PloS one, 10(12). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Haghdoost A, Gohari MA, Mirzazadeh A, Zolala F, and Baneshi MR (2018). A review of methods to estimate the visibility factor for bias correction in network scale-up studies. Epidemiology and health, 40. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hoff PD, Raftery AE, and Handcock MS (2002). Latent space approaches to social network analysis. Journal of the american Statistical association, 97(460):1090–1098. [Google Scholar]
- Jing L, Lu Q, Cui Y, Yu H, and Wang T (2018). Combining the randomized response technique and the network scale-up method to estimate the female sex worker population size: an exploratory study. Public health, 160:81–86. [DOI] [PubMed] [Google Scholar]
- Johnsen EC, Bernard HR, Killworth PD, Shelley GA, and McCarty C (1995). A social network approach to corroborating the number of aids/hiv+ victims in the us. Social networks, 17(3–4):167–187. [Google Scholar]
- Killworth PD, Johnsen EC, McCarty C, Shelley GA, and Bernard HR (1998a). A social network approach to estimating seroprevalence in the united states. Social networks, 20(1):23–50. [DOI] [PubMed] [Google Scholar]
- Killworth PD, McCarty C, Bernard HR, Johnsen EC, Domini J, and Shelley GA (2003). Two interpretations of reports of knowledge of subpopulation sizes. Social networks, 25(2):141–160. [Google Scholar]
- Killworth PD, McCarty C, Bernard HR, Shelley GA, and Johnsen EC (1998b). Estimation of seroprevalence, rape, and homelessness in the united states using a social network approach. Evaluation review, 22(2):289–308. [DOI] [PubMed] [Google Scholar]
- Killworth PD, McCarty C, Johnsen EC, Bernard HR, and Shelley GA (2006). Investigating the variation of personal network size under unknown error conditions. Sociological Methods & Research, 35(1):84–112. [Google Scholar]
- Levin K and Levina E (2019). Bootstrapping networks with latent space structure. arXiv preprint arXiv:1907.10821 [Google Scholar]
- Lin Q, Lunde R, and Sarkar P (2020). On the theoretical properties of the network jackknife In International Conference on Machine Learning, pages 6105–6115. PMLR. [Google Scholar]
- Maltiel R and Baraff AJ (2015). NSUM: Network Scale Up Method. R package version 1.0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Maltiel R, Raftery AE, McCormick TH, and Baraff AJ (2015). Estimating population size using the network scale up method. The annals of applied statistics, 9(3):1247. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McCarty C, Killworth PD, Bernard HR, Johnsen EC, and Shelley GA (2001). Comparing two methods for estimating network size. Human organization, 60(1):28–39. [Google Scholar]
- McCormick TH (2020). The network scale-up method. The Oxford Handbook of Social Networks, page 153. [Google Scholar]
- McCormick TH, Salganik MJ, and Zheng T (2010). How many people do you know?: Efficiently estimating personal network size. Journal of the American Statistical Association, 105(489):59–70. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McCormick TH and Zheng T (2012). Latent demographic profile estimation in hard-to-reach groups. The annals of applied statistics, 6(4):1795. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McCormick TH and Zheng T (2015). Latent surface models for networks using aggregated relational data. Journal of the American Statistical Association, 110(512):1684–1695. [Google Scholar]
- McCormick TH, Zheng T, et al. (2007). Adjusting for recall bias in “how many x’s do you know?” surveys. In Proceedings of the joint statistical meetings [Google Scholar]
- Moody J (2005). Fighting a hydra: A note on the network embeddedness of the war on terror. Structure and Dynamics, 1(2). [Google Scholar]
- R Core Team (2019). R: A Language and Environment for Statistical Computing R Foundation for Statistical Computing, Vienna, Austria. [Google Scholar]
- Rogerson PA (1997). Estimating the size of social networks. Geographical Analysis, 29(1):50–63. [Google Scholar]
- Rohe K et al. (2019). A critical threshold for design effects in network sampling. Annals of Statistics, 47(1):556–582. [Google Scholar]
- Sabin K, Zhao J, Calleja JMG, Sheng Y, Garcia SA, Reinisch A, and Komatsu R (2016). Availability and quality of size estimations of female sex workers, men who have sex with men, people who inject drugs and transgender women in low-and middle-income countries. PLoS One, 11(5). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Salganik MJ, Fazito D, Bertoni N, Abdo AH, Mello MB, and Bastos FI (2011a). Assessing network scale-up estimates for groups most at risk of hiv/aids: evidence from a multiple-method study of heavy drug users in curitiba, brazil. American journal of epidemiology, 174(10):1190–1196. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Salganik MJ, Mello MB, Abdo AH, Bertoni N, Fazito D, and Bastos FI (2011b). The game of contacts: estimating the social visibility of groups. Social networks, 33(1):70–78. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shelley GA, Bernard HR, Killworth P, Johnsen E, and McCarty C (1995). Who knows your hiv status? what hiv+ patients and their network members know about each other. Social networks, 17(3–4):189–217. [Google Scholar]
- Shelley GA, Killworth PD, Bernard HR, McCarty C, Johnsen EC, and Rice RE (2006). Who knows your hiv status ii?: Information propagation within social networks of seropositive people. Human Organization, 65(4):430. [Google Scholar]
- Shelton JF (2015). Proposed utilization of the network scale-up method to estimate the prevalence of trafficked persons. In Forum on Crime & Society, volume 8. [Google Scholar]
- Snidero S, Zobec F, Berchialla P, Corradetti R, and Gregori D (2009). Question order and interviewer effects in cati scale-up surveys. Sociological Methods & Research, 38(2):287–305. [Google Scholar]
- Teo AKJ, Prem K, Chen MI, Roellin A, Wong ML, La HH, and Cook AR (2019). Estimating the size of key populations for hiv in singapore using the network scale-up method. Sexually transmitted infections, 95(8):602–607. [DOI] [PMC free article] [PubMed] [Google Scholar]
- UNAIDS and WHO (2010). Guidelines on estimating the size of populations most at risk to hiv Geneva, Switzerland: World Health Organization, page 51. [Google Scholar]
- Verdery AM, Weir S, Reynolds Z, Mulholland G, and Edwards JK (2019). Estimating hidden population sizes with venue-based sampling: Extensions of the generalized network scale-up estimator. Epidemiology, 30(6):901–910. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zheng T, Salganik MJ, and Gelman A (2006). How many people do you know in prison? using overdispersion in count data to estimate social structure in networks. Journal of the American Statistical Association, 101(474):409–423. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
