Skip to main content
International Journal of Methods in Psychiatric Research logoLink to International Journal of Methods in Psychiatric Research
. 2022 Oct 17;32(2):e1941. doi: 10.1002/mpr.1941

Design and methods of the Longitudinal Eating Disorders Assessment Project research consortium for veterans

Kelsie T Forbush 1,, Trevor J Swanson 1, Melinda Gaddy 2, Mary Oehlert 2, Alesha Doan 1, Robert W Morgan 1, Colin O’Brien 1, Yiyang Chen 1, Kylie Christian 1, Q Chelsea Song 3, David Watson 4, Joanna Wiese 5
PMCID: PMC10242201  PMID: 36251947

Abstract

Introduction

Military service members must maintain a certain body mass index and body fat percentage. Due to weight‐loss pressures, some service members may resort to unhealthy behaviors that place them at risk for the development of an eating disorder (ED).

Objectives

To understand the scope and impact of EDs in military service members and veterans, we formed the Longitudinal Eating Disorders Assessment Project (LEAP) Consortium. LEAP aims to develop novel screening, assessment, classification, and treatment tools for veterans and military members with a focus on EDs and internalizing psychopathology.

Methods

We recruited two independent nationally representative samples of post‐9/11 veterans who were separated from service within the past year. Study 1 was a four‐wave longitudinal survey and Study 2 was a mixed‐methods study that included surveys, structured‐clinical interviews, and qualitative interviews.

Results

Recruitment samples were representative of the full population of recently separated veterans. Sample weights were created to adjust for sources of non‐response bias to the baseline survey. Attrition was low relative to past studies of this population, with only (younger) age predicting attrition at 1‐week follow‐up.

Conclusions

We expect that the LEAP Consortium data will contribute to improved information about EDs in veterans, a serious and understudied problem.

Keywords: assessment, department of defense, disordered eating, eating disorders, longitudinal, mental health, nationally representative, psychiatric disorders, VADIR, veterans


Eating disorders (EDs) are serious psychiatric disorders that negatively impact physical and mental health (Brown & Mehler, 2013; Forney et al., 2016), emotional wellbeing, work productivity, and social relationships (Streatfeild et al., 2021). EDs have a combined prevalence of 13.1%–15% among young women (Allen et al., 2013; Stice et al., 2013) and 3% among young men in the general population (Allen et al., 2013), which is similar to the prevalence of depression or anxiety. EDs affect people of all ethnicities and socioeconomic levels, and occur across the lifespan (Mitchison et al., 2014; van Hoeken & Hoek, 2020). EDs have one of the highest mortality rates among psychiatric disorders (Fichter & Quadflieg, 2016; Iwajomo et al., 2021), with standardized mortality ratios that are two times higher than mortality due to heavy smoking (Chesney et al., 2014). Moreover, EDs are associated with suicide‐related mortality in the USA, with suicide rates similar to (or higher than) depression, opioid addiction, or schizophrenia (Chesney et al., 2014). Due to the effects of EDs on general physical health, civilians with EDs incur additional healthcare costs associated with medical events (e.g., myocardial infarction) or due to comorbid health conditions, such as obesity, which is more common among people with EDs than the general population (Striegel Weissman & Rosselli, 2017). Similar to estimates in civilian populations, veteran patients with an ED incurred 1‐year adjusted total healthcare costs that were $18,152 higher than matched veterans without an ED (Striegel‐Moore et al., 1999), suggesting that there is a critical need to identify and treat EDs in veterans.

Veteran men and women appear to be at risk for the development of an ED (Cuthbert et al., 2020; Masheb et al., 2021). For example, veterans have elevated rates of disordered eating, overweight, and obesity compared to non‐veteran civilians (Cuthbert et al., 2020). Although there is a wide prevalence range, recent studies suggest that 14.1%–32.8% of veteran women and 4.1%–18.8% of veteran men have a current ED (Arditte Hall et al., 2017; Forman‐Hoffman et al., 2012; Masheb et al., 2021). Military personnel must adhere to body mass index (BMI) and body fat percentage requirements, as well as meet physical performance standards that are assessed every six months. Failure to maintain these standards can result in referral to weight‐loss programs and, eventually, discharge from service. The emphasis on body weight within the military may promote unhealthy attempts at weight loss and the development of eating disorders (EDs). For example, being on a weight loss diet while in the military is associated with an increased risk for ED symptoms (Mitchell et al., 2021). If left untreated, EDs result in significantly greater Veterans Affairs (VA) healthcare utilization and physical and psychiatric morbidity and mortality.

Despite the public‐health importance of addressing EDs in active‐duty and veteran populations, few nationally representative studies of veterans have included ED and assessments of related comorbid psychopathology. To address the need to better understand the scope and impact of EDs in military service members and veterans, the Department of Defense's Congressionally Directed Medical Research Program (CDMRP) included EDs as an Area of Encouragement for the first time in 2017. Since 2017, the CDMRP has continued to fund studies on EDs, which led to the development of the Longitudinal Eating Disorders Assessment Project (LEAP) research consortium. LEAP brings together researchers and clinicians across state lines to develop novel screening, assessment, classification, and treatment tools for veterans and military members with a focus on EDs and other forms of internalizing psychopathology. Our consortium also includes researchers with expertise in qualitative methods and public policy, with a focus on identifying formal and informal policies and practices that can potentially inform ED treatment implementation efforts in military‐relevant populations.

In this paper, we report the design, protocol, and methods of the first two datasets from the LEAP research consortium. We also report initial attrition analyses from the first wave of our longitudinal dataset (Study 1). As we describe in the paper, other data waves from Study 1 and Study 2 were ongoing at the time of this writing. Our primary objective was to develop and validate a military‐specific transdiagnostic screening tool for eating and related disorders that can eventually be implemented in healthcare settings. The scientific premise that guided our research was that an integrated transdiagnostic framework for assessing and classifying EDs could lead to improvements in the ability to identify veterans who are most at‐risk for poor post‐discharge psychosocial adjustment and who need a referral to treatment services. Our primary aim was to validate/develop a transdiagnostic ED screening tool. Secondary aims were to (1) evaluate longitudinal relationships among ED symptoms, internalizing, and externalizing psychopathology, and (2) identify barriers to obtaining ED treatment during military service.

We recruited two independent nationally representative samples of veterans from the VA/DoD Identity Repository (VADIR), a VA office with access to DoD records and the only comprehensive registry of military personnel and veterans in the USA. Study 1 was a four‐wave longitudinal survey‐based study that assessed veterans at baseline, 1‐week, 3‐months, and 6‐months. Study 2 was a cross‐sectional study that included surveys, diagnostic interviews, and qualitative interviews. Although the first two studies from the LEAP consortium were primarily focused on EDs and internalizing disorders, the datasets include a wide range of demographic, military‐service related, and other psychopathology constructs that may be of interest to general psychiatric researchers. Given the DoD's focus on publicly available data and our own commitment to open‐science practices, de‐identified data (with personal information removed) will be released to the public when it becomes available in compliance with the General Data Protection Regulation. Thus, the goal of this paper is to provide information on the design, protocol, methods for LEAP studies and initial attrition analyses for the first wave of the Study 1 LEAP dataset.

1. METHODS

1.1. Ethics approval

Ethical approval for the project was provided by the University of Kansas, VA Eastern Kansas Health Care System (VAEKHCS), and the Department of Defense Human Research Protections Office. All participants provided informed consent prior to engaging in any study‐related procedures. The University of Kansas received a HIPAA waiver to recruit veterans who separated from service within the past year from VADIR.

1.2. Study design overview

Study 1 is a four‐wave longitudinal study that included online and paper‐based surveys. Participants were assessed at baseline, 1‐week follow‐up, 3‐month follow‐up, and 6‐month follow‐up. Study 2 is a cross‐sectional study that includes self‐report surveys, structured clinical interviews, and qualitative interviews. In Study 2, veterans will be recruited to participate in a 1‐h structured clinical interview that assesses current and lifetime EDs, mood disorders, generalized anxiety disorder, and post‐traumatic stress disorder. Veterans who were diagnosed with an ED during the structured clinical interview will be invited to participate in an additional qualitative interview. Veterans with EDs will be recruited until we achieve our target sample size for qualitative interviews (N = 50). Qualitative methods will include semi‐structured, open‐ended, in‐depth interviews that range between 45 and 60 min. These exploratory interviews are designed to identify institutional barriers or informal practices in the military that created obstacles for participants seeking ED and trauma‐related care during active duty. Qualitative interviews will be piloted with 10% of the total sample (N = 5) to assess the validity of the questions and modify them according to participants' feedback. Interviews will be recorded and transcribed verbatim. Transcripts will be coded in‐vivo (i.e., using the words of participants) whenever possible, using an inductive, open coding format to catalog dominant and frequent domains (Charmaz, 2006). Taxonomic analysis, which entails developing a classification system to synthesize data, will be used to inventory each domain (Bradley et al., 2007).

1.3. Enrollment status

Baseline and 1‐week follow‐up recruitment for Study 1 is complete, whereas the 3‐ and 6‐month follow‐ups for Study 1 are ongoing. Study 2 launched in March 2022 and recruitment is ongoing (current participant numbers are reported in Table 1).

TABLE 1.

Study 1 baseline sample characteristics

graphic file with name MPR-32-e1941-g008.jpg

1.3.1. VADIR data extraction

Based on power calculations, anticipated attrition, and expected response rates, our Study 1 target sample size was N = 1000 veterans at baseline. For Study 2, our target sample size was N = 400. To achieve our targeted N's, we recruited two separate, non‐overlapping nationally representative samples of veterans from the VADIR who separated from service within the past year. Our rationale for recruiting veterans who discharged within the past year was to collect information as temporally close to their military service as possible to help minimize retrospective recall biases about military‐relevant experiences. VADIR has been used in past studies of mental health to obtain nationally representative samples of veterans (Aronson et al., 2019; Barth et al., 2014; Eber et al., 2013). This repository contains variables related to veterans' military service, demographic characteristics, and contact information. Demographic data included veterans' name, date of birth (used to calculate age), service start and end dates (to verify eligibility), address, telephone number, race, ethnicity, sex, whether the veteran was deployed post‐9/11, branch of service (Army, Navy, etc.), pay category (enlisted, non‐commissioned officer, officer, or warrant officer), personnel category (reserve, active duty, or national guard) and military rank. We requested two random samples of veterans (i.e., one sample for each study), to ensure that each sample was independent from the other sample. Date of birth and service end dates were used to calculate age at discharge.

1.3.2. Data sharing and accessibility

De‐identified and unclassified data collected from this study will be made publicly available through the Center for Open Science (OSF). Our procedures for sharing data will be similar to those used in other large‐scale mental‐health studies, such as the Minnesota Twin Family Study (Iacono et al., 2006). Investigators interested in using the data should contact the Principal Investigator (KTF) to provide a brief description of the intended use and proposed project timeline. These project descriptions will eventually be posted on OSF. Investigators will also be required to sign a data‐use agreement. Such documentation will prevent duplication of research and provide a greater level of protection of intellectual property.

1.3.3. Stratification procedures

Study 1

For Study 1, we obtained data for all 179,111 post‐9/11 veterans separated from military service between October 1, 2018 and September 30, 2019. From this population, we selected a random sample of 4500 veterans stratified by race, ethnicity, personnel category, rank (which we operationalized through pay grade), branch, age at discharge, and deployment status. We oversampled women to recruit a 1:1 ratio of men to women to achieve an adequate sample of women. We achieved an equal sex ratio by dividing the full population into a group of women and a group of men, and then performed the stratification process separately for each group. To perform the stratification, we developed an R package called “sampleVADIR” (Swanson & Forbush, 2021). The stratification procedure implemented by the sampleVADIR function is described below.

First, we removed individuals from the dataset for whom ethnicity, address, or zip code was missing, as well as individuals whose rank was “ROTC.” This left a total of 139,718 individuals in the population from which to sample (24,862 women, and 114,856 men). Before beginning the stratification process, we modified the rank and age at discharge variables into alternative formats to better suit the stratification algorithm. Given that there were 93 possible values for the rank variable (which created too many possible groups to stratify), we converted the rank values into four categories corresponding to pay grades. Specifically, we created groups for enlisted individuals (E), non‐commissioned officers (NCO), warrant officers (W), and commissioned officers (O). Given that age at discharge was defined as a continuous variable, we created four groups associated with quartiles observed in the population. The four quartiles were defined by the ages: (1) 17–25, (2) 25–27, (3) 28–30, and (4) 31–68. The first two quartiles differed slightly for men and women: for women, the upper limit of the first quartile was 24 and, for men, the lower limit of the second quartile was 26. A breakdown of the age at discharge quartiles for both studies is provided in Table 2.

TABLE 2.

Study 1 baseline sample characteristics

graphic file with name MPR-32-e1941-g002.jpg

After preprocessing, the “sampleVADIR” package computed the number of unique strata defined by every combination of the seven variables used in the stratification. For women, there were 1305 unique strata. The proportion of each stratum within the population of women was then computed and multiplied by the number of women to be sampled; this value was then rounded to the nearest integer. This procedure resulted in a sample of 4500 individuals (n = 2258 women and n = 2242 men).

Unexpectedly, approximately 48% of our initial 4500 recruitment mailings were returned due to invalid addresses. Due to the high proportion of invalid addresses, we used the same stratification procedures to select an additional 14,750 cases from the original Study 1 VADIR data. To identify individuals with invalid addresses, we began mailing postcards to let veterans know they would receive an invitation letter. After implementing this procedure, 24% of 14,750 (3546) had invalid addresses. From the 11,204 valid addresses, we selected 10,000 by randomly removing members of overrepresented demographic groups, which mostly included randomly removing veterans who identified as White. Overrepresented groups were identified as those strata that had larger representation in the reduced sample, proportionally, than in the original sample of 14,750. Combining across all Study 1 initial mailings, our incorrect address rate was 38.56%. Of the remaining 8909 individuals that had valid addresses, 2.55% declined to participate and 87.16% did not respond to any mailings. Thus, our rate of enrollment from participants who received mail was 10.29%. Finally, 642 participants completed the 1‐week survey, which was a 70.01% retention rate (or 29.99% attrition rate at 1‐week follow‐up). Demographic and military service characteristics of the individuals who were selected for recruitment are provided in Tables 1 and 2 (4500 from the initial sample and 10,000 from the additional sample).

Study 2

The VADIR variables requested for Study 2 were identical to those requested for Study 1 (see the previous section). For Study 2, we attained a representative sample of 7316 cases from the 119,380 individuals separated from military service between October 1, 2019 and September 30, 2020. Table 2 contains descriptive information regarding the demographic and military service characteristics associated with the total sample of 7316 individuals selected for recruitment for Study 2. The same stratification strategy used to obtain the Study 1 recruitment sample was used to obtain the Study 2 recruitment sample.

1.4. Description of recruitment procedures and materials

After receiving approval from our local and national VA privacy officers, VAEKHCS shared the stratified sample data with the University of Kansas through encrypted email. The stratified sample data were saved on a dual‐authenticated research server at the University of Kansas, and veterans listed in the sample were sent recruitment mailings.

We used a modified version of Dillman's Tailored Design Method (Dillman, 2011; Hoddinott & Bass, 1986) to recruit participants. This three‐stage procedure has been shown to maximize responses to survey‐based research (Dillman et al., 2009). Across this three‐stage procedure, a total of five mailings were sent to participants. Past studies that used the Tailored Design Method to recruit veterans from VADIR achieved response rates of 25%–44% (Coughlin et al., 2011). An initial letter (or postcard), brochure, and decline‐to‐participate card invited veterans to participate in an online Qualtrics® study (Stage 1). The initial letter also included a $1.00 incentive payment—regardless of whether the invited veteran chose to participate in the study—because this technique has been shown to increase survey response rates in previous research (Coughlin et al., 2011). We attempted to send non‐responders a reminder letter two weeks after the initial mailing (Stage 2); however, due to printing delays and COVID‐19 pandemic‐related labor shortages at the U.S. postal service that impacted mail delivery speed, the time between each reminder varied (see Table S1). Next, non‐responders were sent up to three mailings (Stage 3) that contained a reminder letter, two informed consent documents, decline postcard, study questionnaires, and a pre‐paid return envelope. On August 12, 2021 the contents of reminder mailing packets were adjusted due to printing and postage costs. After this date, the first two reminder mailings contained a letter and decline postcard only. However, the final (third) reminder included a letter, two informed consent documents, decline postcard, study questionnaires, and a pre‐paid return envelope.

1.5. Measures

A list of measures is included in Table 3. Internal consistency for each measure is provided in Table 4. Unless otherwise noted, measures were administered in both Studies 1 and 2. Study measures are described below:

TABLE 3.

Longitudinal Eating Disorders Assessment Project assessment schedule

graphic file with name MPR-32-e1941-g003.jpg

TABLE 4.

Internal consistency reliability for selected study measures from Study 1

graphic file with name MPR-32-e1941-g006.jpg

1.5.1. Alcohol Use Identification Test (AUDIT)

The AUDIT is a 10‐item self‐report screen for problematic alcohol use (Babor et al., 1992). The AUDIT has shown high internal consistency across a variety of samples and has a median sensitivity of 0.86 and median specificity of 0.89 (Reinert & Allen, 2002). The AUDIT was administered in Study 2.

1.5.2. Clinical Impairment Assessment (CIA)

The CIA is a 16‐item measure evaluating the severity of psychosocial impairment secondary to an ED (Bohn et al., 2008). This scale has shown evidence for high internal consistency (α = 0.97) and test‐retest reliability (r = 0.86; Bohn et al., 2008).

1.5.3. Demographics, body mass index, and weight history

We adapted the National Survey of Veterans (Department of Veterans Affairs, 2010) to collect basic and military‐specific demographic information from veterans. Additional ad hoc demographic measures included: lifetime history of ED behaviors, body mass index ([BMI] kg/m2), and participants' weight history. We also administered questions adapted from Garber et al. (2008) and the Eating Pathology Symptoms Inventory (Forbush et al., 2013) to assess past ED behaviors to meet military requirements.

1.5.4. Deployment Risk and Resilience Inventory‐2 (DRRI‐2)

We used three scales from the original DRRI‐2 inventory (Vogt et al., 2013): Pre‐deployment Life Events, Combat Experiences, Relationships During Deployment. DRRI‐2 scales have shown evidence for good internal consistency (α > 0.80), criterion validity, and test‐retest validity (α = 0.61–0.94; Maoz et al., 2016; Vogt et al., 2013). The DRRI‐2 scales showed high discriminant validity across scales (Vogt et al., 2013).

1.5.5. Eating Disorder Examination Questionnaire (EDE‐Q)

The EDE‐Q (Fairburn & Beglin, 1994) is a 36‐item questionnaire that assesses ED psychopathology. The EDE‐Q has four subscales: Eating Concern, Weight Concern, Shape Concern, and Restraint. Subscales can be averaged to form a Global Score. The EDE‐Q has demonstrated evidence for high internal consistency, test‐retest reliability, and criterion‐related validity (Berg et al., 2012), although concerns have been raised about factor structure replicability and discriminant validity (Forbush et al., 2013; Thomas et al., 2014).

1.5.6. Eating Pathology Symptoms Inventory (EPSI)

The EPSI (Forbush et al., 2013) is a 45‐item questionnaire designed to measure ED behaviors and cognitions. Initial development and validation of EPSI scales demonstrated evidence for strong internal consistency (α = 0.84–0.89) and test‐retest reliability (r = 0.73). The EPSI has shown excellent convergent and discriminant validity (Forbush et al., 2013, 2014, 2020).

1.5.7. EPSI‐Clinician‐Rated version (EPSI‐CRV)

The EPSI‐CRV (Forbush et al., 2020) is a semi‐structured clinical interview that generates both dimensional scores and current DSM‐5 ED diagnoses. The EPSI‐CRV has shown high inter‐rater reliability, discriminant and convergent validity, and criterion‐related validity (Forbush et al., 2020). The EPSI‐CRV was administered in Study 2.

1.5.8. Externalizing Spectrum Inventory—Brief (ESI‐Brief)

The ESI‐Brief is a 145‐item assessment tool that evaluates three higher‐order dimensions of externalizing psychopathology, including: Callous Aggression, Substance Use, and General Disinhibition, which demonstrated high internal consistency (Patrick et al., 2013). The ESI‐Brief was administered in Study 1.

1.5.9. Generalized anxiety disorder‐Q7

The GAD‐7 (Spitzer et al., 2006) is a seven‐item measure of generalized anxiety with strong support for its construct validity and reliability. Recent research found the GAD‐7 had poor sensitivity and a high false‐positive rate for identifying generalized anxiety disorder and other anxiety disorders (Beard & Björgvinsson, 2014). However, the GAD‐7 performed well as a measure of anxiety symptom severity in a large psychiatric sample (Beard & Björgvinsson, 2014). The GAD‐7 was administered in Study 2.

1.5.10. Inventory of Depression and Anxiety Symptoms‐II (IDAS‐II)

The IDAS‐II is a 99‐item self‐report questionnaire that contains 11 non‐overlapping scales assessing depression, anxiety, mania, obsessive‐compulsive, and trauma‐related psychopathology. The IDAS‐II has shown good‐to‐excellent internal consistency with median α′s ranging from 0.84 to 0.87 across a variety of samples (Watson et al., 2012). IDAS‐II scales have shown strong discriminant, convergent, and criterion validity (Watson et al., 2012).

1.5.11. The Post‐Traumatic Stress Disorder Checklist (PCL‐5)

The PCL‐5 (Blevins et al., 2015) is a 20‐item screening measure. Studies show the PCL‐5 has good stability, internal consistency, test‐retest reliability, and convergent and discriminant validity (Bovin et al., 2016; Wortmann et al., 2016). The PCL‐5 was administered in Study 2.

1.5.12. Patient Health Questionnaire Depression Scale (PHQ‐9)

The PHQ‐9 (Kroenke et al., 2001) is a nine‐item self‐report measure of depression that is well‐validated and reliable. The PHQ‐9 has sensitivity and specificity of 0.88 for identifying major depression in primary care and obstetric‐gynecology clinics (Kroenke et al., 2001). The PHQ‐9 was administered in Study 2.

1.5.13. SCOFF

The SCOFF (Morgan et al., 1999) is a five‐item self‐report screening measure for EDs. The SCOFF was administered in Study 2. Each letter in SCOFF corresponds to one of the items: making oneself sick; losing control over eating; losing at least one stone in weight; feeling fat; and believing food has control over one's life. The SCOFF is a widely utilized instrument, with evidence for good specificity and sensitivity (see Kutz et al., 2020, for a review).

1.5.14. The Structured Clinical Interview for DSM‐5 Disorders (SCID‐5)

The SCID‐5 (First et al., 2015) is a semi‐structured clinical interview that generates current and lifetime psychiatric diagnoses. Studies have demonstrated the specificity/sensitivity, interrater reliability, and predictive validity of the interview (Osório et al., 2019). We administered modules to assess the following lifetime and current psychiatric diagnoses: mood disorders, generalized anxiety disorder, trauma‐ and stress‐related disorders, and eating disorders. The SCID‐5 was administered in Study 2.

1.5.15. Qualitative interview

A qualitative interview will be administered to participants in Study 2. Participants will answer non‐numerical questions concerning experiences in the military. Specifically, they will be asked about institutional and organizational barriers related to seeking or receiving treatment for eating and trauma‐related psychopathology.

1.5.16. World Health Organization Disability Assessment Schedule 2.0 (WHODAS 2.0)

The WHODAS 2.0 (Üstün et al., 2010) is a 36‐item measure of psychiatric impairment. Previous studies have found evidence for strong internal consistency (α = 0.86) and test‐retest reliability (r = 0.98) as well as concurrent and predictive validity (Gold, 2014; Üstün et al., 2010).

2. RESULTS

2.1. Sample representativeness

To assess sample representativeness, we computed correlations by creating a single variable to represent all unique strata within the full population of veteran men and women. Next, we correlated the frequencies of strata observed in the population with the strata we observed in our sample. In both Studies 1 and 2, the correlation of demographic characteristics between the stratified sample and the full population of veterans was 0.95 or higher for each sex; thus, our stratified samples were representative of the full populations from which they were drawn. Further details are provided in Table 4. As a supplementary analysis, we also tested representativeness of our sample across each of the seven categorical demographic variables by comparing the extracted samples to the full VADIR populations for each demographic variable using chi‐square tests (see Supporting Information S1).

2.2. Response bias and attrition analyses

For the Study 1 (longitudinal) sample, we conducted two analyses to investigate the extent to which demographic characteristics that were used to draw stratified samples from VADIR were associated with (1) non‐response to our initial recruitment (response bias analysis), (2) retention from our initial recruitment at 1‐week follow‐up (attrition analysis), (3) missing data in the baseline survey for Study 1, and (4) invalid address. The objectives of these analyses were to assess whether certain demographic groups were significantly less likely to (1) participate (respond), (2) continue participation after completing the baseline survey for Study 1 (attrition), (3) have more than 10% of missing data, and (4) have invalid address. For all analyses, we used a logistic regression model with all seven demographic variables used for stratification, along with sex, as predictors. The first analysis was performed on the full recruitment group for Study 1 (N = 14,500), wherein we used the logistic regression model to predict participation in the baseline survey. For the second analysis, we examined the baseline sample to identify whether any demographic characteristics were associated with participation in the 1‐week follow‐up survey. For the third analysis, we examined the baseline sample to identify whether any demographic characteristics were associated with missing values. For the last analysis, we examined whether demographic characteristics were associated with invalid addresses. Analyses included the following procedures: First, we fitted the saturated model (which included all predictors) and a reduced model, which included all terms except one predictor. Next, we conducted a likelihood‐ratio test (LRT) comparing the saturated model to each reduced model to determine the significance of each variable in predicting participation at (1) baseline and (2) 1‐week follow‐up. We conducted post‐hoc tests for variables with significant LRT results to assess specific levels of each variable that exhibited significant pairwise differences.

Results of the first analysis suggested that several demographic variables were significant predictors of response to the baseline survey, including gender, rank, race, ethnicity, age of discharge (see Figure 1). Men were significantly less likely to respond to our mailings than women (Odds Ratio [OR] = 0.54, 95% CI = [0.46, 0.62]). Rank (operationalized by pay grade) exhibited a statistically significant omnibus effect (Chi‐squared = 26.10, df = 3, p < 0.001), such that enlisted individuals were less likely to participate than non‐commissioned officers (p < 0.001), commissioned officers (p < 0.001), and warrant officers (p = 0.02). Race exhibited a statistically significant omnibus effect (Chi‐squared = 36.26, df = 5, p < 0.001), such that people who identified as Black or African American were significantly less likely to participate than those who identified as White (p < 0.001), Asian (p = 0.03), and multiracial or unknown (p = 0.004). For ethnicity, people not of Hispanic origin were significantly more likely to participate in the baseline survey (OR = 1.30, 95% CI = [1.04, 1.63]). Finally, we found a statistically significant omnibus effect for age of discharge (Chi‐squared = 22.92, df = 3, p < 0.001), such that the fourth quartile was significantly more likely to participate than the first (p < 0.001) and second (p = 0.01) quartiles. Moreover, the second and third quartiles were significantly more likely to participate than the first quartile (p = 0.003, p < 0.001, respectively).

FIGURE 1.

FIGURE 1

Response bias analysis. Plot of odds ratios with 95% confidence intervals for logistic regression model predicting participation in the baseline survey from the group of potential recruits. For each predictor, the associated reference group is listed in parentheses. For instance, we see that the third coefficient from the bottom reflects the odds ratio for males (vs. females), and indicates that males were significantly less likely to participate in the study than females. One coefficient, for warrant officers, was omitted from the plot due to an extremely large confidence interval (resulting from low representation for that group; N = 12). For warrant officers (in comparison with those who were enlisted), the model returned: OR = 4.97, 95% CI = [1.32, 18.80]. The total sample size for the baseline survey was N = 816

For the second analysis, we found that only age at discharge was statistically significantly related to retention in the 1‐week follow up (Chi‐squared = 12.81, df = 3, p = 0.005) (see Figure 2). Individuals whose age at discharge was in the fourth quartile were significantly more likely to participate than the first (p = 0.001), second (p = 0.02) and third (p = 0.01) quartiles.

FIGURE 2.

FIGURE 2

One‐week follow‐up attrition analysis. Plot of odds ratios with 95% confidence intervals for baseline demographic characteristics predicting participation in the 1‐week follow‐up survey. For each predictor, the associated reference group is listed in parentheses. For instance, the second coefficient from the bottom reflects the odds ratio for Black or African American (vs. White) individuals, and indicates that Black or African American participants were significantly less likely to participate in the 1‐week follow‐up than White participants. One case was removed from the analysis, as there was only one individual who identified their race as Native Hawaiian/Pacific Islander in the sample that completed the baseline survey. The total sample size at the 1‐week follow‐up was N = 560

For the third analysis, we found that demographic characteristics were not related to missing data. Thus, there was no evidence of demographic differences between participants who had >10% of missing data and those who did not.

For the fourth analysis, several variables were significant predictors of invalid addresses, including branch, personnel category, age‐at‐discharge, and rank (see Figure 3). Branch exhibited a statistically significant omnibus effect (Chi‐squared = 77.67, df = 3, p < 0.001), such that Navy veterans were more likely to have invalid addresses than Army (p < 0.001), Air Force (p = 0.017), and Marine veterans (p < 0.001). Personnel category exhibited a statistically significant omnibus effect (Chi‐squared = 13.86, df = 2, p < 0.001), with Active Duty individuals being more likely to have invalid addresses than the National Guard (p < 0.001) and Reservists (p = 0.047). Age at discharge had a statistically significant effect (Chi‐squared = 41.09, df = 3, p < 0.001), in which the 1st Quartile was more likely to have invalid addresses than the 2nd Quartile (p < 0.001); the 2nd Quartile was not significantly more likely to have invalid addresses than the 3rd Quartile (p = 0.086); and the 3rd Quartile was more likely to have invalid addresses than the 4th Quartile (p = 0.016). Rank had a significant effect (Chi‐squared = 9.53, df = 3, p = 0.023)—Enlisted individuals were more likely to have invalid addresses than Non‐Commissioned Officers (p = 0.028).

FIGURE 3.

FIGURE 3

Response Bias Analysis due to Invalid Addresses. This figure provides a plot of odds ratios with 95% confidence intervals for VADIR sample characteristics predicting invalid address. For each predictor, the associated reference group is listed in parentheses like Figures 1 and 2

2.2.1. Post‐stratification weights

To control for non‐response bias, we created post‐stratification weights, which can be used in future analyses to adjust for non‐response bias. We applied a survey raking procedure (also known as “sample balancing”) using the R package “anesrake” (Pasek, 2018). Raking is a method used to adjust survey data to ensure the distributions of sample characteristics closely mirror the distributions of population characteristics. In our case, the full population of interest was the group of veterans sampled from the VADIR data who were considered as candidates for participation (e.g., N = 14,500 in Study 1). Although Studies 1 and 2 drew samples from VADIR that represented the total population of recently separated veterans, we observed certain sources of non‐response bias in which some demographic groups were more or less likely to respond to the surveys than other groups. Therefore, we needed to weight the data collected in both studies to match the characteristics of the recruitment groups to ensure that the final collected samples were representative of the population. Given that data collection for Study 2 is ongoing, we only applied this procedure to Study 1.

The raking algorithm implemented in the “anesrake” package (Pasek & Pasek, 2018) used the seven variables used for stratification (and sex) in the total population and evaluated which distributions deviated by 5% or more in the recruited samples. The algorithm iteratively adjusted the frequencies of stratified variables to produce a weight for each case in the sample until the distributions were aligned with the population distributions for each demographic characteristic. Weighted frequencies of demographic variables were then calculated for each level of each variable to assess the effectiveness of the raking procedure. These values are provided for Study 1 in Table 1. This procedure was highly successful, as indicated by an average error rate of less than 0.01%, with the maximum error for a given variable being 0.06%.

3. DISCUSSION

The LEAP Consortium is focused on developing novel methods for screening, classifying, and treating veterans and military members with EDs and other internalizing disorders. The purpose of this paper was to provide information on the design, protocol, methods, response bias, and initial attrition analyses for the first two datasets from the LEAP Consortium. Study 1 is nearly complete and Study 2 began in March 2022. Study 1 is a four‐wave longitudinal survey design and Study 2 is a cross‐sectional study that will include both surveys, structured clinical interviews, and qualitative interviews designed to identify barriers to seeking ED and trauma‐related treatment while in the military. Results showed our recruitment samples were highly representative of the full population of recently separated veterans.

Despite the high correspondence between our recruitment sample of veterans invited to participate and the total population of recently separated veterans, not all invited veterans responded to our surveys. We identified several sources of response bias in our initial survey responses. For example, individuals who identified as men and enlisted personnel were less likely to participate in Study 1 compared to women and non‐commissioned, commissioned, and warrant officers. Individuals in minoritized populations were also less likely to participate in Study 1 compared to veterans who identified as White. Finally, veterans who were 31–68 years of age at discharge were more likely to participate than veterans who were discharged at 17–25 or 25–27. Attrition analyses also revealed that veterans 31–68 years of age were the least likely to drop out of the study. No other demographic characteristics predicted attrition, suggesting that, overall, attrition largely was random rather than related to specific demographic features. To account for response bias, we created sample weights using a validated raking procedure to ensure that researchers using LEAP data in the future will be able to generalize data back to the population from which they were drawn to ensure representative sampling.

Certain limitations of the study should be noted. First, we experienced a higher proportion of returned mail than initially expected and encountered slow response time for mailed surveys due to U.S. postal service delays. These issues resulted in fewer baseline participants in Study 1 than we had initially expected. To address this concern, we recruited additional participants from VADIR and adjusted our protocol to send an initial postcard to check for inaccurate addresses prior to sending our first recruitment invitation letter and enrollment materials. These procedures were successful in ensuring we were able to recruit sufficient numbers of veterans to test our longitudinal research aims. Indeed, we observed 33% invalid addresses—versus 48%—after implementing initial postcards to check for invalid addresses. However, our procedural change resulted in some veterans participating later than the initial sample, often past the 1‐year mark from when they separated from service. For example, the initial sample was sent their first recruitment letter on 9/15/2020 and the additional sample was sent their first recruitment letter on 4/23/2021. We also found demographic differences between individuals who had an invalid address and individuals who did not; however, the potential impact of this bias was minimized by post‐stratification weights that correct for non‐response bias. Second, although we sampled men and women veterans in a 1:1 ratio, women were more likely to enroll in Study 1. We attempted to address this problem by creating post‐stratification survey weights to adjust our Study 1 response data to ensure our final sample matched the full population of veterans recruited from VADIR on demographic characteristics. Third, due to pandemic‐related study issues that prevented staff from in‐person work and unexpected mailing delays, we were delayed in initiating Study 2. To manage our time constraints with budgetary limitations that prevented us from hiring additional staff to increase the rate of weekly interviews, we chose to focus diagnostic interviews on current EDs, mood disorders, generalized anxiety disorder, and post‐traumatic stress disorder because these conditions were the focus of our newly developed self‐reported screening tool. Due to our study design, it was not possible to conduct analyses on the prevalence or correlates of DSM‐5 defined externalizing disorders and certain internalizing disorders. However, researchers who are interested in testing hypotheses related to externalizing psychopathology or additional forms of internalizing psychopathology could use Study 1 or Study 2 self‐reported data from the ESI‐Brief (Patrick et al., 2013) or the IDAS‐II, which comprehensively assess a range of externalizing and internalizing problems, respectively. Fourth, related to our previous concerns, we were unable to assess lifetime diagnoses for most mental‐health disorders, given time and budget constraints. Fifth, Study 2 was ongoing and we were unable to identify (or correct for) non‐response bias in this sample. Future research will focus on creating post‐stratification weights for Study 2, when data collection is complete. Finally, it is possible that there were retrospective recall biases that impacted the reporting of questions about barriers to accessing ED care while serving in the military. We attempted to mitigate this issue by recruiting veterans separated from service within the past year, to minimize recall biases. However, future studies are needed to assess access to ED services among active duty military service members (Table 5).

TABLE 5.

Global correlations between VADIR populations and samples

graphic file with name MPR-32-e1941-g007.jpg

In conclusion, the current study offers a rich source of information about ED and internalizing symptoms and diagnoses, military‐relevant variables, and self‐reported externalizing psychopathology. Study 1 will represent one of the first longitudinal nationally representative samples of US veterans to include EDs and comorbid disorders, which will provide much‐needed information about the temporal relationships among EDs, mood, anxiety, and trauma‐related symptoms. Study 2 will provide prevalence estimates for current and lifetime EDs in veterans and identify barriers to seeking ED and trauma‐informed care. We expect that the LEAP Consortium datasets will, therefore, contribute to improved information about EDs in veterans, an important and understudied problem.

CONFLICT OF INTEREST

We have no conflicts of interest to disclose.

Supporting information

Supporting Information S1

ACKNOWLEDGMENTS

We thank Danielle A. N. Chapa, Sonakshi Negi, Sarah V. Nelson, and Marianna Thomeczek for their work on IRB submissions, survey distribution, and database management. We thank Sarah N. Johnson and Brianne Richson for participant interviewing. The U.S. Army Medical Research Acquisition Activity, 820 Chandler Street, Fort Detrick MD 21702‐5014 is the awarding and administering acquisition office. This work was supported by the Department of Defense, in the amount of $1,721,698, through the Congressionally Directed Medical Research Program's Peer‐Reviewed Medical Research Program, Investigator‐Initiated Award, under Award No. W81XWH‐19‐1‐0207. PI: Forbush. Opinions, interpretations, conclusions and recommendations are those of the author and are not necessarily endorsed by the Department of Defense.

Forbush, K. T. , Swanson, T. J. , Gaddy, M. , Oehlert, M. , Doan, A. , Morgan, R. W. , O’Brien, C. , Chen, Y. , Christian, K. , Song, Q. C. , Watson, D. , & Wiese, J. (2023). Design and methods of the Longitudinal Eating Disorders Assessment Project research consortium for veterans. International Journal of Methods in Psychiatric Research, 32(2), e1941. 10.1002/mpr.1941

DATA AVAILABILITY STATEMENT

De‐identified and unclassified data collected from this study will be made publicly available through the Center for Open Science (OSF). Our procedures for sharing data will be similar to those used in other large‐scale mental‐health studies, such as the Minnesota Twin Family Study (Iacono et al., 2006). Investigators interested in using the data should contact the Principal Investigator (KTF) to provide a brief description of the intended use and proposed project timeline. These project descriptions will eventually be posted on OSF. Investigators will also be required to sign a data‐use agreement. Such documentation will prevent duplication of research and provide a greater level of protection of intellectual property.

REFERENCES

  1. Allen, K. L. , Byrne, S. M. , Oddy, W. H. , & Crosby, R. D. (2013). DSM–IV–TR and DSM‐5 eating disorders in adolescents: Prevalence, stability, and psychosocial correlates in a population‐based sample of male and female adolescents. Journal of Abnormal Psychology, 122(3), 720–732. 10.1037/a0034004 [DOI] [PubMed] [Google Scholar]
  2. Arditte Hall, K. A. , Bartlett, B. A. , Iverson, K. M. , & Mitchell, K. S. (2017). Military‐related trauma is associated with eating disorder symptoms in male veterans. International Journal of Eating Disorders, 50(11), 1328–1331. 10.1002/eat.22782 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Aronson, K. R. , Perkins, D. F. , Morgan, N. R. , Bleser, J. A. , Vogt, D. , Copeland, L. , Finley, E. , & Gilman, C. (2019). Post‐9/11 veteran transitions to civilian life: Predictors of the use of employment programs. Journal of Veterans Studies, 5(1), 14. 10.21061/jvs.v5i1.127 [DOI] [Google Scholar]
  4. Babor, T. F. , de la Fuente, J. R. , Saunders, J. , & Grant, J. (1992). The alcohol use disorders identification test: Guidelines for use in primary health care (WHO Publication No. 92.4). World Health Organization. [Google Scholar]
  5. Barth, S. K. , Dursa, E. K. , Peterson, M. R. , & Schneiderman, A. (2014). Prevalence of respiratory diseases among veterans of operation enduring freedom and operation Iraqi freedom: Results from the national health study for a new generation of US veterans. Oxford University Press. [DOI] [PubMed] [Google Scholar]
  6. Beard, C. , & Björgvinsson, T. (2014). Beyond generalized anxiety disorder: Psychometric properties of the GAD‐7 in a heterogeneous psychiatric sample. Journal of Anxiety Disorders, 28(6), 547–552. 10.1016/j.janxdis.2014.06.002 [DOI] [PubMed] [Google Scholar]
  7. Berg, K. C. , Peterson, C. B. , Frazier, P. , & Crow, S. J. (2012). Psychometric evaluation of the eating disorder examination and eating disorder examination‐questionnaire: A systematic review of the literature. International Journal of Eating Disorders, 45(3), 428–438. 10.1002/eat.20931 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Blevins, C. A. , Weathers, F. W. , Davis, M. T. , Witte, T. K. , & Domino, J. L. (2015). The posttraumatic stress disorder checklist for DSM‐5 (PCL‐5): Development and initial psychometric evaluation. Journal of Traumatic Stress, 28(6), 489–498. 10.1002/jts.22059 [DOI] [PubMed] [Google Scholar]
  9. Bohn, K. , Doll, H. A. , Cooper, Z. , O’Connor, M. , Palmer, R. L. , & Fairburn, C. G. (2008). The measurement of impairment due to eating disorder psychopathology. Behaviour Research and Therapy, 46(10), 1105–1110. 10.1016/j.brat.2008.06.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Bovin, M. J. , Marx, B. P. , Weathers, F. W. , Gallagher, M. W. , Rodriguez, P. , Schnurr, P. P. , & Keane, T. M. (2016). Psychometric properties of the PTSD checklist for diagnostic and statistical manual of mental disorders–Fifth edition (PCL‐5) in veterans. Psychological Assessment, 28(11), 1379–1391. 10.1037/pas0000254 [DOI] [PubMed] [Google Scholar]
  11. Bradley, E. H. , Curry, L. A. , & Devers, K. J. (2007). Qualitative data analysis for health services research: Developing taxonomy, themes, and theory. Health Services Research, 42(4), 1758–1772. 10.1111/j.1475-6773.2006.00684.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Brown, C. A. , & Mehler, P. S. (2013). Medical complications of self‐induced vomiting. Eating Disorders, 21(4), 287–294. 10.1080/10640266.2013.797317 [DOI] [PubMed] [Google Scholar]
  13. Charmaz, K. (2006). Constructing grounded theory: A practical guide through qualitative analysis. Sage. [Google Scholar]
  14. Chesney, E. , Goodwin, G. M. , & Fazel, S. (2014). Risks of all‐cause and suicide mortality in mental disorders: A meta‐review. World Psychiatry, 13(2), 153–160. 10.1002/wps.20128 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Coughlin, S. , Aliaga, P. , Barth, S. , Eber, S. , Maillard, J. , Mahan, C. , Kang, H. , Schneiderman, A. , DeBakey, S. , Vanderwolf, P. , & Williams, M. (2011). The effectiveness of a monetary incentive on response rates in a survey of recent U.S. veterans. Survey Practice, 4, 1–8. 10.29115/sp-2011-0004 [DOI] [Google Scholar]
  16. Cuthbert, K. , Hardin, S. , Zelkowitz, R. , & Mitchell, K. (2020). Eating disorders and overweight/obesity in veterans: Prevalence, risk factors, and treatment considerations. Current Obesity Reports, 9(2), 98–108. 10.1007/s13679-020-00374-1 [DOI] [PubMed] [Google Scholar]
  17. Department of Veterans Affairs . (2010). National survey of veterans, active duty service members, demobilized national guard and reserve members, family members, and surviving spouses. Department of Veterans Affairs. [Google Scholar]
  18. Dillman, D. A. (2011). Mail and internet surveys: The tailored design method – 2007 update with new internet, visual, and mixed‐mode guide. John Wiley & Sons. [Google Scholar]
  19. Dillman, D. A. , Phelps, G. , Tortora, R. , Swift, K. , Kohrell, J. , Berck, J. , & Messer, B. L. (2009). Response rate and measurement differences in mixed‐mode surveys using mail, telephone, interactive voice response (IVR) and the internet. Social Science Research, 38(1), 1–18. 10.1016/j.ssresearch.2008.03.007 [DOI] [Google Scholar]
  20. Eber, S. , Barth, S. , Kang, H. , Mahan, C. , Dursa, E. , & Schneiderman, A. (2013). The national health study for a new generation of United States veterans: Methods for a large‐scale study on the health of recent veterans. Military Medicine, 178(9), 966–969. 10.7205/milmed-d-13-00175 [DOI] [PubMed] [Google Scholar]
  21. Fairburn, C. G. , & Beglin, S. J. (1994). Assessment of eating disorders: Interview or self‐report questionnaire? International Journal of Eating Disorders, 16(4), 363–370. [DOI] [PubMed] [Google Scholar]
  22. Fichter, M. M. , & Quadflieg, N. (2016). Mortality in eating disorders‐results of a large prospective clinical longitudinal study. International Journal of Eating Disorders, 49(4), 391–401. 10.1002/eat.22501 [DOI] [PubMed] [Google Scholar]
  23. First, M. B. , Williams, J. B. W. , Karg, R. S. , & Spitzer, R. L. (2015). Structured clinical interview for DSM‐5—Research version (SCID‐5 for DSM‐5, research version; SCID‐5‐RV). American Psychiatric Association. [Google Scholar]
  24. Forbush, K. T. , Wildes, J. E. , Bohrer, B. K. , Hagan, K. E. , Clark, K. E. , Chapa, D. A. N. , Perko, V. , & Christensen, K. A. (2020). Development and initial validation of the eating Pathology symptoms inventory—Clinician rated version (EPSI‐CRV). Psychological Assessment, 32(10), 943–955. 10.1037/pas0000820 [DOI] [PubMed] [Google Scholar]
  25. Forbush, K. T. , Wildes, J. E. , & Hunt, T. K. (2014). Gender norms, psychometric properties, and validity for the eating pathology symptoms inventory. International Journal of Eating Disorders, 47(1), 85–91. 10.1002/eat.22180 [DOI] [PubMed] [Google Scholar]
  26. Forbush, K. T. , Wildes, J. E. , Pollack, L. O. , Dunbar, D. , Luo, J. , Patterson, K. , Pollpeter, M. , Stone, A. , Bright, A. , & Watson, D. (2013). Development and validation of the eating pathology symptoms inventory (EPSI). Psychological Assessment, 25(3), 859–878. 10.1037/a0032639 [DOI] [PubMed] [Google Scholar]
  27. Forman‐Hoffman, V. L. , Mengeling, M. , Booth, B. M. , Torner, J. , & Sadler, A. G. (2012). Eating disorders, post‐traumatic stress, and sexual trauma in women veterans. Military Medicine, 177(10), 1161–1168. 10.7205/milmed-d-12-00041 [DOI] [PubMed] [Google Scholar]
  28. Forney, K. J. , Buchman‐Schmitt, J. M. , Keel, P. K. , & Frank, G. K. (2016). The medical complications associated with purging. International Journal of Eating Disorders, 49(3), 249–259. 10.1002/eat.22504 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Garber, A. K. , Boyer, C. B. , Pollack, L. M. , Chang, Y. J. , & Shafer, M.‐A. (2008). Body mass index and disordered eating behaviors are associated with weight dissatisfaction in adolescent and young adult female military recruits. Military Medicine, 173(2), 138–145. 10.7205/milmed.173.2.138 [DOI] [PubMed] [Google Scholar]
  30. Gold, L. H. (2014). DSM‐5 and the assessment of functioning: The World Health Organization disability assessment schedule 2.0 (WHODAS 2.0). Journal of the American Academy of Psychiatry & Law Online, 42, 173–181. [PubMed] [Google Scholar]
  31. Hoddinott, S. N. , & Bass, M. J. (1986). The Dillman total design survey method. Canadian Family Physician, 32, 2366–2368. [PMC free article] [PubMed] [Google Scholar]
  32. Iacono, W. G. , McGue, M. , & Krueger, R. F. (2006). Minnesota center for twin and family research. Twin Research & Human Genetics, 9(6), 978–984. 10.1375/twin.9.6.978 [DOI] [PubMed] [Google Scholar]
  33. Iwajomo, T. , Bondy, S. J. , de Oliveira, C. , Colton, P. , Trottier, K. , & Kurdyak, P. (2021). Excess mortality associated with eating disorders: Population‐based cohort study. British Journal of Psychiatry, 219(3), 487–493. 10.1192/bjp.2020.197 [DOI] [PubMed] [Google Scholar]
  34. Kroenke, K. , Spitzer, R. L. , & Williams, J. B. (2001). The PHQ‐9: Validity of a brief depression severity measure. Journal of General Internal Medicine, 16(9), 606–613. 10.1046/j.1525-1497.2001.016009606.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Kutz, A. M. , Marsh, A. G. , Gunderson, C. G. , Maguen, S. , & Masheb, R. M. (2020). Eating disorder screening: A systematic review and meta‐analysis of diagnostic test characteristics of the SCOFF. Journal of General Internal Medicine, 35(3), 885–893. 10.1007/s11606-019-05478-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Maoz, H. , Goldwin, Y. , Lewis, Y. D. , & Bloch, Y. (2016). Exploring reliability and validity of the deployment risk and resilience inventory‐2 among a nonclinical sample of discharged soldiers following mandatory military service. Journal of Traumatic Stress, 29(6), 556–562. 10.1002/jts.22135 [DOI] [PubMed] [Google Scholar]
  37. Masheb, R. M. , Ramsey, C. M. , Marsh, A. G. , Decker, S. E. , Maguen, S. , Brandt, C. A. , & Haskell, S. G. (2021). DSM‐5 eating disorder prevalence, gender differences, and mental health associations in United States military veterans. International Journal of Eating Disorders, 54(7), 1171–1180. 10.1002/eat.23501 [DOI] [PubMed] [Google Scholar]
  38. Mitchell, K. S. , Scioli, E. R. , Galovski, T. , Belfer, P. L. , & Cooper, Z. (2021). Post‐traumatic stress disorder and eating disorders: Maintaining mechanisms and treatment targets. Eating Disorders, 1–15(3), 292–306. 10.1080/10640266.2020.1869369 [DOI] [PubMed] [Google Scholar]
  39. Mitchison, D. , Hay, P. , Slewa‐Younan, S. , & Mond, J. (2014). The changing demographic profile of eating disorder behaviors in the community. BMC Public Health, 14(1), 943. 10.1186/1471-2458-14-943 [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Morgan, J. F. , Reid, F. , & Lacey, J. H. (1999). The SCOFF questionnaire: Assessment of a new screening tool for eating disorders. BMJ, 319(7223), 1467–1468. 10.1136/bmj.319.7223.1467 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Osório, F. L. , Loureiro, S. R. , Hallak, J. E. C. , Machado‐de‐Sousa, J. P. , Ushirohira, J. M. , Baes, C. V. , Apolinario, T. D. , Donadon, M. F. , Bolsoni, L. M. , Guimarães, T. , Fracon, V. S. , Silva‐Rodrigues, A. P. C. , Pizeta, F. A. , Souza, R. M. , Sanches, R. F. , dos Santos, R. G. , Martin‐Santos, R. , & Crippa, J. A. S. (2019). Clinical validity and intrarater and test–retest reliability of the structured clinical interview for DSM‐5–Clinician version (SCID‐5‐CV). Psychiatry & Clinical Neurosciences, 73(12), 754–760. 10.1111/pcn.12931 [DOI] [PubMed] [Google Scholar]
  42. Pasek, J. (2018). anesrake: ANES raking implementation, R package version 0.80.
  43. Pasek, J. , & Pasek, M. (2018). ANES raking implementation and weighted statistics.
  44. Patrick, C. J. , Kramer, M. D. , Krueger, R. F. , & Markon, K. E. (2013). Optimizing efficiency of psychopathology assessment through quantitative modeling: Development of a brief form of the externalizing spectrum inventory. Psychological Assessment, 25(4), 1332–1348. 10.1037/a0034864 [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Reinert, D. F. , & Allen, J. P. (2002). The alcohol use disorders identification test (AUDIT): A review of recent research. Alcoholism: Clinical & Experimental Research, 26(2), 272–279. 10.1111/j.1530-0277.2002.tb02534.x [DOI] [PubMed] [Google Scholar]
  46. Spitzer, R. L. , Kroenke, K. , Williams, J. B. , & Löwe, B. (2006). A brief measure for assessing generalized anxiety disorder: The GAD‐7. Archives of Internal Medicine, 166(10), 1092–1097. 10.1001/archinte.166.10.1092 [DOI] [PubMed] [Google Scholar]
  47. Stice, E. , Marti, C. N. , & Rohde, P. (2013). Prevalence, incidence, impairment, and course of the proposed DSM‐5 eating disorder diagnoses in an 8‐year prospective community study of young women. Journal of Abnormal Psychology, 122(2), 445–457. 10.1037/a0030679 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Streatfeild, J. , Hickson, J. , Austin, S. B. , Hutcheson, R. , Kandel, J. S. , Lampert, J. G. , Myers, E. M. , Richmond, T. K. , Samnaliev, M. , Velasquez, K. , Weissman, R. S. , & Pezzullo, L. (2021). Social and economic cost of eating disorders in the United States: Evidence to inform policy action. International Journal of Eating Disorders, 54(5), 851–868. 10.1002/eat.23486 [DOI] [PubMed] [Google Scholar]
  49. Striegel‐Moore, R. H. , Garvin, V. , Dohm, F.‐A. , & Rosenheck, R. A. (1999). Eating disorders in a national sample of hospitalized female and male veterans: Detection rates and psychiatric comorbidity. International Journal of Eating Disorders, 25(4), 405–414. [DOI] [PubMed] [Google Scholar]
  50. Striegel Weissman, R. , & Rosselli, F. (2017). Reducing the burden of suffering from eating disorders: Unmet treatment needs, cost of illness, and the quest for cost‐effectiveness. Behaviour Research & Therapy, 88, 49–64. 10.1016/j.brat.2016.09.006 [DOI] [PubMed] [Google Scholar]
  51. Swanson, T. , & Forbush, K. (2021). SampleVADIR: Draw stratified samples from the VADIR database. https://CRAN.R‐project.org/package=sampleVADIR
  52. Thomas, J. J. , Roberto, C. A. , & Berg, K. C. (2014). The eating disorder examination: A semi‐structured interview for the assessment of the specific psychopathology of eating disorders. Advances in Eating Disorders: Theory, Research and Practice, 2(2), 190–203. 10.1080/21662630.2013.840119 [DOI] [Google Scholar]
  53. Üstün, T. B. , Chatterji, S. , Kostanjsek, N. , Rehm, J. , Kennedy, C. , Epping‐Jordan, J. , Saxena, S. , von Korff, M. , & Pull, C. (2010). Developing the World Health Organization disability assessment schedule 2.0. Bulletin of the World Health Organization, 88(11), 815–823. 10.2471/blt.09.067231 [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. van Hoeken, D. , & Hoek, H. W. (2020). Review of the burden of eating disorders: Mortality, disability, costs, quality of life, and family burden. Current Opinion in Psychiatry, 33(6), 521–527. 10.1097/yco.0000000000000641 [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Vogt, D. , Smith, B. N. , King, L. A. , King, D. W. , Knight, J. , & Vasterling, J. J. (2013). Deployment risk and resilience inventory‐2 (DRRI‐2): An updated tool for assessing psychosocial risk and resilience factors among service members and veterans. Journal of Traumatic Stress, 26(6), 710–717. 10.1002/jts.21868 [DOI] [PubMed] [Google Scholar]
  56. Watson, D. , O’Hara, M. W. , Naragon‐Gainey, K. , Koffel, E. , Chmielewski, M. , Kotov, R. , Stasik, S. M. , & Ruggero, C. J. (2012). Development and validation of new anxiety and bipolar symptom scales for an expanded version of the IDAS (the IDAS‐II). Assessment, 19(4), 399–420. 10.1177/1073191112449857 [DOI] [PubMed] [Google Scholar]
  57. Wortmann, J. H. , Jordan, A. H. , Weathers, F. W. , Resick, P. A. , Dondanville, K. A. , Hall‐Clark, B. , Foa, E. B. , Young‐McCaughan, S. , Yarvis, J. S. , Hembree, E. A. , Mintz, J. , Peterson, A. L. , & Litz, B. T. (2016). Psychometric analysis of the PTSD Checklist‐5 (PCL‐5) among treatment‐seeking military service members. Psychological Assessment, 28(11), 1392–1403. 10.1037/pas0000260 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supporting Information S1

Data Availability Statement

De‐identified and unclassified data collected from this study will be made publicly available through the Center for Open Science (OSF). Our procedures for sharing data will be similar to those used in other large‐scale mental‐health studies, such as the Minnesota Twin Family Study (Iacono et al., 2006). Investigators interested in using the data should contact the Principal Investigator (KTF) to provide a brief description of the intended use and proposed project timeline. These project descriptions will eventually be posted on OSF. Investigators will also be required to sign a data‐use agreement. Such documentation will prevent duplication of research and provide a greater level of protection of intellectual property.


Articles from International Journal of Methods in Psychiatric Research are provided here courtesy of Wiley

RESOURCES