Skip to main content
PLOS One logoLink to PLOS One
. 2020 Oct 8;15(10):e0239904. doi: 10.1371/journal.pone.0239904

Development and validation of the Multidimensional Internally Regulated Eating Scale (MIRES)

Aikaterini Palascha 1,*, Ellen van Kleef 1,#, Emely de Vet 2,#, Hans C M van Trijp 1,#
Editor: Rodrigo Ferrer3
PMCID: PMC7544044  PMID: 33031400

Abstract

In this paper, we describe the systematic development and validation of the Multidimensional Internally Regulated Eating Scale (MIRES), a new self-report instrument that quantifies the individual-difference characteristics that together shape the inclination towards eating in response to internal bodily sensations of hunger and satiation (i.e., internally regulated eating style). MIRES is a 21-item scale consisting of seven subscales, which have high internal consistency and adequate to high two-week temporal stability. The MIRES model, as tested in community samples from the UK and US, had a very good fit to the data both at the level of individual subscales, but also as a higher-order formative model. High and significant correlations with measures of intuitive eating and eating competence lent support to the convergent validity of MIRES, while its incremental validity in relation to these measures was also upheld. MIRES as a formative construct, as well as all individual subscales, correlated negatively with eating disorder symptomatology and weight-related measures (e.g., BMI, weight cycling) and positively with adaptive behavioral and psychological outcomes (e.g., proactive coping, body appreciation, life satisfaction), supporting the criterion validity of the scale. This endeavor has resulted in a reliable and valid instrument to be used for the thorough assessment of the features that synthesize the profile of those who tend to regulate their eating internally.

Introduction

Internally regulated eating (IRE), which can be broadly defined as eating in response to internal, bodily sensations of hunger and satiation, is considered an adaptive way of eating with positive effects on physical, psychological, behavioral, and dietary outcomes [16]. IRE has been addressed from various specific theoretical perspectives including, but not limited to, those of intuitive eating [7], eating competence [8], and mindful eating [9]. Palascha et al. [10] recently reviewed these various conceptualizations of IRE to conclude that none of them captures IRE style (i.e., the general inclination towards eating in response to internal/physiological signals of hunger and satiation) comprehensively. The authors conceptualized an integrated model with the key dimensions of IRE style and the relationships between them. The Palascha et al. model suggests that five individual-difference characteristics (detailed below) work as necessary and only jointly sufficient conditions for the manifestation of the IRE style.

Existing measures of IRE, such as the Intuitive Eating Scale 2 (IES-2) [11], the Eating Competence Satter Inventory 2 (ecSI-2) [12], the Mindful Eating Questionnaire (MEQ) [13] and the Mindful Eating Scale (MES) [14] have made impactful contributions, but have failed to capture the full complexity of IRE and the inter-connectedness between the characteristics that define the IRE style. Therefore, there is a need for new measures to assess IRE to its full complexity and potential. The Multidimensional Internally Regulated Eating Scale (MIRES) is proposed to quantify the five individual-difference characteristics that collectively form the IRE style. The present paper reveals the systematic development and validation of the MIRES, a short and easily administered 21-item scale.

In this research we followed a stepwise, theory-based and empirically driven process to develop and validate the MIRES (Fig 1). Next to testing the scale’s structure, internal consistency, measurement invariance, and temporal stability, we also examined its content, construct, discriminant, convergent, criterion, and incremental validity. In the next section, we present briefly the conceptual model of the key characteristics of the IRE style, followed by a description of the operationalization of constructs into subscales. For a more complete overview of the conceptual model, including evidence on why each characteristic of IRE style is considered adaptive, see Palascha et al. [10].

Fig 1. Steps in the development and validation of MIRES.

Fig 1

Conceptual definitions and operationalization

Collectively the concept of IRE implies that individuals are sensitive to bodily signals of hunger and satiation, have self-efficacy in using those signals to determine when and how much to eat, trust these bodily signals to guide eating, and have a relaxed and enjoyable relationship with food and eating. Sensitivity to physiological signals of hunger and satiation (SH and SS, respectively) is defined as the ability to sense/perceive and interpret the physiological signals that the body generates in response to hunger and satiation. Self-efficacy in using physiological signals of hunger and satiation (SEH and SES, respectively) is defined as the perception of ease or difficulty in using physiological signals of hunger and satiation to decide when and how much to eat. Internal Trust (IT) refers to the tendency to trust the body’s physiological processes for the regulation of eating. Food Legalizing (FL) is defined as the tendency to have a relaxed relationship with food and particularly a relaxed attitude towards indulgent food. Finally, Food Enjoyment (FE) concerns the tendency to derive pleasure from eating by attending to and appreciating the sensory qualities of the food that is consumed.

IT, FL, and FE are operationalized as uni-dimensional constructs in our model (S1 Fig). Since hunger and satiation are different processes, Sensitivity to hunger signals (SH) and Sensitivity to satiation signals (SS) are operationalized as distinct constructs. The same holds for Self-efficacy in using hunger signals (SEH) and Self-efficacy in using satiation signals (SES). Furthermore, sensitivity and self-efficacy may vary across challenging situations such as when emotional or external cues are salient [1517]. Therefore, we operationalized each of the constructs mentioned above along three dimensions: under 1. neutral conditions, i.e., when individuals are calm, relaxed, and without much distraction (SH: Neutral, SS: Neutral, SEH: Neutral, SES: Neutral), 2. under emotional prompts, i.e., when negative emotions are salient (SH: Emotional, SS: Emotional, SEH: Emotional, SES: Emotional), and 3. under external prompts, i.e., when external influences, such as a distracting environment, are salient (SH: External, SS: External, SEH: External, SES: External). Since individuals may respond differently to positive and negative emotions, we decided to narrow down to negative emotions. Additionally, high-arousal emotions are assumed to have a universal effect by suppressing eating, while there is more variability in how individuals respond to emotions of moderate arousal [16]. Therefore, only moderate arousal emotional states were selected for the emotional context (i.e., sadness, loneliness, boredom). Regarding the external prompts context, there is a variety of external factors that influence our eating in different ways (e.g., portion sizes, mealtime schedules, eating with others, availability of tasty food, eating in a busy or distracting environment). Given this heterogeneity, we decided to select a single external cue, eating under distraction, because it regards a generic cue that is representative of the process by which several external cues influence eating behavior (i.e., when “noise” from the external environment is salient) and is relevant for both hunger and satiation.

Model specification

Since the characteristics of the IRE style are not interchangeable—all of them are necessary for the IRE style to manifest—we treated the IRE style as a formative construct. Formative constructs are formed by the combination of their indicators and causality is assumed to flow from the indicators to the construct [18]. Conversely, a reflective construct exists independently of the indicators that are used to measure it and causality flows from the construct to the indicators. Thus, the IRE style is formed by the totality of its seven defining constructs, while each of these constructs is a reflective one (uni-dimensional or decomposed to measurable sub-dimensions).

Methods

Through interactive discussions within the author team, we generated a pool of 103 items, which were purported to measure the individual-difference characteristics of the IRE style. Existing measures of intuitive eating [11, 19], eating competence [12], mindful eating [13, 14], and interoceptive awareness [20] were used for inspiration during item generation. Researchers in the field of nutrition and experts evaluated and enriched the content of the initial item pool, which then underwent two rounds of pretesting with college samples. This preliminary work helped us to identify the most appropriate and relevant items for the constructs under study, to sort out the internal structure of the scale, to optimize its length, and to identify the most appropriate method for its administration. Starting from the structure obtained from this preliminary work, we examined the scale’s internal consistency, confirmed its internal structure with Confirmatory Factor Analysis (CFA), and tested its two-week temporal stability and several types of validity (i.e., construct, discriminant, convergent, criterion, and incremental) in broad samples of consumers from the UK and US (Table 1). This research was conducted according to the guidelines laid down in the Declaration of Helsinki and complied with the Netherlands Code of Conduct for Research Integrity. Written consent was obtained for all survey participants. Participants who were recruited via market research agencies had previously consented to participate in the panel of the agency. This research was approved by the Social Sciences Ethics Committee of Wageningen University and Research. The data of this project can be found here [21].

Table 1. Overview of sample characteristics.

UK sample (N = 974) UK sub-sample (N = 213) US sample (N = 1200)
Gender
    Males 417 (42.8) 102 (47.9) 590 (49.2)
    Females 557 (57.2) 111 (52.1) 610 (50.8)
Age
    18–24 105 (10.8) 16 (7.5) 183 (15.3)
    25–34 174 (17.9) 27 (12.7) 253 (21.1)
    35–44 214 (22.0) 42 (19.7) 255 (21.3)
    45–54 235 (24.1) 58 (27.2) 277 (23.1)
    55–65 246 (25.3) 70 (32.9) 232 (19.3)
Education level
    Low 94 (9.7) 20 (9.4) 84 (7.0)
    Middle 438 (45.0) 101 (47.4) 360 (30.0)
    High 442 (45.4) 92 (43.2) 756 (63.0)

Values are presented as counts (percentages).

Measures

Internally regulated eating

MIRES was administered with 7-point Likert-type response scales (1 = “Completely untrue for me” to 7 = “Completely true for me”) (see S1 Appendix for information on administration of the MIRES). The MIRES items were developed and tested in the English language. An overview of the initial item pool and the adjustments it was subjected to during the scale development and validation process can be found in S2 Appendix.

A necessary condition for identification of formative models is the addition of at least two reflective measures that are caused directly or indirectly by the formative construct [22]. Thus, to achieve identification when testing the complete formative model we also developed six items that were reflective of the higher-order factor IRE style. We use the abbreviation RI (Reflective items) to refer to these items in the rest of the paper. Cronbach’s alpha for the RI was 0.90 and AVE was 0.61. Uni-dimensionality of the RI factor was supported by the good model fit (χ2 (9) = 110.68, p < 0.001, CFI = 0.98, TLI = 0.96, RMSEA = 0.10, SRMR = 0.03) and the high factor loadings (0.68–0.85).

Intuitive eating

We measured intuitive eating to test the convergent and incremental validity of MIRES. The 21-item IES-2 [11] was used to measure the four constructs of intuitive eating, namely, Unconditional Permission to Eat (UPE), Eating for Physical Rather Than Emotional Reasons (EPR), Reliance on Hunger and Satiety Cues (RHSC), and Body Food Choice Congruence (BFCC). Items were administered on a 5-point scale (1 = “Strongly disagree” to 5 = “Strongly agree”). Cronbach’s alphas were 0.69 (UPE), 0.87 (EPR), 0.93 (RHSC), and 0.88 (BFCC).

Eating competence

We also measured eating competence to test the convergent and incremental validity of MIRES. The 16-item Eating Competence Satter Inventory 2.0 (ecSI-2) was used to measure the four constructs of eating competence [12, 23]; Eating Attitudes (EatAtt), Food Acceptance (FoodAccept), Internal Regulation (IntReg), and Contextual Skills (ContSkills). Items were administered on a 5-point scale (1 = “never” and 5 = “always”) and responses were used as continuous variables in this study. Cronbach’s alphas were 0.88 (EatAtt), 0.75 (FoodAccept), 0.84 (IntReg), and 0.83 (ContSkills).

Eating disorder symptomatology

The Binge Eating Scale (BES) and the Restrictive Eating Scale (RES) of the Multifactorial Assessment of Eating Disorder Symptoms (MAEDS) [24] were used to assess the frequency of manifesting binge eating and restrictive eating behaviors. Items were administered on a 7-point frequency scale (1 = “Never” to 7 = “Always”). Two items from each subscale were dropped before data collection (“I crave sweets and carbohydrates” because it regards a behavior that is non-specific for binge eating and had a low item-total correlation in the original study; “I am too fat” because it reflects a belief rather than a behavior; “I eat 3 meals a day” because it is the only item with negative item-total correlation and because for some people it may seem as a stringent behavior, while for others as an adaptive one; “I hate to eat” because it was deemed extreme and had a low item-total correlation in the original study). Cronbach’s alphas for the adapted scales were 0.91 (BES) and 0.87 (RES). The fit of the RES model was initially unacceptable. Thus, we allowed for correlated error terms between the two items on fasting that have similar wording. BES and RES were measured to assess the criterion and incremental validity of MIRES.

Proactive coping

The 8-item Proactive Coping Scale (PCS) of the Proactive Coping Inventory, as adapted by Gan et al. [25], was used to measure cognitions and behaviors related to self-regulatory goal attainment. Items were administered on a 4-point scale (1 = “Not at all true” to 4 = “Completely true”). The PCS model fit was improved by allowing for correlated error terms between the items that refer to dealing with challenges as there is word congruence among them. We further removed the two reverse-scored items after data collection because of low item-total correlations (0.184 and 0.165, respectively). The adapted PCS had a Cronbach’s alpha of 0.88. PCS was measured to assess the criterion and incremental validity of MIRES.

Adaptive eating behaviors

Two adaptive eating behaviors from the Adult Eating Behavior Questionnaire (AEBQ) were assessed [26]. Satiety responsiveness (SR) assesses with four items the tendency to respond to internal satiety signals. Slowness in eating (SE) measures with four items the tendency to consume meals at a slow pace. Items were administered on a 5-point scale (1 = “Strongly disagree” to 5 = “Strongly agree”). Cronbach’s alphas were 0.81 (SR) and 0.72 (SE). SR and SE were measured to assess the criterion and incremental validity of MIRES.

Body appreciation

Body appreciation was measured with the 10-item Body Appreciation Scale-2 (BAS-2) [27]. The scale assesses the tendency of individuals to accept, respect, and have favorable opinions towards their bodies Responses were measured on a 5-point scale (1 = “Never” to 5 = “Always”). Its Cronbach’s alpha was 0.96. BAS-2 was measured to assess the criterion and incremental validity of MIRES.

Self-esteem

To assess self-esteem, we used the Single-Item Self-Esteem scale (SISE) [28], which consists of a single item “I have high self-esteem” administered on a 5-point scale (1 = “Not very true of me” to 5 = “Very true of me”). Using test-retest data over three points in time and following the procedure suggested by Heise [29], developers have obtained a reliability score of 0.75 for SISE. The scale’s reliability was not estimated in this study due to the lack of repeated measurements. SISE was measured to assess the criterion and incremental validity of MIRES.

Life satisfaction

The 5-item Satisfaction With Life Scale (SWLS) [30] was used to measure global cognitive judgments of one’s life satisfaction. Items were administered on a 7-point scale (1 = “Strongly disagree” to 7 = “Strongly agree”). Cronbach’s alpha was 0.92. SWLS was measured to assess the criterion and incremental validity of MIRES.

Weight-related measures

Current weight and height were reported in pounds and feet/inches, respectively. Values were transformed to kilograms and meters and were used to calculate Body Mass Index (BMI). Highest and lowest weight during the last four years, excluding periods of pregnancy or sickness, was also reported. Based on subtraction of these values a variable called Maximal Weight Change (MWC) was calculated. Individuals whose MWC was <4kg were classified as with stable weight. Individuals whose MWC was ≥4kg were asked additional questions on their weight trajectory and were categorized into 1. those who gained weight (≥4kg increase in weight without significant fluctuations; fluctuations of ≥4kg were considered significant), 2. those who lost weight (≥4kg decrease in weight without significant fluctuations; fluctuations of ≥4kg were considered significant), or 3. those whose weight cycled (weight had fluctuated with gains and losses of ≥4kg). Weight cyclers also reported number of intentional weight losses and unintentional weight gains of ≥4kg during the last four years. Responses were used to calculate a measure of Weight Cycling Severity (WCS). These measures were also measured to assess the criterion and incremental validity of MIRES.

Analysis and results

To confirm the scale’s internal structure with CFA and to test several properties of its subscales (i.e., internal consistency, discriminant validity, measurement invariance, construct validity) we administered MIRES to a nearly representative sample (in terms of gender and age) of UK adults (N = 1380) that was recruited via a market research agency (exclusion criteria were pregnancy and lactation, history of eating disorders, diabetes, or bariatric surgery, and current use of appetite-enhancing or -suppressing medication). Data were checked for violations of normality (acceptable skewness values were below 2 in absolute value and acceptable excess kurtosis values below 3 in absolute value) and presence of multivariate outliers (i.e., values outside the boxplots of the Mahalanobis distances for raw scores and residuals). No violations of normality were observed for the variables. After exclusion of multivariate outliers (N = 20) and those who failed an attention check question (N = 386) the sample was skewed towards females and older individuals (Table 1). Given that 195 parameters were to be estimated in the CFA model, the sample size (N = 974) was adequate to get reliable estimates based on the 5:1 participants-to-parameter ratio [31].

Internal structure and consistency

The Lavaan package [32] in R (version 3.4.1) [33] was used to conduct CFA with the Maximum Likelihood estimation. Adequacy of fit was determined by four indices (CFI > 0.95, TLI > 0.95, RMSEA < 0.06, SRMR < 0.08) [34]. The structure of MIRES was examined in a sequential process in which individual first-order factor models were tested before subscales were combined into higher-order constructs. The multi-factor model including all MIRES subscales provided a very good fit to the data (χ2 (1040) = 2567.43, p < 0.001, CFI = 0.97, TLI = 0.97, RMSEA = 0.04, SRMR = 0.04) and all standardized factor loadings were high (above 0.70) and significant (S1 Table). A number of measurement-model modifications were made when testing this model. First, because the items in the sensitivity and self-efficacy subscales were asked in triple (across three contexts), method effects were accounted for by allowing error terms between identical items to be correlated. Second, because the conceptual distinction between contexts re-appeared in the sensitivity and self-efficacy subscales, we also accounted for context effects by allowing the disturbance terms of the first-order factors referring to the same context to correlate with each other (e.g., SH: Neutral, SS: Neutral, SEH: Neutral, SES: Neutral). Composite reliabilities and Average Variance Extracted (AVE) were calculated according to Fornell and Larcker [35]. Reliabilities of the MIRES first- and second-order factors ranged between 0.84 and 0.96, and AVE was as low as 0.64 and as high as 0.88 (Table 2).

Table 2. Descriptive statistics, composite reliabilities, and AVE for the MIRES first- and second-order factors.

M SD Composite reliability AVE
First-order factors
    IT 4.52 1.68 0.94 0.80
    FL 4.43 1.79 0.91 0.71
    FE 5.34 1.32 0.94 0.75
    SH: Neutral 5.91 1.10 0.88 0.70
    SH: Emotional 5.38 1.48 0.88 0.71
    SH: External 5.32 1.43 0.87 0.70
    SS: Neutral 5.55 1.35 0.91 0.77
    SS: Emotional 4.83 1.73 0.89 0.73
    SS: External 5.09 1.53 0.89 0.72
    SEH: Neutral 5.49 1.34 0.90 0.75
    SEH: Emotional 4.85 1.64 0.94 0.84
    SEH: External 5.00 1.50 0.90 0.74
    SES: Neutral 5.34 1.58 0.96 0.88
    SES: Emotional 4.69 1.87 0.91 0.76
    SES: External 5.03 1.65 0.93 0.82
Second-order factors
    SH 5.54 1.14 0.84 0.64
    SS 5.15 1.39 0.92 0.79
    SEH 5.11 1.31 0.88 0.72
    SES 5.02 1.57 0.93 0.82

IT: Internal trust, FL: Food legalizing, FE: Food enjoyment, SH: Sensitivity to physiological signals of hunger, SS: Sensitivity to physiological signals of satiation, SEH: Self-efficacy in using physiological signals of hunger, SES: Self-efficacy in using physiological signals of satiation, AVE: Average Variance Extracted.

Discriminant validity of constructs

Several alternative models were fitted and compared to show the discriminant validity of the sensitivity and self-efficacy constructs (Table 3). First, to test whether sensitivity and self-efficacy are truly distinct from each other we compared two pairs of alternative models: one for hunger and one for satiation. Starting with hunger, in one model the three SH subscales (SH: Neutral, SH: Emotional, SH: External) loaded on a second-order factor SH and the three SEH subscales (SEH: Neutral, SEH: Emotional, SEH: External) loaded on another second-order factor SEH. In the alternative model, the two second-order factors were collapsed into one factor. The alternative model had significantly lower fit. The same was the case for the distinction between SS and SES.

Table 3. Change in chi square and fit indices between models testing the discriminant validity of MIRES constructs.

Factors* Δχ2 (df)a P value ΔCFI ΔTLI ΔRMSEA ΔSRMR
Sensitivity vs. Self-efficacy
    SH vs. SEH 130.72 (1) < 0.001 -0.009 -0.012 0.01 0.005
    SS vs. SES 116.95 (1) < 0.001 -0.006 -0.008 0.011 0.005
Hunger vs. Satiation
    SH vs. SS 316.95 (1) < 0.001 -0.022 -0.031 0.024 0.016
    SEH vs. SES 455.77 (1) < 0.001 -0.024 -0.034 0.031 0.029
Neutral context vs. Emotional context vs. External context
    SH: Neutral vs. SH:Emotional vs. SH:External 1341.51 (3) < 0.001 -0.235 -0.47 0.276 0.086
    SS: Neutral vs. SS:Emotional vs. SS:External 1005.99 (3) < 0.001 -0.139 -0.278 0.211 0.048
    SEH: Neutral vs. SEH:Emotional vs. SEH:External 1300.46 (3) < 0.001 -0.188 -0.377 0.239 0.065
    SES: Neutral vs. SES:Emotional vs. SES:External 1633.31 (3) < 0.001 -0.158 -0.315 0.267 0.051

SH: Sensitivity to physiological signals of hunger, SS: Sensitivity to physiological signals of satiation, SEH: Self-efficacy in using physiological signals of hunger, SES: Self-efficacy in using physiological signals of satiation.

* In the initial model, factors were distinct. In the alternative model, factors were collapsed into a single factor.

a Alternative model–Initial model.

In a similar way, we tested the discriminant validity of hunger and satiation constructs by comparing two pairs of alternative models: one for sensitivity and one for self-efficacy. The alternative model, in which SH and SS were collapsed into one factor, was significantly worse compared to the model where the two factors were distinct. The same was the case for SEH and SES.

Finally, the conceptual distinction between different contexts of sensitivity and self-efficacy was tested. For each second-order construct (SH, SS, SEH, and SES), we compared the fit of a three-factor model in which each item loaded to its respective context versus an alternative model in which the three factors were collapsed into one factor. In all cases, the fit of the alternative model was significantly worse.

Measurement invariance

Measurement invariance was examined for the items that were asked in triple (across contexts) to test the assumption that each item should have a consistent performance irrespectively of the context in which it is asked. To do this, we constrained the loadings of these items to be equal across the three contexts. The decrease in fit in the constrained model was significant (Δχ2 (24) = 102.502, p < 0.001), however, the changes in fit indices were within the acceptable criteria (ΔCFI = -0.002, ΔTLI = -0.001, ΔRMSEA = 0, ΔSRMR = 0.001) according to Chen’s [36] recommendations for factor loading invariance (ΔCFI ≤ 0.010, ΔRMSEA ≤ 0.015, and ΔSRMR ≤ 0.030).

Construct validity

Since the IRE style is by nature a non-diet eating style, we used independent samples t-tests to compare scores on the MIRES subscales between individuals who said they were currently dieting for weight loss purposes (n1 = 131) and those who said they were not (n2 = 843), as a means of testing the scale for construct validity in a broad sense. Non-dieters scored significantly higher than dieters in all but one MIRES subscales, in line with our expectations (S2 Table). For FE, the mean difference between groups did not reach significance.

Temporal stability

A sub-sample of 679 participants from the UK sample filled in the MIRES for a second time after two weeks. Response rate was 43.2%, but the entire survey was completed by 261 participants. Those who failed the attention check (N = 46) and two multivariate outliers were excluded, leaving a sample of 213 responses for analysis (Table 1). The sample size was adequate to get reliable estimates in models testing the stability of first-order factors, while in models testing the stability of second-order factors the sample was slightly small (4:1 participant-to-parameter-ratio).

No violations of normality were observed for the variables. We used an elaborated procedure of temporal stability assessment as suggested by Steenkamp and van Trijp [37]. Pearson’s correlation coefficients, intra-class coefficients with confidence intervals, and means for the summed scores of factors were also calculated. Stability coefficients of the MIRES first- and second-order factors ranged between .63 and .90 (Table 4). Imposition of constraints on factor loadings did not result in significant decreases in model fit, thus, the meaning of all subscales was stable. Some subscales were further found to be stable in terms of item reliabilities (SS: Neutral and EH: External) and construct reliability (FL, SH: External, SS: Emotional, SS: External, EH: Emotional, and ES: Neutral). Finally, SH: Neutral, SEH: Neutral, and SEH manifested perfect stability as their stability coefficient was not significantly different from unity. Paired samples t-tests indicated that most factor means were stable over time; however, the means of IT, FL, SH: Emotional, and SS: External changed significantly.

Table 4. Stability coefficients, Pearson correlation coefficients, intra-class correlation coefficients, and mean scores for the MIRES first- and second- order factors.

Stability coefficient Pearson’s r ICC (CI)** Mean 1 Mean 2 P value
First-order factors
    IT 0.74 0.69* 0.80 (0.73–0.86) 18.64 20.22 < 0.001
    FL 0.79 0.74* 0.85 (0.80–0.89) 18.39 19.37 0.005
    FE 0.67 0.65* 0.79 (0.72–0.84) 27.00 27.42 0.292
    SH: Neutral 0.66 0.57* 0.73 (0.64–0.79) 17.79 17.90 0.624
    SS: Neutral 0.74 0.69* 0.81 (0.75–0.86) 17.20 17.08 0.556
    SH: Emotional 0.69 0.64* 0.77 (0.70–0.83) 16.63 16.08 0.037
    SS: Emotional 0.83 0.77* 0.87 (0.83–0.90) 15.16 15.07 0.702
    SH: External 0.70 0.62* 0.77 (0.69–0.82) 16.21 15.82 0.133
    SS: External 0.76 0.70* 0.82 (0.76–0.86) 16.08 15.56 0.033
    SEH: Neutral 0.63 0.59* 0.74 (0.66–0.80) 16.84 16.92 0.754
    SES: Neutral 0.76 0.71* 0.83 (0.78–0.87) 16.84 16.87 0.903
    SEH: Emotional 0.65 0.61* 0.76 (0.68–0.82) 15.30 14.88 0.161
    SES: Emotional 0.74 0.71* 0.83 (0.78–0.87) 15.24 14.79 0.125
    SEH: External 0.71 0.65* 0.78 (0.71–0.83) 15.47 15.09 0.148
    SES: External 0.72 0.68* 0.81 (0.75–0.85) 15.96 15.62 0.204
Second-order factors
    SH 0.90 0.75* 0.85 (0.81–0.89) 50.63 49.79 0.102
    SS 0.90 0.83* 0.90 (0.87–0.93) 48.45 47.70 0.145
    SEH 0.83 0.71* 0.83 (0.78–0.87) 47.62 46.89 0.237
    SES 0.85 0.78* 0.88 (0.84–0.91) 48.05 47.27 0.218

IT: Internal trust, FL: Food legalizing, FE: Food enjoyment, SH: Sensitivity to physiological signals of hunger, SS: Sensitivity to physiological signals of satiation, SEH: Self-efficacy in using physiological signals of hunger, SES: Self-efficacy in using physiological signals of satiation.

* p < 0.001.

** Intra-class correlation coefficients using an absolute agreement definition.

Length optimization

In order to further optimize the scale’s length and to have the same number of items per subscale (i.e., three), we decided to drop seven items; four items from the IT subscale, one item from the FL subscale, and two items from the FE subscale. The decision on which items to drop was based on the meaning of items to retain the scale’s content validity [38]; items whose meaning was very similar to other items in their respective subscales were dropped. The three subscales manifested similar properties after the exclusion of items (IT: Stability coefficient = 0.70, r = 0.65, ICC = 0.78 (0.70–0.84), Mean 1 = 14.00, Mean 2 = 15.16, p < 0.001; FL: Stability coefficient = 0.82, r = 0.74, ICC = 0.85 (0.80–0.88), Mean 1 = 13.72, Mean 2 = 14.46, p = 0.005; FE: Stability coefficient = 0.66, r = 0.61, ICC = 0.76 (0.69–0.82), Mean 1 = 16.01, Mean 2 = 16.33, p = 0.204). The final scale consisted of 45 items.

Confirmation of the internal structure of MIRES as a multidimensional, formative model

The 45-item MIRES was further administered to a representative sample of 1251 adults from the US [39] (Table 1; see also S3 Table for some additional characteristics) (recruited via a market research agency) in order to confirm the internal structure of MIRES as a multidimensional formative model and to test the scale’s convergent, criterion, and incremental validity. Exclusion criteria were pregnancy and lactation, because these conditions relate to temporal irregularities in the eating patterns of women. Fifty-one multivariate outliers were excluded leaving 1200 responses for analysis. Based on the recommended 5:1 participants-to-parameter ratio, a sample of 1200 participants would be adequate to give reliable estimates for a model with maximum 240 parameters. All models that we tested had less than 240 parameters to be estimated, thus the sample size was adequate for our analyses. No significant violations of normality were observed for most variables. BMI and MWC had kurtosis values above 3 and the latter also had a skewness value above 2. However, according to Kline’s [40] more relaxed criteria for skewness and kurtosis (<3 and <10, respectively) none of these variables were considered problematic, thus no transformations were conducted.

The MIRES model was subjected to CFA (S2 Fig) with the following additional specifications. The three first-order factors—IT, FL, FE—and the four second-order factors—SH, SS, SEH, SES—loaded to the higher-order IRE style construct as formative indicators (arrows pointing to the higher-order construct). Covariances between all first- and second-order factors with the higher-order formative factor were fixed to zero, as otherwise Lavaan estimates both these covariances and the formative regression coefficients, which seem to be confounded leading to identification problems. To warrant identification, the six RI also loaded to the IRE style construct as reflective indicators (arrows pointing to the six RI).

The model had an excellent fit to the data (χ2 (1130) = 2804.10, p < 0.001, CFI = 0.97, TLI = 0.97, RMSEA = 0.04, SRMR = 0.03). All observed variables served as reliable and significant indicators of their corresponding constructs and all first-order factors loaded highly and significantly to their respective second-order factors (S2 Fig), as was the case in the UK sample. Regression coefficients of the seven formative indicators of the IRE style were not interpreted because their values were influenced by the presence of multi-collinearity among the seven subscales of MIRES (Variance Inflation Factors 1.52–7.85, cut-off <3.3), which are moderately to strongly correlated with each other (S4 Table). High and significant loadings were obtained for the six RI (0.66–0.86) and a large amount of variance in these items was accounted for by the IRE style factor (AVE = 0.82).

Convergent validity

Bivariate correlations of the MIRES total score, RI, and MIRES subscales with the IES-2 and ecSI-2 total scores were substantial and significant (0.32–0.70) (S5 Table). High correlations were particularly observed between certain MIRES subscales and conceptually related constructs of IES-2 and ecSI-2. For example, FL and FE correlated most strongly with the EatAtt (0.56) and ContSkills (0.46) subscales of ecSI-2, respectively. Similarly, SEH and SES correlated most strongly with the RHSC subscale of IES-2 (0.66 and 0.68, respectively).

Criterion validity

The criterion validity of MIRES, IES-2, and ecSI-2 was examined with Structural Equation Modelling (SEM) (for outcomes measured with multiple items) and with linear regression (for the single-item outcomes SISE, BMI, MWC, and WCS). Analyses with MIRES were conducted at the level of a total score (summed score of all items), at the level of the seven MIRES subscales as separate latent constructs (IT, FL, FE, SH, SS, SEH, SES), and at the level of the RI as an independent scale. Analyses for IES-2 and ecSI-2 were conducted only at the level of total scores.

MIRES, as well as its individual subscales, displayed negative associations with binge eating, restrictive eating, BMI, maximal weight change, and weight cycling severity, and positive associations with all adaptive outcomes assessed in this study (Table 5). In general, MIRES, IES-2, and ecSI-2 displayed comparable predictive abilities (S6 Table) and all were better at predicting behavioral and psychological outcomes, compared to physical outcomes. MIRES accounted for a slightly larger amount of variance in RES, SR, and SE compared to the other scales, IES-2 was better at predicting BES, BMI, MWC, and WCS, and finally ecSI-2 was better at predicting PCS, BAS-2, SWLS, and SISE. The RI manifested comparable criterion validity to MIRES. Finally, certain MIRES subscales (FL, SH, SS, SES) achieved higher predictive power compared to the MIRES summed score for certain outcomes (e.g., RES, BES, SR, SE, BMI).

Table 5. Bivariate correlations among all constructs measured in the US sample.

1 2 3 4 5 6 7 8 9 10 11 12 13 14
1. MIRES -
2. IES-2 0.69** -
3. ecSI-2 0.67** 0.60** -
4. BES -0.38** -0.46** -0.16** -
5. RES -0.15** -0.12** -0.12** 0.47** -
6. PCS 0.38** 0.35** 0.44** -0.02 0.14** -
7. SR 0.23** 0.19** 0.15** 0.002 0.34** 0.27** -
8. SE 0.23** 0.21** 0.17** -0.14** 0.16** 0.21** 0.39** -
9. BAS-2 0.49** 0.53** 0.59** -0.26** -0.02 0.52** 0.26** 0.24** -
10. SWLS 0.29** 0.28** 0.40** -0.04 0.04 0.44** 0.24** 0.16** 0.62** -
11. SISE 0.34** 0.35** 0.40** -0.15** 0.003 0.40** 0.19** 0.12** 0.71** 0.60** -
12. BMI -0.15** -0.21** -0.12** 0.20** 0.05 -0.09** -0.12** -0.07** -0.21** -0.10** -0.12** -
13. MWC -0.16** -0.21** -0.16** 0.13** 0.16** -0.08** 0.002 0.005 -0.19** -0.13** -0.12** 0.43** -
14. WCS -0.22** -0.27** -0.09 0.34** 0.27** 0.11* 0.08 -0.005 -0.06 0.01 -0.001 0.24** 0.29** -

MIRES: Multidimensional Internally Regulated Eating Scale, IES-2: Intuitive Eating Scale-2, ecSI-2: Eating Competence Satter Inventory 2.0, BES: Binge Eating Scale, RES: Restrictive Eating Scale, PCS: Proactive Coping Scale, SR: Satiety Responsiveness, SE: Slowness in Eating, BAS-2: Body Appreciation Scale-2, SWLS: Satisfaction With Life Scale, SISE: Single Item Self-Esteem Scale, BMI: Body Mass Index, MWC: Maximal Weight Change, WCS: Weight Cycling Severity.

* Correlation is significant at the 0.05 level.

** Correlation is significant at the 0.01 level.

Incremental validity

The incremental validity of MIRES in relation to IES-2 and ecSI-2 was examined with SEM (for multi-item outcomes) and hierarchical regression analysis (for single-item outcomes). Specifically, we examined whether MIRES accounted for variance in each outcome measure above and beyond the variance accounted for by IES-2 and ecSI-2, respectively. At Step 1, IES-2 was entered as a single predictor of each respective outcome and at Step 2, MIRES was added as a second predictor (in SEM analyses, MIRES was also entered as a predictor in the model at Step 1, but its regression coefficient was fixed at zero). The same procedure was followed with ecSI-2. Changes in beta coefficients were not interpreted because multi-collinearity between these conceptually similar measures was expected to interfere with these estimates. For most outcomes, a significant increase in R2 was observed when MIRES was added in the model (Table 6). Specifically, MIRES accounted for 0.7%-16% additional variance in outcome measures above and beyond IES-2 and ecSI-2. MIRES did not account for a significant increase in explained variance of physical outcomes (BMI [ΔR2 = 0], MWC [[ΔR2 = 0], and WCS [[ΔR2 = 0.002]) above and beyond IES-2, neither for satisfaction with life (ΔR2 = 0) and self-esteem (ΔR2 = 0.005) above and beyond the variance explained for by ecSI-2.

Table 6. Incremental variance in outcome measures accounted for by MIRES.

MIRES vs. IES-2 MIRES vs. ecSI-2
R2 (IES-2) R2 (IES-2 + MIRES) ΔR2 P value R2 (ecSI-2) R2 (ecSI-2 + MIRES) ΔR2 P value
BESa 0.249 0.259 0.010 <0.001 0.032 0.192 0.160 <0.001
RESa 0.021 0.030 0.009 0.001 0.018 0.030 0.012 <0.001
PCSa 0.150 0.191 0.041 <0.001 0.227 0.244 0.017 <0.001
SRa 0.047 0.069 0.022 <0.001 0.028 0.066 0.038 <0.001
SEa 0.048 0.072 0.024 <0.001 0.052 0.074 0.022 <0.001
BAS-2a 0.284 0.316 0.032 <0.001 0.348 0.367 0.019 <0.001
SWLSa 0.085 0.104 0.019 <0.001 0.178 0.178 0.000 0.370
SISEb 0.119 0.138 0.019 <0.001 0.156 0.161 0.005 0.085
BMIb 0.045 0.045 0.000 0.802 0.013 0.043 0.03 <0.001
MWCb 0.043 0.043 0.000 0.485 0.035 0.042 0.007 0.047
WCSb* 0.071 0.073 0.002 0.298 0.007 0.055 0.048 <0.001

MIRES: Multidimensional Internally Regulated Eating Scale, IES-2: Intuitive Eating Scale-2, ecSI-2: Eating Competence Satter Inventory 2, BES: Binge Eating Scale, RES: Restrictive Eating Scale, PCS: Proactive Coping Scale, SR: Satiety Responsiveness, SE: Slowness in Eating, BAS-2: Body Appreciation Scale-2, SWLS: Satisfaction With Life Scale, SISE: Single Item Self-Esteem Scale, BMI: Body Mass Index, MWC: Maximal Weight Change, WCS: Weight Cycling Severity.

a Values obtained with SEM.

b Values obtained with hierarchical regression analysis.

* N = 504.

Testing the properties of the simplified 21-item version of MIRES

Since the 45-item MIRES manifested good psychometric properties, we wanted to examine whether the inclusion of the three contexts (neutral, emotional, external) in the sensitivity and self-efficacy subscales offers predictive advantages compared to just the neutral context. In this way we could ascertain whether a simplified version of the scale (21 items) could still be applicable. To test this empirically we performed SEM and regression analysis (depending on the outcome variable) using either the full subscales (SH, SS, SEH, and SES) including all three contexts each or the neutral counterpart of each subscale to predict each outcome measured in the US sample. The full subscales accounted for 0–8% additional variance, depending on the outcome, compared to their neutral counterparts (S7 Table). In addition, the fit of the 21-item MIRES model was still excellent (χ2 (296) = 1258.161, p < 0.001, CFI = 0.97, TLI = 0.96, RMSEA = 0.05, SRMR = 0.04) (Fig 2), correlations among the MIRES subscales and with IES-2 and ecSI-2 reduced only slightly (S8 and S9 Tables), and the incremental validity of MIRES was still upheld (S10 Table). Thus, despite the fact that the 45-item full version offers some predictive advantages, the simplified version with only 21 items generally upholds the psychometric properties of the full scale.

Fig 2. The multi-dimensional model of internally regulated eating style (simplified version).

Fig 2

All loadings were significant at the 0.01 level. Covariances and disturbance terms of first-order factors are not depicted in the figure for easier readability.

Discussion

Internally regulated eating is an adaptive way of eating that leads to positive physical, psychological, behavioral, and dietary outcomes as shown by the current and previous research [16]. While several attempts have been made to conceptualize and quantify this eating style, none seems to capture the full complexity of this construct. In this paper, we describe the rigorous development and validation of the MIRES, an instrument to assess the individual-difference characteristics that are necessary and jointly sufficient conditions for the manifestation of the IRE style.

Using a bottom-up approach, we showed that all first- and second-order factors of MIRES are measured reliably and a significant amount of variance in the items is accounted for by the corresponding latent factors. All first-order models and the multi-factor model that we tested had very good fit to the data. We confirmed that sensitivity to hunger, sensitivity to satiation, self-efficacy with hunger, and self-efficacy with satiation are distinct constructs, and that the three contexts within each of these subscales are also distinct from each other. Results supported the metric measurement invariance of the items asked across contexts and initial evidence on the construct validity of MIRES was obtained, as non-dieters scored higher in all but one MIRES subscales compared to dieters. Scores on FE did not differ significantly between groups, suggesting that this is perhaps the least determinative characteristic among the ones that form the IRE style. We further showed that all MIRES subscales are stable over a period of two weeks in terms of factor loadings, while even higher levels of stability (in terms of item reliabilities, construct reliabilities, or correlation of the same factor over time) were evidenced for certain subscales. Pearson’s correlations underestimated the true stability of these constructs, while intra-class correlation coefficients overestimated it. Factor means remained stable for most factors except for IT, FL, SH: Emotional, and SS: External. As regards the latter two factors, however, the means of their respective second-order factors (SH and SS) were stable. The change in means in IT and FL, suggests that these subscales show variation over time across the whole sample, which could be systematic (i.e., these subscales measure less stable characteristics) or random (i.e., due to chance). Further studies are required to confirm which of the two plausible explanations is true. Evidence on the multidimensional nature of the MIRES model was also obtained in this study. The convergent validity of MIRES was supported by the moderate to strong correlations with measures of intuitive eating and eating competence. Measures of IRE were generally better at predicting behavioral and psychological outcomes compared to physical outcomes, which is in line with existing evidence [13, 6]. MIRES associated negatively with binge eating, restrictive eating, BMI, maximal weight change, and weight cycling severity, and positively with all adaptive outcomes assessed in this study. This confirms the adaptive nature of the constructs it assesses. The six RI had comparable predictive power to the 45-item MIRES. Furthermore, certain MIRES subscales (FL, SH, SS, and SES) accounted for a larger amount of variance in certain outcomes compared to the MIRES summed score. This further justifies their applicability as independent measures. The incremental validity of MIRES, above and beyond IES-2 and ecSI-2, was supported for most outcome variables measured in this study. Finally, we showed that the simplified 21-item version of MIRES upholds the psychometric properties of the full 45-item scale.

MIRES can be used by researchers and practitioners for a complete assessment of the IRE style as well as of its distinct components. MIRES can be used as an independent variable, moderator, or mediator in future scientific research investigating the role of IRE style in various processes in the eating domain. It can also be used as an outcome variable when assessing the impact of interventions aimed to strengthen IRE. Finally, MIRES can be used as a screening instrument by health practitioners who try to promote IRE among their clients or patients.

While MIRES manifested good psychometric properties, there are limitations that should be addressed. First, we should note that all data presented in this paper are solely based on self-reports. Although self-reports are practical tools for the assessment of personality constructs, they are subject to several types of response bias such as socially desirable responding, acquiescent responding, or extreme responding [41]. Individual responses may also be limited by the lack of sufficient self-awareness or by self-deception effects. Second, identification restrictions are inherent to formative models [42], as is the one presented in this paper. Thus, researchers who are interested in conducting CFA or SEM using the complete formative MIRES model should also measure the six RI that we specifically developed to facilitate model identification. Third, the preliminary work was conducted with college students (18–35 years old) while in later steps we used community samples (18–65 years old); thus, it could be argued that it is not safe to assume the invariance of the model’s internal structure across the scale development and validation process. To test the model for measurement invariance across age groups, subgroups should have at least 980 participants each to allow for reliable estimates to emerge based on the 5:1 participant to parameter ratio. The sample sizes in our study did not allow us to conduct this analysis in the typical stepwise process [43]; however, when we fitted the model in subgroups with all but seven parameters fixed to the values obtained from the full sample (only regression coefficients of the seven formative indicators were left free to be estimated) the model fit was still acceptable (18–34 years: χ2 (1319) = 2467.93, p < 0.001, CFI = 0.95, TLI = 0.95, RMSEA = 0.05, SRMR = 0.03; 35–65 years: χ2 (1319) = 2969.25, p < 0.001, CFI = 0.96, TLI = 0.96, RMSEA = 0.04, SRMR = 0.05) providing, thus, preliminary evidence for the invariance of the model across age groups. Finally, we acknowledge that administration of the full version of MIRES may be more complex than other self-reports because twelve of its items are repeated across three different contexts. Thus, we advise potential users to use the simplified version of the scale that consists of only 21 items.

Next to these limitations, the strengths of this newly developed measure should also be considered. In contrast to what most scale developers do, in this research we were particularly interested in the precise specification of the measurement model. Those who aim to assess the IRE style need to measure the complete set of seven MIRES subscales and calculate a total score, while those who want to focus on a particular characteristic of the IRE style can choose to measure a subscale in isolation and calculate the summed score of items of that particular subscale. The bottom-up approach that we took for the scale’s development and validation (assessing the properties of lower-order factors before moving to higher levels) can give researchers and practitioners confidence on the reliability and validity of the scale’s sub-parts. It should be noted here that using only a subset of subscales would allow conclusions to be drawn only on those particular constructs that are measured and not on the IRE style construct. We further observed strong convergence and comparable criterion validity between MIRES and the six RI. Given that RI is a reliable scale in itself, it could be used as the snap version of MIRES. This adds even more flexibility in the use of the new instrument. Finally, the multidimensional nature of MIRES enables the distinction of several closely related but conceptually distinct features of the IRE style. For example, the distinction between sensitivity to and self-efficacy in using physiological signals of hunger and satiation has been examined very deficiently in existing literature (see e.g., [44]). Therefore, MIRES can be used for a more differentiated assessment of the essentials of the IRE style.

Although we followed a rigorous process for the scale’s development and validation, replication of the current findings in other populations or population segments is needed. For example, the measurement invariance of the model could be tested across sexes, age groups, and other potentially interesting population groups such individuals with overweight or obesity. Once measurement invariance of the model is evidenced, norm scores can be developed for the various subgroups. Moreover, it would be interesting to administer the simplified version of the scale without any introductory text in the sensitivity and self-efficacy subscales in order to ascertain whether this influences how individuals interpret the items. Additional studies could also be conducted to assess the temporal stability of the RI scale and to ascertain whether the change in means over time in two MIRES subscales (IT and FL) that we observed was systematic or random. Future research could also test the face validity of the final MIRES because relevance of items with the construct definitions was assessed only at the very beginning of the scale development process. This would ensure that the retained items still do a good job in reflecting the meaning of the constructs they are purported to measure. Given that a theory-based approach was used in this research, we expect that MIRES will uphold its face validity. Finally, behavioral experiments could provide convincing and invaluable evidence for the construct and predictive validity of MIRES.

Supporting information

S1 Fig. Conceptual model of internally regulated eating style.

The direction of arrows indicates whether a construct is formative—arrows point to the construct—or reflective—arrows point to the dimension.

(TIF)

S2 Fig. The multi-dimensional model of internally regulated eating style (full version).

All loadings were significant at the 0.01 level. Context effects, method effects, covariances between first- and second-order factors, and disturbance terms of first- and second-order factors are not depicted in the figure for easier readability.

(TIF)

S1 Table. Factor loadings for the MIRES first- and second-order factors.<.

/SI_Caption>

(DOCX)

S2 Table. Mean scores on MIRES first- and second-order factors for dieters and non-dieters.

(DOCX)

S3 Table. Additional sample characteristics of the US sample.

(DOCX)

S4 Table. Bivariate correlations of summed scores of MIRES, RI, and MIRES subscales.

(DOCX)

S5 Table. Bivariate correlations of summed scores of MIRES, RI, and MIRES subscales with IES-2 and ecSI-2.

(DOCX)

S6 Table. Standardized regression coefficients (and R2) for the criterion validity of MIRES, IES-2, and ecSI-2.

(DOCX)

S7 Table. Standardized regression coefficients (and R2) for the criterion validity of SH, SS, SEH, SES (full subscales including neutral, emotional, external contexts) vs. the neutral counterpart of each subscale.

(DOCX)

S8 Table. Bivariate correlations of summed scores of MIRES (21 items), RI, and MIRES subscales.

(DOCX)

S9 Table. Bivariate correlations of summed scores of MIRES (21 items), RI, and MIRES subscales with IES-2 and ecSI-2.

(DOCX)

S10 Table. Incremental variance in outcome measures accounted for by MIRES (21 items).

(DOCX)

S1 Appendix. The Multidimensional Internally Regulated Eating Scale (MIRES).

(DOCX)

S2 Appendix. Modifications of the MIRES item pool during the scale development and validation process.

(DOCX)

Acknowledgments

The authors would like to thank Dr. I. van der Lans for his valuable comments on the statistical analyses and the manuscript, as well as all individuals who participated in the studies presented in this paper.

Data Availability

The data of this research have been uploaded in the OSF Repository (Creation date: 06/03/2020, URL: https://osf.io/w3guh/?view_only=406839237c9f4bc09ff154d065c17da3).

Funding Statement

The authors received no specific funding for this work.

References

  • 1.Schaefer JT, Magnuson AB. A review of interventions that promote eating by internal cues. J Acad Nutr Diet. 2014. May;114(5):734–60. Epub 2014 Mar 14. 10.1016/j.jand.2013.12.024 [DOI] [PubMed] [Google Scholar]
  • 2.Van Dyke N, Drinkwater EJ. Review article relationships between intuitive eating and health indicators: literature review. Public Health Nutr. 2014. August;17(8):1757–66. Epub 2013 Aug 21. 10.1017/S1368980013002139 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Clifford D, Ozier A, Bundros J, Moore J, Kreiser A, Morris MN. Impact of non-diet approaches on attitudes, behaviors, and health outcomes: A systematic review. J Nutr Educ Behav. 2015. Mar-Apr;47(2):143–55.e1. 10.1016/j.jneb.2014.12.002 [DOI] [PubMed] [Google Scholar]
  • 4.Bruce LJ, Ricciardelli LA. A systematic review of the psychosocial correlates of intuitive eating among adult women. Appetite. 2016. January;96:454–72. Epub 2015 Oct 22. 10.1016/j.appet.2015.10.012 [DOI] [PubMed] [Google Scholar]
  • 5.Warren JM, Smith N, Ashwell M. A structured literature review on the role of mindfulness, mindful eating and intuitive eating in changing eating behaviours: effectiveness and associated potential mechanisms. Nutr Res Rev. 2017. December;30(2):272–83. Epub 2017 Jul 18. 10.1017/S0954422417000154 [DOI] [PubMed] [Google Scholar]
  • 6.Ulian MD, Aburad L, da Silva Oliveira MS, Poppe AC, Sabatini F, Perez I, et al. Effects of health at every size® interventions on health‐related outcomes of people with overweight and obesity: a systematic review. Obes Rev. 2018. December;19(12):1659–66. Epub 2018 Sep 27. 10.1111/obr.12749 [DOI] [PubMed] [Google Scholar]
  • 7.Tribole E, Resch E. Intuitive eating: A revolutionary program that works. 3rd ed. New York: St. Martin’s Griffin; 2012. [Google Scholar]
  • 8.Satter E. Eating competence: definition and evidence for the Satter Eating Competence model. J Nutr Educ Behav. 2007. Sep-Oct;39(5 Suppl):S142–53. 10.1016/j.jneb.2007.01.006 [DOI] [PubMed] [Google Scholar]
  • 9.Kristeller JL, Wolever RQ. Mindfulness-based eating awareness training for treating binge eating disorder: the conceptual foundation. Eat Disord. 2011. Jan-Feb;19(1):49–61. 10.1080/10640266.2011.533605 [DOI] [PubMed] [Google Scholar]
  • 10.Palascha A, van Kleef E, de Vet E, van Trijp H. Internally Regulated Eating Style: A Comprehensive Theoretical Framework. Br J Nutr. 2020. September 30;1–36. 10.1017/S0007114520003840 [DOI] [PubMed] [Google Scholar]
  • 11.Tylka TL, Kroon Van Diest AM. The Intuitive Eating Scale–2: Item refinement and psychometric evaluation with college women and men. J Couns Psychol. 2013. January;60(1):137–53. 10.1037/a0030893 [DOI] [PubMed] [Google Scholar]
  • 12.Krall JS, Lohse B. Validation of a measure of the Satter eating competence model with low-income females. Int J Behav Nutr Phys Act. 2011. April 7;8:26 10.1186/1479-5868-8-26 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Framson C, Kristal AR, Schenk JM, Littman AJ, Zeliadt S, Benitez D. Development and validation of the mindful eating questionnaire. J Am Diet Assoc. 2009. August;109(8):1439–44. 10.1016/j.jada.2009.05.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Hulbert-Williams L, Nicholls W, Joy J, Hulbert-Williams N. Initial validation of the mindful eating scale. Mindfulness. 2014. December;5(6):719–29. [Google Scholar]
  • 15.Herman CP, Polivy J. Norm-violation, norm-adherence, and overeating. Coll Antropol. 2007. March;31(1):55–62. [PubMed] [Google Scholar]
  • 16.Macht M. How emotions affect eating: a five-way model. Appetite. 2008. January;50(1):1–11. Epub 2007 Jul 25. 10.1016/j.appet.2007.07.002 [DOI] [PubMed] [Google Scholar]
  • 17.Zlatevska N, Dubelaar C, Holden SS. Sizing up the effect of portion size on consumption: a meta-analytic review. J Mark. 2014. May;78(3):140–54. [Google Scholar]
  • 18.Diamantopoulos A, Siguaw JA. Formative versus reflective indicators in organizational measure development: A comparison and empirical illustration. Br. J. Manag. 2006. December;17(4):263–82. [Google Scholar]
  • 19.Hawks S, Merrill RM, Madanat HN. The intuitive eating scale: Development and preliminary validation. Am J Health Educ. 2004. April;35(2):90–9. [Google Scholar]
  • 20.Mehling WE, Price C, Daubenmier JJ, Acree M, Bartmess E, Stewart A. The multidimensional assessment of interoceptive awareness (MAIA). PLoS One. 2012;7(11):e48230 Epub 2012 Nov 1. 10.1371/journal.pone.0048230 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Palascha A, van Kleef E, de Vet E, van Trijp H. MIRES—Development and Initial Validation. 2020 [cited 2020 Sep 4]. Database: OSF [Internet]. Available from: https://osf.io/w3guh/?view_only=406839237c9f4bc09ff154d065c17da3
  • 22.Bollen KA, Davis WR. Causal indicator models: Identification, estimation, and testing. Struct Equ Modeling. 2009. July 14;16(3):498–522. [Google Scholar]
  • 23.Lohse B. The Satter eating competence inventory for low-income persons is a valid measure of eating competence for persons of higher socioeconomic position. Appetite. 2015. April;87:223–8. Epub 2014 Dec 31. 10.1016/j.appet.2014.12.228 [DOI] [PubMed] [Google Scholar]
  • 24.Anderson DA, Williamson DA, Duchmann EG, Gleaves DH, Barbin JM. Development and validation of a multifactorial treatment outcome measure for eating disorders. Assessment. 1999;6(1): 7–20. 10.1177/107319119900600102 [DOI] [PubMed] [Google Scholar]
  • 25.Gan Y, Yang M, Zhou Y, Zhang Y. The two-factor structure of future-oriented coping and its mediating role in student engagement. Pers Individ Dif. 2007. September;43(4):851–63. [Google Scholar]
  • 26.Hunot C, Fildes A, Croker H, Llewellyn CH, Wardle J, Beeken RJ. Appetitive traits and relationships with BMI in adults: Development of the Adult Eating Behaviour Questionnaire. Appetite. 2016. October 1;105:356–63. Epub 2016 May 20. 10.1016/j.appet.2016.05.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Tylka TL, Wood-Barcalow NL. The Body Appreciation Scale-2: item refinement and psychometric evaluation. Body Image. 2015. January;12:53–67. Epub 2014 Oct 21. 10.1016/j.bodyim.2014.09.006 [DOI] [PubMed] [Google Scholar]
  • 28.Robins RW, Hendin HM, Trzesniewski KH. Measuring global self-esteem: Construct validation of a single-item measure and the Rosenberg Self-Esteem Scale. Pers Soc Psychol Bull. 2001. February;27(2):151–61. [Google Scholar]
  • 29.Heise DR. Separating reliability and stability in test-retest correlation. Am Sociol Rev. 1969. February;34(1):93–101. [Google Scholar]
  • 30.Diener ED, Emmons RA, Larsen RJ, Griffin S. The satisfaction with life scale. J Pers Assess. 1985. February;49(1):71–5. 10.1207/s15327752jpa4901_13 [DOI] [PubMed] [Google Scholar]
  • 31.Tinsley HE, Tinsley DJ. Uses of factor analysis in counseling psychology research. J Couns Psychol. 1987. October;34(4):414–24. [Google Scholar]
  • 32.Rosseel Y. Lavaan: An R package for structural equation modeling. J Stat Softw. 2012. May;48(2):1–36. [Google Scholar]
  • 33.R Core Team. R: A language and environment for statistical computing. R Foundation for Statistical Computing; 2018. [cited 2019 Sep 14]. Available from: https://www.R-project.org/ [Google Scholar]
  • 34.Bentler PM. Comparative fit indexes in structural models. Psychol Bull. 1990. March;107(2):238–46. 10.1037/0033-2909.107.2.238 [DOI] [PubMed] [Google Scholar]
  • 35.Fornell C, Larcker DF. Evaluating structural equation models with unobservable variables and measurement error. J Mark Res. 1981. February;18(1):39–50. [Google Scholar]
  • 36.Chen FF. Sensitivity of goodness of fit indexes to lack of measurement invariance. Struct Equ Modeling. 2007. July 31;14(3):464–504. [Google Scholar]
  • 37.Steenkamp JB, Van Trijp HC. The use of LISREL in validating marketing constructs. Int. J. Res. Mark. 1991. November;8(4):283–99. [Google Scholar]
  • 38.Rossiter JR. The C-OAR-SE procedure for scale development in marketing. Int. J. Res. Mark. 2002. December;19(4):305–35. [Google Scholar]
  • 39.Howden LM, Meyer JA; United States Census Bureau. Age and Sex Composition: 2010. 2010 Census Briefs. 2011. Available from: https://www.census.gov/
  • 40.Kline RB. Methodology in the social sciences. Principles and practice of structural equation modeling. 2nd ed. New York: Guilford Press; 2005. [Google Scholar]
  • 41.McDonald JD. Measuring personality constructs: The advantages and disadvantages of self-reports, informant reports and behavioural assessments. Enquire. 2008;1(1):75–94. Available from: https://pdfs.semanticscholar.org/ [Google Scholar]
  • 42.Edwards JR. The fallacy of formative measurement. Organ Res Methods. 2011. April;14(2):370–88. [Google Scholar]
  • 43.Putnick DL, Bornstein MH. Measurement invariance conventions and reporting: The state of the art and future directions for psychological research. Dev Rev. 2016. September;41:71–90. 10.1016/j.dr.2016.06.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Dockendorff SA, Petrie TA, Greenleaf CA, Martin S. Intuitive Eating Scale: An examination among early adolescents. J Couns Psychol. 2012. October;59(4):604–11. Epub 2012 Sep 3. 10.1037/a0029962 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Gian Mauro Manzoni

13 Dec 2019

PONE-D-19-26656

Development and validation of the Multidimensional Internally Regulated Eating Scale (MIRES)

PLOS ONE

Dear Mrs Palascha,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

We would appreciate receiving your revised manuscript by Jan 27 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Gian Mauro Manzoni, Ph.D., Psy.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements:

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at http://www.plosone.org/attachments/PLOSOne_formatting_sample_main_body.pdf and http://www.plosone.org/attachments/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

3. Your ethics statement must appear in the Methods section of your manuscript. If your ethics statement is written in any section besides the Methods, please move it to the Methods section and delete it from any other section. Please also ensure that your ethics statement is included in your manuscript, as the ethics section of your online submission will not be published alongside your manuscript.

4.  Thank you for including your ethics statement:  "The studies were conducted according to the guidelines laid down in the Declaration of Helsinki and complied with the code of conduct of Wageningen University. Written consent was obtained for all survey participants.". 

For studies reporting research involving human participants, PLOS ONE requires authors to confirm that this specific study was reviewed and approved by an institutional review board (ethics committee) before the study began. Please provide the specific name of the ethics committee/IRB that approved your study, or explain why you did not seek approval in this case.

Once you have amended this/these statement(s) in the Methods section of the manuscript, please add the same text to the “Ethics Statement” field of the submission form (via “Edit Submission”).

For additional information about PLOS ONE ethical requirements for human subjects research, please refer to http://journals.plos.org/plosone/s/submission-guidelines#loc-human-subjects-research.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: No

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: No

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors present the results of an effort to develop a new measure of internally regulated eating. They make a reasonable case for the need for such a measure given extant measures of associated constructs. They mine the literature for dimensions that conceptually define internally regulated eating and develop a self-report measure that maps closely onto a domain-space defined by those dimensions. Their work is among the best I have read in terms of scale development and initial validation. Item generation and selection are thoughtful, and the psychometric analyses are thorough and informative. All in all, it is, in my view, a strong manuscript.

Here are three minor considerations:

1. It is not clear to my what is to be gained from the series of one-factor models evaluated as described on pp. 12-13. The problem with these analyses is that do not examine the critical question of whether items written for a particular dimension uniquely indicate the intended dimension. It leads to reporting loadings in Table 1 that appear to be from a single model but in fact are from many small models. All I needed to see was the test of the full model, reported around the middle of p. 13, and my preference would be to see the loadings from that model in a supplemental table.

2. I understand the context distinction, but building it into the measures significantly increases the length and complexity of the measure if used as the authors intend. The question is whether information about internally regulated eating from specific contexts offers predictive advantages in hypothesis tests. If, as an example, one looks at the dieters-nondieters comparisons in Table S1, Cohen's ds are virtually identical for SS and SES and not substantially different for SH and SEH. In terms of the latter two, it is notable that the inconsistent d is for the emotional context, which seems to be operating somewhat differently than the other two here and in other results. Similarly, the stability coefficients in Table 4 are very similar within dimension across context (e.g., .57, .64 and .62 for SH). Scores are collapsed across context within dimension in Tables S4 and S5, making it impossible to determine whether predictive or incremental validity varies as a function of context. In short, I think the authors need to make a stronger case for building context into the measure. My recommendation would be to, if at all possible, drop the distinction.

3. One issue for measures like this is how they should be scored. For the validity analyses, the authors consider both a total score and a score by dimension (ignoring context). My inference is that they would favor either scoring depending on research question. It is worth noting that researchers sometimes use only a subset of subscales from a measure like this. When the first-order factors are reflective indicators of the general factor, a total score can be meaningfully interpreted. In the case of formative indicators, as the authors have cast the first-order factors, that is not the case. The general factor is not fully defined unless all of the formative indicators are included in a composite.

A very minor matter: On p. 33, just before the general discussion, the authors suggest differences that imply comparisons of correlation coefficients, which they did not do. "Outperformed" implies a significantly larger coefficient, which requires a test comparing them. One somewhat related and even more minor matter: Please include the actual coefficients for all pairwise rs in Table 5. The use of "NS" in a table like this is a meta-analyst's nightmare and, generally speaking, unnecessarily hides information that need not be hidden.

These matters are very minor and do not detract from my overall positive view of the manuscript. I congratulate the authors on a fine effort.

Reviewer #2: The authors report on the development of a new measure (MIRES). My first impression is that this is an extremely long paper, for its content. As such, it loses focus and clarity about its aims. By the end, I was not clear what psychological or clinical issue the authors were addressing, or what the implications of the work would be.

More importantly, the presentation failed to meet a number of basic psychometric requirements in developing such a measure.

MAJOR ISSUES

There is a core problem with the reporting of the initial samples used in studies 3-7. The ages of the first two samples are substantially different, making the findings hard to compare. Without age-based norms for the original measure, comparability of scores or of factor structure is not safe to assume.

Studies 1-4 are simply not reported in anything like adequate detail. Where are the initial factor analyses that I presume must have been conducted? What factors emerged, using what methods? Was there replication? Internal consistency? Test-retest reliability? Most importantly, how do the authors justify a sample in studies 3 and 4 that are simply too small to allow for reliable factor derivation (80 items requires a sample of 800 participants PER STUDY to ensure reliable outcomes)? In short, the derivation of a pool of 49 items is just not safe. Nor do we have any reason to assume that the factors that are addressed in studies 5-7 are in any way related to what was found in Studies 1-4.

For Study 5, the construct validity element is not clear. What were the t-values per measure, and what was the P value adopted? Why not use binomial regression to determine the key MIRES variables? And why use one simple measure like this (dieting/not dieting), when understanding such a complex construct? An appropriate index would have been to determine whether the MIRES is a better predictor than other measures of the same constructs, to show that the MIRES is a more clinically useful measure.

If Study 6 really did use the same analyses as Study 5, where is the report of the CFA? If one was conducted, then the sample in Study 6 appears to be underpowered, even by the authors' relatively lax standards from study 6.

In Study 6, the lack of temporal stability severely undermines the test-retest reliability of the measure - correlations on their own are not adequate, if scores change overe time, which they do on two of the top three factors.

Why were 4 items dropped at the end of Study 6? As this was the most underpowered study of the two in this section, it seems unclear why one would make that change.

In Study 7, the authors perform yet another CFA. Why? And what criteria are used for concluding that their measure was better than the others? That was not apparent from Table 5. In short, have the authors really just reinvented the wheel here, using a relatively long, psychometrically weak measure to do a job that was already done just as well by existing measures (in the case of Table 6, it looks as if the IES-2 already does this job for the core weight variables).

The norms in Table 7 are meaningless if there is no validation.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Oct 8;15(10):e0239904. doi: 10.1371/journal.pone.0239904.r002

Author response to Decision Letter 0


6 Mar 2020

General response

We would like to thank both the reviewers and the academic editor for their helpful comments on our paper. We greatly appreciate the time and effort they spent to review our paper. We have attempted to address all comments in the revised manuscript. In addition, we have embedded our responses within the revision letter text.

Journal Requirements:

When submitting your revision, we need you to address these additional requirements:

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at http://www.plosone.org/attachments/PLOSOne_formatting_sample_main_body.pdf and http://www.plosone.org/attachments/PLOSOne_formatting_sample_title_authors_affiliations.pdf

->We have now adapted our title, affiliations, contributorship, tables (and values with decimals in text), and file names to meet PLOS ONE’s style requirements.

2. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

->We have added the following information in the revised cover letter. The data of Studies 1-7 have been uploaded in the OSF Repository (Creation date: 06/03/2020, URL: https://osf.io/w3guh/?view_only=406839237c9f4bc09ff154d065c17da3). For the purpose of this review, we provide the editors and reviewers with the view-only link to the project. The data will become publicly available once the paper is accepted for publication.

3. Your ethics statement must appear in the Methods section of your manuscript. If your ethics statement is written in any section besides the Methods, please move it to the Methods section and delete it from any other section. Please also ensure that your ethics statement is included in your manuscript, as the ethics section of your online submission will not be published alongside your manuscript.

->Our ethics statement can be found in the Methods section of the paper.

4. Thank you for including your ethics statement: "The studies were conducted according to the guidelines laid down in the Declaration of Helsinki and complied with the code of conduct of Wageningen University. Written consent was obtained for all survey participants.".

For studies reporting research involving human participants, PLOS ONE requires authors to confirm that this specific study was reviewed and approved by an institutional review board (ethics committee) before the study began. Please provide the specific name of the ethics committee/IRB that approved your study, or explain why you did not seek approval in this case.

Once you have amended this/these statement(s) in the Methods section of the manuscript, please add the same text to the “Ethics Statement” field of the submission form (via “Edit Submission”).

For additional information about PLOS ONE ethical requirements for human subjects research, please refer to http://journals.plos.org/plosone/s/submission-guidelines#loc-human-subjects-research.

->We have uploaded the retrospective approval from the Social Science Ethics Committee of Wageningen University in the submission system and we have adapted our ethics statement in the manuscript (lines 174-178) as well as in the submission form.

Reviewers' comments:

Reviewer #1: The authors present the results of an effort to develop a new measure of internally regulated eating. They make a reasonable case for the need for such a measure given extant measures of associated constructs. They mine the literature for dimensions that conceptually define internally regulated eating and develop a self-report measure that maps closely onto a domain-space defined by those dimensions. Their work is among the best I have read in terms of scale development and initial validation. Item generation and selection are thoughtful, and the psychometric analyses are thorough and informative. All in all, it is, in my view, a strong manuscript.

->We thank the reviewer for his nice words. We are happy that he/she acknowledges our hard work.

Here are three minor considerations:

1. It is not clear to my what is to be gained from the series of one-factor models evaluated as described on pp. 12-13. The problem with these analyses is that do not examine the critical question of whether items written for a particular dimension uniquely indicate the intended dimension. It leads to reporting loadings in Table 1 that appear to be from a single model but in fact are from many small models. All I needed to see was the test of the full model, reported around the middle of p. 13, and my preference would be to see the loadings from that model in a supplemental table.

->We agree with the reviewer that this part is confusing and perhaps less interesting for the reader. Thus, we removed it from the text (see edits in lines 161-162, 282-301, 375-377) and we added the requested table in the Supplementary material (S1 Table).

2. I understand the context distinction, but building it into the measures significantly increases the length and complexity of the measure if used as the authors intend. The question is whether information about internally regulated eating from specific contexts offers predictive advantages in hypothesis tests. If, as an example, one looks at the dieters-non-dieters comparisons in Table S1, Cohen's ds are virtually identical for SS and SES and not substantially different for SH and SEH. In terms of the latter two, it is notable that the inconsistent d is for the emotional context, which seems to be operating somewhat differently than the other two here and in other results. Similarly, the stability coefficients in Table 4 are very similar within dimension across context (e.g., .57, .64 and .62 for SH). Scores are collapsed across context within dimension in Tables S4 and S5, making it impossible to determine whether predictive or incremental validity varies as a function of context. In short, I think the authors need to make a stronger case for building context into the measure. My recommendation would be to, if at all possible, drop the distinction.

->We agree with the reviewer that the context distinction makes our scale lengthier and more difficult to use. Thus, it makes sense to examine whether the addition of three contexts has predictive advantages over the neutral counterparts of the SH, SS, SEH, and SES subscales. However, we believe that Cohen’s ds and stability coefficients are not appropriate indicators of the contexts’ predictive advantages. To test this empirically we conducted additional analyses with the data of Study 7. Specifically, we tested whether the full subscales (SH, SS, SEH, and SES), which included all three contexts each, account for a larger amount of variance in each outcome measure compared to the neutral counterpart of each subscale. We addressed this issue in lines 773-789 and we added supplementary tables (S7, S8, S9, S10) and a figure (S2 Fig) to present the relevant results. Since the 45-item and 21-item versions of the scale are comparable we leave it to the user to decide which version to use depending on their specific interests (see also edits in lines 848-854 and S1 Appendix).

3. One issue for measures like this is how they should be scored. For the validity analyses, the authors consider both a total score and a score by dimension (ignoring context). My inference is that they would favor either scoring depending on research question. It is worth noting that researchers sometimes use only a subset of subscales from a measure like this. When the first-order factors are reflective indicators of the general factor, a total score can be meaningfully interpreted. In the case of formative indicators, as the authors have cast the first-order factors, that is not the case. The general factor is not fully defined unless all of the formative indicators are included in a composite.

->We thank the reviewer for raising this important issue about the use of our scale. We now clarify this in lines 840-848.

A very minor matter: On p. 33, just before the general discussion, the authors suggest differences that imply comparisons of correlation coefficients, which they did not do. "Outperformed" implies a significantly larger coefficient, which requires a test comparing them.

->We changed the phrasing in lines 768-770.

One somewhat related and even more minor matter: Please include the actual coefficients for all pairwise rs in Table 5. The use of "NS" in a table like this is a meta-analyst's nightmare and, generally speaking, unnecessarily hides information that need not be hidden.

->We replaced the “NS” with the actual values as requested. We also corrected the values in the third column (correlations of IES-2) because we realized that the ones we had included in the original version were from an adapted version of IES-2 (in which all reverse-scored items had been excluded).

These matters are very minor and do not detract from my overall positive view of the manuscript. I congratulate the authors on a fine effort.

->Once again, we thank the reviewer for his particularly useful comments that have helped us to improve our paper and scale substantially.

Reviewer #2: The authors report on the development of a new measure (MIRES). My first impression is that this is an extremely long paper, for its content. As such, it loses focus and clarity about its aims. By the end, I was not clear what psychological or clinical issue the authors were addressing, or what the implications of the work would be.

->We thank the reviewer for the time spent on reviewing our paper and for the detailed comments. In this paper we present a large number of studies and analyses that we conducted for the thorough development and validation of the MIRES. We believe that all the information we provide is necessary and useful for the users of the scale. This prevents us from reducing the length of the paper substantially. However, we have made several efforts (already in the original version of the paper) to help the readers keep up with paper and its aims. Specifically, we use the following structure. First, we introduce the aim of the paper early on, in lines 60-73. Second, under the Methods section we present the overview of studies with their specific aims (lines 154-171), accompanied by Figure 1 (line 180), which is a graphical summary of the whole paper. Third, the results of each of the main studies are summarized in the discussion section of each study, followed by an overall wrap up in lines 801-823 of the General discussion section. Finally, we made some edits in lines 792-800 to remind readers of the main construct that we examine in this research.

More importantly, the presentation failed to meet a number of basic psychometric requirements in developing such a measure.

->We regret to see that the reviewer is not satisfied with the amount of evidence we present on the scale’s psychometric properties. We believe this has to do with the way Studies 1-4 had been presented in the initial version of the manuscript. As we discuss later on, we now provide additional information on this preliminary work.

MAJOR ISSUES

There is a core problem with the reporting of the initial samples used in studies 3-7. The ages of the first two samples are substantially different, making the findings hard to compare. Without age-based norms for the original measure, comparability of scores or of factor structure is not safe to assume.

->Starting with college samples is a common practice in scale development given the low costs and convenience of collecting such data. While many scale developers often settle with such data, we chose to go one step forward and test the properties of our scale in large community samples to increase the external validity of our research. We do not think that comparison of findings between college and community samples is problematic because in our studies we are mainly interested in the covariance between items not the variance within items. While means and standard deviations may vary depending on the sample, the covariance between items is not expected to do so (or at least not substantially).

Furthermore, the reason why we do not provide age-based or gender-based norms is because if we were to do this, we would first have to test our scale for measurement (in)variance across age groups or sexes. Since this would increase the length of the paper even more, we decided to leave this out for now and maybe present these data in another publication.

Studies 1-4 are simply not reported in anything like adequate detail. Where are the initial factor analyses that I presume must have been conducted? What factors emerged, using what methods? Was there replication? Internal consistency? Test-retest reliability? Most importantly, how do the authors justify a sample in studies 3 and 4 that are simply too small to allow for reliable factor derivation (80 items require a sample of 800 participants PER STUDY to ensure reliable outcomes)? In short, the derivation of a pool of 49 items is just not safe. Nor do we have any reason to assume that the factors that are addressed in studies 5-7 are in any way related to what was found in Studies 1-4.

->We agree with the reviewer that these studies had been presented very briefly in the original paper. To enrich this section (see edits in lines 204-232), we added information on the rationale of dropping items, the sample size justification, and the most important findings of these studies, which helped us to bring the model to its current form.

The main reason why we do not present any hard data on these studies is because this preliminary work took place during the initial steps of developing our conceptual model. In the course of these studies our model was adjusted several times since we came across empirical findings we had not initially expected. Thus, should we want to add statistics on these studies we would have to present a long course of model adaptations and this would make the paper unnecessarily lengthier and more difficult to read. It would be challenging to keep the paper in its current concise form with a clear conceptual model to begin with. After all, in studies 5-7, which are the large-scale studies we conducted, we tested all important psychometric properties that are commonly reported in scale development (or even more). Thus, we believe that readers and users have all evidence needed to judge the psychometric quality of the scale.

For Study 5, the construct validity element is not clear. What were the t-values per measure, and what was the P value adopted? Why not use binomial regression to determine the key MIRES variables? And why use one simple measure like this (dieting/not dieting), when understanding such a complex construct? An appropriate index would have been to determine whether the MIRES is a better predictor than other measures of the same constructs, to show that the MIRES is a more clinically useful measure.

->We added the t values in Table S2 as requested (we also found some small inaccuracies in the Cohen’s d column and we corrected them). As we now more clearly explain in lines 364-368, in Study 5 we tried to get preliminary evidence on the scale’s construct validity. Our aim was not to make an accurate prediction of dieting behaviour, but to show that non-dieters score higher in all MIRES subscales than dieters. This supports the very nature of the constructs we investigate in a broad sense, since internally regulated eating is by nature a non-diet eating style. Also, it was not the aim of this study to test the scale for criterion or incremental validity, as is suggested by the reviewer. Evidence on those types of validity are presented in Study 7 (Table S6 and Table 6).

If Study 6 really did use the same analyses as Study 5, where is the report of the CFA? If one was conducted, then the sample in Study 6 appears to be underpowered, even by the authors' relatively lax standards from study 6.

->In lines 401-415 we describe the exact procedure that we followed to analyse the data of Study 6. We now clarify that we conducted individual analyses per subscale. Because we ran and compared up to seven CFA models per subscale (19 subscales), we decided not to report these results in a table, but rather to describe them briefly in lines 416-425, accompanied by Table 4. Since analysis was conducted per subscale the sample size was adequate for testing all first-order factor models. For testing the stability of second order factors the sample was indeed a bit small (4:1 ratio). We addressed this issue in lines 391-394.

In Study 6, the lack of temporal stability severely undermines the test-retest reliability of the measure - correlations on their own are not adequate, if scores change over time, which they do on two of the top three factors.

->The reviewer is right that the scores of two subscales (IT and FL) change over time despite the fact that stability coefficients are adequate. This means that there is variation over time across the whole studied population. The change in means across time is a stricter test of stability than correlations and is not always reported by scale developers. However, we deem this important and so we have reported these results, hoping that readers will appreciate it. We made some edits in lines 456-460 and 813-815 in order to address this issue more explicitly.

Why were 4 items dropped at the end of Study 6? As this was the most underpowered study of the two in this section, it seems unclear why one would make that change.

->As we now more explicitly explain in lines 435-446, we decided to drop four items in order to reduce the length of the scale and to have the same number of items per subscale. The reason why we dropped those items after the end of Study 6 was because we wanted to test the stability of the MIRES subscales first. The deletion of items was not based on findings from Study 6 but on the meaning of items, because all subscales were initially found to be stable on the basis of stability coefficients, which were adequate. The change in means was an extra analysis that we conducted at a later stage when writing the paper. By that time all studies had been completed.

In Study 7, the authors perform yet another CFA. Why? And what criteria are used for concluding that their measure was better than the others? That was not apparent from Table 5. In short, have the authors really just reinvented the wheel here, using a relatively long, psychometrically weak measure to do a job that was already done just as well by existing measures (in the case of Table 6, it looks as if the IES-2 already does this job for the core weight variables).

->As we explain in lines 470-473, in Study 7 we test for the first time the full formative model of MIRES, including the six reflective items.

The criteria we used to decide whether our scale is better than the others included 1. the ability to predict outcome measures – Table S6 and Table 5 (descriptive comparison of scales) and 2. the incremental variance accounted for by MIRES – Table 6 (statistical testing). As can be clearly seen in Table 6, our scale accounts for significant amounts of extra variance in most outcome measures above and beyond the variance accounted for by IES-2 and ecSI-2 (see also edits in lines 715-720, which are intended to make this point clearer). Exceptions are the weight outcomes for which all three measures are relatively weak predictors.

The norms in Table 7 are meaningless if there is no validation.

->Perhaps the reviewer has a different interpretation of norm scores than we do. We do not use norm scores in relation to predictive validity, i.e., expecting that individuals from different norm score categories would behave differently or show different outcomes, which would indeed require further validation. We use them from a descriptive point of view, to allow researchers to use them as reference and to compare between different groups or populations. Also, a subject’s raw score has little meaning, unless we know the subject’s position relative to others in some group/population. Thus, our norm scores provide the basis for comparing different groups or an individual with a group. We now clarify this in lines 742-745. As we explained earlier, we do not provide age-based or gender-based norms because this would require additional analyses that could perhaps be included in another publication. Finally, we re-calculated the norms based only on the data from the US sample because it is more useful to have country-based norms (see edits in lines 736-738 and Table 7).

Attachment

Submitted filename: Response to reviewers.docx

Decision Letter 1

Gian Mauro Manzoni

28 Apr 2020

PONE-D-19-26656R1

Development and validation of the multidimensional internally regulated eating scale (MIRES)

PLOS ONE

Dear Mrs Palascha,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

We would appreciate receiving your revised manuscript by Jun 12 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Gian Mauro Manzoni, Ph.D., Psy.D.

Academic Editor

PLOS ONE

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: (No Response)

Reviewer #2: No

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: (No Response)

Reviewer #2: No

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: (No Response)

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: (No Response)

Reviewer #2: No

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: I do not understand the authors' conclusion that, having spent all that effort to demonstrate that the short and long version of the paper are equivalent, it should be up to the reader to decide which version to use. They present no rationale for having two versions, and I would recommend that they should focus on the shorter version, as that will eventually take a lot less time on the part of patients, clinicians and researchers.

The authors state that they clarity their preferred scoring in lines 840-848, but they do not.

The paper remains excessively long, and the authors do not justify that length in their response letter. I did not find that the very minor changes alluded to made it any clearer what the aims, clinical issue or implications were.

The lack of norms is defended as potentially making the paper longer (which would be more than offset if the authors had actually shortened the paper in response to previous feedback). However, the authors ignore the key issue that was identified - the lack of comparability between samples.

The authors decline to provide key findings in Studies 1-4. The same applies later to other studies. That means that they are not replicable. That is a fundamental error in scientific communication, and I could not support publication of any such work, including this paper.

The authors do not justify the lack of validation for their scores in Table 7, and they remain meaningless as a result.

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Oct 8;15(10):e0239904. doi: 10.1371/journal.pone.0239904.r004

Author response to Decision Letter 1


11 Jun 2020

General response

We want to thank the reviewers for taking the time to review our paper for a second time. We are happy that one of the reviewers was satisfied with the first revision and gave consent for acceptance of our paper. We equally appreciate the effort of the second reviewer to provide us with additional insights for improving our paper. We found some comments challenging, but we feel we have managed to address them all. The paper has been substantially revised in response to the reviewer comments. Most important changes are: (i) providing the details on studies 1-4, which we originally addressed more as pre-studies, but have now become an integral part of the “full history” of the development of the 21-item MIRES scale, (ii) strong recognition of the short version of the scale (21 items instead of 45 items), and (iii) shortening of the paper (from 12375 words to 11953 words) despite adding the requested material on studies 1-4, primarily obtained in the background information to the scale. We are grateful for the reviewer’s critical look at the paper and feel that the paper has benefited greatly from the reviewer’s comments. Please find our specific responses embedded in the text below.

Reviewers' comments:

Reviewer #2:

1. I do not understand the authors' conclusion that, having spent all that effort to demonstrate that the short and long version of the paper are equivalent, it should be up to the reader to decide which version to use. They present no rationale for having two versions, and I would recommend that they should focus on the shorter version, as that will eventually take a lot less time on the part of patients, clinicians and researchers.

In hindsight, we agree with the reviewer that most potential users may prefer to use the short version of the scale and that they should be able to access it quickly. In the revised manuscript, we now appreciate this from the beginning by including the following changes:

• We now mention the short version of the scale in the abstract (lines 30-33) and introduction (lines 92-95) so that readers expect this as the final outcome.

• We have made edits in lines 1031-1035 to contrast the two versions and show the advantage of the short version more explicitly.

• We moved Fig S2, which illustrates the simplified version, into the paper (see Fig 2 in lines 1037-1040) and moved Fig 2, which illustrates the full version, to the supplementary material (see S2 Fig). In the new Fig 2, we replaced the item codes with the actual items so that readers can see directly how the scale looks. S1 Appendix still provides information on how the full version was administered in our studies (to be used as a reference to the version we used in our studies) (lines 525-538). Although we are now directing users to the short version, we believe it is important to present the full version for transparency purposes.

• We explicitly advise potential users to go for the short version of the scale (lines 1107-1109).

• We skipped lines 1125-1130.

2. The authors state that they clarity their preferred scoring in lines 840-848, but they do not.

The main point raised in the first review was that users should not measure only a subset of subscales if they want to make conclusions about IRE style because all subscales of a formative construct need to be measured for the formative construct to be measured. This is the issue we addressed in lines 840-848 of the first revision. In the revised manuscript, we now explicitly refer to the appropriate scoring depending on the case (lines 1114-1119).

3. The paper remains excessively long, and the authors do not justify that length in their response letter. I did not find that the very minor changes alluded to made it any clearer what the aims, clinical issue or implications were.

We agree with the reviewer that the paper is long. We made additional efforts to shorten the paper. As we mentioned in the first revision, we believe all information about the scale’s properties is necessary and useful for the scale’s users; therefore, we tried to cut down parts of the paper that are less critical, i.e., information that can be traced back to existing literature. Effectively, we managed to reduce the length by 422 words and achieved that through the following changes:

• We removed examples of items and evidence on validity for the various measures we used in Study 7 to validate our scale (lines 754-881).

• We shortened the conceptual part by skipping background information and evidence of adaptivity for the IRE characteristics (lines 134-180), because a detailed overview of the conceptual model can be found in our theoretical paper that is freely accessible online. We now present only the definitions of the IRE characteristics and we explain how we operationalize them. In lines 126-128 we refer readers to the complete overview of the model.

• We skipped the description of the temporal stability assessment process because readers can find the full description of the method in the paper of Steenkamp & Van Trijp (lines 659-668).

• We skipped the section on norms (lines 982-1001); but see also point 4 below.

• We also made smaller edits throughout the paper to aid brevity and clarity (e.g., see lines 57-92).

• We skipped the wrap up of findings in the general discussion (lines 1059-1081) and adapted the abstract to contain all this information.

4. The lack of norms is defended as potentially making the paper longer (which would be more than offset if the authors had actually shortened the paper in response to previous feedback). However, the authors ignore the key issue that was identified - the lack of comparability between samples.

Even though we are not entirely sure, we believe that the reviewer aims to communicate that the whole scale development process should be conducted with comparable samples and that should non-comparable samples be used, measurement invariance of the model would have to be tested to ensure that the model holds its properties across the different samples.

We agree that measurement invariance testing is important when new scales are developed and perhaps should be conducted not only for different age groups but also for gender groups or other clinically relevant groups such as people with obesity. We now address the issue of comparability between samples in lines 1092-1105 and we have conducted additional analyses to provide preliminary evidence on the measurement invariance of our model across the age groups of interest (18-34 and 35-65). We realized that the sample sizes in our subgroups (in both studies 5 and 7) are too small to conduct typical measurement invariance testing (at least 980 participants per group are required to test the model), thus, we tried to find an alternative solution. Instead of testing for configural, metric, scalar, and residual invariance in a stepwise process as is usually done, we tried to test our model for full invariance at once by fixing all parameters to the values obtained from the full sample of study 7 and then fitting this model in each of the subgroups (18-34 and 35-65). In this way, the model had less parameters to be estimated and the sample sizes of our subgroups were adequate to provide reliable estimates. To warrant model identification we left only seven parameters free, the regression coefficients of the seven formative indicators. The fit of the model was good in both age groups providing preliminary evidence for the measurement invariance of our model. We further discuss this issue in the directions for future research (lines 1150-1153), where we also make a link to the development of norms.

5. The authors decline to provide key findings in Studies 1-4. The same applies later to other studies. That means that they are not replicable. That is a fundamental error in scientific communication, and I could not support publication of any such work, including this paper.

Per the reviewer’s request we have now included extensive information on these early studies as part of the “full history” of the MIRES development, both conceptually and in interaction with measurement. Studies 1-4 were initially considered pilot studies leading to the 52 initial items of the MIRES scale. These studies have helped us to develop that initial item set based on conceptual and empirical considerations. We absolutely agree with the reviewer on the importance of transparency and reproducibility of research. In the previous version we had focussed on the later stages and outcome of MIRES, and we now shift to describe the full history of the process with a stronger focus also on the early stages of scale development.

To allow readers to understand the full process of scale development that we followed, we added all details of Studies 1-4 in the paper and we made all data of this project freely accessible to the public (see edits in lines 30-37, 98-114, 206-215, 309-508). This will allow other researchers to replicate our work and to build further on it. Taking into account that there is always a degree of subjectivity in the scale development process (e.g., when selecting items to retain or to drop) we believe that readers are now fully equipped to understand the decisions we took during the scale’s trajectory by accessing the full data of the project.

6. The authors do not justify the lack of validation for their scores in Table 7, and they remain meaningless as a result.

The reviewer is right that descriptive kind of norms are less informative compared to validated norms that have been developed in relation to important clinical outcomes. After all, norms should develop from multiple studies and not from a single study. For these reasons and to help reduce the length of the paper we decided to skip the section on norms (lines 982-1001) and to discuss this issue in the discussion section as a promising direction for future research (lines 1152-1153).

Attachment

Submitted filename: Response to reviewers.docx

Decision Letter 2

Rodrigo Ferrer

24 Jul 2020

PONE-D-19-26656R2

Development and validation of the multidimensional internally regulated eating scale (MIRES)

PLOS ONE

Dear Dr. Palascha,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

First of all, allow me to compliment you on your remarkable research, which is undoubtedly a contribution to the assessment of eating behaviors. I have reviewed your paper completely and the responses to the reviewers and I think you have done a good job. However, there is one comment on which I agree with reviewer 2 and that is the one referring to the length of the manuscript, although, more than the length, I have my misgivings about the structure.

In particular, I consider that all the studies are, in fact, part of one same study (with different samples and analyses), whose division into studies makes the manuscript somewhat complicated. I believe that there are parts of the report that are sufficient to be mentioned within the procedure (studies 1 to 4), putting the different pools of items as supplementary material, while, for the remaining studies, it would only be necessary to specify the different samples and the different analyses.

If you do not agree with modifying and unifying the structure of the study, please try to make it as integrated and simple as possible. I know that, at this stage of the work, it can be tedious to make this type of cosmetic transformation, however, it is necessary for me to make this recommendation since, in order for your work to have the scope it deserves, it is not only necessary to have the technical rigour (which your work already has), but it also needs to be friendly and easy to understand for the reader.

Please submit your revised manuscript by Sep 07 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Rodrigo Ferrer, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (if provided):

Dear Authors:

First of all, allow me to compliment you on your remarkable research, which is undoubtedly a contribution to the assessment of eating behaviors. I have reviewed your paper completely and the responses to the reviewers and I think you have done a good job. However, there is one comment on which I agree with reviewer 2 and that is the one referring to the length of the manuscript, although, more than the length, I have my misgivings about the structure.

In particular, I consider that all the studies are, in fact, part of one same study (with different samples and analyses), whose division into studies makes the manuscript somewhat complicated. I believe that there are parts of the report that are sufficient to be mentioned within the procedure (studies 1 to 4), putting the different pools of items as supplementary material, while, for the remaining studies, it would only be necessary to specify the different samples and the different analyses.

If you do not agree with modifying and unifying the structure of the study, please try to make it as integrated and simple as possible. I know that, at this stage of the work, it can be tedious to make this type of cosmetic transformation, however, it is necessary for me to make this recommendation since, in order for your work to have the scope it deserves, it is not only necessary to have the technical rigour (which your work already has), but it also needs to be friendly and easy to understand for the reader.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Oct 8;15(10):e0239904. doi: 10.1371/journal.pone.0239904.r006

Author response to Decision Letter 2


9 Sep 2020

Response to reviewers

Additional Editor Comments (if provided):

Dear Authors:

First of all, allow me to compliment you on your remarkable research, which is undoubtedly a contribution to the assessment of eating behaviors. I have reviewed your paper completely and the responses to the reviewers and I think you have done a good job. However, there is one comment on which I agree with reviewer 2 and that is the one referring to the length of the manuscript, although, more than the length, I have my misgivings about the structure.

In particular, I consider that all the studies are, in fact, part of one same study (with different samples and analyses), whose division into studies makes the manuscript somewhat complicated. I believe that there are parts of the report that are sufficient to be mentioned within the procedure (studies 1 to 4), putting the different pools of items as supplementary material, while, for the remaining studies, it would only be necessary to specify the different samples and the different analyses.

If you do not agree with modifying and unifying the structure of the study, please try to make it as integrated and simple as possible. I know that, at this stage of the work, it can be tedious to make this type of cosmetic transformation, however, it is necessary for me to make this recommendation since, in order for your work to have the scope it deserves, it is not only necessary to have the technical rigour (which your work already has), but it also needs to be friendly and easy to understand for the reader.

Response:

We would like to thank the editor for the time he spent on reading our original paper and the subsequent revisions and for his particularly kind and gratifying words regarding our work. We found the additional suggestions of the editor very useful and we believe the new edits we made have made the paper more “to the point” and easier to read. In line with the editor’s suggestion, we now present studies 1-7 as different parts of a single study with a unified Methods section and a unified Analysis and Results section. We refer to studies 1-4 as the preparatory work of this research (mentioned briefly under the Methods section) that led to the structure we used in large-scale testing of the MIRES. The data of all studies are still available online in the OSF repository (we changed the study numbers to descriptive titles so that readers can link more easily to our paper) and the versions of the MIRES items are now presented within a single document in the supplementary material (S2 Appendix). We also adapted Figure 1 so that it doesn’t refer specifically to studies and doesn’t include the sample sizes but is more descriptive of the properties of MIRES that we tested in this research. Finally, we have added a large paragraph in lines 948-985 of the Discussion section to summarize the main findings of this research because the individual discussions sections we used in the previous version have now been deleted. With these changes we managed to reduce the length of the paper by 2,619 words.

As an additional point, we now cite the latest version of our theoretical paper, which is also at the second round of revision in another journal (Palascha A, van Kleef E, de Vet E, van Trijp H. Internally Regulated Eating Style: A Comprehensive Theoretical Framework. OSF [Preprint, version 2]. 2019 [cited 2020 Sep 4]. Available from: https://osf.io/rmbft/ doi: 10.31219/osf.io/rmbft).

We hope these changes to have produced a paper that meets the criteria for publication at PlosOne, but we are happy to make more changes if the editor deems this necessary. We are looking forward to the editor’s response.

Attachment

Submitted filename: Response to reviewers.docx

Decision Letter 3

Rodrigo Ferrer

16 Sep 2020

Development and validation of the multidimensional internally regulated eating scale (MIRES)

PONE-D-19-26656R3

Dear Dr. Palascha,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Rodrigo Ferrer, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Rodrigo Ferrer

30 Sep 2020

PONE-D-19-26656R3

Development and validation of the multidimensional internally regulated eating scale (MIRES)

Dear Dr. Palascha:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Rodrigo Ferrer

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. Conceptual model of internally regulated eating style.

    The direction of arrows indicates whether a construct is formative—arrows point to the construct—or reflective—arrows point to the dimension.

    (TIF)

    S2 Fig. The multi-dimensional model of internally regulated eating style (full version).

    All loadings were significant at the 0.01 level. Context effects, method effects, covariances between first- and second-order factors, and disturbance terms of first- and second-order factors are not depicted in the figure for easier readability.

    (TIF)

    S1 Table. Factor loadings for the MIRES first- and second-order factors.<.

    /SI_Caption>

    (DOCX)

    S2 Table. Mean scores on MIRES first- and second-order factors for dieters and non-dieters.

    (DOCX)

    S3 Table. Additional sample characteristics of the US sample.

    (DOCX)

    S4 Table. Bivariate correlations of summed scores of MIRES, RI, and MIRES subscales.

    (DOCX)

    S5 Table. Bivariate correlations of summed scores of MIRES, RI, and MIRES subscales with IES-2 and ecSI-2.

    (DOCX)

    S6 Table. Standardized regression coefficients (and R2) for the criterion validity of MIRES, IES-2, and ecSI-2.

    (DOCX)

    S7 Table. Standardized regression coefficients (and R2) for the criterion validity of SH, SS, SEH, SES (full subscales including neutral, emotional, external contexts) vs. the neutral counterpart of each subscale.

    (DOCX)

    S8 Table. Bivariate correlations of summed scores of MIRES (21 items), RI, and MIRES subscales.

    (DOCX)

    S9 Table. Bivariate correlations of summed scores of MIRES (21 items), RI, and MIRES subscales with IES-2 and ecSI-2.

    (DOCX)

    S10 Table. Incremental variance in outcome measures accounted for by MIRES (21 items).

    (DOCX)

    S1 Appendix. The Multidimensional Internally Regulated Eating Scale (MIRES).

    (DOCX)

    S2 Appendix. Modifications of the MIRES item pool during the scale development and validation process.

    (DOCX)

    Attachment

    Submitted filename: Response to reviewers.docx

    Attachment

    Submitted filename: Response to reviewers.docx

    Attachment

    Submitted filename: Response to reviewers.docx

    Data Availability Statement

    The data of this research have been uploaded in the OSF Repository (Creation date: 06/03/2020, URL: https://osf.io/w3guh/?view_only=406839237c9f4bc09ff154d065c17da3).


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES