Skip to main content
Children logoLink to Children
. 2022 Jun 29;9(7):972. doi: 10.3390/children9070972

Effects of Anonymity versus Examinee Name on a Measure of Depressive Symptoms in Adolescents

César Merino-Soto 1, Anthony Copez-Lonzoy 2, Filiberto Toledano-Toledano 3,4,*, Laura A Nabors 5, Jorge Homero Rodrígez-Castro 6, Gregorio Hernández-Salinas 7, Miguel Ángel Núñez-Benítez 8
Editor: Cesar A Soutullo
PMCID: PMC9315511  PMID: 35883956

Abstract

There is evidence in the literature that anonymity when investigating individual variables could increase the objectivity of the measurement of some psychosocial constructs. However, there is a significant gap in the literature on the theoretical and methodological usefulness of simultaneously assessing the same measurement instrument across two groups, with one group remaining anonymous and a second group revealing identities using names. Therefore, the aim of this study was to compare the psychometric characteristics of a measure of depressive symptoms in two groups of adolescents as a consequence of identification or anonymity at the time of answering the measuring instrument. The participants were 189 adolescents from Metropolitan Lima; classrooms were randomly assigned to the identified group (n = 89; application requesting to write one’s own name) or to the anonymous group (n = 100; application under usual conditions), who responded to the Childhood Depression Inventory, short version (CDI-S). Univariate characteristics (mean, dispersion, distribution), dimensionality, reliability, and measurement invariance were analyzed. Specific results in each of the statistical and psychometric aspects evaluated indicated strong psychometric similarity. The practical and ethical implications of the present results for professional and research activity are discussed.

Keywords: adolescents, assessment, anonymous response, validity, depression measure

1. Introduction

According to systematic reviews on the topic of social desirability in the clinical context, a situation that represents a pervasive risk in applied research based on self-report, application for clinical and forensic practice seems to be attributed to the identification of those evaluated [1,2]. To mitigate response biases associated with self-report and to conform to ethical standards, participants are generally asked to respond anonymously. The requirement of response anonymity has a long history in the application of surveys of all types, as well as its implications for its adequacy in the ethical standards of scientific research. However, this strategy is also exposed to particular effects due to its subjective value on the privacy of responses [3], and anonymity has influenced the quality of responses even in ethnocultural research contexts and clinical applications, overestimating scores [4,5,6]. On the other hand, in nonclinical samples, anonymity may reduce the sense of responsibility in the response process [7], even more so in the face of stigmatizing behaviors such as stealing, cheating, psychoactive substance use, and erotophilic behaviors [8,9,10,11,12]. Even in circumstances of respondent quasi-identity, perceived identity protection influences the possible contamination of scores related to gender, age, and place of origin [9]. Finally, response anonymity does not guarantee the absence of careless responses or insufficient effort (C/IE; [13]), given that this type of response is almost always present in anonymous surveys presented in pencil–paper format or on a web platform [13,14,15].

At this point, we arrive at the following question: Would there be an impact on mental health assessments of adolescents if their responses are anonymous or identified? Apparently, this question has not been asked before in the context of screening the adolescent community, and it seems possible that it has not been asked in published research. The anonymous responses of adolescents in assessments for research purposes does not appear to be problematic because no actual or masked identification of the evaluee is usually required; however, in screening assessments for emotional problems within an institution, accurate referral requires identifying the adolescent being assessed to refer him or her to appropriate clinical intervention services [16,17]. The identification of the symptomatology associated with childhood depression is essential to reduce its effects on the mental health of children and adolescents. The USA’s National Institute of Mental Health [18] indicated that in the preadult stage (among children and adolescents), approximately three million individuals suffer from mental disorders [18] and require proper identification to provide them with timely clinical services. This is more sensitive because the adolescent stage is vulnerable to mood alterations, social and school behavioral changes, and transition to new family roles [19,20]. Therefore, in the context of the assessment of children and adolescents, it is necessary to use screening instruments that are widely applied [17] and to use especially short scales, because they reduce irrelevant variance, potentially producing acceptable levels of specificity and sensitivity and thus improving control of Type I and Type II errors in identification and referral to clinical services [21,22,23,24,25]. Like longer measures with more items, short scales have advantages and limitations that the user must weigh out in deciding on their use and the interpretation of their scores. However, in the context of mass use and given the purpose of screening, short scales with good evidence of validity may be the best option.

One of the instruments for the detection of childhood depression is the Child Depression Inventory (CDI; [26]). This measure has a shorter version (Child Depression Inventory-Short, CDI-S) that is used as a screening and treatment follow-up instrument. It takes between 5 and 10 min to administer and even less time to score. Overall, the CDI-S is a cost- and speed-efficient tool for assessing behaviors of low population prevalence [27] and for the assessment of adults with intellectual disabilities [28] and populations with physical disabilities [29], including in the school context [30]. The CDI-S can have better evaluative efficiency than the full version because of its intrinsic and psychometric characteristics, and one of them is its dimensionality. That is, while studies using the full version yield different factorial solutions (between three and eight factors, possibly associated with the analysis strategies applied and the criteria that the different authors applied; [27]), the dimensionality of the brief version seems less problematic due to the reduced number of items. If the internal structure of the CDI is modified in subsequent studies, the problem lies in the instability (a) of the construct to be generalizable across contexts and (b) of the content sampling of the construct as originally planned. Additionally, whenever this structure of the instrument changes, the interpretation is not always statistically or conceptually comparable across groups or studies.

The CDI-S uses the self-report method, and this type of procedure generally presents several challenges that the researcher or clinician must recognize and address. For example, one such limit is social desirability [3,8,12], which tends to interact with the examinee’s perception of threat toward the evaluative situation [31,32]. Therefore, the aim of this research was to analyze the effect of anonymity and subject identification on the psychometric properties (internal structure, reliability, measurement invariance) of the CDI-S scores in the adolescent population. This was conducted in the context of the natural application of screening instruments to detect early symptoms of depression.

2. Materials and Methods

2.1. Participants

In total, 189 Peruvian adolescent students living in Metropolitan Lima were enrolled. All were enrolled in a public, tuition-free educational institution at the secondary elemental level. The majority (69.1%) lived in the same district in which the institution was located. The mean age was 13.23 and ranged from 11 to 17 years (SD = 1.14). The distribution of students in their grades of study was as follows: first (51, 26.0%), second (64, 32.7%), third (69, 35.2%), and fourth (12, 6.1%). The demographic characteristics of the adolescents are shown in Table 1. For the purposes of the study, the participants were divided into two groups, using the classroom as the unit to identify them and randomly assign the modification of the CDI-S filling instructions (see the Procedure section); the groups were identified as group A (those who received the unmodified instructions, n = 100) and group B (modified group, or those who received the modified instructions, n = 89).

Table 1.

Description of participants (n = 196).

N %
Sex
Male 99 50.5
Female 97 49.5
Place of birth
Lima 165 84.2
Other 31 15.8
Family structure
I live with both parents 122 62.2
I live with one of my parents 60 30.6
I live with other people 14 7.1
Mother’s (father’s) education
Less than high school 30 (25) 15.3 (12.8)
Completed high school 91 (81) 46.4 (41.3)
Technical education (1 to 2 years) 20 (13) 10.2 (10.2)
Technical (3 years) 10 (45) 5.1 (6.6)
University 41 (12) 20.9 (23.0)
No information 4.0 (2.0) 2.0 (6.1)

2.2. Instruments

Children’s Depression Inventory-Short (CDI-S; [33]). The Spanish version of the CDI-S [34] was used. This self-report is used to screen for depressive symptoms, derived from the 27 item long version. The CDI-S can be applied to children and adolescents between the ages of 7 and 17 years individually or in groups. The CDI-S consists of 10 items selected by the author as the most representative of the construct, and its format is identical to that of the longer version. Each item has three phrases describing symptoms ranging from less (absence of the symptom) to more intense (severe presence of the symptom). The instructions ask the participants to choose the sentence that best fits how he or she has felt in the last 15 days. Items 2, 4, 5, 6, and 10 are reverse scored. The internal consistency coefficient found in the adaptation of the Hispanic version was 71 [34].

2.3. Ethical Considerations

This study is a part of the research project (HIM/2015/017/SSA.1207; “Effects of mindfulness training on psychological distress and quality of life of the family caregiver”) that was approved on 16 December 2014 by the Research, Ethics, and Biosafety Commissions of the Hospital Infantil de México Federico Gómez National Institute of Health in Mexico City. While conducting this study, the ethical rules and considerations for research with humans currently enforced in Mexico [35] and those outlined by the American Psychological Association [36] were followed. All family caregivers were informed of the objectives and scope of the research and their rights in accordance with the Declaration of Helsinki [37]. The caregivers who agreed to participate in the study signed an informed consent letter. Participation in this study was voluntary and did not involve payment. The caregivers who provided consent for their child to participate completed an informed consent letter. Youth participants provided assent and returned a survey if they wished to participate.

2.4. Procedure

The authorization of the directors of the educational institution was obtained, and the corresponding permissions were requested from the parents, who were informed of the research proposal and the data collection procedures. Once the directors and parents agreed to participate, the instrument was administered during class time. The students who provided assent completed the CDI-S. Classrooms were randomly assigned to groups A and B; these groups had different instructions for filling out the CDI-S: group “A” received instructions to fill out the CDI-S anonymously, while group “B” was asked to give their name in order to have a better identification at the time of collecting the completed questionnaires. The general instruction given to the adolescents emphasized that they could stop responding at any time, without consequence. All information on examinees in both groups was transferred to a database, but the names of the examinees in group B were not entered into this database. When the database was completed, the written names were removed from the paper questionnaires.

2.5. Data Analysis

The analysis consisted of univariate and multivariate analysis phases. First, several statistical aspects of the items, such as distribution, location (mean), dispersion (standard deviation), and floor and ceiling (minimum and maximum frequency of response), were analyzed. The statistical comparison between the distributions of each item was made using the KS-D test [38,39] for two independent samples, and the overlap coefficient (OVL; [40]) was used as a measure of the practical significance of the comparison of two distribution functions that are not necessarily normally distributed [41]; the model was used for different variances to ensure better precision.

The internal structure of the CDI-S was examined by a confirmatory factor analysis, with the maximum likelihood method adjusted for item nonnormality (SB-χ2; [42]), on the matrix of interitem polychoric correlations; given the limited number of response categories, this approach can be a satisfactory estimation method [43,44,45]. The measurement invariance of the items was examined by means of two procedures: the first was the metric congruence of the items [46], in which the factor loadings of the items of each group were compared by means of the congruence coefficient (φ; [46]). The second procedure used differential item functioning analysis (DIF; [47]), with the following specifications: (a) the matching variable was the observed score, θ, and (b) the grouping variable (G) was the status of the group examined, where the reference group was “A” (anonymous group) and the focus group was “B” (provided names). The DIF analysis was implemented with ordinal logistic regression (OLR; [48]), in which each item was assumed to be a dependent and continuous latent variable (Z, standardized in logits). The independent variables were the measured attribute (or observed score, θ), subject grouping (G; group A vs. group B), and attribute–group interaction (θ*G). Each represents a different type of DIF [49,50]. The OLR methodology consists of modeling three equations: one representing the nonuniform DIF (OLR1, Z = β0 + β1θ + β2G + β3θ*G), one for uniform DIF (OLR2, Z = β0 + β1θ + β2G), and another model for representing responses without DIF (OLR3, Z = β0 + β1θ). The stepwise screening strategy [49,50] focused on the evaluation of practical and statistical significance, according to which for each item we first evaluated the difference between the −2 log likelihood (Δχ2, gl = 1, α = 0.05) between OLR1 and OLR2 models for detection of no uniform DIF (null hypothesis: OLR1 = OLR2) and then between OLR3 and OLR2 for detection of uniform DIF. The Bonferroni correction [50,51] was applied to adjust nominal α according to the number of items (0.05/10 = 0.005). Results below this level (αBonferroni = 0.005) identified the impact of the interaction term (θ*G) and therefore the presence of nonuniform DIF. If the previous null hypothesis (nonuniform DIF) is not rejected, the second step tested the uniform DIF by the difference (Δ) of the beta coefficients of the models OLR3 (βθ) and OLR2 (βG). A result ≥10% indicated statistical significance at the nominal level α = 0.20 [50].

Finally, reliability was estimated by the α coefficient [52] and ω [53]; although ω tends to be more appropriate [54], the α coefficient was also reported because it is a measure of score reliability that (a) is still popular in behavioral science research, (b) serves for direct comparison with the Spanish validation study, and (c) allows for comparison with the ω coefficient to assess the impact of possible noncompliance with the basic assumption for using α [54].

3. Results

3.1. Equivalence between Groups

The equivalence of characteristics in both groups was analyzed in an equivalence testing framework [55]. To maximize the sensitivity of the test for equivalence of means, the minimum standardized difference was set at d = 0.10 [56]. This showed that the average ages of the two groups were equivalent, t = 0.90 (gl = 194, p = 0.40). Similarly, statistical and practical equivalence between the two groups was found in the distribution of the following variables: school levels, Mantel–Haenszel χ2 (gl: 1) = 3.20, p = 0.07, γ = 0.178; sex, χ2 (gl: 1)= 5.86, p = 0.48, VCramer = 0.05; family configuration, χ2 (gl: 2) = 5.86, p = 0.05, VCramer = 0.17; place of birth, χ2 (gl: 1) = 1.61, p = 0.20, VCramer = 0.09; mother’s level of education, Mantel–Haenszel χ2 (gl: 1) = 1.80, p = 0.17, VCramer = 0.16; and father’s level of education, Mantel–Haenszel χ2 (gl: 1) = 0.4, p = 0.83, VCramer = 0.13. Considering these results, the sociodemographic equivalence of both groups can be accepted.

3.2. Univariate Analysis

3.2.1. Items Level

The univariate statistics for the items (Table 2) in both groups were similar, and the discrepancies can be established as small. The Pearson correlations of these descriptive statistics (M, SD, g1, g2, floor and ceiling effect) between groups A and B were 0.96, 0.94, 0.95, 0.92, 0.95, and 0.97, respectively. These high magnitudes confirm that the pattern of descriptive statistics at the item level was similar between the compared groups. To verify this more rigorously, the statistics for each item were analyzed individually. In Table 3, the distributional differences in the response range of each item were not statistically significant (KS-D between 0.009 and 0.11), and the degree of overlap (coefficient OVL) between the distributions was greater than 79.2% but approximately 95%, suggesting that the items showed practically overlapping distributions between groups A and B (see Figure 1). Differences in the location or media (d between |0.000| and |0.210|; t-test < 1.50) and variances (FL; Levene [57]; α nominal with Bonferroni correction: 0.05/10 = 0.005) were essentially trivial and not statistically significant. These results, taken together, point to univariate similarity at the item level between the two groups.

Table 2.

Univariate descriptive and CFA for the items CDI-S.

Item Descriptive Statistics CFA
M 95% CI SD g 1 g 2 ritc Floor Ceiling λ h 2
Group A (anonymous)
Item 1 0.094 [0.01; 0.18] 0.353 4.032 16.751 0.43 92.5 1.90 0.740 0.548
Item 2 0.509 [0.38; 0.64] 0.521 0.168 −1.51 0.396 50 0.90 0.604 0.365
Item 3 0.179 [0.08; 0.28] 0.385 1.697 0.895 0.455 82.1 0.00 0.722 0.521
Item 4 0.17 [0.07; 0.27] 0.402 2.224 4.229 0.496 84 0.90 0.792 0.628
Item 5 0.094 [0.01: 0.18] 0.353 4.032 16.751 0.489 92.5 1.90 0.824 0.679
Item 6 0.557 [0.37: 0.74] 0.757 0.946 −0.603 0.313 60.4 16.0 0.377 0.142
Item 7 0.575 [0.45; 0.70] 0.515 −0.097 −1.505 0.235 43.4 0.90 0.448 0.201
Item 8 0.358 [0.23; 0.49] 0.52 1.009 −0.128 0.614 66 1.90 0.794 0.631
Item 9 0.321 [0.19; 0.45] 0.508 1.217 0.407 0.414 69.8 1.90 0.647 0.419
Item 10 0.123 [0.04; 0.20] 0.33 2.334 3.513 0.431 87.7 0.00 0.810 0.656
Group B (name of examinee)
Item 1 0.056 [−0.01; 0.12] 0.23 3.947 13.884 0.212 94.4 0.00 0.617 0.380
Item 2 0.544 [0.40; 0.69] 0.523 0.06 −1.462 0.475 46.7 1.10 0.601 0.362
Item 3 0.189 [0.08; 0.30] 0.394 1.617 0.627 0.247 81.1 0.00 0.575 0.330
Item 4 0.111 [0.02; 0.21] 0.35 3.289 11.138 0.453 90 1.10 0.757 0.574
Item 5 0.067 [−0.01; 0.15] 0.292 4.814 24.931 0.443 94.4 1.10 0.876 0.768
Item 6 0.556 [0.34; 0.77] 0.795 0.981 −0.692 0.086 63.3 18.9 0.156 0.024
Item 7 0.467 [0.33; 0.61] 0.524 0.376 −1.331 0.316 54.4 1.10 0.521 0.271
Item 8 0.356 [0.21; 0.50] 0.547 1.233 0.577 0.455 67.8 3.30 0.580 0.336
Item 9 0.378 [0.25; 0.51] 0.488 0.513 −1.777 0.296 62.2 0.00 0.323 0.104
Item 10 0.178 [0.07; 0.29] 0.413 2.193 4.15 0.35 83.3 1.10 0.685 0.469

g1 and g2: skewness and kurtosis coefficients; ritc: item–test correlation; CFA: factor loadings (λ) of the confirmatory factor analysis; h2: squared factor loading.

Table 3.

Comparative results between groups.

Item Distribution Location (Mean) Dispersion
KS-D OVL t (df) d (95% CI) FL (1, 194)
Item 1 0.019 0.792 0.71 (182.440) 0.099 (−0.18; 0.38) 3.363
Item 2 0.033 0.973 −0.40 (188.645) −0.057 (−0.34; 0.22) 0.007
Item 3 0.009 0.986 −0.18 (187.401) −0.025 (−0.31; 0.26) 0.119
Item 4 0.060 0.913 1.12 (193.870) 0.150 (−0.12; 0.44) 4.385
Item 5 0.019 0.904 0.43 (193.878) 0.061 (−0.22; 0.34) 1.388
Item 6 0.029 0.976 0.009 (185.561) 0.000 (−0.28; 0.28) 0.393
Item 7 0.110 0.916 1.48 (187.786) 0.210 (−0.07; 0.49) 0.272
Item 8 0.017 0.975 0.038 (185.438) 0.000 (−0.28; 0.28) 0.006
Item 9 0.075 0.951 −0.84 (191.038) −0.120 (−0.40; 0.16) 0.660
Item 10 0.044 0.880 −1.11 (169.440) −0.160 (−0.44; 0.12) 4.515

KS-D: D Kolmogorov–Smirnov statistic; OVL: overlap coefficient; t (df): Student’s t-test for comparison of means and degrees of freedom; d: standardized difference; FL: Levene’s F test (gl1, gl2).

Figure 1.

Figure 1

Comparative density distribution of responses to items.

3.2.2. Score Level

Descriptive results in groups A (M = 2.981, SD = 2.629) and B (M = 2.900, SD = 2.312) suggested essential similarity (t = 0.23, gl = 194, p = 0.185), which was verified with a mean equivalence analysis [55] and contrasted against a standardized difference d = 0.10. The difference between variances (robust test, FL; [57]) also pointed to insubstantial differences (FL = 1.170, p = 0.281).

3.3. Internal Structure

3.3.1. Dimensionality

The fit of the items to a unidimensional model was good (p > 0.05, df = 35) for both groups: group A, SB-χ2 = 34.63, CFI = 1.00, RMSEA = 0.00 (90% CI = 0.00, 0.07); group B, SB-χ2 = 36.11, CFI = 0.99, RMSEA = 0.01 (90% CI = 0.00, 0.07). Within each dimension, factor loadings were high but heterogeneous (see Table 2, CFA heading).

3.3.2. Differential Item Functioning: Anonymity vs. Examinee Name

The metric congruence (equality of factor loadings between groups A and B) was φ = 0.989, which presents substantial equality between them [46]. The analysis of the nonuniform and uniform DIF (Table 4) in each item showed the absence of any type of DIF.

Table 4.

Results of differential functioning of items (ordinal logistic regression).

Item Non-Uniform DIF Uniform DIF
PDif.(LL) DIF β1β2 DIF
Item 1 0.226 No −0.0067 No
Item 2 0.224 No −0.0032 No
Item 3 0.461 No 0.0075 No
Item 4 0.363 No −0.0044 No
Item 5 0.981 No −0.0053 No
Item 6 0.918 No 0.0007 No
Item 7 0.278 No 0.0258 No
Item 8 0.778 No −0.0000 No
Item 9 0.712 No 0.0057 No
Item 10 0.702 No 0.0517 No

PDif. (LL): p value of difference between the −2 log likelihood; DIF: differential functioning of items.

3.3.3. Reliability

The α coefficients for groups A and B were 0.741 and 0.633, respectively; the difference between them [58] was not statistically significant, W = 1.417, F(93, 81) = 1.43: ω coefficients were 0.898 and 0.835 for groups A and B, respectively, and can also be considered to be similar.

4. Discussion

The aim of this research was to analyze the effect of anonymity and identification by name of adolescents in a research context. This effect was examined on the statistical and psychometric properties (internal structure, reliability, measurement invariance) of the scores of an abbreviated measure of depressive symptoms, the CDI-S. The findings of the present study are interesting and indicate that there are no effects on the psychometric properties and, consequently, on the interpretation of the CDI-S score. It can be stated that the Type I or II error that could be present in the identification and referral decisions with the CDI-S would probably be less associated with the identification of the assessed person so that the scores obtained are valid. Although a reduction in internal consistency was observed, this was statistically trivial and possibly without relevant effects on the standard error of measurement. This trivial effect was observed mainly for the coefficient ω, while the α coefficient showed a comparatively smaller reduction. It is possible that this difference interacted with one of the assumptions of coefficient α, which is tau equivalence and correlated errors [52], but correlated errors were not detected in the modeling of the dimensionality of the CDI-S in either group. Even with this reduction in internal consistency as measured by the α coefficient, the lack of statistical significance suggests that it may be considered sampling error. A complementary finding is that, in contrast to the Hispanic study by del Barrio et al. [34], here, a single latent dimension was endorsed to the CDI-S and had higher reliability; differences in reliability estimates obtained from two coefficients were also detected (α and ω), which usually represent noncompliance with the tau-equivalence model in the items [54].

The practical implications of the present results point to several potential consequences. First, the clinician using mass screening strategies now has evidence that respondent identification has a trivial effect on score variability. Second, and as a consequence of the above, the clinician can be confident that CDI-S results are possibly less influenced by subject anonymity or identification. Finally, another no less important implication is of an ethical nature; in this context, the clinician must pay attention to the safety of the tests applied and to the identification of respondents. Within an effective strategy to prevent unauthorized dissemination of the test applied and its results, identification of the examinee creates a more challenging situation than anonymous application.

The results should be interpreted in consideration of the specific limitations of the study. First, the sample size in each group limits its statistical power in each of the statistics applied, but even more so in the representativeness of the population to which it can be generalized. This limitation requires a replication study as a necessary condition to verify this effect of anonymity/identification in survey for schooled adolescents, an issue that apparently has not been addressed in previous studies [9]. On the other hand, a balance for this limitation on sample size is that the robust method used here [42] has proven to be effective in estimating the parameters of interest (the factor loadings) and their statistical significance in challenging situations such as small sample size and distributional skewness of the items [43,44,45]. Therefore, the problem of the accuracy of the estimates may have been partially solved. Second, the sample size also prevented further partitioning along of the main study variable (group A and group B), as it meant further reducing the samples compared; for example, the difference between males and females was not examined in interaction with the effect of anonymity, and the extent to which they affect response variability is not known.

With respect to sample size, previous studies have suggested that by applying multiple criteria (absolute number of cases according to expert opinion, the ratio number of cases—number of parameters or number of observed variables, and statistical power), the range of minimum sample sizes varies from 16 to 2760 cases [59]. Other more sophisticated methods also produce divergence (e.g., on the basis of statistical power; [60,61]). Methodological research has shown that aspects such as the size of factor loadings, communality, and the number of dimensions [62,63] are stable criteria. On the basis of opportunity and contextual constraints in the present study and the minimum sample size for estimating the parameters of interest (e.g., factor loadings and communalities; [62,63]), our sample size may be sufficient (approximately 200).

5. Conclusions

The present study shows support for a unidimensional internal structure of the CDI-S in Peruvian adolescents. When specific conditions were imposed on the selection groups (anonymity/identification of the participants), no significant differences were found at the level of internal structure, and therefore both models were acceptable. The CDI-S can be considered a unidimensional measure for use in the general adolescent population (as is the case in our study) since experiencing some condition of dysphoria and/or negative self-esteem does not seem to be differentiable if there is no additive exposure to some clinical condition (e.g., institutionalization, chronic noncommunicable diseases, terminal illnesses); similar findings have been obtained in the literature with other depression assessment instruments (e.g., PHQ-9). Measurement invariance was corroborated, which would imply that the possible impact of anonymity would be closely related to socially inappropriate behaviors. The reliability of the CDI-S scores for both groups was not compromised. This would imply that the measurement bias would not be directly related to an identification condition but rather to other factors already identified in the literature. Due to sample size limitations in the groups of interest, further research is required on other conditions of intergroup variability, such as some sociodemographic variables and mental and physical health conditions versus each condition of anonymity vs. participant name.

Acknowledgments

The authors thank Christian David Ruelas Vega for his support.

Author Contributions

Conceptualization, C.M.-S. and A.C.-L.; methodology, C.M.-S., F.T.-T., G.H.-S., L.A.N. and A.C.-L.; software, C.M.-S. and A.C.-L.; validation, J.H.R.-C. and A.C.-L.; formal analysis, C.M.-S. and A.C.-L.; investigation, F.T.-T., G.H.-S. and L.A.N.; resources, F.T.-T., G.H.-S. and J.H.R.-C.; data curation, C.M.-S. and A.C.-L.; writing—original draft preparation, C.M.-S., A.C.-L. and F.T.-T.; writing—review and editing, C.M.-S., A.C.-L., F.T.-T., J.H.R.-C. and M.Á.N.-B.; visualization, C.M.-S., G.H.-S. and M.Á.N.-B.; supervision, C.M.-S. and L.A.N.; project administration, F.T.-T. and C.M.-S.; funding acquisition, F.T.-T. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Commissions of Research, Ethics and Biosafety (Comisiones de Investigación, Ética y Bioseguridad), Hospital Infantil de México Federico Gómez National Institute of Health.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Conflicts of Interest

The authors declare no conflict of interest.

Funding Statement

This work is one of the results of the research project HIM/2015/017/SSA.1207, “Effects of mindfulness training on psychological distress and quality of life of the family caregiver”. Main researcher: Filiberto Toledano-Toledano, Ph.D. The present research was funded by federal funds for health research and was approved by the Commissions of Research, Ethics and Biosafety (Comisiones de Investigación, Ética y Bioseguridad), Hospital Infantil de México Federico Gómez National Institute of Health. The source of federal funds did not influence the study design, data collection, analysis, interpretation, or decisions regarding publication.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Perinelli E., Gremigni P. Use of social desirability scales in clinical psychology: A systematic review. J. Clin. Psychol. 2016;72:534–551. doi: 10.1002/jclp.22284. [DOI] [PubMed] [Google Scholar]
  • 2.van de Mortel T.F. Faking it: Social desirability response bias in self-report research. Aust. J. Adv. Nurs. 2008;25:40–48. [Google Scholar]
  • 3.Whelan T.J. Anonymity and confidentiality: Do survey respondents know the difference?; Proceedings of the 30th Annual Meeting of the Society of Southeastern Social Psychologists; Durham, NC, USA. 26–27 October 2007. [Google Scholar]
  • 4.Ash P., Abramson E. The effect of anonymity on attitude-questionnaire response. J. Abnorm. Soc. Psychol. 1952;47:722–723. doi: 10.1037/h0061890. [DOI] [PubMed] [Google Scholar]
  • 5.Kraus J. Effect of anonymity on response of adoptive parents to a child-problems questionnaire. Aust. Soc. Work. 1975;28:37–39. doi: 10.1080/03124077508549460. [DOI] [Google Scholar]
  • 6.Rosen N.A. Anonymity and Attitude Measurement. Public Opin. Q. 1960;24:675–679. doi: 10.1086/266983. [DOI] [Google Scholar]
  • 7.Lelkes Y., Krosnick J.A., Marx D.M., Judd C.M., Park B. Complete anonymity compromises the accuracy of self-reports. J. Exp. Soc. Psychol. 2012;48:1291–1299. doi: 10.1016/j.jesp.2012.07.002. [DOI] [Google Scholar]
  • 8.Beatty J.R., Chase S.K., Ondersma S.J. A randomized study of the effect of anonymity, quasi-anonymity, and certificates of confidentiality on postpartum women’s disclosure of sensitive information. Drug Alcohol Depend. 2014;134:280–284. doi: 10.1016/j.drugalcdep.2013.10.016. [DOI] [PubMed] [Google Scholar]
  • 9.Caballero R., Sen S., Nygård J.F. Anticipating anonymity in screening program databases. Int. J. Med. Inform. 2017;103:20–31. doi: 10.1016/j.ijmedinf.2017.04.003. [DOI] [PubMed] [Google Scholar]
  • 10.Durant L.E., Carey M.P., Schroder K.E. Effects of anonymity, gender, and erotophilia on the quality of data obtained from self-reports of socially sensitive behaviors. J. Behav. Med. 2002;25:438–467. doi: 10.1023/A:1020419023766. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.O’Malley P.M., Johnston L.D., Bachman J.G., Schulenberg J. A Comparison of confidential versus anonymous survey procedures: Effects on reporting of drug use and related attitudes and beliefs in a national study of students. J. Drug Issues. 2000;30:35–54. doi: 10.1177/002204260003000103. [DOI] [Google Scholar]
  • 12.Ong A.D., Weiss D.J. The impact of anonymity on responses to sensitive questions. J. Appl. Soc. Psychol. 2000;30:1691–1708. doi: 10.1111/j.1559-1816.2000.tb02462.x. [DOI] [Google Scholar]
  • 13.Meade A.W., Craig S.B. Identifying careless responses in survey data. Psychol. Methods. 2012;17:437–455. doi: 10.1037/a0028085. [DOI] [PubMed] [Google Scholar]
  • 14.Schneider S., May M., Stone A.A. Careless responding in internet-based quality of life assessments. Qual. Life Res. 2018;27:1077–1088. doi: 10.1007/s11136-017-1767-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Ward M.K., Meade A.W. Applying social psychology to prevent careless responding during online surveys. Appl. Psychol. 2018;67:231–263. doi: 10.1111/apps.12118. [DOI] [Google Scholar]
  • 16.Martínez I.M.M. Efectos del anonimato en la comunicación de grupos que utilizan tecnologías asistidas por ordenador. Un estudio cuantitativo y cualitativo. An. Psicol. Ann. Psychol. 2001;17:121–128. [Google Scholar]
  • 17.Matthey S., Petrovski P. The children’s depression inventory: Error in cutoff scores for screening purposes. Psychol. Assess. 2002;14:146–149. doi: 10.1037/1040-3590.14.2.146. [DOI] [PubMed] [Google Scholar]
  • 18.National Institute of Mental Health . Major Depression among Adolescents. National Institute of Mental Health; Bethesda, MD, USA: 2016. [Google Scholar]
  • 19.Bould H., Araya R., Pearson R.M., Stapinski L., Carnegie R., Joinson C. Association between early temperament and depression at 18 years. Depress. Anxiety. 2014;31:729–736. doi: 10.1002/da.22294. [DOI] [PubMed] [Google Scholar]
  • 20.Garber J., Brunwasser S.M., Zerr A.A., Schwartz K.T., Sova K., Weersing V.R. Treatment and prevention of depression and anxiety in youth: Test of cross-over effects. Depress. Anxiety. 2016;33:939–959. doi: 10.1002/da.22519. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Emons W.H., Sijtsma K., Meijer R.R. On the consistency of individual classification using short scales. Psychol. Methods. 2007;12:105–120. doi: 10.1037/1082-989X.12.1.105. [DOI] [PubMed] [Google Scholar]
  • 22.Kruyen P.M., Emons W.H.M., Sijtsma K. Test length and decision quality in personnel selection: When is short too short? Int. J. Test. 2012;12:321–344. doi: 10.1080/15305058.2011.643517. [DOI] [Google Scholar]
  • 23.Kruyen P.M., Emons W.H.M., Sijtsma K. On the shortcomings of shortened tests: A literature review. Int. J. Test. 2013;13:223–248. doi: 10.1080/15305058.2012.703734. [DOI] [Google Scholar]
  • 24.Kruyen P.M., Emons W.H.M., Sijtsma K. Assessing individual change using short tests and questionnaires. Appl. Psychol. Meas. 2014;38:201–216. doi: 10.1177/0146621613510061. [DOI] [Google Scholar]
  • 25.Ziegler M., Kemper C., Kruyen P. Short scales—five misunderstandings and ways to overcome them. J. Individ. Differ. 2014;35:185–189. doi: 10.1027/1614-0001/a000148. [DOI] [Google Scholar]
  • 26.Kovacs M. The children’s depression inventory (CDI) Psychopharmacol. Bull. 1985;21:995–998. [PubMed] [Google Scholar]
  • 27.Volpe R.J., DuPaul G.J. Assessment with brief behavior rating scales. In: Andrews J., Janzen H., Saklofske D., editors. Handbook of Psychoeducational Assessment: Ability, Achievement, and Behavior in Children. Academic Press; San Diego, CA, USA: 2001. pp. 357–387. [Google Scholar]
  • 28.Ailey S.H., Marks B., Heller T. Evaluation of two self-report depression measures for adults with Down syndrome. NADD Bull. 2003;5:71–75. [Google Scholar]
  • 29.de la Vega R., Racine M., Sánchez-Rodríguez E., Solé E., Castarlenas E., Jensen M.P., Engel J., Miró J. Psychometric properties of the short form of the Children’s Depression Inventory (CDI-S) in young people with physical disabilities. J. Psychosom. Res. 2016;90:57–61. doi: 10.1016/j.jpsychores.2016.09.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Davanzo P., Kerwin L., Nikore V., Esparza C., Forness S., Murrelle L. Spanish translation and reliability testing of the Child Depression Inventory. Child Psychiatry Hum. Dev. 2004;35:75–92. doi: 10.1023/B:CHUD.0000039321.56041.cd. [DOI] [PubMed] [Google Scholar]
  • 31.Holden R.R., Magruder C.D., Stein S.J., Sitarenios G., Sheldon S. The effects of anonymity on the holden psychological screening inventory. Pers. Individ. Differ. 1999;27:737–742. doi: 10.1016/S0191-8869(98)00274-8. [DOI] [Google Scholar]
  • 32.Sloan D.M., Marx B.P., Epstein E.M., Lexington J.M. Does altering the writing instructions influence outcome associated with written disclosure? Behav. Ther. 2007;38:155–168. doi: 10.1016/j.beth.2006.06.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Sitarenios G., Kovacs M. Use of the children’s depression inventory. In: Maruish M.E., editor. The Use of Psychological Testing for Treatment Planning and Outcomes Assessment. Lawrence Erlbaum Associates Publishers; Mahwah, NJ, USA: 1999. pp. 267–298. [Google Scholar]
  • 34.del Barrio V., Roa M.L., Olmedo M., Colodrón F. Primera adaptación del CDI- S en población española. Acción Psicol. 2002;3:263–272. doi: 10.5944/ap.1.3.558. [DOI] [Google Scholar]
  • 35.Sociedad Mexicana de Psicología . Código Ético del Psicólogo [Ethical Code of the Psychologist] Trillas; Ciudad de México, Mexico: 2010. [Google Scholar]
  • 36.American Psychological Association . Ethical Principles of Psychologists and Code of Conduct. With the 2016 Amendment to Standard 3.04. American Psychological Association; Washington, DC, USA: 2017. [Google Scholar]
  • 37.World Medical Association World Medical Association Declaration of Helsinki: Ethical principles for medical research involving human subjects. JAMA. 2013;310:2191–2194. doi: 10.1001/jama.2013.281053. [DOI] [PubMed] [Google Scholar]
  • 38.Kolmogorov A. Sulla determinazione empirica di una legge di distribuzione. Inst. Ital. Attuari, Giorn. 1933;4:83–91. doi: 10.12691/ajams-1-1-2. [DOI] [Google Scholar]
  • 39.Smirnov N.V. Estimate of deviation between empirical distributions functions in two independent samples. Bull. Mosc. Univ. 1939;2:3–16. [Google Scholar]
  • 40.Bradley E.L. Overlapping coefficient. In: Kotz E.S., Johnson N.L., Read C.B., editors. Encyclopedia of Statistical Sciences. Wiley; New York, NY, USA: 1985. pp. 546–547. [Google Scholar]
  • 41.Golsdtein R. The overlapping coefficient and an “improved” rank-sum statistic. Stata Techical Bull. 1994;22:12–15. [Google Scholar]
  • 42.Satorra A., Bentler P.M. Corrections to test statistics and standard errors in covariance structure analysis. In: von Eye A., Clogg C.C., editors. Latent Variables Analysis: Applications for Developmental Research. Sage; Thousand Oaks, CA, USA: 1994. pp. 399–419. [Google Scholar]
  • 43.Boomsma A. Reporting analyses of covariance structures. Struct. Equ. Model. 2000;7:461–483. doi: 10.1207/S15328007SEM0703_6. [DOI] [Google Scholar]
  • 44.Lei P.-W., Wu Q. Estimation in structural equation modeling. In: Hoyle R.H., editor. Handbook of Structural Equation Modeling. Guilford Press; New York, NY, USA: 2012. pp. 164–179. [Google Scholar]
  • 45.Tong X., Bentler P.M. Evaluation of a new mean scaled and moment adjusted test statistic for SEM. Struct. Equ. Modeling. 2013;20:148–156. doi: 10.1080/10705511.2013.742403. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Lorenzo-Seva U., ten Berge J.M.F. Tucker’s congruence coefficient as a meaningful index of factor similarity. Methodology. 2006;2:57–64. doi: 10.1027/1614-2241.2.2.57. [DOI] [Google Scholar]
  • 47.Elosua P., Wells C. Detecting DIF in polytomous items using MACS, IRT and ordinal logistic regression. Psicológica. 2013;34:327–342. [Google Scholar]
  • 48.Zumbo B.D. A Handbook on the Theory and Methods of Differential Item Functioning (DIF): Logistic Regression Modeling as a Unitary Framework for Binary and Likert-Type (Ordinal) Item Scores. Directorate of Human Resources Research and Evaluation, Department of National Defense, Sociedad Mexicana de Psicología; Ottawa, ON, Canada: 1999. [Google Scholar]
  • 49.Crane P.K., Gibbons L.E., Jolley L., van Belle G. Differential item functioning analysis with ordinal logistic regression techniques. DIFdetect and difwithpar. Med. Care. 2006;44:S115–S123. doi: 10.1097/01.mlr.0000245183.28384.ed. [DOI] [PubMed] [Google Scholar]
  • 50.Crane P.K., Gibbons L.E., Narasimhalu K., Lai J.S., Cella D. Rapid detection of differential item functioning in assessments of health-related quality of life: The functional assessment of cancer therapy. Qual. Life Res. 2007;16:101–114. doi: 10.1007/s11136-006-0035-7. [DOI] [PubMed] [Google Scholar]
  • 51.Crane P.K., van Belle G., Larson E.B. Test bias in a cognitive test: Differential item functioning in the CASI. Stat. Med. 2004;23:241–256. doi: 10.1002/sim.1713. [DOI] [PubMed] [Google Scholar]
  • 52.Cronbach L.J. Coefficient alpha and the internal structure of tests. Psychometrika. 1951;16:297–334. doi: 10.1007/BF02310555. [DOI] [Google Scholar]
  • 53.McDonald R.P. Test Theory: A Unified Treatment. Lawrence Erlbaum Associates Publishers; Hillsdale, NJ, USA: 1999. [Google Scholar]
  • 54.Merino-Soto C., Dominguez-Lara S. Respuesta a carta al editor. Diferenciando la espada de la mano. Rev. Latinoam. Cienc. Soc. Niñez Juv. 2017;15:629–631. [Google Scholar]
  • 55.Weber R., Popova L. Testing equivalence in communication research: Theory and Application. Commun. Methods Meas. 2012;6:190–213. doi: 10.1080/19312458.2012.703834. [DOI] [Google Scholar]
  • 56.Coe R., Soto C.M. Magnitud del efecto: Una guía para investigadores y usuarios. Rev. Psicol. 2003;21:147–177. doi: 10.18800/psico.200301.006. [DOI] [Google Scholar]
  • 57.Levene H. Robust tests for equality of variances. In: Olkin I., Ghurye S.G., Hoeffding W., Madow W.G., Mann H.B., editors. Contributions to Probability and Statistics: Essays in Honor of Harold Hotelling. Stanford University Press; Menlo Park, CA, USA: 1960. pp. 278–292. [Google Scholar]
  • 58.Merino-Soto C. Diferencias entre coeficientes alfa de Cronbach, con muestras y partes pequeñas: Un programa VB. An. Psicol. 2016;32:587–588. doi: 10.6018/analesps.32.2.203841. [DOI] [Google Scholar]
  • 59.Vargas-Halabi T., Mora-Esquivel R. Sample sizes using structural equation modeling with latent variables: A practical method. Rev. Actual. Investig. Educ. 2017;17:25–60. doi: 10.15517/aie.v17i1.27294. [DOI] [Google Scholar]
  • 60.MacCallum R.C., Browne M.W., Sugawara H.M. Power analysis and determination of sample size for covariance structure modeling. Psychol. Methods. 1996;1:130–149. doi: 10.1037/1082-989X.1.2.130. [DOI] [Google Scholar]
  • 61.MacCallum R., Lee T., Browne M.W. The issue of isopower in power analysis for tests of structural equation models. Struct. Equ. Model. 2010;17:23–41. doi: 10.1080/10705510903438906. [DOI] [Google Scholar]
  • 62.Mundfrom D.J., Shaw D.G., Ke T.L. Minimum sample size recommendations for conducting factor analyses. Int. J. Test. 2005;5:159–168. doi: 10.1207/s15327574ijt0502_4. [DOI] [Google Scholar]
  • 63.Wolf E.J., Harrington K.M., Clark S.L., Miller M.W. Sample size requirements for structural equation models: An evaluation of power, bias, and solution propriety. Educ. Psychol. Meas. 2013;76:913–934. doi: 10.1177/0013164413495237. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.


Articles from Children are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES