Skip to main content
Public Opinion Quarterly logoLink to Public Opinion Quarterly
. 2024 Mar 28;88(1):123–148. doi: 10.1093/poq/nfad060

Using Machine Translation and Post-Editing in the TRAPD Approach: Effects on the Quality of Translated Survey Texts

Diana Zavala-Rojas 1,, Dorothée Behr 2, Brita Dorer 3, Danielly Sorato 4, Veronika Keck 5
PMCID: PMC11014682  PMID: 38617051

Abstract

A highly controlled experimental setting using a sample of questions from the European Social Survey (ESS) and European Values Study (EVS) was used to test the effects of integrating machine translation and post-editing into the Translation, Review, Adjudication, Pretesting, and Documentation (TRAPD) approach in survey translation. Four experiments were conducted in total, two concerning the language pair English-German and two in the language pair English-Russian. The overall results of this study are positive for integrating machine translation and post-editing into the TRAPD process, when translating survey questionnaires. The experiments show evidence that in German and Russian languages and for a sample of ESS and EVS survey questions, the effect of integrating machine translation and post-editing on the quality of the review outputs—with quality understood as texts output with the fewest errors possible—can hardly be distinguished from the quality that derives from the setting with human translations only.

Introduction

Given its large impact on data quality in cross-national studies (Harkness et al. 2010), survey translation, that is, the translation and translation assessment of survey questionnaires, has consolidated as an important area of comparative survey methodology. The Translation, Review, Adjudication, Pretesting, and Documentation (TRAPD) approach to survey translation serves as a methodological gold standard (Harkness 2003), and variants of it are used to translate questionnaires in major multilingual projects such as the European Social Survey (European Social Survey 2024) (ESS), the Eurofound surveys (Eurofound 2017), and the European Values Study (Przepiórkowska and Behr 2017) (EVS). TRAPD has also been applied to the translation of questionnaires in medical and health research (Forsyth et al. 2007) or in market research (Kietzmann et al. 2016; Sha and Lai 2016).

Following the TRAPD method, at the “translation” stage (T in the acronym) two translators produce independent and parallel translations of the source questionnaire into a target language; or the source questionnaire is split among the translators. At least one of the translators is recommended to be a trained and/or professional translator or a linguist, while the second translator may be a social scientist. In a review meeting (R), the translations are discussed by the translators together with a reviewer; at the adjudication stage (A), an adjudicator is responsible for the final decisions on different translation options. Oftentimes, the roles of reviewer and adjudicator, both typically having a background in social sciences, are merged. The translated questionnaire is pretested before fieldwork (P) and the whole process is documented (D), including information on difficult translations, needed deviations, or remaining challenges in the translation.

Team members should combine survey knowledge, translation expertise, knowledge of the culture where the questionnaire will be administered, and knowledge related to the topic of the survey. While parallel translations offer variants to compare, the team-based discussion is at the heart of TRAPD, revisiting the decision-making process for the selection of wording as well as fostering interdisciplinary collaboration between professional translators and social scientists; after all, versions can be discussed taking into account different viewpoints by the different experts (Harkness 2003).

Until recently, the use of machine translation (MT) for survey translation has been discouraged in best-practice guidelines for cross-national survey methodology (Mohler et al. 2016b), so the T step was based on all human translations. Given the importance of survey translation quality in comparative studies, highly defective machine translations—as was still common some years ago—were suggesting neither a higher translation quality nor a more efficient translation process. However, the analysis of natural language by computational means has rapidly evolved in the last decade. And with it, MT quality has substantially improved after the development of artificial neural-network-based engines, neural MT (Way 2020; Nitzke and Hansen-Schirra 2021). Moreover, the availability of online machine translation tools, such as Google Translate and DeepL Translate, has allowed it to become a broadly used internet-based service. Given this availability, survey projects teams may perceive that MT has become a suitable method for questionnaire translation. From the angle of survey methodology, testing for the impact of new procedures before adopting them is fundamental to maintain data quality, as all potential sources of error should be considered, as well as their potential interactions (Smith 2011).

This experimental study integrates MT into the TRAPD approach and tests for its potential effects on translated survey questions. MT will be followed up by post-editing (PE), which is understood as the revision of raw machine translation output (more on this in the Methods section). We implemented a highly controlled experimental TRAPD approach to prevent process-related effects from confounding the comparison of the groups using MT against those using only human translation. In the TRAPD implementation reported here, the review and adjudication steps were merged. For this merged step, we will use the term review/er throughout the text. A difference in this study with respect to the theoretical approach of TRAPD is that we did not pretest our translations.

The main research question in this study is: Would replacing one of the initial human translations at the T step through MT and PE impact on the quality of the review output? If so, are the effects negative or positive for the outputs’ quality? The review output is the translation version resulting from the review discussion. Harkness and other proponents of the TRAPD model argued that translation quality is ensured in the review meeting in which the team discusses the translation options (Harkness, Pennell, and Schoua-Glusberg 2004; Harkness, Villar, and Edwards 2010; Mohler et al. 2016a); therefore, our experiments focus on assessing the effects of MT and PE on the output texts after these review meetings. Our secondary research questions explore group dependencies of the effects of MT and PE: Are effects conditional on the use of full or light PE? And, are Russian and German translations affected differently by MT and PE? These and the stakes of this research will be explained in detail in the next sections. Previewing the outcome, the overall results for integrating MT and PE into TRAPD are positive. Only in the German light-PE treatment group do we see a slightly higher error level compared to the control group, which may be negligible, though.

To the best of our knowledge, there have been no recommendations for the use of MT and PE in the field of academic social sciences surveys. At the same time, we acknowledge that neural MT is slowly entering the field of measurement instruments and is undergoing various tests of applicability also by other researchers (Iwai et al. 2019; Mondal, Mondal, and Mondal 2019).

Machine Translation and Post-Editing

MT is a highly interdisciplinary scientific area, bringing together, among others, linguists, computer scientists, and translation scholars. Until recently, the language-service-provider community’s acceptance of MT was low; acceptance has become more widespread since the emergence of neural MT (Bahdanau, Cho, and Bengio 2015; Moorkens et al. 2018).

Findings from the European Language Industry Survey (2020) (ELIA et al. 2020) show that MT is the main technology trend for the future. Using MT is deemed to speed up the translation process, thus increasing translation volume as well as turnover (Koponen 2016). Texts obtained by MT are considered pretranslations because of subsequent revisions typically needed by a human (Sin-Wai 2017).

In translation practice, machine translation is commonly implemented in combination with post-editing. We distinguish between full and light PE. Full PE covers the production of accurate, comprehensible, and linguistically correct output that is similar or equal to human translation quality. Light PE, being a time-saving procedure, implies producing an output that is accurate and comprehensible, but not necessarily stylistically or grammatically adequate (Massardo et al. 2016). Full PE can be qualitatively as good as a human translation or even better (Daems et al. 2017), even though it can also show deficiencies compared to a human translation (Moorkens et al. 2018; Toral 2019). But would integrating a suboptimal light-PE version into a team review discussion be sufficient to arrive at a good review output? Given these two types of PE, this study tested whether their differentiation would make a difference on the quality of the review output.

MT is not suitable (at the time of writing) for all text types. The usability of MT heavily depends on the text corpora1 that are used for training MT engines (Nitzke and Hansen-Schirra 2021). Online MT tools have not been specifically trained for questionnaires or survey speak. Questionnaires are wording sensitive, and small wording differences may make a difference for respondents’ understanding and survey responding. Thus, we wanted to explore whether MT combined with different forms of PE can be used for the text-type survey questionnaires. The fact that literary translation, another wording-sensitive translation area, is exploring MT and PE (Moorkens et al. 2018) shows that the time is ripe to test MT and PE for survey translation.

Besides text type, the language pair used in machine translation is known to affect translation quality. Research has looked into MT quality for different language combinations (Doherty and O’Brien 2014; Castilho et al. 2017; Popović 2018), as the more distant languages are from each other, the more difficult they tend to be for MT. Would using different language combinations, in our study English-German and English-Russian, lead to different translation quality in the different experimental groups?

Translation Quality and Translation Quality Assessment

Translation quality (TQ) can be approached both from the angle of the process and from the angle of the product (Gouadec 2010). Focusing in the following on the latter one, there is agreement that translation quality assessment is marked by complexity. There is almost never only one correct translation, and what counts as good, acceptable, or unacceptable depends on various factors, such as the text type, the intended purpose of a text, the target population, and further project-related specifications.

For assessing TQ, there are different approaches. Some test the texts on the target population that is intended to understand such texts. In this research, we evaluate the quality of the texts themselves. For the approaches analyzing the texts directly, one can differentiate between holistic and analytic approaches to TQ: holistic approaches evaluate the text as a whole, whereas analytic approaches allow for identifying specific translation problems and assessing the translation in its details (Lommel 2018, p. 122). For analytic quality assessments, error typologies are used (Castilho et al. 2018). These typologies include errors pertaining to domains such as adequacy and fluency, and allow for the possibility to weigh errors according to a severity level. Such models are typically applied by humans; this process is not error-free since subjective judgement can differ, for instance, as to whether an error counts as a lexical or a terminological error or how severe it is.

Clear guidelines, training, and regular practice with an error typology are practices to reduce the impact of subjectivity on error assignment (Saldanha and O’Brien 2014). Error coding schemes are often applied by two humans, enabling a systematic and comparative evaluation. Latest error typologies, such as the Multidimensional Quality Metrics (MQM), are deliberately comprehensive, applicable to both human and machine translation, and can be tailored to an actual project by selecting relevant subsets of errors (Castilho et al. 2018). MQM also represents and integrates most of the other commonly used translation error taxonomies in the industry (Vardaro, Schaeffer, and Hansen-Schirra 2019, p. 7). As such, MQM will be our method of choice for error coding. Alternative approaches include adequacy or fluency ratings as well as ranking of translations (Castilho et al. 2018), but these would not have allowed in-depth investigation of concrete translation problems occurring in the wording of the review outputs.

A further differentiation of translation quality is whether it is evaluated manually, that is, by humans, or by using algorithms. Algorithms produce metrics that compare source and (human- or machine-) translated target texts, such as the Bilingual Evaluation Understudy, or they run automated checks on specific matters, such as grammar, or other error categories (Castilho et al. 2018).

However, in general, while automated TQ checks are faster, cheaper, and more objective, many of them require a reference translation to evaluate the quality of the human and MT outputs. In addition, instead of having an evaluation of a whole sentence, as is common when using algorithms, our focus was also to assess the translations for errors in individual words, compound words, and phrases. Hence, in our study, we will not be using algorithms to assess TQ.

Moving from error coding as used in translation research and industry to cross-cultural survey methodology, TQ for survey translations is understood as retaining functional equivalence to the source, to the format of a source question as well as to its measurement properties; moreover, it requires adhering to the linguistic needs of the target language, and overall maintaining the same stimulus as the source (Harkness, Villar, and Edwards 2010). Equivalence is assessed by testing the translated texts among the target group, for instance, by cognitive pretests or pilot surveys (Harkness, Pennell, and Schoua-Glusberg 2004).2 However, the focus of this study was to evaluate the texts directly, their linguistic content. Thus, we analyze TQ and compare the text outputs in the experiments based on an error typology. We assume that there is an inverse relationship between errors and quality, in which texts presenting fewer errors are of higher quality, and on this basis we compare the errors between the different translated texts, in the concrete case, the review outputs.

Method

Design of the Experiments

We designed and conducted four experiments to test for the impact of MT and PE in the review outputs produced in the TRAPD method. Figures 1–3 illustrate the design of the experiments.

Figure 1.

Figure 1.

Summary of the control group.

Figure 2.

Figure 2.

Summary of the treatment using full post-editing.

Figure 3.

Figure 3.

Summary of the treatment using light post-editing.

The experiments are characterized by Equations 1 to 3. A language-specific control group, G0k, implements the translation step of the TRAPD without using machine translation at all. The initial translations are discussed in a team meeting at the review step where decisions are made about translation options and where the final translation, the review output, is produced. The process is documented by commenting on particular decisions made. The focus of our analysis is the review output, not the participants, because we conducted linguistic analysis directly on the texts to compare the experimental groups. The language pair used in the experiments is indicated by k =1, 2, where 1 indicates English-German and 2 English-Russian. The team in the control group is built up by two human translators, HT10 and HT20, and a reviewer, R10. This control group simulates a generic form of the TRAPD process implemented by survey projects at present. The texts produced in the control group after the review meeting are compared against those produced by two treatment groups, G1k and G2k. In the first treatment, G1k, at the T step, one translation was conducted by a human translator denoted by HT31, and the second translation was obtained when a post-editor, PE11, who received a machine-translated output, conducted full PE. The human and the post-edited translations are discussed in the team review meeting, and the process is documented in the same way as in the control group. The reviewer is denoted as R21. In a second treatment group, G2k, the first translation was also human-produced, denoted as HT42, and the second input translation for the review meeting was obtained by a post-editor using light PE, denoted as PE22. As in the other groups, there is a reviewer, denoted by R32, the translations were discussed in a team review meeting, and the process was documented. Each participant was only assigned to one group and one role.3 Control and treatment groups shared all other features except for the interventions defined here. The subscript j =0, 1, 2, indicates that the participant was part of the control group, the first treatment using full PE or the second treatment using light PE, respectively. In the review session, each group discussed and finalized a set of yk translated segments. These segments correspond to translation units, such as a sentence or a response option, typically delimited by a period, a question mark, a semicolon, or a line break. Each segment is denoted by yijk, with i =1, …, 268, indicating one out of 268 segments of 40 survey questions sampled from the ESS and the EVS questionnaires. In total among the two languages, we analyzed 1,608 text segments. As the focus is to compare the quality of the translations in the control groups and in the treatments, it is the text segments and not the participants which constitute our unit of analysis.

G0k = {HT10, HT20, R10}  yk = y10k, , yi0k  i = 1, , 268, j = 0, k = 1, 2 (1)
G1k = {HT31, PE11, R21}  yk = y11k, , yi1k i = 1, , 268, j = 1, k = 1, 2 (2)
G2k = {HT42, PE22, R32}  yk = y12k, , yi2k i = 1, , 268, j = 2, k = 1, 2 (3)

Participants

Participants had fixed role-background combinations. Human translators HT10, HT31, and HT42 were professional translators, with previous survey translation experience. Human translators HT20, post-editors PE11 and PE22, as well as reviewers R10, R21, and R32 were social scientists, with work experience in the social sciences and with experience in questionnaire design and translation. Participants had native speaker competence of either German or Russian, k. The combination of having both professional translators and social scientists collaborating in interdisciplinary teams mirrors how the TRAPD model is recommended to be set up in survey projects. We used snowball sampling, translator unions, as well as survey projects to recruit potential participants. Based on answers from a recruitment questionnaire, we recruited six professional translators and twelve social scientists.

We matched backgrounds and experiences to have a similar composition of teams in the control groups and in the treatments in terms of backgrounds and skills. The participants were paid for their task. Participants were informed that “the study aim is to integrate machine translation into team-based questionnaire translation procedures and to evaluate the overall process”; they were neither informed of nor requested further details of the study.4Supplementary Material section A provides more detailed information about the participants in the experiments. All participants, according to their roles, received virtual training and written materials on their task. Training covered, for instance, information on the implemented TRAPD model, a translation brief specifying the translation objectives (including the target group and the survey mode), do’s and don’ts in questionnaire translation, and information on the source questionnaire.

Choice of Languages

The source questions were in the English language. German and Russian were chosen as target languages because they are used in several countries in large-scale cross-national survey projects. Both languages stand for different language families, the Proto-Germanic and the Slavic language family, respectively. As is typically the case in translation studies and linguistics, the native tongues of the research team also played a role in the language choice. By choosing German and Russian, it was possible to analyze the data ourselves, understand the contents of the review sessions, and communicate with participants and coders of translation errors.

Instrument: Survey Questionnaire

The selection of the survey questions was done by a combination of random sampling and item selection based on criteria of known translation problems in human questionnaire translation and/or machine translation. Known translation problems included, for instance, challenging terminology; and machine translation issues included, for instance, gender issues. Supplementary Material section B provides details on the sampling of survey items, and about the criteria for the selection of items. This twofold approach ensured randomness in the selection of questions to be translated and coverage of key characteristics and potential translation challenges of a survey questionnaire.

The ESS sampling frame included questionnaires of Round 1 to Round 9 and repeated questions that were administered every round only once, adding to 1,454 questions. From the EVS, the sampling frame included Wave 1 to Wave 5, 1,745 questions in total. We sampled 262 items stratified by wave/round and study. Starting from this random sample, a final set of 40 questions, which constituted the English source of 268 segments, was selected by, first, evaluating each of the questions in the sample against the criteria and, second, ranking questions in terms of importance. A few modifications were introduced to the original source items to create a lab questionnaire. For instance, Don’t know and Refusal categories were harmonized across ESS and EVS items. Supplementary Material section C contains the source questionnaire in English, and the Supplementary Material to this article shows the source and translated questions per language and group.5

Data Collection

Training of participants and data collection took place from September 15 to October 23, 2020. Participants filled in background questionnaires before, during, and after the experiments. As the experiments took place in 2020 and in-person meetings were discouraged or not legally allowed worldwide, the team-based training and review sessions were conducted online using Zoom.

Translation and Documentation Environment

Depending on the role of each participant in the study, individual translation projects with and without MT features, respectively, were set up in the computer-aided translation tool MateCat (Federico et al. 2014). MateCat is open-access; it includes a neural machine translation application and PE features. The team meetings made use of the MateCat environment and spreadsheets with the translations and translator comments, as shown in figure 4.

Figure 4.

Figure 4.

Documentation environment at the team meeting.

Analytical Approach

Error Scheme

The error scheme used in this study is based on the harmonized MQM-DQF translation quality metrics, which combines the MQM framework6 and the Dynamic Quality Framework (DQF).7 While the MQM was set up as a comprehensive and detailed framework, drawing on many different translation metrics, DQF was based on industry best practices and focused on the issues commonly checked by language service providers (Lommel et al. 2015; Lommel 2018). We took a subset of the DQF-MQM and adapted it to the text type of survey questionnaires by selecting, omitting, or adding new subcategories.

The original four severity levels, running from critical to neutral, were adapted to three levels (major, minor, and neutral), and definitions were tailored to the survey context.

  • Major errors: The translation completely changes the meaning, likely misleads the respondent, or provides incorrect, missing, and/or contradictory information.

  • Minor errors: They may affect the respondent’s comprehension of translated text and increase the time required to read and to understand the translation.

  • Neutral errors: They include errors that might make the translation a bit harder to understand, but ultimately do not stop the respondent from overall understanding and using the translation in terms of the measurement goal.

The final error scheme has seven categories: accuracy, fluency, survey-specific terminology/phrases and features, style, locale convention, verity, and other. Each of these categories is subdivided into subcategories that allow a fine-tuned classification of translation errors in the texts. Supplementary Material section D summarizes the error scheme definition. Every single instance of an error was coded, including repetitive errors. The comments (that is, the documentation) made by the teams in the review step were not considered during error coding to ensure an independent evaluation. The way the severity levels were defined included assessing a potential impact of erroneous wordings on respondents in a face-to-face interview. Over- or underrating the impact when choosing severity levels cannot be ruled out. However, since we applied a consistent error coding approach across all review outputs and since coders were not knowledgeable of which review version they were coding, no translation should be put at a disadvantage through the assessment.

Translation Quality Assessment: Error Coding Process

Besides coding errors and severity levels, coders had to apply a specific error coding syntax, linking source text wording to translation errors in the target. For the error coding to achieve high quality, a harmonization approach was applied to reduce subjective or idiosyncratic language understanding (Kuckartz 2014). The setup of the approach drew on experiences from other studies (Daems, De Clercq, and Macken 2017; Koponen and Salmi 2017). Coders were provided with training for their tasks. In each language, two independent coders, not involved in the experiments, coded each of the segments included in the set, y, for translation errors, zm(yij)k. They did not know which group had produced the translation they were coding. They subsequently met with a referee in a harmonization meeting to discuss cases of diverging coding; the referee only got involved when coders could not agree on a final error coding and needed a third person for judgment. A harmonization process was preferred, as there is evidence suggesting that interrater reliability is not an appropriate method in translation quality assessment (Burchardt and Lommel 2014; Jia, Carl, and Wang 2019).8 During the harmonization meetings, special emphasis was placed on ensuring consistent error coding across review outputs in a language, but there was not one reference translation that would serve as a “gold standard.”

The coders were native speakers of the respective target languages, trained translators or translation practitioners, and in three out of four cases highly familiar with survey translation. The lack of familiarity with survey translation of one of the Russian coders was offset by additional training before the task and by learning-by-doing through the large number of harmonization sessions. Error coding was thus done by translation experts who were able to differentiate between necessary deviations in a translation and errors. Errors were only compared across groups within a language. We did not conduct across-language comparisons, being aware that different language pairs may have different propensities for errors.

Statistical Analysis

The main dependent variable is defined as the count of the errors, zm(yij)k, where m =1, 2, …, M is a consecutive natural number counting the errors, and k and yij linking the error to a specific language, segment, and experimental group. The final codes after the harmonization meetings were automatically retrieved from the coding environments and the groups compared using several statistical techniques described below. Using several statistical techniques allowed us to assess the consistency of our results, and if discrepancies would be present (which is not the case), to report inconclusive results. A z-score test statistic for H0: p1 − p2 = 0 was defined as:

z-score = p1¯p2¯0p¯1p¯1n1+1n2 (4)

where p1¯ represents the ratio of the errors in the control group, zm(yi0)k, divided by the total number of segments, n1 = n2 = 268; is the ratio of the errors in the treatment group, zm(yij)k, divided by n2. Finally, p¯ is the total proportion of errors, calculated as the sum of the errors of the control and treatment groups divided by n1 + n2.

And, a Poisson regression is defined as

zmyijkPoissonλm (5)
log(λ)=β0+β1x1 (6)

for m =1, …, M where the expected count of zm is E(Z) = λ and where x1 is a categorical predictor with three values describing the group from which the counts of errors are estimated, among control group, full PE, and light PE. As both treatments share the same control group, multiple comparisons using z-score tests carry the risk of increasing Error Type I. Therefore, Poisson regression for the counts of errors was also estimated. For both the z-score tests and the Poisson regression, we evaluated the differences in errors for the unweighted data and the data weighted by error severity levels, w = ws, for severity levels, s =1, 1.5, 2, representing neutral, minor, and major error, respectively. Additionally, exploratory analysis included the use of chi-square tests defined as:

χ2=(Om - Em)2Em (7)

where O is the observed number of errors and E is the expected number of errors in each error category.

Results

Descriptive Analysis

Table 1 presents the total number of errors (zk) per treatment group, the number of errors weighted by severity levels, and the mean severity level for each group. Neutral errors were assigned a weight of 1, thus they count as one error. An error considered minor was assigned a weight of 1.5, and a major error was assigned a weight of 2. The mean of the severity levels is also depicted in the table; a lower number indicates more neutral errors. As the majority of errors were of neutral severity, weighting them does not change the relative amount of errors in the groups. Although we present our tables with both languages, comparisons are strictly done within a language, that is, we only compare control and treatments within a language.

Table 1.

Translation errors aggregated by experimental group.

Language TRAPD set-up Group Number of errors (zm(yij)k) Mean error severity Errors weighted by severity
German Control group G01 36 1.31 47
German Full PE G11 43 1.23 53
German Light PE G21 79 1.24 98
Russian Control group G02 44 1.5 66
Russian Full PE G12 37 1.45 53.5
Russian Light PE G22 41 1.37 56

Table 2 depicts the number of errors per error type. Treatment groups in the same language are similar to each other, except for the group using light PE in German. A plurality of errors coded for this group were errors of accuracy and they were neutral, having a severity mean of 1.17.

Table 2.

Translation errors aggregated by category and experimental group.

Language TRAPD set-up Group Category Number of errors (zm(yij)k) Mean error severity Errors weighted by severity
German Control group G01 Accuracy 14 1.43 20.0
G01 Fluency 9 1.00 9.0
G01 Style 12 1.42 17.0
G01 Survey specific 1 1.00 1.0
German Full PE G11 Accuracy 18 1.44 26.0
G11 Fluency 5 1.00 5.0
G11 Style 5 1.20 6.0
G11 Survey specific 15 1.07 16.0
German Light PE G21 Accuracy 39 1.17 45.5
G21 Fluency 9 1.17 10.5
G21 Style 16 1.41 22.5
G21 Survey specific 15 1.30 19.5
Russian Control group G02 Accuracy 20 1.75 35.0
G02 Fluency 7 1.29 9.0
G02 Style 13 1.19 15.5
G02 Survey specific 4 1.62 6.5
Russian Full PE G12 Accuracy 13 1.54 20.0
G12 Fluency 3 1.33 4.0
G12 Style 10 1.05 10.5
G12 Survey specific 11 1.73 19.0
Russian Light PE G22 Accuracy 12 1.54 18.5
G22 Fluency 12 1.25 15.0
G22 Style 8 1.19 9.5
G22 Survey specific 9 1.44 13.0

A few examples shall help to better explain the results:

G01:

  • Source: To be a good citizen, how important would you say it is for a person to…

  • Target: Um ein gutter Bürger zu sein, wie wichtig ist es Ihrer Meinung nach, dass eine Person…

  • Error: Over-translation—severity level neutral: It would have been more appropriate to add the female term for citizen as well, currently only the male form (“Bürger”) is used.

G22:

  • Source: Not at all

  • Target: Нисколько

  • Error: Scales Inconsistency—severity level minor: The word “Нисколько” is difficult to interpret out of context. The option “Вообще не выполняете” (Never do it) would be better and also fit to the question text.

G21:

  • Source: Now suppose two people from different race or ethnic groups each appear in court, charged with an identical crime they did not commit.

  • Target: Nehmen wir an, zwei Menschen unterschiedlicher ethnischer Herkunft erscheinen vor Gericht und werden einer gleichen Straftat angeklagt, die sie nicht begangen haben.

  • Error: Register—severity level major: “Ethnisch” on its own is difficult to understand for certain groups in society.

A χ2 test of independence was performed to examine the error categories of accuracy, fluency, style, and survey-specific errors, and the groups per language. Tables 3 and 4 summarize the χ2 test. The relation between the error categories and experimental groups was significant in the experiments using the German language, χ2 (6, N =158) = 19, p =0.004. These results are driven mainly by fewer errors than expected in the control group and a larger number of observed errors in the full-PE group.

Table 3.

Chi-square test: observed and expected counts, English-German.

Error category G01 G11 G21 Case
Accuracy 14.00 18.00 39.00 Observed
Accuracy 16.18 19.32 35.50 Expected
Fluency 9.00 5.00 9.00 Observed
Fluency 5.24 6.26 11.50 Expected
Style 12.00 5.00 16.00 Observed
Style 7.52 8.98 16.50 Expected
Survey specific 1.00 15.00 15.00 Observed
Survey specific 7.06 8.44 15.50 Expected

Note: χ2 test in German language, χ2(6, N =158) = 19, p =0.004.

Table 4.

Chi-square test: observed and expected counts, English-Russian.

Error category G02 G12 G22 Case
Accuracy 20.00 13.00 12.00 Observed
Accuracy 16.23 13.65 15.12 Expected
Fluency 7.00 3.00 12.00 Observed
Fluency 7.93 6.67 7.39 Expected
Style 13.00 10.00 8.00 Observed
Style 11.18 9.40 10.42 Expected
Survey specific 4.00 11.00 9.00 Observed
Survey specific 8.66 7.28 8.07 Expected

Note: χ2 test in Russian language χ2(6, N =122) = 11.96, p =0.062.

The χ2 test for the experiments using the Russian language is not significant at a p < 0.05; nevertheless, we interpret the results as the p-value is just above this threshold χ2 (6, N =122) = 11.96, p =0.062. These results are driven mainly by a difference in the expected and observed survey-specific errors in the control group, and in the expected and observed fluency errors in the light-PE group. This translates into a very low number of observed survey-specific errors in the control group and a larger number of observed fluency errors in the light-PE group. However, in the Russian language experiment, these differences are not significant enough to have an effect in the z-score test or in the regression coefficients that compare the control group against the treatments.

Main Results

We compared the control group against the treatments in two ways. First, we used z-score tests for two population proportions. This test assesses whether two groups differ significantly on some single characteristic, in this case the number of errors. A first test compared the proportion of errors in the text segments by the control group with the group using full PE. A second test compared the proportion of errors in the control group against the group using light PE. The second strategy to compare the groups was to run a Poisson regression for the counts of errors. Both z-score tests and Poisson regressions show similar and consistent results.

Z-Score Test

Table 5 summarizes the results of the z-score tests with findings described by language below. Overall, experiments in German and Russian show positive results toward integrating the use of machine translation and full PE into TRAPD, when the review output is considered. Using MT and full PE yields equivalent results to an all-human translation setting. When integrating the use of machine translation and light PE into TRAPD, results are positive in the Russian language but negative in the German language, where the review output had a larger proportion of errors. The mean severity of the errors indicated in table 1 for this group is 1.17; it indicates that this method resulted in many neutral errors.

Table 5.

Z-score test results.

Description Group p 1 p 2 z-score p.value CI low CI high Errors weighted by severity
German, full PE G11 0.13 0.16 0.53 0.46 −0.09 0.04 No
G11 0.18 0.20 0.31 0.58 −0.09 0.05 Yes
German, light PE G21 0.13 0.29 19.53 0.00 −0.23 −0.09 No
G21 0.18 0.37 23.64 0.00 −0.27 −0.11 Yes
Russian, full PE G12 0.16 0.14 0.52 0.47 −0.04 0.09 No
G12 0.25 0.20 1.42 0.23 −0.03 0.12 Yes
Russian, light PE G22 0.16 0.15 0.06 0.81 −0.05 0.08 No
G22 0.25 0.21 0.86 0.35 −0.04 0.11 Yes

Note: Method: two-sample test for equality of proportions with continuity correction, two-tail tests.

Confidence Interval (CI), comparisons against control groups, G01 and G02.

Experiments in the German Language

The errors in the control group using only human translation are not significantly different to the errors in the group using machine translation and full PE. This result does not change when the errors are weighted by severity level. This implies that with respect to the errors in the review output, both translation outputs are of the same quality. In the case of the group using machine translation and light PE, there is a significant difference in the proportion of errors in comparison with the control group, and this difference is maintained when the errors are weighted by severity. The translation has more errors when machine translation and light PE were used.

Experiments in the Russian Language

In the case of the experiments in Russian, the difference in errors in the group using only human translation are not statistically significant in comparison with the group using machine translation and full or light PE.

Regression Analysis

Table 6 summarizes results of the Poisson regression models. Overall, running regression models confirms the results of the z-score tests; there is no significant impact of the group in the count of the errors in the Russian language, whereas in the German language, there is an effect of the group using machine translation and light PE.

Table 6.

Results of the Poisson regression models.

Description Term Estimate Standard error z-value p. value CI low CI high Errors weighted by severity
Intercept 3.58 0.17 21.50 0.00 3.24 3.89 No
German, full PE G11 0.18 0.23 0.79 0.43 −0.26 0.63 No
German, light PE G21 0.79 0.20 3.91 0.00 0.40 1.19 No
Intercept 3.85 0.15 26.40 0.00 3.55 4.12 Yes
German, full PE G11 0.12 0.20 0.60 0.55 −0.27 0.52 Yes
German, light PE G21 0.73 0.18 4.14 0.00 0.39 1.09 Yes
Intercept 3.78 0.15 25.10 0.00 3.47 4.07 No
Russian, full PE G12 −0.17 0.22 −0.78 0.44 −0.61 0.26 No
Russian, light PE G22 −0.07 0.22 −0.33 0.74 −0.50 0.36 No
Intercept 4.19 0.12 34.04 0.00 3.94 4.42 Yes
Russian, full PE G12 −0.21 0.18 −1.14 0.25 −0.57 0.15 Yes
Russian, light PE G22 −0.16 0.18 −0.90 0.37 −0.52 0.19 Yes

Experiments in the German Language

The expected mean of errors in the group using only human translation is significantly different to the errors in the group using machine translation and light PE, but it is not significantly different to the group using full PE. This result does not change when the errors are weighted by severity level. The incidence rate ratio is defined as follows for the unweighted data: f (0.79) = e0.79 = 2.20, with s.e. = 0.44. And for the weighted data, f (0.79) = e0.79 = 2.09, with s.e. = 0.37. The incidence rate indicates that errors occurred more often when light PE was used; by changing from the control group to this treatment, the error count is expected to increase by approximately 2 units.

Experiments in the Russian Language

In the case of the experiments in the Russian language, the expected count of errors does not change when the group changes from control to machine translation and full PE or light PE. This means that there are no statistically significant differences in the number of translation errors in texts produced in an all-human setting in comparison to texts produced by settings that integrate machine translation and a form of PE at the initial translation step.

Discussion and Conclusion

Over the past years, neural MT has increased the quality of MT outputs overall and has made PE more efficient. However, the quality of MT and its usability in a specific translation situation (still) depends on the text type, the available MT tool, that is, their suitability for the text type, and the language combination. With these stakes in mind, we set up a research study to assess whether the quality of the review output in the TRAPD method is affected by introducing MT and PE at the translation stage and whether these effects increase or decrease output quality. Overall, the results of this study are encouraging for the use of MT and PE within the TRAPD approach. The experiments reported here, for the German and Russian languages and for a sample of ESS and EVS survey questions, show evidence that the translation quality of the review output is hardly affected by introducing MT and PE at the initial translation stage. The effect of including MT and PE in the treatments was barely quantifiable when comparing it to the control group, which was set up to use only human translation at the initial translation stage.

The secondary research questions asked whether Russian and German translations are affected differently and if differences are conditional on the use of full or light PE. We found that the effects of integrating MT and PE into TRAPD are different between Russian and German, and yes, there are different effects depending on the type of PE used. In the Russian language, the quality of review outputs resulting from the MT and PE treatments cannot be distinguished from the quality that derives from using human translation only. In the German language, there is an increase in the number of errors when light PE is used at the initial translation stage. However, the predicted increase of such errors is 2 units.

This cannot be considered an increase of such magnitude that rules out the use of the method completely. Furthermore, the analyses, conducted both without and with weights according to error severity levels, came to the same conclusions, which is positive for the use of MT and PE when integrated into the TRAPD model. Differences between the Russian and German settings, and here in particular related to the light-PE treatment group, may be explained by different dynamics in the review discussions and/or by different quality levels of the initial translations, including the PE versions. Research taking these aspects into account is currently underway, pointing, for instance, to less extensive review discussions in the German light-PE group. In the experiments, MT and PE were implemented by participants whose background is social sciences. This shows positive evidence that MT and PE can be used in the TRAPD approach by team members whose background is not professional translation, even though at this stage we cannot conclude how post-editors perceived and performed the actual PE task.

A few notes of caution seem apt. Error counts should be put into a larger context: the review meeting should not be the final step in the translation process. In a real-life setting, there would have been room for additional proofreading beyond the context of the review discussions or the possibility to clarify (source text) issues with the developers or further colleagues. Moreover, teams would certainly have picked up further errors when implementing the questionnaire in a survey tool and testing it in quantitative and qualitative pretests (as the full TRAPD model recommends). A few reviewers’ comments point in this direction, indicating for some segments that pretesting would be useful here. Survey practitioners wishing to explore the use of MT and PE in TRAPD should take into account that the mechanisms of the effects of MT and PE can be different in other languages and settings. After all, MT quality is usually better for language combinations where MT engines have been trained on large corpora. For smaller, rarer language combinations, MT quality may be problematic (Nitzke and Hansen-Schirra 2021).

Survey practitioners should also take into consideration that this was a very controlled environment and that the team meetings were implemented with rigor. While with the research presented here, we can draw conclusions on the overall team approach and how MT and PE works within this setting, we cannot yet draw conclusions on steps prior to the review output. However, this study shows how MT may be incorporated in the TRAPD approach: always accompanied by post-editing9 of the raw MT outputs, preferably by full PE; with a rigorous selection of the team members, providing them with training on the workings of MT and the challenges of PE. The team should combine professional translators and social scientists. One of the translations should be done by a professional translator with survey questionnaire experience.

Importantly, the review meeting should be implemented with rigor, including the creation of comprehensive documentation. PE itself does not come without challenges; for instance, smooth wording enabled by the neural-based engines may disguise errors, which may then remain undetected. PE requires “MT literacy” to understand the working of the method and consequently to benefit from it (O’Brien and Ehrensberger-Dow 2020). The activity of PE is regarded as different from translating and thus requires different skills and consequently different training (Guerberof Arenas and Moorkens 2019). Finally, an important remark regarding data security: if online tools are used to obtain the machine translations, the questionnaire texts are shared with the MT providers; therefore, confidentiality and intellectual property should be considered.

Future research will tackle the quality of the raw machine translation output, the role of the initial translations in the review output, as well as dynamics in the different team discussions. Moreover, we encourage study replication in other language combinations. Researchers may also move one step further by testing the use of two versions using MT and PE in a review discussion and then observe the outcome compared to an all-human condition.

To sum it all up, the text outputs analyzed in this article are those considered final after the review step in the TRAPD approach. Harkness and other proponents of TRAPD argued that the review sessions in which the team discusses translation options are fundamental to translation quality. The findings of the present study point to the fact that the review meeting is a very important aspect of the TRAPD method since it seems to compensate for differences in the initial translations. The potential effect of different methods used to produce the parallel translations in the T step, in this article human translation and machine translation in combination with (full or light) PE, do not or hardly remain in the review outputs. Overall, our findings align with the trend of other text types: in some languages MT is achieving a level of maturity such that it can be considered for integration in the translation workflows. At the same time, more research is needed to understand the details of MT and PE steps themselves for the text-type questionnaire.

Supplementary Material

nfad060_Supplementary_Data

Footnotes

1

A text corpus, pl., corpora, is a collection of structured texts or linguistic data.

2

Braun and Johnson (2010) provide an overview of statistical methods suitable for detecting absence of equivalence in the survey data, which may be caused by translation errors.

3

Participants in the review meetings in treatments where machine translation was used did not know which of the versions were originated by a machine translation tool. During the review meetings, the version discussed is the post-edited one, not the raw machine-translated output. Similarly, in the case of human translation, the final version created by translators is commented in the review meeting, not the first draft text at the initial translation step.

4

All participants were given information on data protection and asked for their written informed consent.

5

A repository with all the data, source texts and translations, error coding environment and commented scripts in R and Python, and additional information needed to replicate this article is available at https://osf.io/vre5p/?view_only=001b1d4c6ece4c45a7d7340b72dbdba1.

6

Developed by the EU-funded QTLaunchPad project.

7

Developed by the Translation Automation User Society (TAUS).

8

An estimate of intercoder reliability was estimated using the intraclass correlation coefficient. This statistic is available in the online repository at https://osf.io/vre5p/?view_only=001b1d4c6ece4c45a7d7340b72dbdba1.

9

General guidance on PE can be found in Massardo et al. (2016) and Nitzke and Hansen-Schirra (2021).

Contributor Information

Diana Zavala-Rojas, Principal Investigator of the European Social Survey ERIC, Universitat Pompeu Fabra, Barcelona, Spain; and Deputy Director of the Research and Expertise Centre, Survey Methodology in the Political and Social Sciences Department, Universitat Pompeu Fabra, Barcelona, Spain.

Dorothée Behr, Head of Team, Cross-Cultural Survey Methods, Survey Design and Methodology Department, GESIS—Leibniz Institute for the Social Sciences, Mannheim, Germany.

Brita Dorer, Head of the Translation Workpackage, European Social Survey ERIC, Mannheim, Germany; and Senior Researcher, Survey Design and Methodology Department, GESIS—Leibniz Institute for the Social Sciences, Mannheim, Germany.

Danielly Sorato, Researcher, Research and Expertise Centre for Survey Methodology, Political and Social Sciences Department, Universitat Pompeu Fabra, Barcelona, Spain; and PhD Candidate, Department of Translation and Language Sciences, Universitat Pompeu Fabra, Barcelona, Spain.

Veronika Keck, Senior Client Training Consultant, The Nielsen Company (Germany) GmbH (NielsenIQ), Frankfurt am Main, Germany.

Supplementary Material

Supplementary Material may be found in the online version of this article: https://doi.org/10.1093/poq/nfad060.

Funding

This work was funded from the European Union’s Horizon 2020 Research and Innovation Programme (2014-2020) under Grant Agreement No. 823782. PI at Universitat Pompeu Fabra: Diana Zavala-Rojas; PI at GESIS: Dorothée Behr.

Data Availability

Replication data and documentation are available at https://osf.io/vre5p/?view_only=001b1d4c6ece4c45a7d7340b72dbdba1.

References

  1. Bahdanau Dzmitry, Cho Kyung Hyun, Bengio Yoshua. 2015. “Neural Machine Translation by Jointly Learning to Align and Translate.” In 3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings, edited by Yoshua Bengio and Yann LeCun, 1–15. San Diego, CA: ICLR 2015. http://arxiv.org/abs/1409.0473.
  2. Braun Michael, Johnson Timothy P.. 2010. “An Illustrative Review of Techniques for Detecting Inequivalences.” In Survey Methods in Multinational, Multiregional, and Multicultural Contexts, edited by Janet A. Harkness, Michael Braun, Brad Edwards, Timothy P. Johnson, Lars Lyberg, Peter Ph. Mohler, Beth-Ellen Pennell, Tom W. Smith, 373–93. Hoboken: John Wiley & Sons, Inc. 10.1002/9780470609927.ch20. [DOI] [Google Scholar]
  3. Burchardt Aljoscha, Lommel Arle. 2014. “Practical Guidelines for the use of MQM in Scientific Research on Translation Quality.” Preparation and Launch of a Large-Scale Action for Quality Translation Technology, Report, 19. https://cordis.europa.eu/project/id/296347. Date accessed 8 February 2024.
  4. Castilho Sheila, Doherty Stephen, Gaspari Federico, Moorkens Joss. 2018. “Approaches to Human and Machine Translation Quality Assessment: From Principles to Practice.” In Translation Quality Assessment, edited by Moorkens Joss, Castilho Sheila, Gaspari Federico, Doherty Stephen, 9–38. Cham: Springer. [Google Scholar]
  5. Castilho Sheila, Moorkens Joss, Gaspari Federico, Sennrich Rico, Sosoni Vilelmini, Georgakopoulou Yota, Lohar Pintu, Way Andy, Valerio Miceli Barone Antonio, Gialama Maria. 2017. “A Comparative Quality Evaluation of PBSMT and NMT using Professional Translators.” In Proceedings of MT Summit XVI, the 16th Machine Translation Summit, edited by Sadao Kurohashi, Pascale Fung. Nagoya: MTSummit. https://aclanthology.org/2017.mtsummit-papers.10/.
  6. Daems Joke, De Clercq Orphée, Macken Lieve. 2017. “Translationese and Post-Editese: How Comparable is Comparable Quality?” Linguistica Antverpiensia 16:89–103. 10.52034/lanstts.v16i0.434. [DOI] [Google Scholar]
  7. Daems Joke, Vandepitte Sonia, Hartsuiker Robert J., Macken Lieve. 2017. “Translation Methods and Experience: A Comparative Analysis of Human Translation and Post-Editing with Students and Professional Translators.” Meta 62:245–70. 10.7202/1041023ar. [DOI] [Google Scholar]
  8. Doherty Stephen, O’Brien Sharon. 2014. “Assessing the Usability of Raw Machine Translated Output: A User-Centered Study Using Eye Tracking.” International Journal of Human-Computer Interaction 30:40–51. 10.1080/10447318.2013.802199. [DOI] [Google Scholar]
  9. ELIA, EMT, EUATC, FIT-Europe, GALA, and LIND. 2020. “European Language Industry Survey 2020: Before & After COVID-19.” http://fit-europe-rc.org/en/resources/publications/.
  10. Eurofound. 2017. Sixth European Working Conditions Survey—Overview Report (2017 Update). Technical report. Luxembourg. https://www.eurofound.europa.eu/en/publications/2016/sixth-european-working-conditions-survey-overview-report Date accessed 8 February 2024.
  11. European Social Survey. 2024. Round 12 Survey Specification for ESS ERIC Member, Observer and Guest Countries. London: City University London. [Google Scholar]
  12. Federico Marcello, Bertoldi Nicola, Cettolo Mauro, Negri Matteo, Turchi Marco, Trombetti Marco, Cattelan Alessandro, Farina Antonio, Lupinetti Domenico, Martines Andrea. 2014. “The Matecat Tool.” In Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: System Demonstrations, edited by Lamia Tounsi and Rafael Rak, 129–32. Dublin: Dublin City University and Association for Computational Linguistics.
  13. Forsyth Barbara H., Kudela Martha Stapleton, Levin Kerry, Lawrence Deirdre, Willis Gordon B.. 2007. “Methods for Translating an English-Language Survey Questionnaire on Tobacco Use into Mandarin, Cantonese, Korean, and Vietnamese.” Field Methods 19:264–83. 10.1177/1525822X07302105. [DOI] [Google Scholar]
  14. Gouadec Daniel. 2010. “Quality in Translation.” Handbook of Translation Studies 1:270–75. [Google Scholar]
  15. Guerberof Arenas Ana, Moorkens Joss. 2019. “Machine Translation and Post-Editing Training as Part of a Master’s Programme.” Jostrans: The Journal of Specialised Translation 31:217–38. [Google Scholar]
  16. Harkness Janet A. 2003. “Questionnaire Translation.” In Cross-Cultural Survey Methods, edited by Harkness Janet A., van de Vijver F. J. R., Mohler P. P., 35–56. Hoboken: Wiley & Sons. [Google Scholar]
  17. Harkness Janet A., Braun Michael, Edwards Brad, Johnson Timothy P., Lyberg Lars E., Mohler Peter Ph., Pennell Beth-Ellen, Smith Tom W.. 2010. “Comparative Survey Methodology.” In Survey Methods in Multinational, Multiregional, and Multicultural Contexts, edited by Harkness Janet A., Braun Michael, Edwards Brad, Johnson Timothy P., Lyberg Lars E., Mohler Peter Ph., Pennell Beth-Ellen, Smith Tom W., 1–16. Hoboken: John Wiley & Sons, Inc. 10.1002/9780470609927.ch1. [DOI] [Google Scholar]
  18. Harkness Janet A., Pennell Beth-Ellen, Schoua-Glusberg Alisu. 2004. “Survey Questionnaire Translation and Assessment.” In Methods for Testing and Evaluating Survey Questionnaires, edited by Stanley Presser, Jennifer M. Rothgeb, Mick P. Couper, Judith T. Lessler, Elizabeth Martin, Jean Martin, and Eleanor Singer, 453–73. John Wiley & Sons, Inc. 10.1002/0471654728.ch22. [DOI] [Google Scholar]
  19. Harkness Janet A., Villar Ana, Edwards Brad. 2010. “Translation, Adaptation, and Design.” In Survey Methods in Multinational, Multiregional, and Multicultural Contexts, edited by Harkness Janet A., Braun Michael, Edwards Brad, Johnson Timothy P., Lyberg Lars E., Mohler Peter Ph., Pennell Beth-Ellen, Smith Tom W., 115–40. Hoboken: John Wiley & Sons, Inc. 10.1002/9780470609927.ch7. [DOI] [Google Scholar]
  20. Iwai Ritsuko, Kawahara Daisuke, Kumada Takatsune, Kurohashi Sadao. 2019. “Applying Machine Translation to Psychology: Automatic Translation of Personality Adjectives.” In Proceedings of Machine Translation Summit XVII Volume 2: Translator, Project and User Tracks, edited by Mikel L. Forcada, Andy Way, John Tinsley, Dimitar Shterionov, Celia Rico, and Federico Gaspari, 23–29. Dublin: European Association for Machine Translation.
  21. Jia Yanfang, Carl Michael, Wang Xiangling. 2019. “Post-Editing Neural Machine Translation Versus Phrase-Based Machine Translation for English–Chinese.” Machine Translation 33:9. 10.1007/s10590-019-09229-6. [DOI] [Google Scholar]
  22. Kietzmann Diana, Wiehn Sophia, Kehl Doris, Knuth Daniela, Schmidt Silke. 2016. “Migration Background and Overall Satisfaction with Pre-Hospital Emergency Care.” Applied Nursing Research: ANR 29:96–100. 10.1016/j.apnr.2015.05.009. [DOI] [PubMed] [Google Scholar]
  23. Koponen Maarit. 2016. “Is Machine Translation Post-Editing Worth the Effort? A Survey of Research into Post-Editing and Effort.” The Journal of Specialised Translation 25:131–48. [Google Scholar]
  24. Koponen Maarit, Salmi Leena. 2017. “Post-Editing Quality: Analysing the Correctness and Necessity of Post-Editor Corrections.” Linguistica Antverpiensia, New Series—Themes in Translation Studies 16. https://lans-tts.uantwerpen.be/index.php/LANS-TTS/article/view/439/394.
  25. Kuckartz Udo. 2014. Qualitative Inhaltsanalyse. Weinheim und Basel: Methoden, Praxis, Computerunterstützung. [Google Scholar]
  26. Lommel Arle. 2018. “Metrics for Translation Quality Assessment: A Case for Standardising Error Typologies.” In Translation Quality Assessment, edited by Moorkens Jess, Castilho Sheila, Gaspari Federico, Doherty Stephen, 109–27. Cham: Springer. [Google Scholar]
  27. Lommel Arle, Görög Attila, Melby Alan, Uszkoreit Hans, Burchardt Aljoscha, Popović Maja. 2015. Harmonised Metric. Technical Report, Deliverable 3.1. https://ec.europa.eu/research/participants/documents/downloadPublic?documentIds=080166e5a0f575d3&appId=PPGMS.
  28. Massardo Isabella, van der Meer Jaap, O’Brien Sharon, Hollowood Fred, Aranberri Nora, Drescher Katrin. 2016. MT Post-Editing Guidelines. Vol. 42. Amsterdam, The Netherlands: TAUS Signature Editions. [Google Scholar]
  29. Mohler Peter Ph., Dorer Brita, De Jong Julie, Hu Mengyao. 2016a. “Adaptation.” InGuidelines for Best Practice in Cross-Cultural Surveys, 2nd ed., Chapter 8, edited by Survey Research Center, 378–91. Ann Arbor, MI: Survey Research Center, Institute for Social Research, University of Michigan. http://ccsg.isr.umich.edu/index.php/chapters/adaptation-chapter. [Google Scholar]
  30. Mohler Peter Ph., Dorer Brita, De Jong Julie, Hu Mengyao. 2016b. “Translation.” In Guidelines for Best Practice in Cross-Cultural Surveys,2nd ed., Chapter 7, edited by Survey Research Center, 233–377. Ann Arbor, MI: Survey Research Center, Institute for Social Research, University of Michigan. http://ccsg.isr.umich.edu/index.php/chapters/translation-chapter. [Google Scholar]
  31. Mondal Himel, Mondal Shaikat, Mondal Sarika. 2019. “Feasibility of using “Google Translate” in Adaptation of Survey Questionnaire from English to Bengali: A Pilot Study.” Indian Journal of Social Psychiatry 35:119. [Google Scholar]
  32. Moorkens Joss, Toral Antonio, Castilho Sheila, Way Andy. 2018. “Translators’ Perceptions of Literary Post-Editing Using Statistical and Neural Machine Translation.” Translation Spaces 7:240–62. 10.1075/ts.18014.moo. [DOI] [Google Scholar]
  33. Nitzke Jean, Hansen-Schirra Silvia. 2021. A Short Guide to Post-Editing. Translation and Multilingual Natural Language Processing. Vol. 16. Berlin: Language Science Press. 10.5281/zenodo.5646896; https://library.oapen.org/handle/20.500.12657/52585. [DOI] [Google Scholar]
  34. O’Brien Sharon, Ehrensberger-Dow Maureen. 2020. “MT Literacy—A Cognitive View.” Translation, Cognition & Behavior 3:145–64. 10.1075/tcb.00038.obr; https://www.jbe-platform.com/content/journals/10.1075/tcb.00038.obr. [DOI] [Google Scholar]
  35. Popović Maja. 2018. “Language-Related Issues for NMT and PBMT for English–German and English–Serbian.” Machine Translation 32:237–53. 10.1007/s10590-018-9219-5. [DOI] [Google Scholar]
  36. Przepiórkowska Danuta, Behr Dorothée. 2017. European Values Study Translation Guidelines .Technical Report. Mannheim, Germany: GESIS. [Google Scholar]
  37. Saldanha Gabriela, O’Brien Sharon. 2014. Research Methodologies in Translation Studies, 292. New York: Routledge. 10.4324/9781315760100. [DOI] [Google Scholar]
  38. Sha Mandy, Lai Jennie. 2016. “A Case Study of Improving and Evaluating Consumer Survey Translation.” Translation & Interpreting 8:86–100. [Google Scholar]
  39. Sin-Wai Chan. 2017. The Future of Translation Technology: Towards a World without Babel, 1st ed. Routledge Studies in Translation Technology. New York: Routledge. 10.1007/s10590-017-9199-x. [DOI] [Google Scholar]
  40. Smith Tom W. 2011. “Refining the Total Survey Error Perspective.” International Journal of Public Opinion Research 23:464–84. [Google Scholar]
  41. Toral Antonio. 2019. “Post-Editese: An Exacerbated Translationese.” http://arxiv.org/abs/1907.00900.
  42. Vardaro Jennifer, Schaeffer Moritz, Hansen-Schirra Silvia. 2019. “Translation Quality and Error Recognition in Professional Neural Machine Translation Post-Editing.” Informatics 6:41. 10.3390/informatics6030041. [DOI] [Google Scholar]
  43. Way Andy. 2020. “Machine Translation: Where Are We at Today?” In The Bloomsbury Companion to Language Industry Studies, 1st ed., edited by Angelone Erik, Ehrensberger-Dow Maureen, Massey Gary, 311–32. Bloomsbury Companions. London: Bloomsbury Academic. http://www.bloomsburycollections.com/book/the-bloomsbury-companion-to-language-industry-studies/ch14-machine-translation-where-are-we-at-today/. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

nfad060_Supplementary_Data

Data Availability Statement

Replication data and documentation are available at https://osf.io/vre5p/?view_only=001b1d4c6ece4c45a7d7340b72dbdba1.


Articles from Public Opinion Quarterly are provided here courtesy of Oxford University Press

RESOURCES