Skip to main content
PLOS One logoLink to PLOS One
. 2021 Jul 9;16(7):e0254006. doi: 10.1371/journal.pone.0254006

How do researchers approach societal impact?

Benedikt Fecher 1,2,*,#, Marcel Hebing 1,3,#
Editor: Laurentiu Rozylowicz4
PMCID: PMC8270479  PMID: 34242271

Abstract

Based on a communication-centered approach, this article examines how researchers approach societal impact, that is, what they think about societal impact in research governance, what their societal goals are, and how they use communication formats. Hence, this study offers empirical evidence on a group that has received remarkably little attention in the scholarly discourse on the societal impact of research—academic researchers. Our analysis is based on an empirical survey among 499 researchers in Germany conducted from April to June 2020. We show that most researchers regard societal engagement as part of their job and are generally in favor of impact evaluation. However, few think that societal impact is a priority at their institution, and even fewer think that institutional communication departments reach relevant stakeholders in society. Moreover, we show that researchers’ societal goals and use of communication formats differ greatly between their disciplines and the types of organization that they work at. Our results add to the ongoing metascientific discourse on the relationship between science and society and offer empirical support for the hypothesis that assessment needs to be sensitive to disciplinary and organizational context factors.

Introduction

Until the 1970s there was no doubt among policymakers that public investment in research would have a positive societal impact. The social contract for science at that time meant that science was granted an unusual degree of autonomy in return for widely diffused benefits to society and the economy [1]. It was only from the late 1980s onward that researchers were increasingly expected to account for their achievements in evaluation exercises [24]. Initially, these focused on intra-scientific (and often bibliometric) impact. In the last decade, however, policymakers have begun to focus more on the societal impact of research and hence on what academic research offers for the economy, society, culture, public administration, health, environment, and overall quality of life [2, 5, 6]. Noteworthy examples indicating the gradual shift from scientific to societal impact in research governance include the Research Excellence Framework (REF) in the United Kingdom or the Excellence in Research for Australia Framework (ERA) [79]. In the Netherlands, a region with some of the most developed examples of impact governance, the Association of Universities in the Netherlands (VSNU), the Netherlands Organisation for Scientific Research (NWO), and the Royal Netherlands Academy of Arts and Sciences (KNAW) have implemented guidelines for the evaluation and improvement for research, the Standard Evaluation Protocol (SEP). The SEP is used by research institutions to evaluate research units and focuses strongly on relevance for society as well as research quality [10, 11]. Societal impact is also a key component in European research funding [1214].

In Germany, where the present study was conducted, there is no evaluation exercise comparable to the REF, ERA, or SEP. The topic has, however, been prominently discussed: Referring to the announcement that the next EU Framework Programme for Research and Innovation will focus more on the societal benefits of research, the Alliance of Science Organisations stated in 2018 that any such evaluation should be differentiated and tailored to the demands of science [15]. In 2019, the Federal Ministry of Education and Research (BMBF) published a policy paper in which it stipulated that societal impact must become part of the academic reputation logic [16]. In 2020, it set up a think tank to work out how societal impact could be evaluated [17]. Also in 2020, the Alliance of Science Organizations issued an agreement containing four fields of action; this highlighted, among other things, scientific freedom and the need to anticipate disciplinary differences. The German Council of Sciences and Humanities, an advisory body to the German federal government, called for “more recognition for knowledge and technology transfer” [18]. The German Rectors’ Conference, the umbrella organization of German universities, decided at its general assembly (November 14, 2017) that knowledge transfer is a priority for universities [19]. Besides these developments at the level of science governance in Germany, there are a multitude of institutional initiatives and strategy developments underway. For instance, the Leibniz Association, a union of 96 nonuniversity research institutes, adopted a new mission statement for the transfer of scientific knowledge to society, the economy, and politics in 2019 [20].

Although there is no evaluation exercise in Germany comparable to the REF in the UK, it is evident that the topic has gained momentum in Germany in recent years as well. In light of complex societal challenges and the further integration of German research bodies into the European research area, the societal impact of research will likely become an even more prominent concern in research governance in the near future. If, as stated above, the old social contract granted science relative freedom in return for widely diffused benefits for society, the new social contract for science imposes accountability for the freedom granted to science [2123]. This development has been criticized by the scientific community. For example, many associations for the social sciences and humanities in Germany expressed concern that the BMBF initiative does not take into account the current state of science communication research [24].

If societal impact is to be a new paradigm of science governance, it is important to better understand how societal impact emerges. In this article, we address this question from the perspective of researchers, that is, we ask what their views on societal impact are, what their goals are, and what formats they use to achieve them. Our analysis is based on a survey among 499 researchers in Germany conducted from April to June 2020. Here, we report the key results of this study and reflect on potential implications for science governance. Our results add to the ongoing scholarly discourse on societal impact by offering empirical evidence on a group that has received surprisingly little attention in the scholarly study of societal impact—academic researchers.

Societal impact in research

Two scholarly discourses are particularly relevant for the subject of this study: 1) the discourse in communication and science and technology studies (STS) on the relationship between science and society and 2) the discourse on societal impact measurement in scientometrics and evaluation research. In the following, we will delineate these two discourses and reflect on their implications for this study.

Relationship between science and society: From deficit to dialog

Before the first large-scale impact agendas were implemented, scholars in STS critically examined the nature and role of science in society, drawing on novel concepts of academic knowledge creation such as “Mode 2” [25, 26], “academic capitalism” [27], “post-normal science” [28] or “Triple Helix” [29]. Although these concepts differ in their objectives, they generally assume a form of scientific value creation that is no longer self-sufficient and is increasingly interwoven with society. While these conceptualizations do not employ societal impact as an evaluative paradigm, they have paved the way for new thinking about the relationship between science and society. These explorations of the role of science in society were, however, primarily theoretical in nature. Later, entire lines of (communication) research addressed how the public deals with science, for example, the public understanding or awareness of science (PUS, PAwS), scientific literacy, or more recently the public engagement with science and technology (PEST) and the science of science communication [3033]. In general, research in this area has shifted away from viewing society as an inactive recipient of knowledge—for example, in the so-called “deficit model of science communication” [34]—and towards envisaging more complex and interactive forms of knowledge creation and dissemination [3, 3537].

While there is a considerable body of literature on how the public perceives research, less attention has been paid to the institutional conditions for engagement and how researchers themselves deal with the public. Regarding these institutional conditions, scholars have observed a certain decoupling of central communication infrastructures at institutions and the researchers working there [38, 39]. Others have pointed to the increased legitimation pressure exerted by research organizations and the increase in PR and marketing [40]. When it comes to researchers’ dealings with the public, scholars have focused on a) the relationship between science and specific publics—for instance, the media [41, 42] or politics [41], b) the relationship between science and the broader public [41, 4345] or c) the communication practices of single disciplines [46, 47]. Here, recurring themes include researchers motivations for engaging with the public [48, 49], teaching and training [50, 51], and institutional conditions [52, 53]. Much research on the interfaces between science and society consists of various relevant case studies. Yet, there is a lack of comparable empirical evidence on one particularly decisive group of actors in the dialogical rationale, that is, academic researchers. This explains why, in this study, we focus on researchers and how they approach societal impact.

Measuring societal impact: From economic to broader societal benefit

In scientometrics and the wider field of evaluation research, the shift towards societal impact as an assessment paradigm in science governance has been accompanied by critical reflections. Martin, for example, asked pointedly whether the creation of the REF would create a “Frankenstein monster,” because the costs of conducting the evaluation might outweigh the benefits [8]. Others have argued that evaluations of societal impact are prone to methodological shortcomings [23, 54] and might have unintended behavioral effects [55, 56]. In addition, there are two further points of criticism concerning the concept of societal impact: On the one hand, critics point to the inadequate representation of the complexity of science, for example, because impact logics of the natural sciences are used as a yardstick for evaluations [6, 57]. In contrast, impact assessments have been described as failing to do justice to transfer activities in the social sciences and humanities (SSH) [5860]. In recent years, new approaches have been developed that specifically address the SSH [6, 61, 62]. On the other hand, criticism is directed towards the representation of society, whose benefit has often been reduced to economic indicators (e.g., revenue, jobs) and not the broader societal impact [5, 63]. However, some evaluation exercises have broadened their scope to include the wider benefits to society [2].

Recent models for societal impact have tried to incorporate dialog into their conceptions of societal impact. An example of this is Jong et al.’s concept of productive interactions, which proceeds from the assumption that current productive interactions between researchers and societal stakeholders will improve the probability of future societal impact [64]. In this regard, interactions are deemed productive if encounters between researchers and societal stakeholders lead to knowledge that is academically sound and socially valuable [62, 64]. The concept indicates that two quality regimes—a scientific and a societal—are important for deciding on the productiveness of a dialog. It should be noted that science is part of society and both the sciences and their publics have differing ideas about robustness and usefulness [32]. The notion of societal impact as an (effect of) dialog is useful in the context of this study, because it allows us to move from pure utilization and reception research to examining the anticipated societal impact of researchers. Drawing from the wider field of evaluation research, we examine how academic researchers anticipate the broader societal impact of their research.

Conceptualization of societal impact in this study

Building on the discourse in evaluation research, we refer to broader societal impact, i.e., the benefits that research holds for the economy, individual wellbeing, the environment, and culture [2, 14]. According to Bornmann, three main strands of societal impact can be distinguished: First, societal impact as a product (i.e., as an artifact that embodies scientific knowledge), second, societal impact as use (i.e., the adoption of academic knowledge by societal stakeholders), and third, societal impact as benefits (i.e., the effects of the use of research) [2]. Here we focus on the latter—we consider the desired societal benefits that researchers associate with their work (goals, RQ2) and the formats they use to communicate about their research (formats, RQ3). To further differentiate societal benefits and formats, we conducted a qualitative prestudy (see Method section). To gain a better understanding of how researchers perceive societal impact as a paradigm in research governance, we included questions to elicit respondents’ opinions on the role of societal impact at their institutions and in their work, on whether societal impact should be given more weight in evaluations, and on whether their institutional communication departments are reaching relevant societal stakeholders [38,39] (opinions, RQ1).

Explanatory dimensions

Drawing on De Jong and colleagues and in line with the metascientific discourse on the conception of the relationship between science and society, we see societal impact as the effect of interaction between scientific and societal stakeholders [61, 62, 64]. We hence used a framework for communication inspired by Cohn’s concept of theme-centered interaction [65, 66] and Luhman’s notion of meaning [67, 68] when deriving our explanatory variables. We differentiated between three dimensions of explanatory variables:

  • The content dimension is defined by the researcher’s disciplinary background and their self-assessment as to whether their research is applied or basic. It can be assumed that researchers’ approaches to societal impact varies between disciplines [5860]. We assumed that a researcher’s disciplinary background would affect their choice of societal goals. We further expected to find differences between researchers who considered themselves applied and those who considered themselves basic researchers—societal impact might play a greater role for applied researchers [6973]. We therefore assumed that applied researchers would be more supportive of societal impact evaluations and would also be more active in using communication formats.

  • The organizational dimension is defined by the type of research organization (i.e., universities, universities of applied sciences, nonuniversity research institutions). The organizational factors influencing societal impact have, with notable exceptions [52, 53], been little researched so far. Here, we were particularly interested in differences between various types of public research institutions in Germany (see Method section below). We assumed that researchers at independent institutes would be more active than university researchers in using communication formats because they usually do not have teaching obligations. We further expected researchers who worked at applied science universities to be more active in using collaboration formats.

  • The individual dimension is defined by the sociodemographic factors of age, gender, and academic status. Gender and age differences in connection with human agency have been widely researched in the social sciences, also as they pertain to scholarly communication [7480]. How these relate to researchers’ engagement with society is still little understood. Academic status may have an influence on the use of communication formats, in the sense that high-status academics may be most active in advisory roles.

Research questions

By looking at these dimensions, we aimed to understand how these factors might influence the interaction between scientific and nonscientific actors in terms of the opinions researchers hold, the societal goals they pursue, and the formats they use to communicate about their research. Three research questions guided our analysis and structured the presentation of the results; when analyzing each question, we use our three dimensions of explanatory variables (i.e., content, organization, individual) to structure and compare the results.

  • RQ1: What are researchers’ opinions on societal impact? What differences can be identified along the three dimensions (i.e., content, organization, individual)?

In particular, we were interested in researchers’ perspectives on a) societal engagement as a part of scientific activity, b) whether societal impact should be considered more in research evaluation, c) the performance of institutional communication departments in reaching relevant societal stakeholders, and d) the importance of knowledge transfer at the institutions.

  • RQ2: Which societal goals do researchers aim to achieve with their research? How do these differ along the three dimensions (i.e., content, organization, individual)?

We used the 13 goals that we identified by coding the REF use cases as a framework (e.g., supporting legislative decision making, driving technical innovation, preserving cultural heritage, or protecting the environment).

  • RQ3: Which formats do researchers use to achieve societal impact? How do these differ along the three dimensions (i.e., content, organization, individual)?

Again, we used our coding of the REF use cases as a framework and took a close look at educational offerings, consulting, events, PR, social media, and collaborations.

Method

The institutional review board and data protection officer of the Alexander von Humboldt Institute for Internet and Society approved this study. Informed written consent was obtained from the participants in this study. The data was analyzed anonymously. To answer the research questions, we conducted a survey from April to June 2020 among 499 scientists in Germany. In the following, we describe the instrument and its distribution, the sample design, and the analysis of the data. The survey instrument, the aggregated data, and the prestudy (qualitative analysis of REF case studies) can be found on the project’s website (https://www.impactdistillery.com/2020-impact-survey/). The survey was part of a BMBF-funded research project addressing the question of societal impact and its measurability in the social sciences and humanities (https://www.wihoforschung.de/de/impaqt-2631.php).

Instrument

We designed a standardized instrument, consisting of six sections (A: sociodemographics, B: work context, C: knowledge transfer, D: teaching, E: research, F: general questions, G: personality). Here, we report on the largest section (C: knowledge transfer), which covers our three RQs, and use questions from sections A, B, and F for the explanatory analysis.

RQ1 was covered by a set of items (C1) addressing researchers’ attitudes towards societal impact—i.e., their opinions regarding the importance of knowledge transfer as part of scientific activity and at their institutions, their opinions towards a stronger weighting of societal impact in research evaluation, and the performance of their institution’s communication department in reaching relevant societal stakeholders. We used a 5-point Likert scale (from strongly disagree to strongly agree) to elicit agreement or disagreement with a statement.

RQ2 (goals) and RQ3 (formats) were implemented as multiple-choice questions (C3 and C5). The answer categories were based on a structuring content analysis of the REF impact case studies, which we carried out as a prestudy in spring 2019. In 2014, the UK became the first country to assess the societal impact of research as part of a national assessment. The REF evaluates societal impact via case studies, which are narratives that describe how research conducted at a higher education institution created a wider societal benefit. Of course, REF impact case studies are written to succeed in the evaluation in question [81]. Nevertheless, as researchers’ expressions of impact, they provide a suitable textual data basis for developing goal and format categories [82, 83]. The REF case studies are available in a public database (https://impact.ref.ac.uk/casestudies/). The case studies were analyzed by two coders separately and frequently discussed in the research group in order to achieve intercoder reliability. The main categories of the content analysis were reflected in the questions on societal goals (i.e., What societal impacts do you want to achieve with your research?; 13 items—RQ2) and communication formats (i.e., Which transfer formats have you already used to communicate your results?; 6 items—RQ3). Each question allowed further responses in an open text field—from the few additional responses and the high response rate, we concluded that the identified categories were robust.

To ask about sociodemographic factors (sections A & B), we re-used questions that had already proved useful in a previous survey [77, 80]. The range of disciplines we asked about (A5) was based on the classification of the German Research Foundation [84].

Pretests

To evaluate the quality and reliability of the survey, we conducted two rounds of pretests. The first round involved topic experts (i.e. fellow meta-researchers) and methodological experts who reviewed both the content and the design of the survey. The second pretest round involved researchers from different disciplines who completed the survey, focusing on its usability and providing us with a first dataset for preliminary analysis and optimization. As a result, we revised descriptions and shortened the survey.

Sample design and distribution

We designed a semi-convenience sample, which means that in principle any researcher with the link to the survey could participate. However, we personally invited certain researchers so that we could adequately cover the explanatory dimensions (i.e., the different organizational settings, disciplines, and career stages in Germany). With this in mind, we applied the following distribution strategy: We contacted the faculty heads of 60 German universities and 60 universities of applied science (Fachhochschulen; in the following, we use the short form “applied universities”) and asked them to distribute the survey to researchers in their faculties. We selected the universities and applied universities based on the number of students and chose the 20 largest and the 20 smallest as well as 20 medium-sized ones. Additionally, we contacted the directors of each institute within the biggest German non-university research organizations (in the following shortened to “independent institutes”), i.e., the Max Planck Society, the Leibniz Association, the Helmholtz Association, and the Fraunhofer Gesellschaft. We also contacted the German Research Association’s (DFG) graduate schools. Despite these efforts, our sample was not probabilistic and we assumed a certain self-selection bias due to the topic of the survey. Second, because the sample consisted of researchers in Germany, the transferability of the results to other research and innovation systems is limited.

The distinction between different types of research organizations (i.e., universities, applied universities, independent institutes) is important for this study. Independent institutes are characteristic of the German research system. They are typically independent of universities and focus on specific fields of research; scientists at these institutions are not obliged to teach. These institutes are typically organized within the Max Planck Society, the Helmholtz Association of German Research Centres, the Fraunhofer Society, and the Leibniz Association. Researchers from independent institutes should in theory be able to devote more resources to transfer activities than researchers at universities and applied universities, as this latter group has teaching obligations. Applied universities are Germany-specific tertiary education and research institutions. They are rather transfer-oriented and usually specialized in certain fields (e.g., arts, technology, or business). Researchers from applied universities typically have the most extensive teaching obligations. However, due to their applied approach, societal impact should presumably play a more central role for researchers at these kinds of institutions.

We conducted the online survey from April to June 2020. Participants were invited with an initial email and one reminder, both including a request to distribute the survey among colleagues.

Sample description

Overall, 841 people started the survey, 534 of whom completed it (63%). In this paper, we focus on those who stated that their primary work location is Germany, leaving us with 499 valid (59%) cases to analyze.

Fig 1 provides an overview of the sample. Our distribution system had the desired effect—respondents from all disciplines and institutions took part. Regarding sociodemographic characteristics, the genders were equally distributed (50% male, 49% female, and 1% others). For ages 18 to 29 there were almost twice as many female as male researchers in the sample, while there were almost twice as many males as females in the 50 to 59 age group. For the 60+ group, there were approximately three times more male than female researchers in the sample (we excluded others due to the small sample size of 1%). This, however, corresponded roughly to the general gender distribution in German academia [85, 86].

Fig 1. Description of the sample using the explanatory variables that are at the core of our analysis.

Fig 1

(1) To what extent do researchers from the three disciplinary groups (natural sciences, social sciences, and humanities) consider their work as basic or applied research? (2) Distribution of the disciplinary groups across the prevailing institutional types. (3) Distribution of the researchers’ sex across age groups. (4) Distribution of disciplinary groups for male and female researchers. The number of participants with other sex was so small that they were not included in these graphs.

60% of our participants worked at universities, 28% at independent institutes, 10% at applied universities, and only 2% at other forms of institutions; hence, when analyzing the organizational dimension, we focused on the three largest types of organizations. Furthermore, we found that 34% were PhD candidates, 25% were postdocs, 11% were academics without a PhD, 25% were professors and 5% were others. Regarding the disciplinary background, we used the classification of the German Research Association (DFG) and merged the disciplines into three groups: 41% of the respondents were from the natural sciences, 35% were from the social sciences, and 26% were from the humanities [84].

Analysis

We found very little empirical research that looked at how researchers approach societal impact. Therefore, we conducted an exploratory analysis and focused on descriptive methods. Most of the results are presented as bivariate distributions that cross-tabulate the three dimensions from our conceptual framework with our research questions. For RQ1 (opinions on science communication), we asked respondents about evaluations because these would have practical implications for their working lives. With this particular dependent variable, we calculated a regression model (OLS) to analyze multivariate effects in more detail.

Some of the questions we analyzed were implemented as a 5-point Likert scale (strongly disagree, disagree, neutral, agree, strongly agree). In general, we considered agreement as the last two answer categories of the scale (agree + strongly agree) and report on this.

Results

In the research section we present the results for the three research questions alongside the three dimensions of the independent variables, i.e., the content dimension, the organizational dimension, and the dimension of a researcher’s individual characteristics.

RQ1—opinion: Individual commitment without an institutional mandate

Overall, 89% of the respondents agree that public engagement is part of scientific activity. 53% agree that societal impact should be given more weight in research evaluations. Only 27% of respondents think that knowledge transfer plays an important role at their institution and that their communication department is reaching relevant stakeholders in society.

Across all disciplinary groups, the respondents largely agree that public engagement is part of scientific activity (from 86% in the natural sciences, 91% in the humanities, to 93% in the social sciences). Respondents vary in their opinions on whether societal impact should be given more weight in research evaluations: Only 40% of the respondents from the natural sciences agree compared to 65% for the social sciences and 58% for the humanities. Applied researchers are more in favor of societal relevance being given greater consideration in evaluations than basic researchers (see Fig 2). Only 15% of respondents from the humanities are convinced that their institutional communication departments are reaching relevant stakeholders in society, compared to 30% of natural scientists and 31% of social scientists.

Fig 2. Researchers’ opinions on four core statements regarding societal impact.

Fig 2

The responses are differentiated by how much the researchers consider their work to be basic or applied research and the three disciplinary groups. The original question text was “How strongly do the following statements apply to you?.” We combined the approving answer categories to calculate the approval rates.

Researchers at applied universities agree that societal relevance should have more weight in evaluation more often than those at independent institutes and universities—the approval rates are 62%, 49%, and 53% respectively. Researchers at universities show the lowest agreement to the statement that knowledge transfer plays an important role at their institution: Only 19% approve compared to 36% at applied universities and 37% at independent institutes. Furthermore, researchers at universities particularly disagree with the statement that their communication departments are able to reach relevant stakeholders in society: Only 15% approve compared to 28% at applied universities and 44% at independent research institutes.

Both male and female researchers regard public engagement as part of scientific activity (89% and 90%). However, female researchers agree that societal relevance should be part of research evaluation more often than male researchers (62% compared to 44%). There are noteworthy differences among academic status groups: 60% of doctoral researchers but only 42% of postdocs and 47% of professors in our sample agree that societal relevance should have more weight in evaluation. Age does not seem to affect opinions on whether public engagement is part of scientific work.

Because of the differences outlined above, we decided to conduct a regression analysis (see Table 1). This confirms the descriptive observations: Humanities scholars (p = 0.01) and social scientists (p<0.01) are significantly more likely to agree that societal relevance should be part of research evaluation compared to natural scientists. Furthermore, female researchers (controlled for discipline), applied researchers, and younger researchers are significantly more in favor of including societal relevance in research evaluation.

Table 1. OLS Regression on the researchers opinion regarding evaluation.

Dep. Variable: Evaluation R-squared: 0.185
Model: OLS Adj. R-squared: 0.175
Method: Least Squares F-statistic: 18.39
No. Observations: 494 Prob (F-statistic): 2.68e-19
Df Residuals: 487 Log-Likelihood: -760.08
Df Model: 6 AIC: 1534.
Covariance Type: nonrobust BIC: 1564.

How does the discipline, gender, age, a, applied-research focus, and the opinion that science communication is an academic task influence a researcher’s opinion on whether or not societal relevance should be taken into account when evaluating research. The regression model was calculated using ordinary least squares. The reference category for the discipline is the natural sciences.

This shows that researchers are generally in favor of societal impact and regard public engagement as part of scientific activity. Fewer researchers, but still many, agree that societal impact should have more weight in research evaluations. However, this assessment differs depending on the researcher’s discipline, their applied or basic research focus, their gender, and their age. It is noteworthy that researchers (especially at universities) do not think that knowledge transfer plays an important role at their institution and that institutional communication departments are reaching relevant societal stakeholders. From this, we infer that societal impact is understood as an individual task for which there is no institutional mandate.

RQ2—goals: Disciplines define societal goals

Overall, researchers’ most important societal goal is to contribute to education (69%), followed by stimulating public discourse (55%). Equal proportions of respondents—37% each—regard contributing to informed political decision-making and to the physical and mental wellbeing of the population as a societal goal associated with their work. Given the choice of the 13 societal goals identified in the qualitative coding exercise, less than 10% of respondents selected contributing to national and/or international security and creating an entertainment offering; we hence excluded these in reporting. Fig 3 provides an overview of the societal goals the respondents could choose from by disciplinary group.

Fig 3. Research goals by disciplinary groups.

Fig 3

The radar chart illustrates which goals researchers from our three disciplinary groups pursue when communicating with nonscientific audiences. The numbers are based on the question “What social effects do you most likely want to achieve with your research? With my research, I would like to …”. The original question contained two more options (“… contribute to national and/or international security.” and “… create an entertainment offering.”), but these were chosen by less than 10% of the respondents and are therefore excluded.

It is striking that the disciplinary groups have quite different impact profiles: Scholars from the humanities tend to have culture- and discourse-oriented goals, social scientists have discourse-, social-justice-, and policy-oriented goals, and natural scientists have technology-, health-, and environment-oriented goals. Making a contribution to education is by far the most common impact goal across all disciplines (87% for the humanities, 70% for the social sciences, 58% for the natural sciences). For humanities scholars and social scientists, stimulating and supporting public discourse is among the top societal goals (80% for humanities scholars and 74% for social scientists). Furthermore, 62% of humanities scholars consider preserving cultural heritage a goal of their activities; 54% of social scientists consider strengthening the position of disadvantaged groups to be an impact goal. The main goal for natural scientists is to drive technical innovation (59%), followed by making a contribution to education (58%), and contributing to the physical and mental wellbeing of the population (46%). The differences between applied and basic researchers in terms of their societal goals are negligible.

When looking at the different types of organizations, it is noticeable that researchers from applied universities are most economically oriented: 41% of the researchers from universities of applied sciences indicate that they want to contribute to economic value creation (19% at nonuniversity institutions and 17% at universities). In contrast, researchers from independent institutes are most the policy-oriented ones: 54% of the researchers from independent institutes aim to contribute to political decision-making (31% at universities and 37% at applied universities).

Young researchers (age group 18–29 years, mostly PhD students) reported less often that they aim to contribute to public discourse (37% in comparison to 65% for 30–39-year-olds) or political decision-making (21% in comparison to 48% for 40–49-year-olds). When it comes to gender, male researchers are more inclined to pursue goals that are also related to their disciplines and vice versa. For example, technical innovation is a goal for 43% of male researchers and for 20% of female researchers. Female researchers are more interested in supporting minorities (41%) than male researchers (25%). The academic status of the respondents did not have any noteworthy effects on their goals.

From this, we conclude that societal goals are best explained by the disciplinary backgrounds of researchers. While natural scientists chose goals in the spheres of environment and health as well as innovation and economic value creation, social scientists’ goals centred more on civic society and social justice as well as politics and public discourse. Humanities scholars also chose discourse-oriented goals and focused on societal goals in the cultural sphere. Despite the differences in the definition of societal goals, it is noteworthy that making a contribution to education ranks among the top three goals for every disciplinary group. It is also evident that researchers from applied universities tend to be more economically oriented, while researchers from independent institutes tend to be more policy-oriented.

RQ3—formats: University researchers are the least active

The only transfer format that more than half of the respondents (68%) had used in the past are events (i.e., including public lectures, exhibitions, expert panel discussions). The second most used communication format is public relations (i.e., comment pieces in newspapers or interviews)—45% of respondents have used these in the past. This is followed by educational offerings in schools and for civil society (43%), social media communication (38%), advisory formats (33%), and collaborations with nonscientific partners (33%).

It is evident that humanities scholars especially use social media to communicate about their research: 55% of humanities scholars approve of such methods compared to 33% of natural scientists and 38% of social scientists. Social scientists have the most experience with advisory formats: 50% have used them compared to 28% of humanities scholars and 24% of natural scientists. Comparing applied and basic researchers, we noted that basic researchers hardly ever offer advisory formats (14% for basic researchers vs. 47% for applied researchers) or collaborations (17% for primarily basic researchers vs. 39% for primarily applied researchers).

Researchers from applied universities are the primary users of advisory formats: 57% of researchers at applied universities have used advisory formats compared to 40% of researchers at independent institutes and only 26% of the researchers at universities. Only 28% of the researchers from universities have used collaboration formats, compared to 53% of researchers from applied universities and 35% of researchers from independent institutes. As Table 2 shows, university researchers have remarkably low scores on every communication format.

Table 2. Based on the question “which transfer formats have you already used to communicate your results?” The table indicates which transfer formats researchers from the three prevailing institutional types use.

University Applied university Independent institute
Education offerings (e.g., for schools, civil society groups) 43.6% 59.2% 39.3%
Advisory formats (e.g., reports for politicians/public administration/companies/NGOs) 25.8% 57.1% 40.7%
Events (e.g., public lectures, exhibitions, expert panel discussions) 61.4% 71.4% 79.3%
Public relations (e.g., through comments in newspapers, interviews, appearances in TV programs) 38.6% 59.2% 55.0%
Social media communication (e.g., podcasts, Twitter) 31.9% 22.5% 54.3%
Collaborations with nonscientific partners (e.g., citizen science, industry partnership) 27.9% 53.1% 35.0%

Concerning researchers’ individual characteristics, the analysis shows that social media platforms are used more by younger scientists. Less surprisingly, the older a researcher is, the more likely it is that he or she has ever used a format. Regarding gender differences, more male than female researchers have used advisory formats: 43% of male researchers have done so compared to 24% of female researchers.

To conclude, we find that researchers use a variety of formats to communicate results. There are, however, some interesting differences in usage: Humanities scholars particularly tend to communicate their research via social media. Social scientists are most experienced when it comes to advisory formats. Researchers from applied universities are especially likely to have used advisory and collaboration formats. Younger researchers use social media to communicate about their research more than older researchers, and male researchers particularly tend to use advisory formats. In general, it is noteworthy that university researchers are the least active group for almost any of the formats.

Discussion and conclusions

Societal impact, and hence what academic research offers the economy, society, culture, public administration, health, environment and overall quality of life, is gaining in importance in science governance [2, 5, 6]. Likewise, the topic is on the top of the agenda of prominent policy makers and research organizations in Germany, where our study took place. If science is to be assessed based on its contribution to society, the conditions under which social impact arises should be clear. This article contributes to this discourse by addressing researchers’ perspectives on societal impact, that is, their opinions on societal impact (RQ1), the societal goals they associate with their research (RQ2), and the formats they use to engage with society (RQ3). For this reason, we conducted a survey among 499 researchers in Germany from April to June 2020.

Regarding researchers’ opinions, it is remarkable that the majority of researchers (89%) consider societal engagement to be part of scientific activity. More than half of the researchers (53%) agree that societal impact should be given more weight in evaluations. Even though the majority of researchers regard public engagement as part of scientific work, they are not equally positive about whether societal impact should have more weight in evaluations. One reason for this discrepancy may be that researchers fear that evaluations will lead to additional work or that they will not adequately record their transfer activities [23, 60, 64]. In addition, it is striking that only 27% of the respondents assume that knowledge transfer plays an important role at their institution; also 27% n believe that the institutional communication department is managing to reach relevant stakeholders in society. Humanities scholars (15%) and university researchers (15%) particularly doubt that their communication departments are reaching relevant societal stakeholders. This mirrors previous findings suggesting a certain decoupling between central transfer infrastructures and researchers [38, 39] and leads us to hypothesize that there is a certain mismatch between individual and institutional commitment.

Regarding the societal goals that researchers associate with their work, it is noteworthy that contributing to education is by far the most important goal (picked by 69% of the respondents) across all groups, followed by stimulating public discourse (55%) and contributing to informed political decision-making (37%). Moreover, it is apparent that societal goals are subject to disciplinary considerations: We show that scholars from the humanities have culture and discourse-oriented goals, social scientists have discourse-, policy- and social-justice-related goals and natural scientists have innovation- and health-related goals. This supports many theoretical reflections about the epistemic conditions for societal impact and the different roles that disciplines occupy in society [36, 56, 87]. We also find that young researchers (age 18–29) are less keen on stimulating public discourse than older ones (37% in comparison to 65% for 30–39-year-old researchers) or informing political decision-making (21% for young researchers compared to 48% for 40–49-year-old researchers). This can likely be explained by the experience that older researchers have acquired, which might make them especially well placed to influence public and political decision-making.

As far as the formats used for societal outreach are concerned, the most commonly used ones are events (used by 65% of the respondents), followed by public relations via media (45%), and educational formats for schools and civil society groups (43%). Humanities scholars especially use social media to communicate results (55% compared to 33% of natural scientists and 38% of social scientists). Basic researchers do not use advisory formats (14% vs. 47%) or collaborations (17% vs 39%) anything as often as applied researchers. Note that university researchers report remarkably low usage of any communication format compared to researchers at applied universities or at independent institutes. This might be explained by the fact that university researchers have teaching obligations not faced by those at independent institutes and thus less time for engagement activities. However, researchers from applied universities typically have a higher teaching workload. As far as researchers’ individual characteristics are concerned, it is notable that social media is used more by younger researchers, indicating that social media will likely become more important as a means of engagement.

While our results are mainly of a descriptive nature and we do not make normative assumptions about the subject of our research (i.e. “it is good to have more impact”), we can still draw some practical conclusions: First, considering the discontent with institutional communication departments, it might be worthwhile to implement decentralized support structures on the mesolevel of research organizations. This could more adequately address the complexities of the sciences and their many publics [5, 6, 57, 63]. The findings further suggest that, where applicable, organizational factors (e.g., institutional investments in transfer, training offerings, support infrastructures) should be more strongly incorporated into assessments of societal impact—for example. through formative evaluations [88]. Second, our results suggest that it is strongly advisable that evaluation exercises are responsive to disciplinary differences. For example, if economic and technical impact were the sole basis for assessing societal impact, social sciences and humanities scholars would be discriminated against [6, 63]. Our framework for societal goals and our results can also be the basis for disciplinary self-understanding (e.g., in learned societies), in that they can stimulate a normative discussion about good transfer and its evaluation. Third, considering the comparatively low importance of social media as a means of communicating about research, care should be taken not to overuse online discourse as a way of easily generating impact proxies. Moreover, our findings contribute to an informed discussion in science governance about the constraints of impact evaluation and might help impact officers and communication professionals at universities to reflect on strategies.

Our results add to the ongoing scholarly discourse on societal impact. We think that our results could bring two relevant but still separated discourses closer together, that is, the discourse on impact evaluation and the discourse on science communication/public engagement. It makes sense for critical evaluation research to make use of empirical work on the exchange between science and society. Vice versa, it makes sense to examine how evaluation policies for societal impact might affect researchers’ communication behavior. We further provide initial evidence on potentially relevant research perspectives. This concerns the organization of societal impact at scientific institutions. In addition, we suggest that national innovation systems should be studied comparatively in order to understand the impact of policies and to study more closely the relationship between transfer practices and societal impact in specific disciplines.

Acknowledgments

We would like to thank Natalya Sokolovska, Sascha Schönig, and Elias Koch for their help in coding the REF impact case studies. We would also like to thank Gert G. Wagner and Ricarda Ziegler for their friendly feedback.

Data Availability

The data underlying this study are available on Kaggle (DOI: 10.34740/kaggle/dsv/2366092).

Funding Statement

B.F. received funding by the German Federal Ministry of Education and Research (01PW18008A; 01PW18008B; https://www.wihoforschung.de/de/impaqt-2631.php). The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Brooks H. Lessons of History: Successive Challenges to Science Policy. In: Cozzens SE, Healey P, Rip A, Ziman J, editors. The Research System in Transition. Dordrecht: Springer Netherlands; 1990. pp. 11–22. doi: 10.1007/978-94-009-2091-0_2 [DOI] [Google Scholar]
  • 2.Bornmann L. What is societal impact of research and how can it be assessed? a literature survey. J Am Soc Inf Sci Technol. 2013;64: 217–233. doi: 10.1002/asi.22803 [DOI] [Google Scholar]
  • 3.Sokolovska N, Fecher B, Wagner GG. Communication on the Science-Policy Interface: An Overview of Conceptual Models. Publications. 2019;7: 64. doi: 10.3390/publications7040064 [DOI] [Google Scholar]
  • 4.Weingart P. Impact of bibliometrics upon the science system: Inadvertent consequences? Scientometrics. 2005;62: 117–131. doi: 10.1007/s11192-005-0007-7 [DOI] [Google Scholar]
  • 5.Godin B, Doré C. Measuring the impacts of science, beyond the economic dimensions. 2005. Available: http://www.csiic.ca/ PDF/Godin_Dore_Impacts.pdf Göransson, B., Maharajh, R., & S [Google Scholar]
  • 6.Reale E, Avramov D, Canhial K, Donovan C, Flecha R, Holm P, et al. A review of literature on evaluating the scientific, social and political impact of social sciences and humanities research. Res Eval. 2018;27: 298–308. doi: 10.1093/reseval/rvx025 [DOI] [Google Scholar]
  • 7.Australian Research Council. State of Australian University Research 2018–19. 2018 [cited 17 Dec 2020]. Available: https://dataportal.arc.gov.au/ERA/NationalReport/2018/
  • 8.Martin BR. The Research Excellence Framework and the “impact agenda”: are we creating a Frankenstein monster? Res Eval. 2011;20: 247–254. doi: 10.3152/095820211X13118583635693 [DOI] [Google Scholar]
  • 9.Sousa SB, Brennan JL. The UK Research Excellence Framework and the Transformation of Research Production. In: Musselin C, Teixeira PN, editors. Reforming Higher Education. Dordrecht: Springer Netherlands; 2014. pp. 65–80. doi: 10.1007/978-94-007-7028-7_4 [DOI] [Google Scholar]
  • 10.van der Meulen B, Rip A. Evaluation of societal quality of public sector research in the Netherlands. Res Eval. 2000;9: 11–25. doi: 10.3152/147154400781777449 [DOI] [Google Scholar]
  • 11.VSNU, NWO, KNAW. Standard Evaluation Protocol 2015–2021: Protocol for Research Assessments in the Netherlands. 2016. Available: https://www.vsnu.nl/files/documenten/Domeinen/Onderzoek/SEP2015-2021.pdf
  • 12.European Commission, editor. Open innovation, open science, open to the world: a vision for Europe. Luxembourg: Publications Office of the European Union; 2016. [Google Scholar]
  • 13.European Commission. Press release—Scientific data: open access to research results will boost Europe’s innovation capacity. 2012. Available: http://europa.eu/rapid/press-release_IP-12-790_en.htm
  • 14.European Commission. Assessing Europe’s university-based research: expert group on assessment of university-based research. Publications Office; 2010. Available: https://data.europa.eu/doi/10.2777/80193
  • 15.Allianz der Wissenschaftsorganisationen. Stellungnahme zum 9. EU-Rahmenprogramm für Forschung und Innovation Horizon Europe (FP9). München; 2018. Available: https://www.leopoldina.org/fileadmin/redaktion/Publikationen/Allianz/201806_Stellungnahme_9FRP_der_Allianz.pdf
  • 16.BMBF. Grundsatzpapier des -Bundesministeriums für Bildung und Forschung zur Wissenschaftskommunikation. Berlin: Federal Ministry of Education and Research; 2019. p. 7. [Google Scholar]
  • 17.BMBF. Karliczek: Wissenschaftskommunikation stärker in der Wissenschaft verankern. 2020. Available: https://www.bmbf.de/de/karliczek-wissenschaftskommunikation-staerker-in-der-wissenschaft-verankern-12620.html
  • 18.German Research Council. Mehr Anerkennung für Wissens- und Technologietransfer (transl. More recognition for knowledge and technology transfer). 2016. Available: https://www.wissenschaftsrat.de/download/archiv/pm_2816.pdf?__blob=publicationFile&v=2
  • 19.German Rectors’ Conference. Transfer und Kooperation als Aufgaben der Hochschulen (transl. Transfer and cooperation as tasks of the universities). 2017 [cited 17 Dec 2020]. Available: https://www.hrk.de/positionen/beschluss/detail/transfer-und-kooperation-als-aufgaben-der-hochschulen
  • 20.Leibniz Association. » Transfer in der Leibniz- Gemeinschaft entfaltet sich in seiner vollen Bandbreite vom Technologietransfer bis hin zur Gesellschafts- und Politikberatung. «. 2019; 28.
  • 21.Martin BR. The Changing Social Contract for Science and the Evolution of the University. Science and Innovation. Edward Elgar Publishing; 2003. p. 2831. doi: 10.4337/9781781950241.00011 [DOI] [Google Scholar]
  • 22.Gibbons M. Science’s new social contract with society. Nature. 1999;402: C81–C84. doi: 10.1038/35011576 [DOI] [PubMed] [Google Scholar]
  • 23.Lauronen J-P. The dilemmas and uncertainties in assessing the societal impact of research. Sci Public Policy. 2020;47: 207–218. doi: 10.1093/scipol/scz059 [DOI] [Google Scholar]
  • 24.DG PuK, DGS. Positionierung der Deutschen Gesellschaft für Publizistik und Kommunikationswis‐ senschaft (DGPuK) und der Deutschen Gesellschaft für Soziologie (DGS) zum Grundsatzpapier des Bundesministeriums für Bildung und Forschung (BMBF) zur Wissenschaftskommunikation1. 2020. Available: https://www.dgpuk.de/sites/default/files/Stellungnahme_WissKomm_DGPuK_DGS_2020_20200821.pdf
  • 25.Gibbons M, editor. The new production of knowledge: the dynamics of science and research in contemporary societies. London; Thousand Oaks, Calif: SAGE Publications; 1994. [Google Scholar]
  • 26.Nowotny H, Scott P, Gibbons M. “Mode 2” Revisited: The New Production of Knowledge. Minerva. 2003;41: 179–194. doi: 10.1023/A:1025505528250 [DOI] [Google Scholar]
  • 27.Slaughter S, Leslie LL. Academic capitalism: politics, policies, and the entrepreneurial university. Baltimore: Johns Hopkins University Press; 1997. [Google Scholar]
  • 28.Dankel DJ, Vaage NS, van der Sluijs JP. Post-normal science in practice. Futures. 2017;91: 1–4. doi: 10.1016/j.futures.2017.05.009 [DOI] [Google Scholar]
  • 29.Leydesdorff L, Etzkowitz H. The Triple Helix as a model for innovation studies. Sci Public Policy. 1998. [cited 17 Sep 2020]. doi: 10.1093/spp/25.3.195 [DOI] [Google Scholar]
  • 30.Bonfadelli H, Fähnrich B, Lüthje C, Milde J, Rhomberg M, Schäfer MS, editors. Forschungsfeld Wissenschaftskommunikation. Wiesbaden: Springer Fachmedien Wiesbaden; 2017. doi: 10.1007/978-3-658-12898-2 [DOI] [Google Scholar]
  • 31.Bucchi M, Trench B. Routledge handbook of public communication of science and technology. Place of publication not identified: TAYLOR & FRANCIS; 2019. [Google Scholar]
  • 32.Scheufele DA. Beyond the Choir? The Need to Understand Multiple Publics for Science. Environ Commun. 2018;12: 1123–1126. doi: 10.1080/17524032.2018.1521543 [DOI] [Google Scholar]
  • 33.Fischhoff B. The sciences of science communication. Proc Natl Acad Sci. 2013;110: 14033–14039. doi: 10.1073/pnas.1213273110 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Simis MJ, Madden H, Cacciatore MA, Yeo SK. The lure of rationality: Why does the deficit model persist in science communication? Public Underst Sci. 2016;25: 400–414. doi: 10.1177/0963662516629749 [DOI] [PubMed] [Google Scholar]
  • 35.Greenhalgh T, Jackson C, Shaw S, Janamian T. Achieving Research Impact Through Co‐creation in Community‐Based Health Services: Literature Review and Case Study. Milbank Q. 2016;94: 392–429. doi: 10.1111/1468-0009.12197 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Nowotny H. The Increase of Complexity and its Reduction: Emergent Interfaces between the Natural Sciences, Humanities and Social Sciences. Theory Cult Soc. 2005;22: 15–31. doi: 10.1177/0263276405057189 [DOI] [Google Scholar]
  • 37.Olsson AK, Bernhard I, Arvemo T, Lundh Snis U. A conceptual model for university-society research collaboration facilitating societal impact for local innovation. Eur J Innov Manag. 2020;ahead-of-print. doi: 10.1108/EJIM-04-2020-0159 [DOI] [Google Scholar]
  • 38.Peters HP. Gap between science and media revisited: Scientists as public communicators. Proc Natl Acad Sci. 2013;110: 14102–14109. doi: 10.1073/pnas.1212745110 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Schwetje T, Hauser C, Böschen S, Leßmöllmann A. Communicating science in higher education and research institutions: An organization communication perspective on science communication. J Commun Manag. 2020;24: 189–205. doi: 10.1108/JCOM-06-2019-0094 [DOI] [Google Scholar]
  • 40.Kohring M, Marcinkowski F, Lindner C, Karis S. Media orientation of German university decision makers and the executive influence of public relations. Public Relat Rev. 2013;39: 171–177. doi: 10.1016/j.pubrev.2013.01.002 [DOI] [Google Scholar]
  • 41.Besley JC, Nisbet M. How scientists view the public, the media and the political process. Public Underst Sci. 2013;22: 644–659. doi: 10.1177/0963662511418743 [DOI] [PubMed] [Google Scholar]
  • 42.Könneker C, Niemann P, Böhmert C. External Science Communication by High Profile Digital Native Researchers: Engagement and Attitudes. ECREA; 2018; Lugano. [Google Scholar]
  • 43.Bentley P, Kyvik S. Academic staff and public communication: a survey of popular science publishing across 13 countries. Public Underst Sci. 2010;20. 10.1177/0963662510384461 [DOI] [Google Scholar]
  • 44.Poliakoff E, Webb TL. What Factors Predict Scientists’ Intentions to Participate in Public Engagement of Science Activities? Sci Commun. 2007;29: 242–263. doi: 10.1177/1075547007308009 [DOI] [Google Scholar]
  • 45.Valinciute A. Lithuanian scientists’ behavior and views on science communication. Public Underst Sci. 2020;29: 353–362. doi: 10.1177/0963662520907001 [DOI] [PubMed] [Google Scholar]
  • 46.Post S. Communicating science in public controversies: Strategic considerations of the German climate scientists. Public Underst Sci. 2016;25: 61–70. doi: 10.1177/0963662514521542 [DOI] [PubMed] [Google Scholar]
  • 47.The Royal Society. Sciennce Communication: Survey of factors affecting science communication by scientists and engineers. 2006. Available: https://royalsociety.org/~/media/Royal_Society_Content/policy/publications/2006/1111111395.pdf
  • 48.Martín-Sempere MJ, Garzón-García B, Rey-Rocha J. Scientists’ motivation to communicate science and technology to the public: surveying participants at the Madrid Science Fair. Public Underst Sci. 2008;17: 349–367. doi: 10.1177/0963662506067660 [DOI] [Google Scholar]
  • 49.van de Burgwal LHM, Hendrikse R, Claassen E. Aiming for impact: Differential effect of motivational drivers on effort and performance in knowledge valorisation. Sci Public Policy. 2019;46: 747–762. doi: 10.1093/scipol/scz027 [DOI] [Google Scholar]
  • 50.Besley JC, Tanner AH. What Science Communication Scholars Think About Training Scientists to Communicate. Sci Commun. 2011;33: 239–263. doi: 10.1177/1075547010386972 [DOI] [Google Scholar]
  • 51.Friesike S, Fecher B, Wagner GG. Teach young scientists the importance of societal impact for research. Nature. 2018;554: 300–300. doi: 10.1038/d41586-018-02066-2 [DOI] [PubMed] [Google Scholar]
  • 52.Gerrits RG, Mulyanto J, Wammes JD, van den Berg MJ, Klazinga NS, Kringos DS. Individual, institutional, and scientific environment factors associated with questionable research practices in the reporting of messages and conclusions in scientific health services research publications. BMC Health Serv Res. 2020;20. doi: 10.1186/s12913-019-4859-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Marcinkowski F, Kohring M, Fürst S, Friedrichsmeier A. Organizational Influence on Scientists’ Efforts to Go Public: An Empirical Investigation. Sci Commun. 2014;36: 56–80. doi: 10.1177/1075547013494022 [DOI] [Google Scholar]
  • 54.Huutoniemi K. Interdisciplinarity as Academic Accountability: Prospects for Quality Control Across Disciplinary Boundaries. Soc Epistemol. 2016;30: 163–185. doi: 10.1080/02691728.2015.1015061 [DOI] [Google Scholar]
  • 55.Dahler-Larsen P. The Evaluation Society. 1st ed. Stanford University Press; 2011. doi: 10.1007/s12115-010-9383-6 [DOI] [Google Scholar]
  • 56.Weingart P. Science and the media. Res Policy. 1998;27: 869–879. doi: 10.1016/S0048-7333(98)00096-1 [DOI] [Google Scholar]
  • 57.Donovan C, Gulbrandsen M. Introduction: Measuring the impact of arts and humanities research in Europe. Res Eval. 2018;27: 285–286. doi: 10.1093/reseval/rvy019 [DOI] [Google Scholar]
  • 58.Ochsner M, Hug S, Galleron I. The future of research assessment in the humanities: bottom-up assessment procedures. Palgrave Commun. 2017;3: 17020. doi: 10.1057/palcomms.2017.20 [DOI] [Google Scholar]
  • 59.Hug SE, Ochsner M, Daniel H-D. Criteria for assessing research quality in the humanities: a Delphi study among scholars of English literature, German literature and art history. Res Eval. 2013;22: 369–383. doi: 10.1093/reseval/rvt008 [DOI] [Google Scholar]
  • 60.Benneworth P. Tracing how arts and humanities research translates, circulates and consolidates in society. How have scholars been reacting to diverse impact and public value agendas? Arts Humanit High Educ. 2015;14: 45–60. doi: 10.1177/1474022214533888 [DOI] [Google Scholar]
  • 61.Molas-Gallart J, Tang P. Tracing “productive interactions” to identify social impacts: an example from the social sciences. Res Eval. 2011;20: 219–226. doi: 10.3152/095820211X12941371876706 [DOI] [Google Scholar]
  • 62.Spaapen J, van Drooge L. Introducing ‘productive interactions’ in social impact assessment. Res Eval. 2011. Available: https://academic.oup.com/rev/issue/20/3 doi: 10.3152/095820211X12941371876580 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Benneworth P, Olmos-Peñuela J. Reflecting on the Tensions of Research Utilization: Understanding the Coupling of Academic and User Knowledge. Sci Public Policy. 2018. [cited 5 Jun 2019]. doi: 10.1093/scipol/scy021 [DOI] [Google Scholar]
  • 64.de Jong S, Barker K, Cox D, Sveinsdottir T, Van den Besselaar P. Understanding societal impact through productive interactions: ICT research as a case. Res Eval. 2014;23: 89–102. doi: 10.1093/reseval/rvu001 [DOI] [Google Scholar]
  • 65.Cohn RC. Von der Psychoanalyse zur themenzentrierten Interaktion: von der Behandlung einzelner zu einer Pädagogik für alle. 19. Auflage. Stuttgart: Klett-Cotta; 2018. [Google Scholar]
  • 66.Scharer M. Theme-Centered Interaction by Ruth C. Cohn: An Introduction. In: Meyerhuber S, Reiser H, Scharer M, editors. Theme-Centered Interaction (TCI) in Higher Education. Cham: Springer International Publishing; 2019. pp. 57–95. doi: 10.1007/978-3-030-01048-5_2 [DOI] [Google Scholar]
  • 67.Dobusch L, Kremser W, Seidl D, Werle F. A communication perspective on open strategy and open innovation. Management for schung. 2017;27: 5–25. doi: 10.1365/s41113-017-0015-6 [DOI] [Google Scholar]
  • 68.Luhmann N. Social Systems. Stanford, CA: Stanford University Press; 1995. [Google Scholar]
  • 69.Becher T. The significance of disciplinary differences. Stud High Educ. 1994;19: 151–161. doi: 10.1080/03075079412331382007 [DOI] [Google Scholar]
  • 70.Hamlyn B, Hanson T, Mortimer E, Lewis H, O’Donoghue E, Burchell K. Factors affecting public engagement by researchers A study on behalf of a Consortium of UK public research funders (Technical Report). TNS BMRB; Policy Studies Institute; 2015. [Google Scholar]
  • 71.Heimeriks G, van den Besselaar P, Frenken K. Digital disciplinary differences: An analysis of computer-mediated science and ‘Mode 2’ knowledge production. Res Policy. 2008;37: 1602–1615. doi: 10.1016/j.respol.2008.05.012 [DOI] [Google Scholar]
  • 72.Lewis JM, Ross S, Holden T. The how and why of academic collaboration: disciplinary differences and policy implications. High Educ. 2012;64: 693–708. doi: 10.1007/s10734-012-9521-8 [DOI] [Google Scholar]
  • 73.Pettibone L, Vohland K, Ziegler D. Understanding the (inter)disciplinary and institutional diversity of citizen science: A survey of current practice in Germany and Austria. van den Besselaar P, editor. PLOS ONE. 2017;12: e0178778. doi: 10.1371/journal.pone.0178778 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Duch J, Zeng XHT, Sales-Pardo M, Radicchi F, Otis S, Woodruff TK, et al. The Possible Role of Resource Requirements and Academic Career-Choice Risk on Gender Differences in Publication Rate and Impact. Perc M, editor. PLoS ONE. 2012;7: e51332. doi: 10.1371/journal.pone.0051332 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Fecher B, Friesike S, Hebing M, Linek S, Sauermann A. A Reputation Economy: Results from an Empirical Survey on Academic Data Sharing. DIW Berl Discuss Pap. 2015;1454. Available: 10.2139/ssrn.2568693 [DOI] [Google Scholar]
  • 76.Holden C. General Contentment Masks Gender Gap in First AAAS Salary and Job Survey. Science. 2001;294: 396–411. doi: 10.1126/science.294.5541.396 [DOI] [PubMed] [Google Scholar]
  • 77.Linek SB, Fecher B, Friesike S, Hebing M. Data sharing as social dilemma: Influence of the researcher’s personality. Lozano S, editor. PLOS ONE. 2017;12: e0183216. doi: 10.1371/journal.pone.0183216 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Moss-Racusin CA, Dovidio JF, Brescoll VL, Graham MJ, Handelsman J. Science faculty’s subtle gender biases favor male students. Proc Natl Acad Sci. 2012;109: 16474–16479. doi: 10.1073/pnas.1211286109 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.National Research Council, Committee on Women in Science, Engineering, and Medicine, Statistics C on N, Policy and Global Affairs, Division of Behavioral and Social Sciences and Education, Committee on Gender Differences in the Careers of Science, Engineering, and Mathematics Faculty. Gender Differences at Critical Transitions in the Careers of Science, Engineering, and Mathematics Faculty. Washington, D.C.: National Academies Press; 2010. Available: http://www.nap.edu/catalog/12062 [Google Scholar]
  • 80.Fecher B, Friesike S, Hebing M, Linek S. A reputation economy: how individual reward considerations trump systemic arguments for open access to data. Palgrave Commun. 2017;3: 17051. doi: 10.1057/palcomms.2017.51 [DOI] [Google Scholar]
  • 81.Reichard B, Reed MS, Chubb J, Hall G, Jowett L, Peart A, et al. Writing impact case studies: a comparative study of high-scoring and low-scoring case studies from REF2014. Palgrave Commun. 2020;6: 31. doi: 10.1057/s41599-020-0394-7 [DOI] [Google Scholar]
  • 82.Mayring P. Qualitative Content Analysis. Forum Qual Sozialforschung Forum Qual Soc Res. 2000;1. Available: http://www.qualitative-research.net/index.php/fqs/article/view/1089 [Google Scholar]
  • 83.Kuckartz U. Qualitative text analysis: a guide to methods, practice & using software. 2013.
  • 84.DFG. DFG-Fachsystematik der Wissenschaftsbereiche, Fachgebiete, Fachkollegien und Fächer für die Amtsperiode 2016–2019. 2017. Available: https://www.dfg.de/download/pdf/dfg_im_profil/gremien/fachkollegien/amtsperiode_2016_2019/fachsystematik_2016–2019_de_grafik.pdf
  • 85.Destatis. Personal an Hochschulen—Fachserie 11 Reihe 4.4–2018. In: Statistisches Bundesamt [Internet]. 2019 [cited 18 Dec 2020]. Available: https://www.destatis.de/DE/Themen/Gesellschaft-Umwelt/Bildung-Forschung-Kultur/Hochschulen/Publikationen/Downloads-Hochschulen/personal-hochschulen-2110440187004.html
  • 86.DFG. Chancengleichheits-Monitoring 2018. Bonn; 2017.
  • 87.Bornmann L. Do altmetrics point to the broader impact of research? An overview of benefits and disadvantages of altmetrics. J Informetr. 2014;8: 895–903. doi: 10.1016/j.joi.2014.09.005 [DOI] [Google Scholar]
  • 88.Murray HG. The impact of formative and summative evaluation of teaching in North American universities. Assess Eval High Educ. 1984;9: 117–132. doi: 10.1080/0260293840090204 [DOI] [Google Scholar]

Decision Letter 0

Laurentiu Rozylowicz

24 Mar 2021

PONE-D-21-03474

How Do Researchers Achieve Societal Impact? Results of an Empirical Survey Among Researchers in Germany

PLOS ONE

Dear Dr. Fecher,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Reviewers agreed that the paper is valuable; however, they suggest several revision points. I advise you to incorporate as many suggestions as possible, and when not possible, please explain why. Please enhance the method section by clearly explain the statistics and the survey. A more detailed introduction and discussion are needed.

Please submit your revised manuscript by May 08 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Laurentiu Rozylowicz, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2/ Please improve statistical reporting and refer to p-values as "p<.001" instead of "p=.000". Our statistical reporting guidelines are available at https://journals.plos.org/plosone/s/submission-guidelines#loc-statistical-reporting

3. Please include a copy of Table 3 which you refer to in your text on page 19.

4. Please ensure that you refer to Figure xxxxx in your text as, if accepted, production will need this reference to link the reader to the figure.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The research topic of the study is relevant both academically and socially. The authors bring out the important question how researchers understand their efforts to engage in non-academic organizations and institutions and their possibilities to achieve social impact via their research organizations. They point out rather interesting results. Statistical research on research impact is clearly underdeveloped and needs more studies like this. However, the manuscript has several major problems in methodical soundness and in the theoretical framework, which is not connecting to the results or to the academic discussion about research impact.

There are four major problems that the paper needs to address:

1. The whole analysis section is missing regarding the statistical analysis. The reader cannot see how the authors conducted the study. There is only one section where the authors uses a regression analysis, while the whole analysis is descriptive. It needs a good reason to ground such an analytical choice. The statistical description requires re-phrasing and the tables need reformulations. It is rather difficult to understand the authors’ writing through these expressions and the table.

2. Because the questionnaire of this research is self-designed, there should be description of its validity and reliability, which are now missing.

3. Because the method, analysis and results are insufficiently presented, I strongly recommend to ask an expert in statistics to re-check and revise these sections. I believe such expertise would help to improve the readability, validity and reliability of this article.

4. The theoretical framework relating to science communication in the manuscript does not connect to the results of the study. The discussion section lacks correspondence to academic discussion about research impact. There are several theoretical approaches to social impact of research or science communication and the authors should point out what point of view their hypothesis and results are supporting. The interpretation of the results lacks a meaningful theoretical framework, although the study presents a framework for science communication. I recommend to structure a hypothesis vis-à-vis science communication and interpret the results by this framework for the study.

I suggest some minor recommendations here:

- I recommend that a professional language service would proof read this article.

- The title of the paper corresponds weakly to the study, as the paper studies people’s opinions and understanding of their engagement, not how they actually achieve impact.

- The study utilizes the categories of the Research Excellence Framework (REF) from the United Kingdom. However, the REF cannot be argued to be without normative assumptions when it is a result of policies. The utilization in analysis needs some theoretical justification for using such criteria for impact.

- Basic and applied distinction is not always clear and divergent disciplines may have differing opinions about this. Use of such distinction needs theoretical support.

- The questionnaire has a section considering respondent’s personality. I cannot see this data in the paper and the meaning of such data for the paper.

- What do the authors mean by: "...by the socio-demographic factors: gender, status, and age."? In addition what “status”, do you mean employment status? Please, be more specific of these terms.

- Expression “In relation to the engagement with society” is unclear. It can be read that research is not part of society.

- Did you select the universities according to your criteria for the selection among all universities in Germany?

- Your description of semi-convenient/convenient sample is not consistent in the paper.

- The result section is rather long, and some parts are quite repetitive. Shortening it could make it more reader-friendly.

- Description of the REF coding needs to be placed in the method section.

- Table 2 title needs to be more specific.

- Table 3 is missing completely.

- What does “sustainable strategy for impact” mean in this context?

- What is specific impact in relation to disciplines? I don't think reference to a policy paper is relevant in discussion part. It should have discussion with the science communication literature or impact literature.

- In the appendix figure, there is a wrong title for the bar chart: basic vs. applied.

Reviewer #2: Thanks for the opportunity to review your manuscript entitled "How do researchers achieve societal impact? Results of an empirical survey among researchers in Germany" in which you have identified and argued for an important research gap. I really enjoyed reading the manuscript and my main comments and suggestions below are merely about elaborating and developing your text a bit more.

Consider the following in your revision of the paper:

INTRODUCTION

In the section SOCIETAL IMPACT IN RESEARCH p 10 ff you may problematize and position your study

more by (i) a deeper description existing research of the different concepts of academic knowledge

creation see e.g., Olsson et al 2020, (2020). A conceptual model for university-society

research collaboration facilitating societal impact for local innovation. European Journal of Innovation

Management. https://doi-org.ezproxy.server.hv.se/10.1108/EJIM-04-2020-0159 and (ii) earlier

research on measuring and defining research impact e.g., De Jong, et al (2014) “Understanding

societal impact through productive interactions: CT research as a case”, Research Evaluation, Vol. 23

No. 2, pp. 89-102; Greenhalgh, T., Jackson, C. et al (2016), “Achieving research impact through co-

creationin community-based health services: literature review and case study”, The Milbank

Quarterly, Vol. 94 No. 2, pp. 392-429; Matsumoto, M., et al (2010), “Development of a model to estimate the economic impacts of R&D output of public research institutes”, R & D Management, Vol. 40 No. 1, pp. 91-100. Issues of co-creation in university-society collaboration is also something you may find interesting see e.g., Larsson, J. and Holmberg, J. (2018), “Learning while creating value for sustainability transitions: the case of challenge lab at Chalmers university of technology”, Journal of Cleaner Production, Vol. 172, pp. 4411-4420, and Olsson et al 2020 as above.

AIM – there is a slight difference between the aim on page 5 and page 19.

RQ1 includes the words engagement, evaluation and press departments that need further explanation and linked to earlier research in the introduction section.

This is also reoccurring in the Method section p. 7 as the importance of communication and

performance of communication department- is there any earlier research on these aspects then it should be mentioned in the Introduction.

P. 6 define status in the individual dimension.

P. 12 The status groups -presented differently throughout the paper. A logical presentation of the groups is recommended along with an early definition in the paper of how you define a ‘researcher’.

The age of 18 seem to be very young to be included in the group of researchers?

Methodology and results are presented and illustrated in a proper way.

Fig 1 Institutional types- may the categories by merged into three categories instead?

The manuscript ends with a Discussion – add Conclusion in the heading or add another subsection entitled Conclusion. There are very few references in your Discussion. Return to your earlier refences in the Introductions section and compare your findings to earlier research.

Good luck with your revision!

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Juha-Pekka Lauronen

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Jul 9;16(7):e0254006. doi: 10.1371/journal.pone.0254006.r002

Author response to Decision Letter 0


8 May 2021

Dear Dr. Rozylowicz, dear colleagues,

we were delighted about the positive and helpful reviews we received. We substantially revised the manuscript accordingly.

Sincerely,

Benedikt Fecher

Attachment

Submitted filename: response_PONE-D-21-03474.docx

Decision Letter 1

Laurentiu Rozylowicz

28 May 2021

PONE-D-21-03474R1

How Do Researchers Approach Societal Impact?

PLOS ONE

Dear Dr. Fecher,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

The revision greatly improved the paper, and both reviewers had positive comments. However, there is a need for some minor adjustments for greater clarity and better reproducibility. Please try to include the suggestions in your paper, and, if not suitable, please answer in detail why. Also, please proofread the paper carefully before submitting it.

Please submit your revised manuscript by Jul 12 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Laurentiu Rozylowicz, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: PONE-D-21-03474R1

How Do Researchers Approach Societal Impact?

PLOS ONE

Review

The

The paper and its methodical choices have improved significantly, and that is why, I suggest only some minor revisions by pointing out small possible improvements and recommendations. The study design still has some minor questions related to validity and reliability, which could be clarified and elaborated more. The findings and conclusions of the study are a bit trivial because they do not debate openly with the previous literature and the background literature of the study. Connection to the previous debate could elevate the substantial value of the paper.

1. Pretests: The authors explain that they have done pretests for reliability, but they do not elaborate what result the test provided for reliability, though they explain what they did based on those tests.

2. In the sample design, the authors describe that they selected the universities based on number of students. Why do they not rely it on researchers, teachers and other staff of scholars when the study is about researchers’ attitudes?

3. In the page 19, the expression sounds ambiguous: “Fewer researchers, but still many,…”

4. In the results section (RQ 1) the authors conclude that: “It is noteworthy that researchers (especially at universities) do not think that knowledge transfer plays an important role at their institution and that institutional communication departments are reaching relevant societal stakeholders. From this, we infer that societal impact is understood as an individual task for which there is no institutional mandate.

It is a bit unclear in this argument whether it means understanding of these institutions or the understanding of the researchers in the survey. If this means only the research staff’s attitudes, I suggest to formulate the conclusion differently, so that it is clear that it describes only the staff’s believes. Of course these two dimensions could be compared preferably.

This conclusion of neglected organizational aspects of impact requires information about the actual knowledge transfer activities of the survey institutions, and how much emphasis they have on impact evaluation and practices. Now the authors provide information about general policy recommendations and considerations, but this does not describe what kind of practices the universities have regarding impact requirements, their organization and evaluation, and their relationship with the survey attitudes.

Perhaps, the authors could provide some basic information of how impact is considered in the university and research work in Germany.

5. The claim “but these were chosen by less than 10% of the respondents

and are therefore excluded” is mentioned twice in the same context.

6. In the results (RQ2), the authors claim that “It is striking that the disciplinary groups have quite different impact profiles: Scholars from the humanities tend to have culture- and discourse-oriented goals, social scientists have discourse-, social-justice-, and policy-oriented goals, and natural scientists have technology-, health-, and environment-oriented goals.

This claim is one of the main results in the paper. However, there is nothing striking in this argument according to common knowledge of disciplinary practices and recent studies on research impact and public engagement. Perhaps, the authors could emphasize how these findings support previous understanding of the characteristics of these disciplines and their public engagement.

7. Then the authors conclude that: “From this, we conclude that societal goals are best explained by the disciplinary backgrounds of researchers.”

Because several authors have already pointed out this argument in recent literature (e.g. Benneworth) and in older classics (e.g. Weiss), the authors could elaborate this statement by linking it to recent literature, which is worried about impact policies neglecting such disciplinary goals and their importance.

8. The results (RQ3) bring out education offerings as an important perception of impact. I was wondering why this does not seem to consider the education of the institutions themselves, as they provide a great number of new graduates for employment each year.

9. Usually some sort of limits of the study should be brought out in the conclusions.

10. The conclusions seem a bit unfinished. Maybe the main conclusion is that disciplines have had their own impact goals in relation to contemporary society all along, and now all the responsibilities are suddenly pressured against researchers without proper institutional structures. On the other hand, what are the impact goals of the organizations and do they correspond to disciplines’ goals? As I pointed out in the sixth comment, linking the findings to recent studies on the disputes about impact goals and neglection of disciplinary goals, to older discussion about the difference between research utilization of various academic fields and to the structural/organizational levels of research impact could benefit and increase the value of the paper.

10. The authors point out: certain mismatch between individual and institutional commitment. However, what are the actual institutional demands from individuals?

11. I find this interesting: "Humanities scholars especially use social media to communicate results."

Perhaps, the authors could speculate a bit why the survey has such a surprising result.

12. This is unclear: researchers have teaching obligations not faced by those at independent institutes and thus less time for engagement activities. However, researchers from applied universities typically have a higher teaching workload.

What does this mean regarding engagement activities?

13. I find this statement unnecessary, as the paper clearly has a normative tone: While our results are mainly of a descriptive nature and we do not make normative assumptions about the subject of our research (i.e. “it is good to have more impact…

14. “The findings further suggest that, where applicable, organizational factors (e.g., institutional investments in transfer, training offerings, support infrastructures) should be more strongly incorporated into assessments of societal impact”

I don’t understand how the authors came to this conclusions. Why should they be part of assessments of impact? Why is impact assessment the perfect way to promote public engagement? There are also other ways to organize and promote knowledge in society.

15. This gives a misleading impression that innovation is same than impact: “In addition, we suggest that national innovation systems should be studied comparatively in order to understand the impact of policies and to study more closely the relationship between transfer practices and societal impact in specific disciplines.”

Some minor notions:

- “An example of this is Jong et al.'s concept of productive interactions…”

Productive interactions is a concept of Spaapen and van Drooge (2011).

- “According to Bornmann, three main strands of societal impact can be distinguished:”.

This is probably de Jong et al. (2014) interpretation based on Bornmann’s conceptualization.

- The instrument: “Nevertheless, as researchers’ expressions of impact, they provide a suitable textual data basis for developing goal and format categories”

Some studies (Watermeyer et al.) have shown that the REF case studies construct unsophisticated imaginaries of impact. This problematics could be mentioned regarding the evaluation competition.

Reviewer #2: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Juha-Pekka Lauronen

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Jul 9;16(7):e0254006. doi: 10.1371/journal.pone.0254006.r004

Author response to Decision Letter 1


14 Jun 2021

Again, thank you for this valuable and constructive feedback which we feel added substantially to the quality of the paper!

Attachment

Submitted filename: Response_to_reviewers_PONE-D-21-03474.docx

Decision Letter 2

Laurentiu Rozylowicz

18 Jun 2021

How Do Researchers Approach Societal Impact?

PONE-D-21-03474R2

Dear Dr. Fecher,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Laurentiu Rozylowicz, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Laurentiu Rozylowicz

1 Jul 2021

PONE-D-21-03474R2

How do researchers approach societal impact?

Dear Dr. Fecher:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Laurentiu Rozylowicz

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: response_PONE-D-21-03474.docx

    Attachment

    Submitted filename: Response_to_reviewers_PONE-D-21-03474.docx

    Data Availability Statement

    The data underlying this study are available on Kaggle (DOI: 10.34740/kaggle/dsv/2366092).


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES