Skip to main content
Open Research Europe logoLink to Open Research Europe
. 2022 Apr 25;1:71. Originally published 2021 Jun 18. [Version 2] doi: 10.12688/openreseurope.13621.2

Measuring the impact of publicly funded open innovation programmes: the case of Data Market Services Accelerator

Maria Priestley 1,a, Elena Simperl 1, Cristina Juc 2, María Anguiano 3
PMCID: PMC10445846  PMID: 37645209

Version Changes

Revised. Amendments from Version 1

The new version of the article contains the following updates: - Our claims about the generalisability of the framework and the strength of claims regarding our results have been reduced. We have also highlighted that our manuscript focuses on measuring the performance of only one programme – the DMS Accelerator. - More references have been added to support our claims, particularly in the Introduction section. - We have provided more examples of open innovation, and have justified our decision to describe DMS and other similar initiatives as examples of open innovation, rather than relying on traditional notions of acceleration and incubation. We have also clarified that the scope of our paper focuses on initiatives that serve startups and SMEs, which are only some of the various possible stakeholders that can be targeted by open innovation initiatives. - We have obtained additional data, which has increased our survey sample size from 20 to 47 respondents across the three DMS cohorts. We have also reported the new results using actual numbers (rather than percentages) for transparency. - We have now added a section on the “Limitations and future work”.

Abstract

One of the current goals of the European Commission is to stimulate the development and uptake of data and AI technologies in the economy. Substantial funding is being invested in open innovation programmes that help startups and small-medium enterprises (SMEs) to assimilate the latest technical and regulatory trends, and to share their innovations with other organisations. In order to assess the efficacy and impact of such initiatives, their specific social and economic objectives must be taken into consideration. As an example of how this can be done in practice, our paper presents the motivating objectives and methodological approaches that were used to assess the impact of the Data Market Services Accelerator (DMS), an EU-funded initiative for data-centric companies. We evaluated the performance of the programme in terms of its effect on the market, fundraising capabilities of companies, innovation, and socio-economic aspects. In addition to assessing how DMS was able to meet its intended objectives, our examination also underscored current challenges related to specific outcomes that are meaningful to the European Commission, including data standardisation and long-term legal strategy. We conclude the paper with a series of recommendations to support the impact assessment efforts of other similar innovation programmes.

Keywords: open innovation, data ecosystems, business accelerators, impact assessment, startups, SMEs

Executive summary

This paper discusses current challenges related to the measurement of impact in open innovation programmes for datacentric startups and SMEs. Drawing on our experience from the Data Market Services (DMS) Accelerator, we discuss some of the approaches and methods that can help to assess progress towards the specific objectives of publicly funded data innovation programmes in Europe.

Publicly funded data innovation programmes embody a novel form of support for startups and SMEs. While these initiatives share traits in common with traditional business accelerators and incubators in terms of the services offered, public funders strive to fulfil a unique set of objectives. In addition to the financial return generated by individual firms, such initiatives have longer-term goals related to the creation of employment and regional development. Another goal particular to big data initiatives is the creation of data value chains by means of interoperable products and services. Ambitions of this kind require specialist training in legal strategy and standardisation. However, the substantial time and resource investment required from startups to participate in this long-term vision is not always aligned with short-term business goals and profitability. Impact assessments must therefore consider performance at the individual company and ecosystem levels, as well as deriving metrics for ex-ante analysis of wider socio-economic impact.

Based on a review of previous impact assessment approaches, we used a mixed methods framework to evaluate the impact of DMS Accelerator along the following dimensions:

  • Market

  • Funding

  • Innovation

  • Socio-economic aspects

Our article provides an initial assessment exercise demonstrating how this framework has been applied in practice at DMS, where methods such as surveys, performance metrics and workshops have been used to obtain monitoring data. Based on insights gathered from startups and programme partners throughout the three-year duration of the programme, we derive methodological recommendations and best practices.

The results of our self-assessment highlighted topical challenges that are relevant to other open innovation programmes that specialise in supporting data-centric startups and SMEs. We found that the DMS Accelerator showed rapid development in dimensions related to the market and fundraising, but was less impactful in some aspects of innovation and socio-economic development. Specifically, there was relatively low interest in standardisation and long-term legal strategy among startups, as well as few opportunities for them to communicate or learn about each other’s inventions. It was therefore difficult to justify our expected impact of improved standardisation and interoperability in the context of cross-sector data applications and technology convergence. We propose that this could be addressed in future by adjusting the startup selection procedure to prioritise companies that have a stronger interest in long-term development, as well as the motivation to improve their collaboration capabilities through standardisation or Intellectual Property Rights (IPR). Additionally, a more compelling value proposition can be developed to communicate the benefits of standardisation and IPR to startups who are not yet aware of opportunities to become part of data value chains. By promoting services related to standards and legal strategy, publicly funded programmes such as DMS Accelerator can help startup founders to understand the value of these practices and incorporate them into their business.

Introduction

Many societies across the world are undergoing a transformation through technologies affiliated with AI 1 . This field promises to revolutionise diverse areas of economic activity, including the provision for energy, mobility, food, and healthcare 2 . These developments rely on data as a critical asset, with data-driven innovation becoming a top priority for technology strategists and policy makers in Europe and globally.

Deriving sustainable benefits from data requires a continuous supply of innovative ideas about how best to solve social, economic and environmental challenges. While this can be a difficult task for mature stakeholders whose business processes are already firmly set, startups and SMEs have the agility to explore, generate, and deploy new radical ideas in the economy 3 . At the same time, smaller companies face the greatest barriers (in terms of knowledge and resources) to the adoption of digital innovation practices that are advocated by the European Commission, including standardisation and intellectual property rights 4, 5 . In order to help nascent firms to access the specialist expertise and resources required to overcome these challenges and bring new discoveries to market, corporate and public stakeholders have been active in channeling resources towards them via data-centric business support programmes.

Business accelerators and incubators have been some the most common ways of delivering the support services mentioned above. These programmes started gaining popularity several decades ago, before the emergence of data-centric technologies, as a means of providing nascent firms with advice, services, financing (on occasion), and facilities to help them develop and launch their business 6, 7 . Subtle differences in the objectives and modes of service delivery have informed the naming conventions of these initiatives. For example, accelerators have traditionally been known to quickly move startups from one stage to the next, while incubators have sought to develop longer-term capabilities that constitute self-sustaining, mature businesses 6 . Additionally, incubators have tended to target companies from the local community, while accelerators worked with firms at national and international levels 6 .

Although the labels "accelerator" and "incubator" continue to be used by data-centric programmes, the distinctions between them are less clear-cut in the specific context of data-centric initiatives funded by the EU. An ecosystem of such projects has emerged from a Public Private Partnership (PPP) between the European Commission and the Big Data Value Association (BDVA), with examples including the Data Market Services Accelerator (DMS), European Data Incubator (EDI), EuRopEAn incubator for trusted and secure data value Chains (REACH), MediaFutures, and EUHubs4Data. These support programmes are alike in offering services that address short-term acceleration in addition to long-term sustainable innovation goals, which traditionally characterised incubators in the past. Data-centric accelerators and in cubators are also similar in their adoption of an online delivery format, where a diverse range of applicants and service providers are assembled from different geographic regions, which contrasts with traditional incubators that operate locally on site. It is becoming apparent that despite their self-selected labels, EU-funded accelerators and incubators that specialise in the data economy have a convergence of traits that transcend previous notions of these labels. We therefore consider it constructive to refer to them as open innovation programmes, a higher-level concept that accommodates their differences and decentralised stakeholder configurations.

“Open innovation” is a paradigm that describes the “purposive inflows and outflows of knowledge [between organisations] to accelerate internal innovation and expand the markets for external use of innovation” 8 . Instead of keeping the internal development of products and services within a single firm, open innovation draws upon the collaborative capacity and collective expertise of a wider variety of participants. In addition to multi-stakeholder incubation and acceleration programmes such as those mentioned above, contemporary examples of open innovation include innovation contests, knowledge transfer partnerships, corporate business labs and hubs, and science parks 9 .

Applied to the areas of data and AI, open innovation programmes in Europe encourage the transfer of datasets, services, and ideas between parties in multiple ways. On one level, knowledge and resources are exchanged between the collaborative arrangement of public and private stakeholders who develop joint training, mentoring and networking curricula. For example, standardisation bodies and law firms contribute guidance on data compliance, interoperability and intellectual property issues, while accelerators support incoming firms in developing their business strategy and investment plans, and dissemination managers assist their marketing strategy. At another level, these consortia provide guidance to participating startups and SMEs on how to develop their own infrastructures, business models and competencies to reuse data and integrate data products with external organisations, helping them to become agents of open innovation in their own right 3 . While these efforts, and open innovation more broadly, benefit a variety of stakeholders (e.g. large industry, universities and other research institutions), our article focuses on open innovation programmes targeted specifically towards startups and SMEs as priority stakeholders identified by the European Commission. Given that the funds available for these interventions are a limited resource, effective monitoring and allocation of funds requires programme managers and funders to be able to measure the specific socio-economic benefits achieved by each programme.

Our article explores the question of how impact can be assessed in open innovation programmes that support datacentric startups and SMEs. Motivated by our work at the DMS Accelerator, we review previous approaches to impact assessment, their position in the wider economic context, and what distinguishes data and AI programmes from other types of business support initiatives in the European tech scene. We then present the methodology which was developed and adopted at DMS to monitor the performance and subsequent impact of our services. This paper aims to share our learning with the wider community, discuss best practices, and provide recommendations for policy makers and managers of other similar initiatives.

DMS and its mandate as part of the BDVA ecosystem

DMS Accelerator is one of the open innovation programmes funded by the European Commission to address the major hurdling blocks faced by startups and SMEs specialising in data. Selected companies are invited to draw from a service catalogue that helps them with fundraising, acceleration, standards, legal, promotion, and data skills. These services are delivered by a consortium of 10 organisations that include European accelerators, dissemination managers, standardisation bodies and universities. Overall, the programme services 150 startups and SMEs over the span of three cohorts, one per year. Unlike other programmes that offer seed funding, the DMS offer consists entirely of services that are delivered free of charge.

DMS is part of a broader ecosystem of projects associated with the Big Data Value Association (BDVA). Since 2014, the BDVA has partnered with the European Commission to support the development of jobs, products, and services around Big Data in Europe. In addition to regional development, this partnership intends to support data markets and data value chains. Both of these rely on standardisation as a mechanism to support interoperability between existing data sources and services, providing a foundation for future recombinant innovations that address previously unforeseen uses 5 . For instance, standardised data formats reduce the time and effort required by an external user to incorporate the resource into their own solution, and associated services (e.g., Application Programming Interfaces [APIs]) make it possible for others to benefit from the solution without having to reinvent the same service. Together, compliance with common data standards adds value to the data ecosystem by supporting long-term innovation and interoperability. This is necessary formeeting the European Commission’s desire to achieve technology convergence and new cross-sector applications (e.g., in relation to Cloud, Internet of Things [IoT], and Privacy Preserving Technologies).

One of the difficulties in achieving the above goals is that the economic benefits of standardisation may not be immediate for individual startups. Researchers have highlighted that AI companies face a trade-off between pursuing standardisation for the benefit of compatibility with other platforms versus a better fit and flexibility within their business 10, 11 . Taking the time to learn and implement best practices in relation to data management requires organisations to endure short-term costs, which may be felt most acutely by startups and SMEs 4 . The full impact of programmes such as DMS must therefore take into account not only the financial success generated by firms who participate in these efforts, but also their actualised and potential impact on the wider entrepreneurial ecosystem over the long-term. Outcomes such as the creation of jobs, data skills, collaborations, products, and reusable data sources lead to economic benefits for other stakeholders that can be monetised at future points in time.

Another aspect of interoperability relates to the creation of common data spaces and common practices for sharing data between different stakeholders. A vision of pan-European data sharing spaces has been supported by the BDVA as a private counterpart of BDV PPP 12 . Their role is to guide public and private investment towards fair, secure, and legal governance frameworks such that large-scale data can be connected and valorised. In order to exploit the analytic and economic value of distributed data assets within the bounds of the EU regulatory landscape, small businesses are supported in their compliance with the General Data Protection Regulation (GDPR) and adoption of Intellectual Property Rights (IPR). While some programmes, such as DMS, deliver this support entirely through training and mentoring services, other programmes embed practical experimentation and collaboration opportunities with external stakeholders. For example, initiatives such as DataPitch and the European Data Incubator (EDI) entailed ’challenges’ where startups were matched with corporate data providers to develop new data-driven business models and solutions. Endeavours of this kind have culminated in practical learning and recommendations, with examples including the DataPitch Data Sharing Toolkit and the Legal and Privacy Toolkit, which aim to support subsequent data collaborations and build trust in industries affiliated with data and AI.

Different funding objectives require different impact metrics

There are various kinds of initiatives whose impact assessment approaches can be drawn upon and compared to identify relevant impact metrics for programmes such as DMS. While open innovation programmes share many characteristics in common with traditional business incubators and accelerators, they also cater to specific aspects of the data economy and collective objectives that are more aligned with social innovation programmes. Additionally, the focus on data-driven innovation at DMS resonates with contests and datathons that have acquired a unique set of evaluation metrics. We discuss these approaches below and identify a collection of pertinent criteria to inform the design of our impact assessment methodology.

The similarity of our data-centric open innovation programmes to traditional business incubators and accelerators offers an opportunity to explore how the impact of such programmes has been measured in the past. The classification of different programmes and evaluation of their impact is an ongoing challenge that has been identified in the literature 13, 14 . Programmes differ in their business models, sources of funding, and services offered. Accordingly, numerous authors have highlighted the need for impact measures to take into account their varying objectives. Clarysse and Van Hove 15 identified three emerging archetypes in European accelerators: ecosystem builders (publicly funded), investors (privately funded), and matchmakers (hybrid funding). The distinction between them is summarised in Table 1.

Table 1. Key archetypes in accelerators.

Adapted from Clarysse and Van Hove 15 .

Archetype: Private funding Investor-led Hybrid funding Matchmaker Public funding Ecosystem
Strategic
focus:
Select attractive investment propositions
and turn early–stage projects into profitable
businesses.
Typically non-profit orientation. Corporates
match startups to their own customers or
stakeholders.
Stimulate startup activity
within specific regions or
technological domains.
Programme
package:
Mentorship from serial entrepreneurs and
business angels; often sector specific.
Coaching and mentorship from internal
experts; especially helping startups to
navigate corporate customers.
Mentorship from serial
entrepreneurs and business
developers; most developed
curriculum.
Funding of
startups:
Seed funding offered in exchange for
equity.
Typically no seed funding. Various funding structures
and revenue models.
Selection: Favour ventures in later stages with a
proven track record.
Favour ventures in later stages with a
proven track record.
Favour ventures in very early
stages.
Impact
assessment:
Revenue No hard KPIs Employment

Sources of funding introduce complexity into the identification of assessment criteria for programmes such as DMS. Traditionally, accelerators were funded by investment from business angels, venture capital funds, or corporate venture capital with the intention of capitalising on profitable startups 15 . The return on investment was the main business model that catalysed the growth of such accelerators, and so impact was measured by the return on investment from startups. The working capital of more recent programmes has relied on shareholders such as investors, corporates and public authorities whose missions are less concerned with short-term financial gain. Instead, these programmes strive to create economic benefits on the wider entrepreneurial ecosystem and in accordance with the sustainable development goals (SDGs).

Publicly funded innovation programmes related to data and AI typically fit across the matchmaking and ecosystem building archetypes presented in Table 1. Examples of past programmes include ODINE, DataPitch, STARTS, and Future Internet (e.g., FINODEX). Current programmes include DMS, REACH, Media Futures, and EUHubs4Data.

Previous systematic studies into the performance of different accelerator models have relied on metrics such as the funding raised and valuation attained by the companies, revealing that graduates of publicly sponsored programs tend to raise significantly lower sums of capital post-accelerator 16 . Clarysse and Van Hove 15 highlight that accelerators financed under the objectives of regional development and employment struggle to be profitable in the short or even medium term. These programmes’ selection criteria and success in meeting socio-economic objectives are not aligned to the creation of profit.

While previous publicly-funded programmes have nonetheless monitored firm-level financial performance as part of their impact assessment, many have also traced the impact on employment, society, and the environment. This has been particularly true of social innovation programmes such as those from the past EU call for “Collective Awareness Platforms for Sustainability and Social Innovation” (CAPS). Within this call, the Impact Assessment for Social Innovation (IA4SI) project defined a mixed methods framework encompassing social, economic, political, and environmental impacts 17 . Although this example offers a comprehensive methodological structure, it is useful to consider other frameworks that can better accommodate the focal intention of DMS in terms of technological and data-driven entrepreneurship.

Initiatives that have a stronger focus on data innovation offer insight into some of the additional metrics that have been useful in the past. In particular, hackathons and short innovation contests hosted by platforms such as InnoCentive, Top-Coder, and Kaggle provide well-defined challenges where data scientists, researchers, and developers compete to solve complex problems presented by industry or public stakeholders. Such experiments have contributed to the procurement of novel solutions in fields such as health, criminology, and search technology. The impact of innovation contests has been assessed by frameworks such as the ICAPT (Innovation Contests as an Alternative Procurement Tool), which compares the cost of solutions procured through contests relative to the estimated cost of developing an equivalent solution by traditional methods 18 . This framework also evaluates qualitative benefits such as project awareness and best practices for management. It has been found that contests are cheaper 19 , and that their collaborative format enables solutions to be discovered much quicker and through a more diverse range of participants 20 . These past findings imply the utility of impact metrics that assess the cost, speed, and diversity of innovation contests and research experimentation.

The discussion above drew on a range of innovation initiatives whose impact assessment criteria can be relevant to programmes such as DMS. However, we are faced with the challenge of evaluating a programme that sits at the intersection of financial, social, and technical objectives. The desire to support the profitability of individual firms is combined here with a vision for sustainable long-term development and technological innovation, requiring a unified impact assessment framework. There are a number of past and current programmes that tackled similar transversal objectives in their impact assessments:

  • Open Data Incubator Europe (ODINE) 1 was a programme for open data entrepreneurs that reported impact in terms of the number of incubated ideas, return on investment, engagement, jobs, and geographic representation.

  • Data Pitch matched data providers with startups who worked to address open innovation challenges. Their impact assessment reported on the number of data-driven businesses established, cross-sector and cross-border collaborations, financial impact, and the creation of big-data use cases that drive investment 21 . Following its completion, the programme organisers developed additional resources that could be used by other organisations. For example, the Data Sharing Toolkit helps organisations to generate value by allowing third parties specifically permissioned access to private datasets.

  • Science, Technology and the Arts (STARTS) was a residential innovation programme designed to increase the impact of artists in high-tech scientific environments 22 . They committed to deliver a certain number of residencies, a global methodology and tools to promote collaborative work, as well as knowledge to evaluate success factors in future initiatives.

  • Future Internet (FI) encompassed a number of accelerators working in Internet-enabled innovation 23 . The project developed analytic methods and tools to perform an ex-ante socio-economic impact analysis. Their frame-work contained several assessment areas and KPIs including:

    • Market - customers, revenue, and geographies (footprint in EU economy)

    • Socio-economic - social, scientific, and macro-economic consequences (e.g., wider perception of AI)

    • Innovation - types of technology solution, intellectual property (IPR)

    • Funding - funding requested and quality of financial plan

  • DataBench is currently creating a benchmarking process for organisations that develop Big Data Technologies (BDT) 24 . The framework measures technology development activity against parameters of high business relevance, seeking to demonstrate industrial significance and return on investment.

Methodological implications

What unites the programmes listed in the previous section is that their impact assessment approaches drew on a variety of metrics to capture financial, social, and technological outcomes. Additionally, they presented nuanced methodological considerations in terms of benchmarking and assessing qualitative impacts. In the following paragraphs we discuss what the quantitative and qualitative impacts commonly entail, and suggest ways of synthesising these strengths into a generalisable impact assessment approach.

When it comes to quantitative metrics, a number of earlier programmes have been able to conduct rigorous evaluations of impact through counterfactual analysis. For example, Data Pitch compared their outcomes with what would have been achieved by startups in the absence of the programme 21 . This kind of analysis is conceptually similar to benchmarking approaches such as Data Bench, where the impact assessment is based on relative comparisons. However, it is important to consider that not every programme will have the resources to benefit from counterfactual analysis or benchmarking tools. In the present case, we propose the comparison of measurements across multiple cohorts of the same programme as a simpler way to assess impact in terms of temporal improvement and agility in the specific attributes relevant to that programme.

When evaluating qualitative impacts, previous data innovation programmes were alike in reporting on their methodological contributions and learning. Publicly funded accelerators are characterised by high quality self-assessment and transparent reporting of outcomes, often sharing expertise that has an impact on the wider BDVA ecosystem. Due to the limited duration of publicly-funded programmes (DMS lasts three years) it is not usually possible to measure the actualised impacts and value chains that accrue over the longer term. However, resources that have been released into the ecosystem for future use are one of the ways of ensuring lasting impact. We therefore propose viewing methodological contributions as one of the manifestations of socio-economic impact. In the case of the DMS Accelerator, one of the defining qualities of the programme is that it attracts startups based purely on its services, without offering seed funding. The methods that were developed for selecting and onboarding startups who are motivated and engaged by this service proposition could therefore be an impactful resource for other accelerators whose funding model is similar to that of DMS.

In addition to revealing impact assessment approaches that are held in common by multiple programmes, our overview also highlighted challenges for developing a unified impact assessment methodology. Some past programmes such as Data Pitch were structured in such a way that startups were systematically matched with data providers, making it possible to demonstrate well-defined data experiments and collaborations at the broader socio-technical level. However, because this indicator relied on the specific format of the programme, it would be difficult to replicate the same assessment in another programme such as DMS. One of our intentions in this article is to identify a flexible and generalisable framework, which is applicable to DMS but not endogenous to it.

We suggest that the main impact categories defined previously by the FI-Impact project can be used as the basis for such a framework. Specifically, the categories of 1) Market, 2) Funding, 3) Innovation, and 4) Socio-economic aspects provide a comprehensive yet generalisable representation of the main areas of impact applicable across publicly funded data innovation programmes. In addition to assessing impact on the entrepreneurial ecosystem in terms of sales and investment, the dimensions also accommodate technical advances and qualitative socio-economic impacts. For the purposes of flexibility, the specific metrics within each impact category can be defined in accordance with the monitoring opportunities of particular programmes. A variety of qualitative and quantitative methods can be combined to assess each impact dimension during successive years of the programme. In the remainder of this article, we demonstrate how the four-dimensional impact framework has been implemented in practice at DMS Accelerator, and the results gained from our experience.

Methods

The DMS programme is unique in its focus on service provision, rather than pre-defined data challenges or collaborations that have characterised other similar programmes in the past. Our methodological framework therefore has an emphasis on the evaluation of services (in terms of engagement and satisfaction), while at the same time incorporating financial successes of firms, new products, collaborations, and public awareness metrics. DMS services are already classified into categories (Acceleration, Promotion, Fundraising, Standards and Legal, and Data Skills) that lend themselves to the FI-Impact dimensions (Market, Funding, and Innovation). Moreover, the expected impacts of the DMS programme that are outlined in the grant agreement also map onto all four dimensions, including Socio-Economic impact. These expected impacts are presented in Table 2 (first column).

Table 2. Impact assessment methodology proposed for DMS.

Dimension
Followed by expected and additional impacts as
specified in the DMS grant agreement
Individual company
metrics
Collective metrics / KPIs
Market
DMS expected impact:
    •  At least 50 clients (e.g., start-ups, SMEs) served annually
in partner finding, matchmaking, venture capital raising,
training, and coaching

    •  Reach 2,000 companies out of which 150 will be
serviced

    •  Increase their sales capacity by 15%

•  Sales capacity

•  Revenue

•  New clients

•  Number of contacted and
recruited companies

•  Companies at different stages of
development

•  Representation of different
industry sectors and countries

•  Engagement in acceleration
webinars
Funding
DMS additional impact:

    •  New rounds of private capital reaching 5m euros

    •  Additional public funds reaching 1m

    •  Capacity for future investment and partnerships offered
by the consortium and Advisory Board of investors

•  Additional funding gained

•  Engagement in fundraising
webinars

•  Meetings with investors
Innovation
DMS expected impact:

    •  Offer training on standards and legal issues to 150
companies

    •  Improved standardisation and interoperability especially
in the context of cross-sector applications and
technology convergence (data, Cloud, IoT, connectivity)

•  New products, datasets,
services

•  Patents

•  Engagement in standards & legal
webinars

•  Engagement in data skills courses

•  Mentoring sessions in standards &
legal
Socio-economic
DMS expected impact:

    •  Success stories as a result of services offered
    •  Dissemination and exposure of success stories

DMS additional impact:

    •  At least 200 new jobs requiring data skills will be created
in the portfolio of companies

•  Jobs created

•  Gender composition of
teams

•  Collaborations

•  Success stories

•  Social media followers

•  Website visitors

•  Audience at events

•  Self-organised events

•  Methodology shared with other
programmes
Method of assessment: Surveys supplemented with
desk research
Programme monitoring metrics and
workshops with programme partners

The four-dimensional impact criteria help to address the tension between the short-term and long-term economic impacts of data innovation programmes. At DMS, this is achieved through metrics that monitor impact at two levels of analysis: 1) the success of companies who complete the programme, and 2) impact on the collective ecosystem (the entrepreneurial environment as well as on other open innovation programmes). We evaluate these outcomes using a mixed-methods approach. Quantitative metrics are derived from programme monitoring activities and close-ended survey questions completed by startups when they leave the project. Qualitative approaches were additionally used to gain a more nuanced understanding of these metrics. For example, open-ended survey questions were used to assess the ways, if any, that participation in DMS contributed to the successes of startups in specific dimensions. We also conducted workshops with key partners from DMS to examine the collective impact of the programme and observations that were not covered by routinely monitored KPIs.

To summarise, the results presented in this article are based on programme monitoring metrics, surveys completed by startups, and workshops with DMS partners. Ethical considerations surrounding these data sources are discussed below, followed by descriptions of how each type of data was gathered.

Ethics statement

Our investigation reuses data that were collected by the DMS Accelerator as part of its routine monitoring activities. The data are fully anonymised and we have obtained verbal consent from the DMS partners for our secondary use of the data. When collecting information, DMS informs participating companies that the gathered data may be used for research and reporting purposes. Additional consent to openly publish the data was not obtained due to the inclusion of commercially sensitive information that needs to remain private as outlined in the data availability statement. As our study is low risk and does not use personal data, approval from the King’s College London ethics committee was not required. Our study was done in consultation with the Information Compliance team at King’s College London.

Routine monitoring metrics were sourced from records kept throughout the programme, where available. This includes engagement and dissemination KPIs reported by DMS in its periodic reports from the years 2020 and 2021, as well as webinar and course engagement statistics. Our analysis includes all data that fall within the scope of the expected and additional impacts presented in the DMS grant agreement, as listed in the first column of Table 2. For the purposes of brevity, we have excluded information relating to specific events and webinars conducted by DMS. Instead, the metrics are reported here in aggregate form to reflect the average attendance and satisfaction in each service category.

Survey responses were gathered using an ’Impact Survey’ delivered to all three cohorts of graduates from the DMS programme. The first two cohorts were surveyed twice, once upon leaving the programme and then one year later using a follow-up survey. Blank templates of the survey are available as described in the Data Availability statement 25 . The survey contained a selection of multiple-choice, closed-ended, and open-ended questions related to topics that are summarised in the middle column of Table 2. The closed-ended questions were compulsory, while the open-ended questions that followed them were optional. Participants who did not provide written answers for a particular question were not included in the results reported for that question. The questions were slightly different in each DMS cohort in order to account for differences in the amount of time that had elapsed since they completed the programme. The coordinator of DMS Accelerator distributed these surveys to each cohort using an email that invited the companies to submit their responses in a Google Form. Reminder emails were also sent, and DMS mentors encouraged companies to respond where possible. The response rate was 26% (13 out of 50) from cohorts one and three, and 42% (21 out of 50) from cohort two. Due to the temporal setup of the impact survey, the responses of cohorts one and two reflect outcomes that were achieved within 11 months after the startups left the programme, whereas the responses of Cohort three reflect their immediate outcomes.

Workshop outcomes are based on two two-hour workshops conducted with DMS partners. One of these was the Interim workshop which took place on 10th December 2019, in Leipzig, Germany. This was attended by 13 participants from nine partner organisations who had already been invited to be part of the DMS Consortium meeting on the same date. The workshop used a structure that alternated between small group discussions and plenary discussions, forming three 30 minute blocks where participants discussed topics related to startup selection, service provision, and acceleration. Participants were provided with prompts inviting them to comment on what went well, what could have gone better, and what could be done to improve the programme. The results of this workshop were summarised in a confidential Workshop Report deliverable at DMS. Subsequent actions were also taken by the relevant DMS partners to modify the programme’s startup selection and acceleration process in accordance with the recommendations raised by the consortium. In the second year, DMS organised a White Paper workshop on 12th November 2020. 11 participants attended from seven DMS partner organisations, who were invited to participate in a Microsoft Teams meeting. As with the previous workshop, this setup alternated between small group discussions in breakout rooms and joint plenary discussions. There were three 30 minute blocks where participants discussed the most and least successful aspects of the programme, how impact should be defined and assessed, and ways to make the programme more successful in future. The outcomes of this workshop informed the design of the DMS Impact Survey and the content of this paper. A summary of the structure of both workshops is available in the Extended Data repository 25 .

Results

Our findings are grouped into sections according to the four dimensions of our impact assessment framework: Market, Funding, Innovation, and Socio-economic aspects. The results for each dimension are a triangulation of the methods described above, in the form of monitoring activities, Impact Surveys completed by startups and workshop participation from programme partners.

Market

In the dimension of the market, our methodology sought to measure the footprint created by the DMS programme on the EU economy. As part of this, we examined the diversity of companies serviced by the project, their engagement with entrepreneurial training and their own capacity to serve new clients. Our assessment of market impact is summarised in Table 3.

Table 3. Assessment of market impact.

Cohort One Cohort Two Cohort Three
Impact Survey
answers from
startups
(13 responses in total, 7 surveyed
immediately and 6 one year later using a
follow-up survey)
      •  Sales capacity (assessed only in
follow-up survey) – 3 companies
reported an increase in sales
capacity.

      •  Revenue – 6 companies reported an
increase in revenue.

      •  New clients – 7 companies gained
potential new clients.

Open-ended answers indicated that the
effect of DMS on market impact was
achieved by means of improved public
image (e.g., through promotional videos
and improved negotiation capacity).
(21 responses in total, 15 surveyed
immediately and 6 one year later)

      •  Sales capacity – 9 companies
reported an increase in sales
capacity.

      •  Revenue (assessed only
in follow-up survey)
– 6 companies reported an
increase in revenue.

      •  New clients – 3 companies
gained new clients as a result
of joining DMS.

Open-ended answers mentioned
improved company profile and
marketing, increased reach and
partnerships, improvement in
knowledge, and selling proposition.
(13 responses, surveyed
immediately)

      •  Sales capacity – 6
companies reported an
increase in sales capacity.

      •  Revenue was not included
in the survey, as we did
not expect an immediate
impact.

      •  New clients – 4 companies
gained new clients.

Open-ended answers
mentioned improved strategy,
visibility in the market, better
lead generation and training.
Programme
Monitoring


      •  Contacted companies: 690

      •  Company stages: 15 scaling, 30
validating, five
establishing

      •  Countries represented in portfolio:
16

      •  31 mentoring sessions for 10
startups

Participation in webinars:

      •  10 entrepreneurial webinars, five
participants on average

      •  Two promotional webinars, three
participants on average


      •  Contacted companies: 1,172

      •  Company stages: 15 scaling,
30 validating, five establishing

      •  Countries represented in
portfolio: 20

      •  63 mentoring sessions for 23
startups

Participation in webinars:

      •  11 entrepreneurial webinars,
24 participants on average,
average rating 4.17/5

      •  Three promotional webinars,
14 participants on average,
average rating 4.15/5


      •  Contacted companies:
1,653

      •  Company stages: 15
scaling, 30 validating, five
establishing

      •  Countries represented in
portfolio: 17

      •  104 mentoring sessions
for 28 startups

Participation in webinars:

      •  14 entrepreneurial
webinars, 24 participants
on average, average
rating 4.15/5

      •  Three promotional
webinars, 9 participants
onaverage, average rating
4.27/5
Workshops with
partners
DMS partners felt that the selection
process wassuccessful in attracting a
diverse range of companies,representing
a strong portion of the EU market.
However, the service offering was not
communicatedclearly in relation to the
needs and business models of startups.
Low participation in webinars was flagged
as an important challenge to address
Based on feedback after cohort one,
DMS improved its communication
of services and onboarding process
for startups, leading to increased
engagement with all training.
More personalised mentoring
was provided. The market impact
of startups also benefited from
corporate videos created by DMS.

There are a number of ways through which DMS ensures market diversity, and these are part of the routinely monitored KPIs. The first intention is to purposefully accept companies from different stages of development, with 15 scaling, 30 validating, and five establishing companies selected each year. Between the first and subsequent years of the programme, an increase was achieved in the pool of contacted companies, the number of countries represented in the selected portfolio and the representation of AI, ML, and other industry sectors ( Figure 1).

Figure 1. DMS applications in 2019 (left) and 2020 (right) classified by industry tags.

Figure 1.

After its first year, programme monitoring and workshop activities highlighted a dramatic increase in the level of engagement with DMS promotional and entrepreneurial services. For example, the number of webinar participants in the Acceleration category increased fivefold from cohort one to cohort two. While it is difficult to assess the extent to which these and other DMS services directly influenced the market success of companies, the results of the Impact Survey from all three cohorts indicate that 18 companies (approximately 45% of those who responded to this question) saw an increase in sales capacity. In the optional question that followed, 10 of these respondents provided a quantitative estimate of their increase, amounting to an average of 36% increased sales capacity per company. A further three respondents mentioned qualitative improvements in general process development, product development and pilot studies that helped to improve their market impact. Besides sales capacity, 14 companies (30% of all respondents) reported gaining new clients, with approximately 3–4 new clients on average among the 12 respondents who provided detail. Open-ended survey answers showed that the training and promotional services offered by DMS helped to improve their companies’ image and reach. Workshops with partners reflected these comments and established that one of the main sources of such impact came from professional videos that were created for the start-ups, who were then able to publish them on their own social media or website.

Together, the above findings show progress towards meeting the expected market impacts of the programme, especially in KPIs related to the diversity of the audience and the number of startups served. The target of increasing startups’ sales capacity by 15% has been more difficult to confirm due to low survey response rates as shown in Table 3. Among those who responded, there appears to be a split where half of the companies saw no increase at all, while others reported a substantial increase. This difference in outcomes is consistent with the results observed in other accelerators, whose overall impact can be traced to a much smaller portion of serviced companies that become highly successful 16 .

Funding

In the dimension of funding, we assessed the amount of funding generated by DMS startups since starting the programme. We were also interested in the ways in which DMS services contributed to this success. Our assessment of fundraising impact is summarised in Table 4.

Table 4. Assessment of fundraising impact.

Cohort One Cohort Two Cohort Three
Impact Survey
answers from
startups
(13 responses in total, 7 surveyed
immediately and 6 one year later using
a follow-up survey)
      •  4 companies gained additional
funding, collectively reporting
845K Euros.
Open-ended survey answers indicated
that startups’ fundraising activities
benefited from getting acquainted with
DMS accelerators, learning about IP
protection, and receiving promotional
videos.
(21 responses in total, 15 surveyed
immediately and 6 one year later)
      •  9 companies gained additional
funding, collectively reporting
3.6M Euros.
Open-ended answers indicated
that the startups gained confidence
in pitching and acquired a
better understanding of funding
opportunities. The Pitch Day and
TNW feature also initiated some
introductions for them.
(13 responses, surveyed
immediately)
      •  2 companies gained
additional funding
collectively reporting 625K
Euros.
Open-ended answers indicated
that the startups learnt where
to seek funding opportunities,
how to understand their product,
build financial models, and how
to make them work.
Programme
Monitoring

      •  40 meetings facilitated with
investors
      •  Seven fundraising webinars, five
participants on average

      •  41 meetings facilitated with
investors
      •  Seven fundraising webinars,
18 participants on average,
average rating 4.28/5

      •  9 meetings facilitated with
investors
      •  Seven fundraising
webinars, 10 participants
on average, average rating
4.10/5
Workshops with
partners
DMS partners reported that investor
pitch decks and matchmaking
strategies could be more personalised
and tailored to the needs of the
beneficiaries.
Fundraising went well this year.
A high number of meetings with
investors was facilitated and clear
startup success stories identified by
DMS mentors.

While the programme has no concrete KPIs related to fundraising among startups, several desired impacts were included in the grant agreement. Specifically, DMS aimed for its startups to raise new rounds of private capital reaching 5m Euros and additional public funds reaching 1m. Capacity for these investments and partnerships was supported by the consortium and Advisory Board of investors.

Impact assessments conducted using surveys suggested that approximately 5M Euros had been raised by 15 companies (32% of respondents). Based on the available data, the proportions of successful DMS graduates seems to be roughly consistent with meta-analyses US accelerators, where 23% of companies were found to be successful in raising significant funds after completing the programme 16 .

Due to the low survey response rates, the topic of funding was investigated with additional desk research using Crunchbase to trace the total funds achieved by DMS graduates since leaving the programme. This revealed that more than 55M in private funding and more than 1.5M in public funding had been raised in reality. When asked about the ways in which DMS contributed to their fundraising activity, companies that responded to the impact survey commented on their improved confidence in pitching and a better understanding of funding opportunities, as well as their improved image developed through the promotional services of DMS. Alongside this, programme monitoring statistics showed improved engagement with services related to fundraising after the first cohort, where the number of participants in fundraising webinars had more than doubled after the second year of the programme. Through workshops with DMS partners, we also learnt that programme mentors provided personalised guidance and facilitated additional meetings with investors, which contributed to the acquisition of funds and identification of clear success stories among startups. Although investor matchmaking played an important role throughout the entire programme, some data about investor meetings during the third DMS cohort was lost due to a change in the service delivery format and tracking mechanisms.

Innovation

In the dimension of innovation, our methodology sought to capture novel technology solutions and intellectual property as indicators of impact. This dimension encompassed the role of data regulations, legal strategy, standardisation, and interoperability in supporting data value chains. Our assessment of innovation impact is summarised in Table 5.

Table 5. Assessment of innovation impact.

Cohort One Cohort Two Cohort Three
Impact Survey
answers from
startups
(13 responses in total, 7 surveyed
immediately and 6 one year later
using a follow-up survey)

      •  6 companies developed new
products or services

Open-ended answers indicated
that startups used DMS feedback
to refine their product, improve
their GDPR compliance and put
data management at the core of
their business.
(21 responses in total, 15 surveyed immediately
and 6 one year later)

      •  11 companies developed new products or
services

Open-ended answers reported that DMS
helped startups with GDPR compliance, product
development strategy and data science skills.
(13 responses, surveyed
immediately)

      •  5 companies developed
new products or
services

Open-ended answers
reported that DMS helped
startups with their IPR
strategy, innovative ideas
andfunding awareness.
Programme
Monitoring
      •  Eight mentoring sessions
in standards and legal
requested by seven startups

      •  Three Data Skills courses, 18
participants average

      •  Seven Standards & Legal
webinars, nine participants
on average, 3.8/5 average
rating
      •  10+ mentoring sessions in standards and
legal

      •  Three Data Skills courses, 18 participants
average

      •  Seven Standards & Legal webinars, nine
participants on average, 3.8/5 average
rating
      •  15+ mentoring sessions
in standards and legal

      •  Two Data Skills courses,
23 participants average

      •  Seven Standards &
Legal webinars, seven
participants on average,
4.7/5 average rating
Workshops
with partners
New services were proposed in
relation to intellectual property
rights (IPR) and IP mentoring.
Mentoring sessions on GDPR and IPR received
excellent feedback. However, delivery was
constrained by limited resources and the scope
of these services was not clearly distinguished
from consultancy.
Low commitment and awareness of
standardisation among SMEs made it difficult to
engage them with services related to standards.

DMS fulfilled its expected impact of delivering training on standards and legal issues related to data. Programme monitoring activities showed that engagement with Data Skills courses had tripled in the span of three years. Survey responses from startups also indicated that the portion of startups who developed new products or services had almost doubled in the second year, but declined again in the third cohort, which may be attributed to the shorter length of time covered by the survey. The products included web services, mobile and cloud platforms for various sectors, demonstrating potential alignment with the DMS expected impact of fostering cross-sector applications, and technology convergence. In their written answers, the survey respondents reported benefiting from DMS services related to compliance with GDPR and product development strategy. This feedback was reflected in the comments shared by DMS partners during workshops, where they reported that mentoring sessions on GDPR and IPR received excellent feedback.

In addition to positive feedback, workshops with partners revealed a number of limitations related to the uptake of standards and legal services. While mentoring in GDPR was in high demand, other DMS partners observed a low commitment and awareness of topics related to standardisation and long-term legal strategy (e.g,. intellectual property) among startups. Although services related to standards and legal issues experienced an increase of interest between cohorts, this service category had much lower participation compared to entrepreneurial and fundraising training.

Socio-economic aspects

In the socio-economic dimension, our methodology sought to measure the direct and indirect consequences of DMS in terms of societal and macro-economic forces. The impacts considered here include employment and the capacity to improve the wider perception of AI among European citizens. Our assessment of socio-economic impact is summarised in Table 6.

Table 6. Assessment of socio-economic impact.

Cohort One Cohort Two Cohort Three
Impact Survey
answers from
startups
(13 responses in total, 7 surveyed
immediately and 6 one year later
using a follow-up survey)
      •  8 companies reported
that their team has grown,
collectively generating 22
new jobs.
      •  4 companies reported
a change in the gender
composition of their team,
with an average increase of
14% in women.
      •  9 companies pursued
new collaborations or
partnerships.
Open-ended answers reported
benefiting from the networking
opportunities provided by DMS,
but reported needing more
personalised support with
connecting to investors and
collaborators.
(21 responses in total, 15 surveyed
immediately and 6 one year later)
      •  13 companies reported that
their team has grown, collectively
generating 50 new jobs.
      •  8 companies reported a change
in the gender composition of
their team, with an average
increase of 0.47% in women (5
teams increased, 2 decreased).
      •  20 companies pursued new
collaborations or partnerships.
Open-ended answers reported that
DMS helped to reach new audiences
and markets, but improvements could
be made in the amount of personalised
support, connecting to investors and
more interaction between startups.
(13 responses, surveyed immediately)
      •  5 companies reported that their
team has grown, collectively
generating 6 new jobs (9 jobs
created, 3 lost).
      •  2 companies reported a change
in the gender composition of
their team, with an increase of
20% in women reported by one
of the companies.
      •  11 companies pursued new
collaborations or partnerships.
Open-ended answers reported that
DMS helped to represent and build a
community around their startup, as
well as helping them to develop their
workforce.
Programme
Monitoring
      •  Social media followers: 879
      •  Website visitors per month:
400+
      •  Audience at events: 1,600
(Pixels Camp) & 15,492 (TNW)
      •  Self-organised events: 11
      •  Social media followers
(cumulative): 2,215
      •  Website visitors per month:
2,000+
      •  Audience at events: 13,240 (at
TNW, Pixels Camp was cancelled
due to COVID-19)
      •  Self-organised events: 5
Methodology shared with other
programmes:
      •  Form for engagement with
companies
      •  Impact assessment approach
(this paper)
      •  Social media followers
(cumulative): 2,571
      •  Website visitors per month:
3,000+
      •  Audience at events: 7,500 (at
TNW, Pixels Camp was cancelled
due to COVID-19)
      •  Self-organised events: 14
Methodology shared with other
programmes:
      •  Form for engagement with
companies
      •  Assessment of startup needs
      •  Selection practices of data-
centric programmes
Workshops
with partners
Mentoring sessions, bootcamps
and events provided opportunities
for startups to interact directly
with the beneficiaries of the
programme, and this was positively
appreciated on both ends.
Promotional campaign of the
programme was good, but few
opportunities were available for
communication between startups.
Visibility of the project across the
European ecosystem could also be
improved.

Responses to the Impact Survey showed that 26 companies (55% of the respondents) had created new jobs within the duration of the programme or shortly after, collectively reporting 78 new jobs and an average of 2–3 new team members per company. The position types were specified for 43 of the new roles, with approximately 24 roles (52% of the reported) being in technical and data-related fields, 12 roles (26%) in business development, 4 roles (10%) in marketing, and the remainder related to finance, sales and HR. While these observations contribute towards the DMS desired goal of creating 200 new jobs requiring data skills, the magnitude of progress towards that number is difficult to assess due to the low survey response rates from startups after completing the programme.

DMS monitors additional social impacts such as the gender composition of startup teams, their collaborations, and published success stories. Impact Surveys showed that, compared to the first cohort, each of these areas demonstrated an improvement in the second and third cohorts. A small portion of the respondents (14 companies, around 30% of respondents) experienced a change in the gender composition of their team, with the representation of women increasing by an average of 14% per company. A significant portion of the respondents (40 companies, 85% of respondents) reported pursuing new collaborations. These startups reported benefiting from the networking opportunities provided by DMS, but they also requested more personalised support in regards to networking with investors and interacting with other DMS startups. The latter point was also highlighted by DMS partners during workshops, where communication opportunities between startups were identified as a weakness in the current service offering.

At the collective level of analysis, programme monitoring statistics indicated a good promotional campaign, presence at events and a large audience. Through feedback acquired from mentors and Impact Surveys, a number of success stories were identified among startups who benefited from the programme. The dissemination of these stories was increased through the promotional reach of the programme.

In addition to the social impact of DMS in terms of public engagement, we were also interested in assessing its position in relation to other similar programmes. Workshops with DMS partners revealed that there was room to increase the programme’s visibility and create closer partnerships with other open innovation programmes in Europe, especially on regional and local levels. In addition to approaching these connections directly, the outputs of the programme could be used to gain exposure. In particular, deliverables such as the startup selection and onboarding procedure ( DMS deliverable D2.3), assessment of startup needs ( D3.5), and whitepapers on impact assessment ( D4.5, this paper) and selection practices (D4.6) are publicly available, such that other similar programmes can benefit from relevant parts of our methodology..

The reasons for increased engagement

A major finding observed across all dimensions of our results was the increase in engagement with services beginning with the second cohort of the programme. The reasons behind this success were revealed by the workshops that were conducted with DMS partners.

During the Interim workshop following the first DMS cohort, the project partners expressed that the programme had already established a quick and agile selection process that attracted high numbers of applications and met the programme KPIs. The partners contributed to this success by leveraging their local networks to promote the project in wider events (e.g., ’The Next Web’ [TNW] conference), but a number of challenges were identified in regards to the startup selection and onboarding procedure.

One of the most challenged criteria in the applications of the first cohort was related to the management of diversity and inclusivity by each startup. It was suggested that this question discouraged applications from companies who have not yet achieved gender equality. For this reason, the question was rephrased in the next intake to accommodate social and environmental responsibility more broadly. The application form was also extended by asking startups to include a pitch deck and to provide more open-ended responses about the needs and vision of their company. These adjustments helped to ensure that the selected companies in the second and third DMS cohorts were better aligned with the services offered by the programme.

Another challenge identified during the first Interim workshop was a lack of information communicated to startups about the services available at DMS. Moreover, partners reported that startups requested additional personalised support as a way to improve the programme. In response to these limitations, the programme management and promotion team provided better communication during the second and third open calls about the value proposition of DMS and the services on offer. The onboarding process was also updated to include a needs assessment and personal calls that helped to build rapport with individual startups. This was accompanied by additional personalised mentoring during cohorts two and three.

Taken together, the first workshop with DMS partners served the purpose of generating insight into the reasons behind participants’ low engagement with the programme, enabling the programme to respond and multiply its impact on startups in the next two cohorts. The second workshop re-evaluated these amendments to reveal how they contributed to the increased engagement that was observed across all DMS services in the subsequent iterations of the programme.

Discussion

Building on prior literature and impact assessment methods in previous open innovation programmes, this article presented a general methodological framework and applied it in the context DMS Accelerator. We have demonstrated how dimensions related to the market, fundraising, innovation and socio-economic aspects can be assessed through a combination of qualitative and quantitative indicators coming from surveys, workshops and monitoring data.

Although it was possible to demonstrate the positive impact of DMS in each of our evaluated dimensions, the magnitude of impact was different in each area, and some outcomes were difficult to map onto what was proposed in the DMS grant agreement. We summarise these findings below and present recommendations for other similar programmes.

The impact created by DMS

In relation to the market, DMS monitoring data showed high diversity in the developmental stages, geographic origins and industry sectors of companies that were serviced by the programme. Evidence of this was observed in the descriptive data accompanying open call applications and the portfolio of selected companies. The companies captured by the programme represented different levels of maturity (45 scaling, 90 validating, and 15 establishing), they came from different geographic regions (28 countries), and represented around 14 industry sectors. Besides descriptive data about the market coverage of programme participants, the programme also provided marketing training and services that enabled participants to independently exert an impact on the market. This was demonstrated by high participation and engagement with entrepreneurial services, as well as the graduating firms’ reported impacts on their sales capacity and client base. Together, these findings give us confidence that DMS has improved the entrepreneurial capacity of selected companies in Europe.

In the dimension of fundraising, DMS was able to demonstrate strong outcomes in the number and size of investments generated by graduating firms. Much of this information came from external platforms such as Crunchbase, as internal surveys conducted by DMS did not have sufficiently high response rates and temporal tracking of the investments achieved by participants. Nonetheless, survey monitoring data showed that at least some of their successes were attributed to the training and guidance that had been acquired by startups through DMS 25 .

It was more challenging to trace impact in the areas of innovation and socio-economic development. While startups demonstrated substantial contributions in terms of new products, services, data skills, and the creation of jobs, there was low interest in data standardisation and long-term legal strategy. It was therefore difficult to justify the DMS expected impact of "improved standardisation and interoperability in the context of cross-sector data applications and technology convergence". This finding was revealed through the evaluation of engagement with different DMS services over time, which is discussed in the next section.

Methodological insights

The specific methods used in our monitoring approach included surveys to startups, programme monitoring metrics and workshops. When used on their own, each of these methods has certain limitations. For instance, the Impact Survey suffered from low response rates among startups, while workshops with partners were limited in representing the subjective viewpoints of the people involved in running the programme. Monitoring methods such as webinar statistics and impact surveys are also likely to have represented different audiences. While any team member could participate and rate webinars, the end-of-year Impact Survey was mostly completed by a single company representative, who may not necessarily be aware of the skills and competencies acquired by all members of personnel who participated in the programme. By triangulating the various pieces of information available to us, we were able to create the in-depth and critical assessment necessary to evaluate and improve the impact of DMS. Although the applicability of our mixed-methods approach to other similar initiatives was not tested here, we believe that this methodology can be replicated using resources and practices that are already likely to be a part of other programmes’ monitoring approach.

In addition to methodological triangulation, our findings were classified into four dimensions of impact which made it possible to identify specific areas for improvement. The observed increase in engagement across the four service dimensions was not equal, with some areas demonstrating significantly higher engagement than others. In particular, services related to the market and fundraising showed fivefold and threefold increases respectively. Engagement with innovation services had also increased, but with a slightly lower magnitude compared to the other dimensions, suggesting that this is area could benefit most from further improvement.

Although a large volume of new products and services was generated by startups in the second cohort, programme monitoring and workshops with partners indicated that there was low interest in training related to long-term innovation strategy and standardisation. Other data competencies such as GDPR and data skills received more interest. The low appeal of standards training complements a separate observation made by programme partners during workshops, where they felt that the focus on data in the Data Market Services accelerator is not something that we are sufficiently promoting and showcasing. This was reflected in the needs presented by startups entering the programme, who were most interested in general business competencies such as fundraising, promotion, and international mobility ( DMS D3.5).

Possible ways of resolving this gap in the future could involve adjustments to the startup selection procedure to prioritise those that have a stronger motivation to adopt data standards. Additionally, it will be important to develop a more compelling value proposition that communicates the benefits of standardisation and intellectual property to startups who may not yet be aware of opportunities for building data value chains. To support active engagement in this area, we can consider another current limitation related to the lack of communication between participants of the programme. Both issues could be solved by establishing “communities of practice” around data innovation and the adoption of standards.

As discussed in the background literature, standardisation can be an arduous and costly process for businesses. The challenges faced by DMS are likely to be shared by other similar programmes that specialise in data. We hope that this paper will start a discussion about the best ways to monitor and foster the impact of data innovation programmes in Europe.

Limitations and future work

An important limitation of our study is that it focused on measuring the impact of only one initiative. This presented difficulties in the form of small sample sizes during data collection, as well as limited representation of open innovation programmes more broadly. As noted in the introduction of this article, open innovation occurs not only through business accelerators and incubators, but also via contests, knowledge transfer partnerships, corporate labs and hubs, and science parks. In order to confirm if and how our impact assessment approach is generalisable to other programmes, additional studies would be needed to replicate our conceptual and methodological framework in different settings.

Another limitation of our study relates to theoretical literature. Our study sits in the specific context of data-centric open innovation programmes that are publicly funded. Given that these types of programmes are still somewhat new in relation to the wider open innovation movement, there are not as many scientific publications that cater to the unique characteristics and impact assessment practices of data innovation and data entrepreneurship programmes. Our paper contributed a small step towards this expanding discussion, and we are hopeful that other similar studies are already adding greater scientific rigour to the question of how impact can be measured in data-centric programmes.

Best practice and recommendations

In the beginning of this article, we highlighted the diversity of innovation programmes that have proliferated in the European data economy. Due to their varied funding models and objectives, their impact cannot be measured according to the same criteria. While our specific findings may be relevant to other publicly funded data initiatives, what we wish to highlight here is the generalised methodology that has helped us to derive actionable insights. We would like to propose the following recommendations to other initiatives similar to DMS who may be seeking to monitor and demonstrate their impact to the funding body:

  • Use a variety of methods to measure impact. Diverse monitoring approaches help to buffer against the limitations of each method. Specifically, there can be limitations of sample size (e.g., low survey response rates), depth of data (quantitative vs. qualitative), and subjective bias (service providers vs. recipients). We suggest that a combination of methods such as surveys, quantitative monitoring tools, and workshops can be synthesised to accommodate the perspectives of service recipients as well as providers. Additionally, different methods can be reconciled to capture individual as well as collective outcomes (in terms of the team members, companies, and ecosystems served by the programme).

  • Measure impact along multiple dimensions. By classifying monitoring activities into multiple impact areas (e.g., market, funding, innovation, and socio-economic), it is possible to draw comparisons and identify areas that are most in need of improvement, so that resources can be targeted efficiently towards those services. For data programmes in particular, we recommend that special attention is required to assess the role of standardisation and legal strategy in fostering innovation through data value chains.

  • Monitor changes in impact over time. It is expected that the impact of a programme will increase as it assimilates the learning derived from successive iterations of service delivery. The magnitude of change can serve as an indicator of the agility with which the programme is able to respond to dynamic economic circumstances. Metrics derived as part of routine KPI monitoring activities can be compared between cohorts and interpreted through discussions with programme partners who were involved in service delivery. This can help to identify the actions and strategies that contributed to particular outcomes. In the case of DMS, we learnt that a substantial increase in engagement with services has been achieved, which could be traced back to the self-assessment methodology. Through workshops with partners and surveys completed by startups, DMS was able to identify the precise changes required to improve the selection and onboarding procedures for startups, as well as the services offered to them, such that the companies in the subsequent cohorts could benefit maximally from the programme.

In conclusion, our article has discussed some of the challenges of assessing the impact of open innovation programmes related to data and AI. We have proposed a generalised methodological framework that combines multiple monitoring tools and utilises the direct experiences of stakeholders to evaluate and develop actionable recommendations that can improve the social and economic impact such programmes.

Data and software availability

Underlying data

The main contribution of our article is methodological. All data underlying the methodological results are available as part of the article and no additional source data are required.

The underlying survey data cannot be shared openly due to small sample sizes and the respondents’ public affiliation with DMS Accelerator. In order to prevent the possibility of de-anonymisation and to protect commercially sensitive data, access to these data is controlled and classed as confidential as outlined in the ethics statement.

The original routine monitoring and workshop reports cannot be published because they are classified as “Confidential” in the project. A redacted version of the reports can be provided upon request.

To request access to the above survey and workshop data, please email the authors (maria.1.priestley@kcl.ac.uk and manguiano@zabala.es). Maria Priestley and María Anguiano will evaluate each request and provide access to trusted parties who want to use the data for justified and compatible purposes (e.g. colleagues who want to undertake a comparative research study or a review).

Extended Data

Zenodo Repository: DMS Accelerator monitoring data 25 . https://doi.org/10.5281/zenodo.6401138

This project contains the following extended data:

  • DMS Impact Survey template

  • DMS Needs Analysis Survey template

  • DMS Workshop template

  • DMS service engagement data

Data are available under the terms of the Creative Commons Attribution 4.0 International license (CC-BY 4.0).

Acknowledgements

We are grateful to Dr Gefion Thuermer for facilitating a workshop that informed the creation of this article. We would also like to thank all partners of DMS Accelerator for their active participation and valuable insights. The individuals named here have consented to their inclusion in these acknowledgements.

Funding Statement

This project is funded by the European Union’s Horizon 2020 Programme under Grant Agreement Nº. 825014 (Supporting the European data market providing free support services to data-centric SMEs and start-ups [Data Market Services / DMS Accelerator]).

The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

[version 2; peer review: 1 approved, 1 approved with reservations]

Footnotes

References

  • 1. Trajtenberg M: Ai as the next gpt: a political-economy perspective. Technical report, National Bureau of Economic Research,2018. 10.3386/w24245 [DOI] [Google Scholar]
  • 2. Jordan MI, Mitchell TM: Machine learning: Trends, perspectives, and prospects. Science. 2015;349(6245):255–260. 10.1126/science.aaa8415 [DOI] [PubMed] [Google Scholar]
  • 3. De Marco CE, Martelli I, Di Minin A: European smes’ engagement in open innovation when the important thing is to win and not just to participate, what should innovation policy do? Technol Forecast Soc. 2020;152:119843. 10.1016/j.techfore.2019.119843 [DOI] [Google Scholar]
  • 4. Gupta K: The role of smes and startups in standards development. Available at SSRN 3001513,2017;1–10. 10.2139/ssrn.3001513 [DOI] [Google Scholar]
  • 5. EC: Making the most of the eu’s innovative potential: An intellectual property action plan to support the eu’s recovery and resilience. Directorate-General for Internal Market, Industry, Entrepreneurship and SMEs, GROW,2020. Reference Source [Google Scholar]
  • 6. Dempwolf CS, Auer J, D’Ippolito M: Innovation accelerators: Defining characteristics among startup assistance organizations. Small Bus Admin. 2014;1–44. 10.13140/RG.2.2.36244.09602 [DOI] [Google Scholar]
  • 7. Cohen S, Hochberg YV: Accelerating startups: The seed accelerator phenomenon.2014;1–16. 10.2139/ssrn.2418000 [DOI] [Google Scholar]
  • 8. Chesbrough H: Open innovation: Where we’ve been and where we’re going. Res Technol Manag. 2012;55(4):20–27. 10.5437/08956308X5504085 [DOI] [Google Scholar]
  • 9. Blume T: New Taxonomy for Corporate Open Innovation Initiatives. Springer,2020. 10.1007/978-3-658-27349-1 [DOI] [Google Scholar]
  • 10. Blind K: The impact of standardization and standards on innovation. NESTA Working Paper 13/15,2013. Reference Source [Google Scholar]
  • 11. Canhoto AI, Clear F: Artificial intelligence and machine learning as business tools: A framework for diagnosing value destruction potential. Bus Horiz. 2020;63(2):183–193. 10.1016/j.bushor.2019.11.003 [DOI] [Google Scholar]
  • 12. de Vallejo IL, Scerri S, Tuikka T: Towards a european data sharing space: Enabling data exchange and unlocking ai potential. BDVA Position Paper,2019. Reference Source [Google Scholar]
  • 13. Bone J, Gonzalez-Uribe J, Haley C, et al. : The impact of business accelerators and incubators in the uk.2019. Reference Source [Google Scholar]
  • 14. Galiyeva N, Fuschi DL: A research proposal for measuring the effectiveness of business incubators. Int J Organ Innov. 2018;5(3):32–46. Reference Source [Google Scholar]
  • 15. Wright M, Clarysse B, Van Hove J: A look inside accelerators. Nesta report,2015. Reference Source [Google Scholar]
  • 16. Cohen S, Fehder DC, Hochberg YV, et al. : The design of startup accelerators. Res Policy. 2019;48(7):1781–1797. 10.1016/j.respol.2019.04.003 [DOI] [Google Scholar]
  • 17. Passani A, Spagnoli F, Bellini F, et al. : Collective awareness platform for sustainability and social innovation (caps). In Organizational Innovation and Change. Springer,2016;103–114. 10.1007/978-3-319-22921-8_9 [DOI] [Google Scholar]
  • 18. Paik JH, Scholl M, Sergeev R, et al. : Innovation contests for high-tech procurement. Res Technol Manag. 2020;63(2):36–45. 10.1080/08956308.2020.1707007 [DOI] [Google Scholar]
  • 19. Calandrelli ED: An evaluation of short innovation contest implementation in the federal context. PhD thesis, Mas-sachusetts Institute of Technology,2013. Reference Source [Google Scholar]
  • 20. Bender E: Challenges: crowdsourced solutions. Nature. 2016;533(7602):S62–S64. 10.1038/533S62a [DOI] [PubMed] [Google Scholar]
  • 21. Landzaat W, Godel M, Perkins R: Data pitch evaluation. London Economics report,2019. Reference Source [Google Scholar]
  • 22. Henchoz N, Puissant PX, Leal AS, et al. : Artist residencies for innovation: development of a global framework. In ACM SIGGRAPH 2019 Art Gallery.2019;1–5. 10.1145/3306211.3320140 [DOI] [Google Scholar]
  • 23. FI-Impact: D2.1 impact assessment guidebook. Future Internet Impact Assurance -Project number 632840,2014. Reference Source [Google Scholar]
  • 24. Ivanov T, Eichhorn T, Berre AJ, et al. : Building the databench workflow and architecture. In International Symposium on Benchmarking, Measuring and Optimization. Springer,2019;165–171. 10.1007/978-3-030-49556-5_17 [DOI] [Google Scholar]
  • 25. Priestley M: Dms accelerator monitoring data. Zenodo[extended data].2021. 10.5281/zenodo.4905499 [DOI] [Google Scholar]
Open Res Eur. 2022 May 23. doi: 10.21956/openreseurope.15814.r29098

Reviewer response for version 2

Gernot Pruschak 1

Dear authors,

Thank you for taking the time and updating the manuscript. The authors put in substantial efforts in revising the manuscript and detailly responded to all my comments. I especially congratulate the authors for increasing the sample size.

Nevertheless, whereas the authors addressed most of my comments successfully, there still exists room for improvements in two areas: 1) Terminology and 2) Embodiment in the (Scientific) Literature. However, both of these issues can be solved via the same mean:

The authors claim that data-centric innovation is a relatively new phenomenon, a statement that is definitely true. However, at the same time the authors put in lengthy efforts in elaborating that the focus of their research are not accelerators and incubators but instead open innovation programs. If this is the case, they might also embody their research in the wider open innovation literature. As a case in point, there exists ample literature discussing open data, especially in the public domain (e.g., governments and universities). It would be therefore great if the authors could situate their research more in context with recent developments in the open data literature. This would boost the scientific background of the paper and make it clear to readers why the focus is on open innovation and not on accelerators and incubators.

If applicable, is the statistical analysis and its interpretation appropriate?

No

Is the study design appropriate and is the work technically sound?

Partly

Is the work clearly and accurately presented and does it engage with the current literature?

No

Are the conclusions drawn adequately supported by the results?

No

Are sufficient details of methods and analysis provided to allow replication by others?

Partly

Are all the source data and materials underlying the results available?

Yes

Reviewer Expertise:

Open Innovation, Open Science, Open Data, Quantitative Research, Surveys, Descriptive Statistics, Parametric Statistics

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however I have significant reservations, as outlined above.

Open Res Eur. 2022 Jul 27.
Maria Priestley 1

Dear Dr. Pruschak,

Thank you very much for your additional comments on our revised article. We have now uploaded a new version which incorporates your helpful feedback. The terminology and positioning of our article in the wider scientific literature has been clarified using the following changes:

  • The Abstract has been updated to specify that this work builds on earlier open innovation initiatives related to open data.

  • A new paragraph has been added to the Introduction section, which now includes a short discussion of the open data movement and references to research in this area.

We would like to thank you once again for your time reviewing our work. We hope that the new version of the article meets your expectations.

Best wishes,

Maria

Open Res Eur. 2022 Apr 25. doi: 10.21956/openreseurope.15814.r29097

Reviewer response for version 2

Francesco Mureddu 1

I am satisfied with the amendments

If applicable, is the statistical analysis and its interpretation appropriate?

Yes

Is the study design appropriate and is the work technically sound?

Yes

Is the work clearly and accurately presented and does it engage with the current literature?

Yes

Are the conclusions drawn adequately supported by the results?

Yes

Are sufficient details of methods and analysis provided to allow replication by others?

Yes

Are all the source data and materials underlying the results available?

Yes

Reviewer Expertise:

Impact assessment and evaluation, innovation, policy analysis.

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

Open Res Eur. 2021 Aug 24. doi: 10.21956/openreseurope.14692.r27372

Reviewer response for version 1

Gernot Pruschak 1

Dear authors,

Thank you for giving me the opportunity to review your manuscript entitled “Measuring the impact of publicly funded open innovation programs: the case of Data Market Services Accelerator.” This research topic is highly relevant due to the strong increase in publicly funded open innovation initiatives. Thus, it is of utmost importance to generate standard measurement criteria to assess their effectiveness and efficiency. In addition, I appreciate the idea of employing the Data Market Services Accelerator as a case study to test the implementation of the metrics.

However, whereas the research idea is great, there exist room for improvements in the research execution. I outline my concerns below:

Standardized measurement scale:

The authors` “intention in this article is to identify a flexible and generalizable framework” for measuring the impact of publicly funded open innovation initiatives (p. 7). Yet the results as exhibited in Tables 4 through 7 solely highlight criteria relevant for the DMS. Based on the claims in the introduction (and also the lengthy elaborations on why open data requires standardization), I was surprised to not see a list of criteria applicable to various open innovation programs as the result. In fact, the results part contains the findings from the measures for the DMS. Consequently, there exist two ways forwards: Either the authors reduce their claims in the introduction and transparently depict that the manuscript focuses on reporting the effectiveness and efficiency of the DMS or the authors tone down the DMS related work but instead present the general toolkit as their main results.

Relevance and coherence of arguments:

The manuscript includes (especially in the introduction but also in other parts) lengthy elaborations on topics only peripherally related to measuring open innovation programs. In turn, at other parts the authors jump from one thought to another without providing logical connections. For example, the sub-chapter “Open innovation programmes for startups and SMEs” does not include any example of open innovation programmes (Kindly note that this only represents one of multiple appearances of this problem). Instead, it elaborates in general what open innovation is and how it can be seen in the context of AI. To overcome this issue, I would suggest the authors to rework the whole manuscript and cut all paragraphs that do not directly relate to the main theme. This especially applies to the introduction which is considerably long and contains several paragraphs which are not necessary from a scientific point of view but instead are written in similar fashion with consulting reports which must contain certain words (AI, standardization, Big Data, …) and phrases. On this note, I would also like to remind the authors that good scientific practice implies that we include all information sources as citation. In fact, the first two pages of the manuscript contain only four citations. There exist numerous paragraphs with statements including important contents that do not include references. To make this more comprehensible for readers please include citations for every idea that was not discovered by yourself. As an approximate guideline I would like to give you the following advice: Every sentence containing a statement (not based on your research) that is more specific than “the sky is blue” should include a citation

Accelerator, incubator and open innovation:

“Support programmes for startups and SMEs have typically taken the form of accelerators and incubators. More recently, open innovation programmes have emerged at the intersection of these two paradigms” (p. 2). Whereas the introduction states that this research focuses on the intersection and interplay of accelerators and incubators, the methods and results sections do only focus on accelerator. Again, please ensure coherency throughout the manuscript. On this note, I also doubt that the developed measures can apply to all open innovation programs as open innovation does not only take place within startups and SMEs but also includes stake holders coming from large industry, universities and other research institutions. Consequently, I would suggest toning down the focus on open innovation and instead rely on a more nuanced approach towards entrepreneurship.

Methods:

The survey sample consists of six respondents for the first cohort and 20 respondents for the second cohort. Nevertheless, the authors argue as if the sample size would be sufficiently large by employing percentage numbers and inducing from the survey results that the DMS worked successful. Being a quantitative researcher myself, I strongly doubt the expressiveness of these two surveys. Consequently, I would advise to refrain from using percentages but instead including the actual numbers (e.g. 2 companies instead of 33.3% of our respondents) to transparently depict the small sample size. Moreover, please refrain, especially in the discussion section, from making strong claims based on the survey and workshop results.

Discussion:

The discussion section does not include limitations and suggestion for future research. Please add them. Moreover, readers might want to know the theoretical and scientific implications of your work. In addition, the discussion section should not present or discuss new results. Therefore, I would suggest moving those parts of the “Increasing engagement” sub-chapter highlighting new results (e.g. results from the first workshop) to the results section. Last, the authors argue in the beginning of the discussion that their research shows differences in regard to geographic origins and industry sectors. However, the results section does not split the findings into geographical regions or industry. Please only include elaborations on findings in the discussion section that were also shown in the results section.

Summing up, I agree with the importance and the potential of the research topic. Yet I believe the manuscript requires substantial improvement work before it can be considered for indexing. Therefore, my suggestion would be a major revision.

If applicable, is the statistical analysis and its interpretation appropriate?

No

Is the study design appropriate and is the work technically sound?

Partly

Is the work clearly and accurately presented and does it engage with the current literature?

No

Are the conclusions drawn adequately supported by the results?

No

Are sufficient details of methods and analysis provided to allow replication by others?

Partly

Are all the source data and materials underlying the results available?

Yes

Reviewer Expertise:

Open Innovation, Open Science, Open Data, Quantitative Research, Surveys, Descriptive Statistics, Parametric Statistics

I confirm that I have read this submission and believe that I have an appropriate level of expertise to state that I do not consider it to be of an acceptable scientific standard, for reasons outlined above.

Open Res Eur. 2022 Apr 4.
Maria Priestley 1

Dear Dr. Pruschak,

We would like to thank you for the detailed feedback which you provided on our manuscript last year. Over the past few months, we have conducted a major revision of the paper in response to your suggestions.

A full description of the changes can be found in a separate document linked here.

The main changes can be summarised as follows:

  • Our claims about the generalisability of the framework and the strength of claims regarding our results have been reduced. We have also highlighted that our manuscript focuses on measuring the performance of only one programme – the DMS Accelerator.

  • More references have been added to support our claims, particularly in the Introduction section.

  • We have provided more examples of open innovation, and have justified our decision to describe DMS and other similar initiatives as examples of open innovation, rather than relying on traditional notions of acceleration and incubation. We have also clarified that the scope of our paper focuses on initiatives that serve startups and SMEs, which are only some of the various possible stakeholders that can be targeted by open innovation initiatives.

  • We have decided to retain information about the characteristics of data-centric programmes (including words such as “AI”, “standardisation” and “big data”), as these concepts are an essential part of the specialised technical training and advisory services offered by data-centric open innovation programmes. By including these terms in the article, we are able to engage with audiences outside academia who work in data-oriented initiatives.

  • We have obtained additional data, which has increased our survey sample size from 20 to 47 respondents across the three DMS cohorts. We have also reported the new results using actual numbers (rather than percentages) for transparency.

  • We have now added a section on the “Limitations and future work”.

Once again, thank you for taking the time to consider our submission. Your insights are greatly appreciated. We hope that you will find the updated article more rigorous and worthy of indexing.

Open Res Eur. 2021 Aug 19. doi: 10.21956/openreseurope.14692.r27403

Reviewer response for version 1

Francesco Mureddu 1

The present article discusses the question of how impact can be assessed in open innovation programmes related to data and AI. Drawing on the authors’ experience of publicly-funded data incubators in Europe, they review previous approaches to impact assessment, their position in the wider economic context, and what distinguishes data and AI programmes from other types of accelerators in the European tech scene. In that regard, the authors propose a methodology to assess their impact and apply it to one such programme, Data Market Services (DMS). The methodology presented here is being used to monitor and adjust the governance of our services based on work done in the first two years of the programme.

The paper is very interesting and informative, rigorous and actually very useful also for my professional activity. It is flawless, and I recommend indexing as it is.

If applicable, is the statistical analysis and its interpretation appropriate?

Yes

Is the study design appropriate and is the work technically sound?

Yes

Is the work clearly and accurately presented and does it engage with the current literature?

Yes

Are the conclusions drawn adequately supported by the results?

Yes

Are sufficient details of methods and analysis provided to allow replication by others?

Yes

Are all the source data and materials underlying the results available?

Yes

Reviewer Expertise:

Impact assessment and evaluation, innovation, policy analysis.

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

Open Res Eur. 2022 Apr 4.
Maria Priestley 1

Dear Dr. Mureddu,

We would like to thank you very much for your encouraging feedback on our manuscript last year. We wanted to let you know that over the past few months we have made some revisions to the article in response to the second reviewer’s comments.

A full description of the changes can be found in a separate document linked here.

The main changes can be summarised as follows:

  • Our claims about the generalisability of the framework and the strength of our claims regarding the results have been reduced. We have also highlighted that our manuscript focuses on measuring the performance of only one programme – the DMS Accelerator.

  • More references have been added to support our claims, particularly in the Introduction section.

  • We have provided more examples of open innovation, and have justified our decision to describe DMS and other similar initiatives as examples of open innovation, rather than relying on traditional notions of acceleration and incubation. We have also clarified that the scope of our paper focuses on initiatives that serve startups and SMEs, which are only some of the various possible stakeholders that can be targeted by open innovation initiatives.

  • We have obtained additional data, which has increased our survey sample size from 20 to 47 respondents across the three DMS cohorts. We have also reported the new results using actual numbers (rather than percentages) for transparency.

  • We have now added a section on the “Limitations and future work”.

 Once again, thank you very much for taking the time to consider our submission. We hope that the updated version still resonates with your experience and remains useful to your work.

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Data Availability Statement

    Underlying data

    The main contribution of our article is methodological. All data underlying the methodological results are available as part of the article and no additional source data are required.

    The underlying survey data cannot be shared openly due to small sample sizes and the respondents’ public affiliation with DMS Accelerator. In order to prevent the possibility of de-anonymisation and to protect commercially sensitive data, access to these data is controlled and classed as confidential as outlined in the ethics statement.

    The original routine monitoring and workshop reports cannot be published because they are classified as “Confidential” in the project. A redacted version of the reports can be provided upon request.

    To request access to the above survey and workshop data, please email the authors (maria.1.priestley@kcl.ac.uk and manguiano@zabala.es). Maria Priestley and María Anguiano will evaluate each request and provide access to trusted parties who want to use the data for justified and compatible purposes (e.g. colleagues who want to undertake a comparative research study or a review).

    Extended Data

    Zenodo Repository: DMS Accelerator monitoring data 25 . https://doi.org/10.5281/zenodo.6401138

    This project contains the following extended data:

    • DMS Impact Survey template

    • DMS Needs Analysis Survey template

    • DMS Workshop template

    • DMS service engagement data

    Data are available under the terms of the Creative Commons Attribution 4.0 International license (CC-BY 4.0).


    Articles from Open Research Europe are provided here courtesy of European Commission, Directorate General for Research and Innovation

    RESOURCES