- Open Access
Does health intervention research have real world policy and practice impacts: testing a new impact assessment tool
Health Research Policy and Systems volume 13, Article number: 3 (2015)
There is a growing emphasis on the importance of research having demonstrable public benefit. Measurements of the impacts of research are therefore needed. We applied a modified impact assessment process that builds on best practice to 5 years (2003–2007) of intervention research funded by Australia’s National Health and Medical Research Council to determine if these studies had post-research real-world policy and practice impacts.
We used a mixed method sequential methodology whereby chief investigators of eligible intervention studies who completed two surveys and an interview were included in our final sample (n = 50), on which we conducted post-research impact assessments. Data from the surveys and interviews were triangulated with additional information obtained from documentary analysis to develop comprehensive case studies. These case studies were then summarized and the reported impacts were scored by an expert panel using criteria for four impact dimensions: corroboration; attribution, reach, and importance.
Nineteen (38%) of the cases in our final sample were found to have had policy and practice impacts, with an even distribution of high, medium, and low impact scores. While the tool facilitated a rigorous and explicit criterion-based assessment of post-research impacts, it was not always possible to obtain evidence using documentary analysis to corroborate the impacts reported in chief investigator interviews.
While policy and practice is ideally informed by reviews of evidence, some intervention research can and does have real world impacts that can be attributed to single studies. We recommend impact assessments apply explicit criteria to consider the corroboration, attribution, reach, and importance of reported impacts on policy and practice. Impact assessments should also allow sufficient time between impact data collection and completion of the original research and include mechanisms to obtain end-user input to corroborate claims and reduce biases that result from seeking information from researchers only.
Since the 1980s there has been a growing expectation that health research will have direct social and economic utility and impact [1–5]. Health intervention research, in particular, which uses scientific methods to produce knowledge about treatments, services, programs, or strategies that aim to protect, promote, or improve health, is assumed to hold immediate promise for influencing and improving future policy and practice [2, 6]. Research funding submissions require applicants to predict the practical benefits that might flow from their planned studies [7, 8]. However, a large proportion of funded research fails to translate into real world solutions [6, 9–12].
Over 90 different terms are used to describe research impact on policy and practice , including translation, diffusion, adoption, adaptation, uptake, exchange, research utilization, and research implementation. How and to what extent research is translated into policy and practice is emerging as an important field of research [14, 15]. A key area within this field is assessment of impact, or how to measure the dividends from research [14, 16].
In practice, assessments of research impacts are mostly commissioned by governments to determine the public benefit from research spending [4, 15, 17]. Governments are increasingly signaling that research metrics of research quality are insufficient to determine research value because they say little about real-world benefits of research [3, 4, 7].
To date, gold standards in the assessment of research impact combine case study narratives with relevant qualitative and quantitative indicators [4, 16–18]. The impact assessment literature universally calls for expanded measures to better assess the nature and quality of real world impacts, as well as better predictive measures of longer term benefits [4, 14]. “The holy grail is to find short term indicators that can be measured before, during or immediately after the research is completed and that are robust predictors of the longer term impact … from the research” .
In this paper, we build on the existing literature to propose an expanded ‘impact assessment’ framework, and apply a model which builds on current best practice to 5 years (2003–2007) of intervention research funded by Australia’s peak health and medical research funding agency, the National Health and Medical Research Council (NHMRC) [15, 18–24].
Our primary aims were to i) pilot a modified impact assessment process and ii) determine what proportion of intervention projects had any demonstrable impact on subsequent policy or practice in ‘real world’ settings after the research undertaken had concluded and to group these projects according to the magnitude their impacts. The project was approved by the University of Sydney Human Research Ethics Committee (15003).
A number of approaches have been developed to describe the impacts of research . The Payback framework [19, 24] and its adaptation into the Canadian framework  are the most widely used . These approaches and the literature underpinning them indicate that it is important for impact assessment to distinguish between different types and stages of impact and to draw on multiple sources of data. A conceptual framework can help organize data collection, analysis, and reporting to promote clarity and consistency in the impact assessments made. It is important to acknowledge that such a model facilitates impact assessment rather than being a precise model of how research impact occurs .
The primary focus of our study was to examine the policy and practice impacts of research that occur in the ‘real world’ after the research had concluded. We combined and adapted the Payback and Canadian frameworks to produce a conceptual model that would best fit our purpose. We grouped different types of impacts into four levels of impact that might arise from intervention research: i) scholarly outputs; ii) translational outputs; iii) policy or practice impacts; and iv) long-term population outcomes. Each impact level was populated with sub-categories or indicators to further facilitate assessment of the type of impacts that occur at each level (Figure 1). For comparison with our approach, the Payback categories are included in Figure 1 alongside the impact levels we describe.
We considered that the scholarly outputs, such as publications, acquisition of new research funding, and research capacity building, were not real-world impacts as we had defined them and therefore they are not reported here; others have made a similar distinction .
Translational outputs were defined as those activities that occur beyond research publication and are designed to facilitate uptake of study findings in real-world settings. These outputs may be activities undertaken by researchers, their institutions, or government programs to facilitate knowledge uptake such as implementation protocols, training workshops, and information exchange meetings. They may also be part of a general dissemination strategy conducted at the end of a study. These outputs may or may not lead to policy or practice impacts and long-term population changes. While some impacts at this level, particularly those related to information packaging for implementation, are included as part of our impact assessment here, we considered these to be of lesser importance than policy or practice impacts.
Policy or practice impacts were defined as demonstrable changes, or benefits to products, processes, policies, and or practices, that occur after a research project has concluded. These impacts are concrete, measurable changes in policy or practice such as a new government policy, a change in organizational or clinical practice, a health education campaign or related new funding that can be attributed to the research intervention in question. Impacts at this level could also include stopping or changing existing interventions following demonstration of intervention ineffectiveness. Policy or practice impacts can be widespread or localized, and may benefit specific or general populations. Impacts at this level are our primary focus in this paper.
Finally, we defined long-term population outcomes as changes in health behaviors and health outcomes, such as disease incidence, prevalence, or other health indicators, or as improvements in social or economic outcomes. Such changes rarely occur quickly, and they may be difficult to attribute to health intervention research, let alone to single studies [9, 15]. Due to these attribution issues and the distal nature of population outcomes, outcomes at this level were not assessed in our research.
As in other frameworks [20, 24], our model (Figure 1) reflects a sequence of impacts (while recognizing that this will often not occur) and distinguishes immediate research outputs from impacts which tend to occur later and beyond the research setting or context. Models with sequential stages assume that an impact or output at one stage may, or may not, lead to increasingly concrete and widespread impacts over time.
Our sample of intervention research studies was generated from a list provided by the NHMRC of 721 primary research grants that were potentially intervention-related (as identified by the NHMRC from their records) which commenced between 2003 and 2007. Studies were included in our sample if they met our definition of intervention research (i.e., any form of trial or evaluation of a service, program, or strategy aimed at disease, injury, or mental health prevention, health promotion or psychotherapeutic intervention, conducted with general or special populations, or in clinical or institutional settings), if the study commenced before 2007, and if the study data had been collected and analyzed before we began our data collection in 2012. Clinical trials of potentially prescribable drugs, vaccines, and diagnostic tests were excluded because of the very different trajectories that such therapeutic goods are required to navigate before being authorized for use by the Australian Therapeutic Goods Administration. We determined if studies met our inclusion criteria initially by reviewing grant documents and publications, after which 83 studies were included in our sample, and then by surveying the chief investigators (CIs), after which 13 studies from our initial sample were found to be ineligible, leaving 70 eligible studies. We chose to include intervention studies commencing between 2003 and 2007 in our sample because this 5-year window provided a balance between allowing projects sufficient time post-study completion to have realized real-world impacts by the time of our data collection in 2012 and to minimize recall bias about grants that finished too long ago.
Our study used a mixed methods sequential methodology. CIs were invited to complete two surveys and an interview (Figure 2). Only those CIs who completed both surveys and the interview were included in our final sample (n = 50).
Preparing case studies to determine impacts
In order to determine if the studies had included real-world impacts relevant to our study we collected data from multiple data sources (Figure 2 and described below) which were triangulated to produce case studies. The case studies were reviewed by two authors and either classified as having at least one impact or no impacts.
An online survey (see http://hdl.handle.net/2123/9864) was sent to study CIs to elicit their views about the potential influence and significance of their research findings and any impacts of their research in the real world following the research project. They were also asked questions on contextual factors and research characteristics known to be influential in practice uptake.
Independent confirmation of results
Information about the research findings of each study was considered to be essential in determining the extent to which the reported post-study impacts could be attributed to the specific research projects under consideration. Rather than only relying on outcomes reported by the CIs, which may have been subject to responders’ bias and selective recall , the information related to the study findings included in the case studies was subject to a separate verification process. This involved developing a summary for each study based on published results related to the principal study outcomes which was then reviewed by a panel of authors. The panel classified interventions according to whether there were statistically significant changes on principal outcomes: those that did, those with ‘mixed’ results (in cases where there were significant changes for some but not all principal outcomes or if unintended or ancillary results were emphasized over the primary focus of the original research questions), and those where the intervention did not produce statistically significant changes. In some cases, no published results related to the principal outcomes could be found. The panel was not asked to make judgments in relation to the quality of the research, the appropriateness of the research methods, or the importance of the findings. In this assessment and verification step, ‘statistical significance’ provided a simple and objective way of considering study outcomes, and was not intended as a measure of the clinical or societal value of the outcome.
Semi-structured ‘conversational’ interviews were customized for each study based on responses to the online survey. The interviewers sought to obtain more information about any claimed real-world impacts of the study and to explore the CIs perceptions of what had helped or hindered the uptake of their intervention. Interviews covered the following themes: i) implications: what are they and have any come into effect?; ii) impacts: what occurred, and why do you think some impacts occurred, while others did not?; iii) engagement with potential end-users (research team and other stakeholders) before, during, and after the study; iv) communication before, during, and after the completion of the study; v) contribution to knowledge: what was the nature of any contribution?; and vi) follow-up: what occurred following the post-research dissemination?
Document analysis to corroborate reported impacts
The impacts reported by the CIs were corroborated, where possible, by completing an internet search using Google to search for relevant web pages, newsletters, media releases, or other documents.
Piloting the impact scoring tool
We sought to pilot an impact scoring process so that we could group and compare studies according to the magnitude of their impact. This process included the following steps:
Preparing impact assessment summaries
The detailed case studies that included real-world impacts relevant to our study (n = 19) were used to develop a two to three page impact assessment summary for consideration by an expert panel. A common format was used to allow comparison between cases. The summaries included the study aims and research question/s, a description of the intervention, the study findings, post-study impacts that potentially met our definition of policy and practice impacts (only impacts that had actually occurred rather than those that could potentially occur in the future were included), evidence of independent corroboration and attribution of impacts, any contextual factors that may have had an influence on impact, and a summary of information and evidence related to each of our scoring criteria (corroboration, attribution, reach, and importance) for each of the reported impacts (see scoring of impacts below).
Convening an expert panel
An expert panel made up of 12 experienced intervention researchers from the fields of public health, health policy, and clinical medicine, 6 of whom were external to the project, was convened to review and assess the impact assessment summaries and provide an overall assessment of the policy and practice impact of each of study. Four panelists were former or current government health policy makers. As the studies being assessed were heterogeneous in terms of the topic area they addressed, it was not possible to convene a panel made up of content experts. Panelists were therefore selected because of their intervention research expertise and their knowledge of how evidence is translated into policy and practice. None of the panelists had been involved in the studies being assessed.
To introduce the session, the panel was briefed on our study methods. They were also introduced to our conceptual framework and provided with our definition of policy and practice impacts. Each case study was then presented in turn. After each presentation, panelists were given an opportunity to ask clarifying questions before scoring impact for that case. Panel members were provided with the case summaries in advance and hard copies of these were made available to the panelists for reference during the panel proceedings. Panel proceedings were completed over the course of a single day.
Scoring of impacts
The conceptual framework (Figure 1) was used to focus the panel’s attention on the impacts of interest for our study during the scoring process. Only impacts that were consistent with the categories of interest within this model were scored. The combined impacts of each case were scored using a modification of the scoring tool used by the Australian Excellence in Innovation (EII) trial  and the UK’s Research Excellence Framework trials [27, 28]. The original processes considered intervention reach combined with significance of impact; however, a recommendation from the EII trial was that reach and impact significance should be judged separately to avoid biasing small studies . The need for also considering corroboration and attribution of claimed impacts has been repeatedly emphasized in the literature [4, 14, 16, 17, 22].
Based on these recommendations we developed a scoring system that included four dimensions of impact, namely corroboration, attribution, reach, and importance. Importance was used as it was felt the term significance may be confused with the statistical significance of the study findings rather than the importance of the reported impacts. Our scoring system consisted of a series of questions considered by the panelists when scoring each dimension of impact and a rating system based on a Likert scale of 1 to 9 for scoring each dimension (Table 1). We also provided panelists with specific instructions in relation to scoring each dimension. For example, we instructed the panel to judge reach of the impacts against what was possible for the relevant target population, not against other interventions with different potential, and to judge importance of the impacts claimed by referring to the definition and examples in the scale. Using this information as a guide, panelists were asked to score the overall impact of each study, independently, using hand held keypad transmitters.
The results from the panel were analyzed by examining the distribution of scores as group means and standard deviations for each question, as well as a measure of spread of responses (coefficient of variation, standard deviation/mean). The distribution was also categorized into tertiles: low, medium, and high impacts.
A total of 50 CIs completed both surveys and an interview, providing a response rate of 71% (50 out of a possible 70 eligible studies). Of the 20 CIs that did not respond, 3 did not return Survey 1 (83 sent; 7 confirmed as ineligible), 11 did not respond after receiving Survey 2 (73 sent; 6 confirmed as ineligible), and 6 did not respond to a request for interview (56 invitations sent).
Number of studies with impacts and other study characteristics
Thirty-one studies (62%) were classified as having no real-world impacts and 19 (38%) were assessed as having at least one impact.
The 50 studies included in our overall sample represent a diverse range of intervention research projects. Just over a third (36%) were primary prevention or health promotion interventions; 24% were early intervention or screening interventions; and 40% were interventions related to the treatment or management of an illness, disease, or disorder. Table 2 provides an overview of the topic areas addressed by the studies within each of these intervention groups for impact and no impact categories. The proportion of primary prevention or health promotion interventions in the no impact group (n = 15; 48%) was greater than that for the impact group (n = 3; 16%).
The funding period for the studies in our final sample ranged from 1 to 4 years, with the mean funding period being 2 years. A large proportion (n = 20; 40%) of studies in our sample did not commence until 2007, the final year in our study period (2003–2007). For most of the studies (88%; n = 44), the grant funding period concluded by 2009, while the funding period for all of the studies (n = 50) had concluded by 2011. The pattern in terms of number of studies by year in which the funding was completed was similar for the impact and no impact categories.
Results of the impact scoring tool pilot
Spread of impact scores
For the studies assessed as having impacts (n = 19), the mean impact scores (n = 12 raters) and 95% confidence intervals for each of the impact dimensions within our scoring system (corroboration, attribution, reach, and importance) are provided in Table 3. Responses for each dimension were ranked and divided into tertiles and summarized across dimension in the far right column, as low (n = 6), medium (n = 7), and high (n = 6) impact groups. The four dimensions were not summed to provide a total score for each study as we considered that each dimension should remain as a separate consideration. However, there was general concordance in the ranking of each study across dimensions. The mean scores across cases showed substantial variation, which allowed scores to be categorized into tertiles and these were examined for differences in magnitude of impact between studies Figure 3.
Distribution of responses among raters
The coefficient of variation between raters’ scores for each grant within each impact dimension was generally small, indicating that there was a high degree of consistency of scores between raters. The variation for ratings was less than 0.25 for all of the projects in the high (n = 6) and medium level (n = 7) impact groups. However, it ranged from 0.39 to 0.41 in the low impact group (n = 6), indicating that there was less agreement between raters for these studies, which tended to have less corroborative evidence about impact, or were assessed as being closer to traditional research metric impacts than to policy and practice impacts. This showed greater spread in the distribution of panel responses in a subset of grants appraised.
Type of impacts reported
The type of impacts reported for the 19 studies classified as having impacts are summarized by impact score group in Table 4. Some studies reported impacts of more than one impact type (e.g., policy and service change). Half of the impacts (n = 21) reported were in the form of translational outputs (e.g., intervention resources, including websites, publications, and manuals for end users or training). Although potentially useful resources for policy and practice, resources were not necessarily proof of use and we considered them to be the lowest level of impact. Where these impacts were endorsed by professional bodies or had significant reach, the impacts were considered to be of higher magnitude. The remaining impacts (n = 21) were classified as policy and practice impacts which included clinical practice changes (n = 6), service changes (n = 9), organizational changes (n = 1), commercialization of products or services (n = 1), and policy changes (n = 4). As these all require a more substantial degree of personal or organizational change, they were considered to equate to a higher level of impact.
Within the group of studies (n = 6) assessed by the panel to have high impacts, there was one commercial product launch, three policy changes, four changes to service delivery, one organizational change, and one change to clinical practice. All of the studies in the high and medium impact group had at least one impact that fell within the policy and practice impact category, while only four out of six of the studies in the low impact group had impacts of this type. None of the studies in the low and medium impact groups had more than one impact within the policy and practice category, two studies within the high impact group had three impacts each of this type, three had two impacts each of this type, and two had only one impact of this type. Many of the projects classified as having medium (n = 7) and low impact (n = 6) reported impacts that were anecdotal, with weak evidence that was not easily corroborated.
There have been a number of studies that have examined the impacts of a set of publically funded intervention studies, where the unit of measure is an individual study. Our study differs from these in terms of its specific focus on intervention research in clinical or community settings and on measurable policy and practice impacts rather than scholarly outputs or longer term population outcomes [9, 29–31]. We chose to focus on policy and practice impacts as there is a need for reliable measures of impact to provide sound information about translation beyond the research setting, and to counter-balance the tendency to focus on research metrics as a sole indicator of impact. Other studies of health research impacts have defined their scope in terms of the content area (e.g., breast cancer, stroke) [21, 25, 26, 32, 33], or assessed the impacts of a program of research or research institution as a whole [22, 23].
We found that single intervention research studies can and do have concrete and measurable post-research real-world impacts, with 38% of the studies in our sample demonstrating some impacts on policy and practice. The impacts were often multiple and diverse, covering all of the categories of interest within our research impacts model (Figure 1). We also found that the magnitude of impact varied between studies. Studies received lower impact scores where their impacts involved the development of resources and training, rather than concrete changes to policy and practice and/or because evidence to corroborate the researcher’s claims about impact was weak or could not be found. When the studies were divided into tertiles, three almost equal groups (high, medium, and low impact) were formed. In this paper we do not discuss why some studies had impact and others did not. This will be the subject of another paper.
Other research has found that individual studies from a range of research types, including basic, applied, and clinical research, can have wider impacts (outside of research settings) [9, 21, 29]. However, it is difficult to compare our findings with these previous studies due to differences in study methodologies. For example, another Australian study that examined the impacts of intervention research found moderate to high policy and practice impacts were scored in 10 out of 17 cases (59%) . The remaining cases were scored as having limited impact, but the scoring system did not allow a score of zero to be recorded for the impact category. The study also used a broader definition of policy and practice impacts than we did and reported on a more homogeneous sample of interventions. In another study examining the impacts of cardiovascular research (both basic and clinical), all of the 14 clinical cases in the sample had impacts beyond the research setting . In this study, the timeframe from when the original studies were completed to impact data collection was between 15 and 20 years.
The method we piloted resulted in a good level of agreement between raters about the extent of the post-research policy and practice impacts for most of our case studies; there was more variability in scoring of the low impact cases. Our method also provided an estimate of the magnitude of the impact, which is important in order to compare impacts across heterogeneous studies . However, it is possible that not enough time had elapsed for all of the studies in our sample and subsequent impacts may yet occur. It has been suggested that it can take a considerable amount of time (up to 17 years) for evidence to be translated into practice [33, 34]. A considerable proportion (40%) of the studies in our sample commenced in 2007, 5 years before our data collection began in 2012, and nearly half (48%) of the studies were completed in the 3 years prior to our data collection. It is also possible that some of the studies in our sample did not have impacts because they were only single studies. Ideally, policy and practice change should result from a summarized body of knowledge in the form of systematic reviews or research synthesis, rather than single studies alone. It may be that it is more appropriate to apply an impact assessment process to a researcher’s body of work rather than a single research study. This would capture impacts that cannot be attributed to a single research study and allow sufficient time for impacts to occur. We suggest it is important to make explicit the purpose of any research impact assessment process to determine whether the unit of analysis is to be that of a single study, a researcher’s body of work, the research institution, or of a synthesis of all of the published evidence on a given topic.
While others have scored the impacts of research [9, 21, 22, 28, 31], our scoring system is unique in that it involved scoring four separate dimensions of impact, namely corroboration, attribution, reach, and importance. This system sought to overcome some of the issues with attribution and corroboration that had been encountered during previous assessment processes, as well as the need to distinguish reach from significance so as not to downplay the potential impact of smaller studies, or studies with small target groups [4, 14, 16, 17, 22]. However, on reflection, panel members tended to score the importance of the impacts highly across all dimensions, compared to the benchmark examples provided in the scoring sheet. For example, we used the impact of human papilloma virus vaccine research [35, 36] as an example of a study that would be given a high score of nine for importance because of its global implications for prevention of a serious and prevalent disease. None of the research studies in our sample had impacts as significant as this example, but some panel members nevertheless scored some projects at eight for importance. The reasons for this are unclear, although it may be that something of a ceiling effect operates, with raters scoring all projects with impacts above a certain level in terms of importance on a similarly high scale. While the scores may be comparable within any given process, further guidance to panels about appropriate magnitude of scores should be provided to support between-panel comparison. It may also be necessary to review individual panel member scores during the panel process to ensure that all panel members are scoring impacts based on similar considerations .
Having a system for scoring cases according to the level of corroborating evidence and degree of evidence for attribution was beneficial. However, it was not possible to find supporting information of acceptable quality for all of the studies in our sample. There was a greater degree of variability in panel scores where limited supporting information was available or it was absent. To improve reliability of scoring, it may be necessary to mark case summaries without adequate supportive evidence as ‘unclassified’  and not score these cases until supporting evidence becomes available. Another way to verify claims made by CIs would be to supplement publicly available information with third party interviews of end-users. This is a resource-intensive process  and for this reason few examples where interviews or surveys of end users are used as part of impact assessment processes can be found [9, 37]. To improve the feasibility of this approach, we suggest that end user input should ideally be sought post-panel assessment, so that resources are expended on impacts of potential significance.
Reliance on only the perspectives of researchers is problematic for other reasons. Researchers may sometimes be unaware of post-research impacts of their research, because research impact is seldom a key research performance measure, so they do not actively track uptake, or because they consider that their role finishes with publication . This may result in research impacts being under-reported [26, 29, 30, 32]. Researchers may also differ in the way they conceptualize and explain the impact of their research compared to other groups . In addition, societal importance is a very difficult concept to judge. Therefore, contributors to impact assessment research and impact assessment panels should include as wide a cross section of viewpoints as possible [3, 4, 22]. While our panel included both researchers and practitioners (policy and clinical), limited resources meant that it was still predominantly made up of researchers. The Australian EII trial exemplified end-user participation, with expert panels comprising 70% end users . We recommend that future studies include assessment panels made up of a predominance of non-researchers, and a high mix of different stakeholders.
There is a growing sense outside, and increasingly inside, the research sector that health intervention research is not an ‘end in itself’, and needs to have demonstrable public benefit. In order to demonstrate this benefit, we need to have a means of measuring the impacts of research. Moreover, if such methods are to be widely used in practice by research funders and academic institutions to assess research impacts, the right balance between comprehensiveness and feasibility must be struck. This study builds on current best practice for assessing real-world policy and practice impacts and demonstrates a systematic and multidimensional approach to impact assessment. The findings of this study could help funding systems determine how to assess impact in the future; however, the methods we employed were resource intensive. Further research to refine the process so that it may be more feasibly applied on a routine basis is warranted.
Australian Excellence in Innovation (EII) trial
National Health and Medical Research Council.
Denholm EM, Martin WJ: Translational research in environmental health sciences.Transl Res 2008, 151:57–58. 10.1016/j.trsl.2007.09.005
Erno-Kjolhede E, Hansson F: Measuring research performance during a changing relationship between science and society.Res Eval 2011, 20:131–143.
Frodeman R, Holbrook JB: Science’s social effects.Issues Sci Technol 2007, 23:28–30.
Bornmann L: Measuring the societal impact of research.Eur Mol Biol Organ J 2012, 13:673–676.
Martin BR: The research excellence framework and the ‘impact agenda’: are we creating a Frankenstein monster?Res Eval 2011, 20:247–254. 10.3152/095820211X13118583635693
Grimshaw JM, Eccles MP, Lavis JN, Hill SJ, Squires JE: Knowledge translation of research findings.Implement Sci 2012, 7:1748–5908.
Holbrook JB: Re-assessing the science–society relation: The case of the US National Science Foundation’s broader impacts merit review criterion (1997–2011). In Peer Review, Research Integrity, and the Governance of Science – Practice, Theory, and Current Discussion. Edited by: Frodeman RJ, Holbrook B, Mitcham C, Xiaonan H. Dalian: People’s Publishing House and Dalian University of Technology; 2012:328–362.
Henshall C: The impact of payback research: developing and using evidence in policy.Res Eval 2011, 20:257–258. 10.3152/095820211X13118583635873
Milat A, Laws R, King L, Newson R, Rychetnik L, Rissel C, Bauman A, Redman S, Bennie J: Policy and practice impacts of applied research: a case study analysis of the New South Wales Health Promotion Demonstration Research Grants Scheme 2000–2006.Health Res Policy Sys 2013, 11:5. 10.1186/1478-4505-11-5
Glasgow RE, Emmons KM: How can we increase translation of research into practice? Types of evidence needed.Annu Rev Public Health 2007, 28:413–433. 10.1146/annurev.publhealth.28.021406.144145
Bhattacharyya OK, Estey EA, Zwarenstein M: Methodologies to evaluate the effectiveness of knowledge translation interventions: a primer for researchers and health care managers.J Clin Epidemiol 2011, 64:32–40. 10.1016/j.jclinepi.2010.02.022
Brownson RC, Kreuter MW, Arrington BA, True WR: Translating scientific discoveries into public health action: how can schools of public health move us forward?Public Health Rep 2006, 121:97–103.
Buykx P, Humphreys J, Wakerman J, Perkins D, Lyle D, McGrail M, Kinsman L: ‘Making evidence count’: a framework to monitor the impact of health services research.Aust J Rural Health 2012, 20:51–58. 10.1111/j.1440-1584.2012.01256.x
Buxton M: The payback of ‘Payback’: challenges in assessing research impact.Res Eval 2011, 20:259–260. 10.3152/095820211X13118583635837
Banzi R, Moja L, Pistotti V, Facchini A, Liberati A: Conceptual frameworks and empirical approaches used to assess the impact of health research: an overview of reviews.Health Res Policy Syst 2011, 9:26. 10.1186/1478-4505-9-26
Donovan C: State of the art in assessing research impact: introduction to a special issue.Res Eval 2011, 20:175–179. 10.3152/095820211X13118583635918
Grant J, Brutscher P-B, Kirk SE, Butler L, Wooding S: Capturing Research Impacts: A Review of International Practice. Documented Briefing. Santa Monica, CA: Rand Corporation; 2010.
Hanney SR, Gonzalez-Block MA, Buxton MJ, Kogan M: The utilisation of health research in policy-making: concepts, examples and methods of assessment.Health Res Policy Syst 2003, 1:2. 10.1186/1478-4505-1-2
Hanney S, Packwood T, Buxton M: Evaluating the benefits from health research and development centres: a categorization, a model and examples of application.Evaluation 2000, 6:137–160. 10.1177/13563890022209181
Graham KER, Chorzempa HL, Valentine PA, Magnan J: Evaluating health research impact: development and implementation of the Alberta innovates – health solutions impact framework.Res Eval 2012, 21:354–367. 10.1093/reseval/rvs027
Wooding S, Hanney S, Pollitt A, Buxton J: Project Retrosight. Understanding the Returns from Cardiovascular and Stroke Research: Policy Report. Cambridge: RAND Europe; 2011. [http://www.rand.org/pubs/monographs/MG1079.html] Accessed Oct 2014
Group of Eight: Excellence in Innovation: Research Impacting our Nation’s Future - Assessing the Benefits. Adelaide: Australian Technology Network of Universities; 2012. [https://go8.edu.au/programs-and-fellowships/excellence-innovation-australia-eia-trial] Accessed Oct 2014
Higher Education Funding Council for England, Scottish Funding Council, Higher Education Funding Council for Wales, Department for Employment and Learning Northern Ireland: Research Excellence Framework Impact Pilot Exercise Findings of the Expert Panels. London: Research Excellence Framework; 2010. [http://www.ref.ac.uk/pubs/refimpactpilotexercisefindingsoftheexpertpanels/] Accessed Oct 2014
Buxton M, Hanney S: How can payback from health services research be assessed?J Health Serv Res 1996, 1:35–43.
Hanney SR, Grant J, Wooding S, Buxton MJ: Proposed methods for reviewing the outcomes of health research: the impact of funding by the UK‘s ’Arthritis Research Campaign’.Health Res Policy Syst 2004, 2:4. 10.1186/1478-4505-2-4
Hanney SR, Watt A, Jones TH, Metcalf L: Conducting retrospective impact analysis to inform a medical research charity’s funding strategies: the case of Asthma UK.Allergy Asthma Clin Immunol 2013, 9:17. 10.1186/1710-1492-9-17
Higher Education Funding Council for England: Decisions on Assessing Research Impact. London: Research Excellence Framework; 2011. [http://www.ref.ac.uk/pubs/2011-01/] Accessed Oct 2014
Higher Education Funding Council for England: Panel Criteria and Working Methods. London: Research Excellence Framework; 2012. [http://www.ref.ac.uk/pubs/2012-01/] Accessed Oct 2014
Hanney S, Buxton M, Green C, Coulson D, Raftery J: An assessment of the impact of the NHS Health Technology Assessment Programme.Health Technol Assess 2007, 11:iii-iv. ix-xi, 1–180
Johnston SC, Rootenberg JD, Katrak S, Smith WS, Elkins JS: Effect of a US National Institutes of Health programme of clinical trials on public health and costs.Lancet 2006, 367:1319–1327. 10.1016/S0140-6736(06)68578-4
Oortwijn WJ, Hanney SR, Ligtvoet A, Hoorens S, Wooding S, Grant J, Buxton MJ, Bouter LM: Assessing the impact of health technology assessment in The Netherlands.Int J Technol Assess Health Care 2008, 24:259–269.
Pollitt A, Wooding S, Hanney S, Buxton M, Grant J: Project Retrosight. Understanding the Returns from Cardiovascular and Stroke Research: Methodology Report. Cambridge: RAND Europe; 2011. [http://www.rand.org/pubs/technical_reports/TR925.html] Accessed Oct 2014
Donovan C, Butler L, Butt AJ, Jones TH, Hanney SR: Evaluation of the impact of national breast cancer foundation-funded research.Med J Aust 2014, 200:214–218. 10.5694/mja13.10798
Grant J, Cottrell R, Cluzeau F, Fawcett G: Evaluating “payback” on biomedical research from papers cited in clinical guidelines: applied bibliometric study.BMJ (Clin Res Ed) 2000, 320:1107–1111. 10.1136/bmj.320.7242.1107
Garland SM, Hernandez-Avila M, Wheeler CM, Perez G, Harper DM, Leodolter S, Tang GW, Ferris DG, Steben M, Bryan J: Quadrivalent vaccine against human papillomavirus to prevent anogenital diseases.N Engl J Med 2007, 356:1928–1943. 10.1056/NEJMoa061760
Tay S-K: Cervical cancer in the human papillomavirus vaccination era.Curr Opin Obstet Gyn 2012, 24:3–7. 10.1097/GCO.0b013e32834daed9
Barber R, Boote JD, Parry GD, Cooper CL, Yeeles P, Cook S: Can the impact of public involvement on research be evaluated? A mixed methods study.Health Expect 2012, 15:229–241. 10.1111/j.1369-7625.2010.00660.x
Kwan P, Johnston J, Fung AY, Chong DS, Collins RA, Lo SV: A systematic evaluation of payback of publically funded health and health services research in Hong Kong.BMC Health Serv Res 2007, 7:121. 10.1186/1472-6963-7-121
This work was supported by funding from the National Health and Medical Research Council of Australia (Grant #1024291). The authors would like to thank the study participants for their time and contribution to this research.
The authors declare that they have no competing interests.
The study was co-conceived by all of the authors except RN. All of the authors except RN participated in the design of the study. GC and JS led data collection and analysis. All of the other authors participated in the analysis and interpretation of the data. GC and RN led the drafting the manuscript. All of the other authors contributed to the drafting of the manuscript. All of the authors have read and approved the final manuscript.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made.
The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.
To view a copy of this licence, visit https://creativecommons.org/licenses/by/4.0/.
The Creative Commons Public Domain Dedication waiver (https://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Cohen, G., Schroeder, J., Newson, R. et al. Does health intervention research have real world policy and practice impacts: testing a new impact assessment tool. Health Res Policy Sys 13, 3 (2015). https://doi.org/10.1186/1478-4505-13-3
- Intervention research
- Research impact
- Research translation