- Open Access
- Open Peer Review
Indicators for tracking programmes to strengthen health research capacity in lower- and middle-income countries: a qualitative synthesis
Health Research Policy and Systems volume 12, Article number: 17 (2014)
The monitoring and evaluation of health research capacity strengthening (health RCS) commonly involves documenting activities and outputs using indicators or metrics. We sought to catalogue the types of indicators being used to evaluate health RCS and to assess potential gaps in quality and coverage.
We purposively selected twelve evaluations to maximize diversity in health RCS, funders, countries, and approaches to evaluation. We explored the quality of the indicators and extracted them into a matrix across individual, institutional, and national/regional/network levels, based on a matrix in the ESSENCE Planning, Monitoring and Evaluation framework. We synthesized across potential impact pathways (activities to outputs to outcomes) and iteratively checked our findings with key health RCS evaluation stakeholders.
Evaluations varied remarkably in the strengths of their evaluation designs. The validity of indicators and potential biases were documented in a minority of reports. Indicators were primarily of activities, outputs, or outcomes, with little on their inter-relationships. Individual level indicators tended to be more quantitative, comparable, and attentive to equity considerations. Institutional and national–international level indicators were extremely diverse. Although linkage of activities through outputs to outcomes within evaluations was limited, across the evaluations we were able to construct potential pathways of change and assemble corresponding indicators.
Opportunities for improving health RCS evaluations include work on indicator measurement properties and development of indicators which better encompass relationships with knowledge users. Greater attention to evaluation design, prospective indicator measurement, and systematic linkage of indicators in keeping with theories of change could provide more robust evidence on outcomes of health RCS.
The need for all countries to generate and use health research in order to inform practice and policy decisions has become increasingly accepted over the last decade . However, there remain gaps in the production of health research, particularly in many low- and middle-income countries (LMICs) . Profiles to assess LMIC capacity for equity-oriented health research have been developed , resources assembled for health research capacity strengthening (RCS) , and ways forward proposed by leading African health researchers  and health systems organizations [6, 7]. RCS has been defined as a “process of individual and institutional development which leads to higher levels of skills and greater ability to perform useful research” . Experience has accumulated among those engaged in RCS for development in general , including case studies of health RCS , yet the heterogeneity and complexity of health RCS initiatives have hindered systematic assessments of effectiveness . As one author has noted, “We are at the early stages of knowing how best to identify, target and affect the many factors that are important for stronger research capacity. Furthermore, as RCS initiatives become more wide-ranging and complex, they become more difficult to monitor and evaluate.…. There is a clear need for improved strategies and the development of a tried and tested framework for RCS tracking” .
Organisations that fund and manage research capacity strengthening initiatives, both nationally, e.g., the UK Collaborative on Development Sciences , and internationally, e.g., the ESSENCE on Health Research Initiative , have responded by collaborating to identify common issues relating to evaluating RCS programmes. Critical to making sense of RCS outcomes is the need to be explicit about the pathway by which change is to be brought about, i.e., the theory of change . Indicators of the steps along the pathway from activities through outputs to outcomes can be linked within frameworks for evaluation of health RCS .
Currently, indicators or metrics are in widespread use in health programmes to monitor performance, measure achievement, and demonstrate accountability . Generally accepted criteria for development evaluation involve use of Specific, Measurable, Attainable, Realistic and Timely (SMART) indicators . Research impact evaluators have suggested including indicators not only of knowledge production and capacity development, but also of changes in health system policies, programs, and practices .
In the research reported here, we investigated reports of health RCS evaluations held by funders as a potentially rich source of untapped information. Our objective was to describe the design of health RCS evaluations, the nature of the indicators used, and the linkages among activities, outputs, and outcomes. We sought evidence to underpin the design of rigorous health RCS evaluations and the choice of indicators to be used for tracking progress and impacts, in ways which can transparently demonstrate value to all health RCS stakeholders – funders, research organizations, researchers, trainees, and research users.
We adopted a qualitative approach to report identification, evaluation quality appraisal, indicator extraction, and synthesis. We consulted with stakeholders from LMIC health research funding agencies as part of a knowledge user oriented process . Formal approval was obtained from the University of Toronto Health Sciences Research Ethics Board (#26837).
Our experience in reviewing proposals for health RCS, conducting and evaluating it, and searching the peer-reviewed literature led us to expect evaluations of a range of initiatives, from discreet projects, through portfolios of projects, to integrated long-term programmes. We initially consulted with funding agency members of the ESSENCE on Health Research initiative regarding report availability. Using a snowballing process, we invited other funders of LMIC research, known to support health RCS, to contribute health RCS evaluation reports. Eleven of 31 funding agencies contacted agreed to provide such reports, from which two authors agreed upon 54 reports of relevant health RCS evaluations that were publicly available, written in English, and produced since 2000. Each report was read by a pair of reviewers to assess the type of health RCS, funders, countries, detail available , and approach to evaluation. Sometimes more than one report was involved in evaluation of a health RCS initiative. Applying maximum variety sampling , we purposively selected 18 reports of 12 evaluations.
Because of the growing emphasis on evidence of effectiveness, we appraised the quality of the evaluations. We derived the following quality appraisal questions from the Development Assistance Committee standards  and applied them to each evaluation report:
Was the purpose of the evaluation clearly stated?
Was the methodology described (including the analysis)?
Were the indicators made explicit and justified?
In the methodology, we were particular interested in design, indicator measurement and collection, and bias. Our appraisal of the quality of indicators mentioned in the reports drew from SMART criteria (p10, Sn 2.9 of OECD standards) . Two reviewers appraised each evaluation independently, providing brief justifications for their appraisals.
We conducted a systematic framework analysis on the evaluation reports , extracting text relating to indicators used, and the context of that use. Many reports contained narrative descriptions of an activity, output, or outcome, which implied the nature of a corresponding indicator, while fewer explicitly defined indicators. Both descriptions and definitions were extracted and coded according to the categories in the ESSENCE Planning, Monitoring and Evaluation matrix  or to new categories that emerged. In order to promote learning and consistency in the extraction process, members of the research team each coded at least three reports, published by at least two funders, and relating to at least two evaluations, with each report being coded independently by two researchers. Discussion on coding of a smaller initial set fostered a common approach prior to coding all reports. Two authors independently extracted text from each evaluation report, checking consistency and resolving discrepancies through discussion, if necessary, by bringing in a third reviewer. We stopped when no new insights emerged from analysis of additional reports.
We reviewed extracted material and created additional categories as needed. Given the importance of pathways consistent with theories of change, we attempted to identify and document links between aims and indicators and from activities through to outputs and outcomes. Clear descriptions of these links were unfortunately rare within any one evaluation. Therefore, we brought together examples of indicators and their contexts from several different evaluations, in order to illustrate the potential for such linkages. At several stages throughout the project our interim findings were discussed with the ESSENCE on Health Research initiative steering committee. Their feedback helped us to focus our analysis, and to validate and interpret our results .
The 12 evaluations were of health RCS initiatives covering the wide range we had expected. They were of different durations, conducted by different kinds of evaluators, at different stages of the initiative, and using a variety of evaluation approaches (Table 1) [23–40].
Quality of the health RCS evaluation designs
All evaluations had clear statements of their purpose or objectives, often with explicit terms of reference appended to the reports (see Quality Appraisal of Evaluations – Illustrative Examples below). Most evaluations used mixed method designs and drew on existing data or prior reports, often supplemented with site visits and/or interviews. The degree of complexity of the various evaluations reflected the complexity of the health RCS initiative; for example, the design of an evaluation concerning individuals in scholarship programs [23, 24] was simpler than that used to evaluate changes in health economics capacity across an entire region . Variability in evaluation design also related partly to the stage of the evaluation: a review early in the project cycle  was less complex than that of a long-running program undergoing a final stage review [38, 39]. Several evaluators were constrained by the lack of a clear monitoring and evaluation framework , to help them orient their observations, and by the short time frame allowed for their review . Though some reports were able to use historical comparisons [38, 39], the majority were not able to draw on any baseline data [29–32], and only one evaluation considered (but did not use) a ‘control’ comparison . These constraints limited assessment of change, its attribution to the health RCS programme, and potential estimates of effectiveness.
Quality appraisal of evaluations – illustrative examples
Purpose of evaluation clearly stated
To appraise Swedish International Development Agency’s support to capacity building in the sub-Saharan Africa region. The most important purpose from the evaluators’ point of view was to provide stakeholders with the opportunity to learn about and develop the ongoing project .
To assess implementation and preliminary outcomes, focusing on awardees careers; to guide a future outcome evaluation .
To assess European and Developing Countries Clinical Trials Partnership (EDCTP) programme performance, including economic, social, and environmental impacts; address the role of EDCTP in the broader international research and development agenda; learn lessons and make recommendations for future initiatives .
Explicit evaluation design
A feasibility study, including pilot tests, guided the evaluation survey design .
Data collection clearly described and validity checked
Used qualitative interview recording, transcribing, and thematic coding. A self-assessment tool was used for research competency but its provenance was not explained .
Interviews to solicit information on factors influencing post-grant careers; interviewees selected to balance gender, research interest, and nationality .
Online surveys for awardees and unsuccessful applicants .
Indicators explicit and justified
Each bibliometric indicator provided insights into research quality, i.e., quantity of papers, citation rates, impact factor; norm-referencing .
Indicators were stipulated in an evaluation framework and designed with stakeholders using intervention logic .
Evaluation used EDCTP’s indicators, but limited by absence of any a priori formulated measurable indicators for the expected outcome set at the start of the programme .
Biases and limitations discussed
The lack of a uniform monitoring and evaluation framework and reporting system resulted in collection of different types of data, and therefore different insights and conclusions .
Limitations of using a self-assessment survey [25, p. 14] and the subjectivity of the evaluations and learning .
Variables (e.g., linguistic, internet access) and potential biases in responses, recall, and classification were taken into account .
Consideration was given to the feasibility of a comparative evaluation design and the need for longer-term and more rigorous design to assess outcomes and impact of Global Research Initiative Program .
Lack of pre-determined measurable indicators and independently verifiable data necessitated an opinion-based retrospective evaluation .
The evaluations surveyed individual grantees, institutional representatives, or relevant key informants at national and international levels. Questionnaires were all crafted specifically for the evaluation, i.e., no existing instruments with known properties or prior validation were used, yet only one evaluation report describing a formal pilot to test the questionnaire was used . Half of the evaluations explicitly addressed potential biases or other threats to the integrity of the evaluation, with some noting low response rates. Some cited the importance of site visits and other means to triangulate reports from grantee respondents, and a few described iteratively re-visiting different groups to obtain feedback on their emerging findings.
Indicators used for tracking progress in health RCS initiatives
The extent of indicator description and depth of justification for the choice of indicators varied widely between evaluation reports. Indicators were often linked to specific objectives, attainable and realistic for the programme, and many were timely for programme monitoring (SMART criteria). Developing indicators that were ‘measurable’ seemed to be more of a challenge. Some reports referred to “measuring progress using testable goals – relevance, governance, efficiency and effectiveness” , but did not provide an explicit definition of these terms. A few evaluations did include indicators which involved considerable measurement work, e.g., on bibliometric indicators of the quality of research performed by grantees . Other evaluations explicitly linked indicators to intervention logic frameworks (see Quality Appraisal of Evaluations – Illustrative Examples above). Although no single evaluation provided enough information to enable us to describe an explicit pathway of activities to outputs and outcomes, it was possible to link common indicators across evaluations. Specific examples follow to illustrate the context in which the indicators were used at different levels.
Individual level indicators
Indicators relating to training in research skills for researchers and also other personnel, such as data managers and laboratory staff, were common (see Table 2). Indicators of training in areas relevant to professional skills (e.g., research management) and of training quality (e.g., PhD education) meets international standards  and researcher/student satisfaction, were noted in some reports [34–36].
Some reports included indicators that assessed equality of award allocation by utilising data which had been disaggregated by gender, nationality, country income level, discipline, and level of award [34, 38, 40]. One evaluation specifically focused on how gender-related indicators could become more part of research funding, training, and reporting . Other equity-related disaggregations, e.g., by socio-economic status within the country or potentially excluded groups, such as ethnic minorities or aboriginal peoples , were only apparent in one evaluation [28, p. 85]. From a North–South equity perspective, one evaluation noted “Most project coordinators and project leaders or principal investigators are African researchers (55.5%), with good representation of female researchers: 40% in AIDS projects and 25% in TB and Malaria projects” but the benchmark for such judgements was not clear .
Some evaluations included indicators for trainee mentoring, noting low mentor to trainee ratios, with intense competition for senior supervisors among the many research projects funded by international donors . Job outcome indicators were pertinent to several health RCS activities and were very much affected by context, e.g., the lack of career opportunities/structures for post-doctoral students resulting in a high proportion of PhD graduates not continuing active research careers [35, 36].
Institutional level indicators
Several evaluations linked support for individual grantees to institutional research strengthening [35, 36, 40]. Others retrospectively analysed funding allocations according to location and characteristics of the recipient institutions . Indicators used in an evaluation of a PhD and MSc Fellowship programme identified a lack of institutional guidance on criteria for student selection and a lack of linkage between training support and engagement in research [35, 36]. Other indicators focused on the institutional capacity to mentor more junior researchers (as distinct from supervision of research students), to help returning graduates start research (e.g., an improved information and communication technology system to facilitate communications with colleagues globally), and to support active investigators (Table 3).
Indicators related to research infrastructure and management activities focused on ‘hard’ infrastructure (e.g., libraries, lab equipment) and ‘softer’ systems (e.g., implementation of new routines, policies, resource allocations and systems of compensation , leadership of funding proposals to attract research funding and implement and manage research [35, 36], and the development of organisational learning mechanisms) . A few evaluations highlighted missed opportunities such as the limited sharing of donated equipment, materials, and techniques [35, 36]. Indicators of institutional collaborations used in some evaluations included local ownership , regional partnerships [26, 27], and enhanced visibility of the institution in the national and international research communities .
National/international level indicators
Indicators of activities with national policy makers, regional organizations, or networks captured components of the system in which individual and institutional health RCS were embedded (Table 4) [6, 14]. Stakeholder engagement and research uptake indicators included systematic identification of potential users of research for early engagement [34, 40], a comprehensive communication strategy , and appropriate tailor-made tools for dissemination of research [33, 38]. Indicators of the capacities of research users and policy makers to utilise research information were rare but included skills in acquiring research information, assessing its quality, and using it for decision-making . Involvement of non-scientific communities was used as an indicator of embedding research partnerships within public health structures . Important indicators of national research capacity were the commitment of Ministries of Health to research and the development of national research councils with explicit national research priority-setting processes and legal frameworks for research [27, 34].
Development of trans-disciplinary platforms and networks of researchers or institutions were key indicators of the ability to assemble a critical mass of researchers [26, 27, 40]. The promotion of the financial sustainability of research capacity within a country or region, sometimes through involvement of private partners, was another indicator . Network leaders were identified as graduates from previously funded programmes, a long term indicator of one programme’s impact . An indicator of stability of co-operation across partner institutions was lack of dependence on specific individuals in a context of high personnel turnover [26, 34]. Whether a network was dysfunctional or smoothly run, whether feelings of injustice and insecurity were developing , and the existence of rules around ‘competition-collaboration’  were all indicators of the quality of network functioning. One evaluation of a programme which funded multiple networks noted that most networks did not have information ‘at their fingertips’ and some could not obtain output level data. It urged the use of a more formal monitoring and evaluation framework grounded in each individual programme’s theory-in-use and program logic .
Our systematic analysis of diverse international health research funders’ evaluations uncovered a broad set of indicators including metrics available to measure return on investment in health research . Many of the evaluations used a subset of indicators among those identified by the ESSENCE on health research initiative including curricula developed, courses run, researchers trained, scientific collaborations initiated, and partnerships strengthened. Given the global focus on health equity , the rarity of disaggregation of indicator data according to equity categories was concerning. The Ford Foundation’s work on active recruitment of those from disadvantaged backgrounds , and NIH-FIC’s Career Track’s inclusion of ‘minority type’ (Celia Wolfman, personal communication) hold promise.
Missing in the evaluations were some important constructs relevant to health RCS, particularly ongoing relationships among RCS stakeholders to facilitate conduct and use of research . Further, nomenclature was highly variable for the national/international level – terms included societal, macro, environment, and network, perhaps reflecting the systems nature of much health RCS . Use of the term ‘local’ to describe that which is not global is not particularly helpful in thinking about the scale of health RCS efforts, as it can refer to scales from a community, through municipalities, districts, provinces or regions within a country, and nations to multi-country regions, e.g., East Africa. Greater attention should be paid to clarifying scale, perhaps separating out three components – provincial-national research environment, international-global research environment, and research networks – in order to facilitate greater clarity of relationships between indicators and consistency in cross case comparisons [6, 7].
Comments on quality of indicators were rare, despite Development Assistance Committee standards, and only a few individual indicators met most of the SMART criteria . The quality challenges may reflect the division of responsibility for collecting indicator-related data among funders, institutions and researchers implementing health RCS, and evaluators. They may also reflect the limited investment of time and resources in evaluations, relegating them to more of a milestone monitoring role than a key ingredient for determination of equity, effectiveness, or efficiency. Each stakeholder may be interested in different indicators on account of their differing roles in assessing research impact . Stakeholders should therefore be involved in early planning regarding the selection and quality of indicators to be used .
Health RCS contribution assessment
Virtually all evaluations were retrospective in nature, with only a few [33, 40] engaging in the kind of forward planning to promote applicability of indicator selection over time and rigour of evaluation designs . Few evaluations systematically considered assumptions, pre-conditions, or measurement challenges, confounders or co-interventions, all of which are needed to clarify causality. Explicit use of theories of change  with delineation of pathways linking indicators within explicit frameworks  was rare, perhaps because of the limited attention to mechanisms by which health RCS initiatives might effectively address problems identified and bring about the hoped-for changes . Such gaps undercut assessment of the contribution of health RCS programmes to longer term impacts .
Limitations of our study
Not all health RCS funders whom we approached provided reports. Further, we could not undertake detailed analysis on a large number of evaluations due to the labour intensive nature of data extraction and analysis. Nevertheless, the evaluations we did analyse covered a broad range of countries, types of health RCS initiatives, international funders, and contexts. Many common themes emerged during our analysis, particularly later in our analysis process, as we reached saturation, suggesting that incorporating additional evaluations would not yield substantially new information. Allocation of extracts from evaluation narratives and indicators to the various framework analysis categories was occasionally only resolved through discussion. Most evaluations captured only one point in the life cycle of a health RCS initiative – only two tracked health RCS longitudinally [23, 24, 38]. Similarly, only a few encompassed the contributions of a range of health development efforts, research programs, and RCS initiatives, to the gradual emergence of a health research system, as has been possible in case studies taking a longer term view .
Directions for evaluation of health RCS
The strengths and weaknesses of the health RCS evaluations which we analyzed likely reflect those in the broader field of evaluation of research for development. Certain development funders are committed to “strengthening the evidence base for what works or does not work in international development as well as developing and strengthening evaluation research capacity within the UK and internationally” . Where health RCS is integrated within a research program, an adequate proportion of the program budget should be allocated to quality evaluation, e.g., US federal guidelines suggest 3% to 5% for evaluation activities . Rigorous evaluation design could draw on development evaluation efforts by organizations such as the International Initiative for Impact Evaluation (http://www.3ieimpact.org) and the Network of Networks on Impact Evaluation (http://nonie2011.org/). Building on the mixed methods work synthesized here, systematic attention to indicator framing, selection, measurement, and analysis, could occur while maintaining flexibility and revisiting indicators as health RCS proceeds [46, 51]. We have formulated these potential directions as a set of recommendations for which different stakeholders in health RCS could show leadership (Table 5).
Our research has synthesized new knowledge about evaluation designs and associated indicators that can be tracked in different contexts for different health RCS initiatives, tailored to the particular aims of an initiative. The use of more rigorous designs and better measurement within clearer evaluation frameworks should produce the kinds of robust evidence on effectiveness and impacts that are needed to better justify investments in health RCS.
European and Developing Countries Clinical Trials Partnership
Low- and middle-income countries
Research capacity strengthening
Specific, Measurable, Attainable, Realistic and Timely.
Hanney SR, González-Block MA: Organising health research systems as a key to improving health: the World Health Report 2013 and how to make further progress. Health Res Policy Syst. 2013, 11: 47-10.1186/1478-4505-11-47.
McKee M, Stuckler D, Basu S: Where there is no health research: what can be done to fill the global gaps in health research?. PLoS Med. 2012, 9: e1001209-10.1371/journal.pmed.1001209.
Tugwell P, Sitthi-Amorn C, Hatcher-Roberts J, Neufeld V, Makara P, Munoz F, Czerny P, Robinson V, Nuyens Y, Okello D: Health research profile to assess the capacity of low and middle income countries for equity-oriented research. BMC Public Health. 2006, 6: 151-10.1186/1471-2458-6-151.
Whitworth J, Sewankambo NK, Snewin VA: Improving implementation: building research capacity in maternal, neonatal, and child health in Africa. PLoS Med. 2010, 7: e1000299-10.1371/journal.pmed.1000299.
Fonn S: African PhD research capacity in public health: raison d'être and how to build it. Global Forum Update on Research for Health. Edited by: Malin S. 2005, London: Pro-Brook Publishers, 80-83.
Changing Mindsets: Research Capacity Strengthening in Low- and Middle-Income Countries. Edited by: Ghaffar A, IJsselmuiden C, Zicker F. 2008, COHRED, Global Forum for Health Research, and UNICEF/UNDP/World Bank/WHO Special Programme for Research and Training in Tropical Diseases (TDR), http://www.who.int/tdr/publications/documents/changing_mindsets.pdf Accessed 26 Apr 2013
Bennett S, Paina L, Kim C, Agyepong I, Chunharas S, McIntyre D, Nachuk S: What must be done to enhance capacity for Health Systems Research?. Background Papers Commissioned by the Symposium Secretariat for the First Global Symposium on Health Systems Research. 2010, Montreux, Switzerland: World Health Organization, 16-19.http://healthsystemsresearch.org/hsr2010/index.php?option=com_content&view=article&id=77&Itemid=140,
Golenko X, Pager S, Holden L: A thematic analysis of the role of the organisation in building allied health research capacity: a senior managers’ perspective. BMC Health Serv Res. 2012, 12: 276-10.1186/1472-6963-12-276.
Vogel I: Research Capacity Strengthening: Learning from Experience. 2011, UKCDS, http://www.ukcds.org.uk/sites/default/files/content/resources/UKCDS_Capacity_Building_Report_July_2012.pdf Accessed 27 Dec 2013
Cole DC, Kakuma R, Fonn S, Izugbara C, Thorogood M, Bates I: Evaluations of health research capacity development: a review of the evidence. Am J Trop Med Hyg. 2012, 87 (5 suppl 1): 801-
Wirrman-Gadsby E: Research capacity strengthening, donor approaches to improving and assessing its impact in low- and middle-income countries. Int J Health Plann Mgmt. 2011, 26: 89-106. 10.1002/hpm.1031.
ESSENCE on Health Research: Planning, Monitoring and Evaluation Framework for Capacity Strengthening in Health Research. 2011, Geneva: ESSENCE on Health Research, http://www.who.int/tdr/publications/non-tdr-publications/essence-framework/en/index.html Accessed 26 Apr 2013
Centre for Theory of Change: What is Theory of Change?. http://www.theoryofchange.org/what-is-theory-of-change/#2 Accessed 18 Jan 2014
Boyd A, Cole DC, Cho D-B, Aslanyan G, Bates I: Frameworks for evaluating health research capacity strengthening: a qualitative study. Health Res Policy Syst. 2013, 11: 46-10.1186/1478-4505-11-46.
Hales D, UNAIDS: An Introduction to Indicators. UNAIDS Monitoring and Evaluation Fundamentals. 2010, Geneva: UNAIDS, 14-http://www.unaids.org/en/media/unaids/contentassets/documents/document/2010/8_2-Intro-to-IndicatorsFMEF.pdf Accessed 26 Apr 2013
Organization for Economic Co-operation and Development: Development Assistance Committee Quality Standards for Development Evaluation. 2010, http://www.oecd.org/dac/evaluationofdevelopmentprogrammes/dcdndep/44798177.pdf Accessed 26 Apr 2013
Banzi R, Moja L, Pistotti V, Facchini A, Liberati A: Conceptual frameworks and empirical approaches used to assess the impact of health research: an overview of reviews. Health Res Policy Syst. 2011, 9: 26-10.1186/1478-4505-9-26.
Keown K, Van Eerd D, Irvin E: Stakeholder engagement opportunities in systematic reviews: knowledge transfer for policy and practice. J Continuing Education Health Professions. 2008, 28: 67-72. 10.1002/chp.159.
Maessen O: Intent to Build Capacity through Research Projects: An Examination of Project Objectives, Abstract and Appraisal Documents. 2005, Ottawa: International Development Research Centre, http://www.idrc.ca/uploads/user-S/11635261571Intent_to_Build_Capacity_through_Research_Projects.pdf Accessed 26 Apr 2013
Lincoln YS, Guba EG: Naturalistic Inquiry. 1985, Newbury Park, CA: Sage Publications
Pope C, Ziebland S, Mays N: Qualitative research in health care: analysing qualitative data. BMJ. 2000, 320: 114-116. 10.1136/bmj.320.7227.114.
Saini M, Shlonsk A: Systematic Synthesis of Qualitative Research: A Pocket Guide for Social Work Research Methods. 2012, Oxford: Oxford University Press, 240-
Day R, Stackhouse J, Geddes N: Evaluating Commonwealth Scholarships in the United Kingdom: Assessing Impact in Key Priority Areas. 2009, London: Commonwealth Scholarship Commission in the UK, http://cscuk.dfid.gov.uk/wp-content/uploads/2011/03/evaluation-impact-key-report.pdf Accessed 26 Apr 2013
Ransom J, Stackhouse J, Groenhout F, Geddes N, Darnbrough M: Evaluating Commonwealth Scholarships in the United Kingdom: Assessing Impact in the Health Sector. 2010, London: Commonwealth Scholarship Commission in the UK, http://cscuk.dfid.gov.uk/wp-content/uploads/2011/03/evaluation-impact-health-report.pdf Accessed 26 Apr 2013
Podems D: Evaluation of the Regional Initiative in Science and Education. 2010, OtherWise; Cape Town: The Carnegie-IAS Regional Initiative in Science and Education (RISE)
Health Research for Action (HERA): Review of DANIDA-Supported Health Research in Developing Countries: Main Report Volume One and Annexes Volume Two. 2007, Reet: HERA, http://www.enrecahealth.dk/pdfs/20070309_DANIDA-Research_FINAL_report_volume_I.pdf Accessed 26 Apr 2013
Mills A, Pairamanzi R, Sewankambo N, Shasha W: Mid-Term Evaluation of the Consortium for National Health Research, Kenya. 2010, Nairobi: Consortium for National Health Research, [made available by the funders]
van Velzen W, Kamal R, Meda N, Nippert I, Piot P, Sauer F: Independent External Evaluation Report of the European and Developing Countries Clinical Trials Partnership (EDCTP programme). 2009, The Hague: EDCTP, http://ec.europa.eu/research/health/infectious-diseases/poverty-diseases/doc/iee-report-edctp-programme_en.pdf Accessed 26 Apr 2013
Sachdeva N, Jimeno C, Peebles D: Gender Evaluation of Governance Equity and Health Program: Final Report on Phase I; Institutional Assessment. 2008, Ottawa: International Development Research Centre
Sachdeva N, Peebles D: Gender Evaluation of Governance Equity and Health Program: Final Report; Summary and Way Forward/Next Steps. 2008, Ottawa: International Development Research Centre
Sachdeva N, Peebles D, Tezare K: Gender Evaluation of Governance Equity and Health Program: Final Report on Phase 2; Projects Review. 2008, Ottawa: International Development Research Centre
Peebles D, Sachdeva N: Gender Evaluation of Governance Equity and Health Program: Final Report on Phase 3 – Gender Training Workshop. 2008, Ottawa: International Development Research Centre
Zuckerman B, Wilson A, Viola C, Lal B: Evaluation of the Fogarty International Research Collaboration Awards (FIRCA) Program: Phase II Outcome Evaluation. 2006, Bethesda: National Institutes of Health
Srivastava CV, Hodges A, Chelliah S, Zuckerman B: Evaluation of the Global Research Initiative Program (GRIP) for New Foreign Investigators. 2009, Bethesda: National Institutes of Health
Agyepong I, Kitua A, Borleffs J: Mid-Term Evaluation Report, NACCAP Programmes First Call: INTERACT. 2008, The Hague: NACCAP
Vullings W, Meijer I: NACCAP Mid-Term Review: Information Guide and Appendices. 2009, Amsterdam: Technopolis Group
Erlandsson B, Gunnarsson V: Evaluation of HEPNet in SSA. 2005, Stockholm: Sida, http://www.sida.se/Publications/Import/pdf/sv/Evaluation-of-HEPNet-in-SSA.pdf Accessed 26 Apr 2013
Minja H, Nsanzabana C: TDR Research Capacity Strengthening Programmes: Impact Assessment Study Review for the Period 2000–2008 (Draft). 2010, Basel: Swiss Tropical & Public Health Institute (SwissTPH)
Minja H, Nsanzabana C, Maure C, Hoffmann A, Rumisha S, Ogundahunsi O, Zicker F, Tanner M, Launois P: Impact of health research capacity strengthening in low- and middle-income countries: the case of WHO/TDR programmes. PLoS Negl Trop Dis. 2011, 5: e1351-10.1371/journal.pntd.0001351.
Pederson JS, Chataway J, Marjanovic S: CARTA (Consortium for Advanced Research Training in Africa): The Second Year. 2011, Cambridge: RAND Europe, [made available through the funder and contractors]
Östlin P, Schrecker T, Sadana R, Bonnefoy J, Gilson L, Hertzman C, Kelly MP, Kjellstrom T, Labonté R, Lundberg O, Muntaner C, Popay J, Sen G, Vaghri Z: Priorities for research on equity and health: towards an equity-focused health research agenda. PLoS Med. 2011, 8: e1001115-10.1371/journal.pmed.1001115.
Volkman TA, Dassin J, Zurbuchen M: (Eds): Origins, Journeys and Returns: Social Justice in International Higher Education. 2009, New York: NY Social Science Research Council & Columbia University Press
Coen SE, Bottorff JL, Johnson JL, Ratner PA: A relational conceptual framework for multidisciplinary health research centre infrastructure. Health Res Policy Syst. 2010, 8: 29-10.1186/1478-4505-8-29.
de Savigny D, Adam T: Systems Thinking for Health Systems Strengthening. 2009, Geneva: Alliance for Health Policy and Systems Research/WHO
Kuruvilla S, Mays N, Pleasant A, Walt G: Describing the impact of health research: a research impact framework. BMC Health Serv Res. 2006, 6: 134-10.1186/1472-6963-6-134.
Bates I, Boyd A, Smith H, Cole DC: A practical and systematic approach to organisational capacity strengthening in the health sector in Africa. Health Res Policy Syst. 2014, 12: 11-10.1186/1478-4505-12-11.
Kok MO, Schuit AJ: Contribution mapping: a method for mapping the contribution of research to enhance its impact. Health Res Policy Syst. 2012, 10: 21-10.1186/1478-4505-10-21.
Kok MO, Rodrigues A, Silva A, de Haan S: The emergence and current performance of a health research system: lessons from Guinea Bissau. Health Res Policy Syst. 2012, 10: 5-10.1186/1478-4505-10-5.
DfID, UK: Call for Proposals for Contract on Centre of Excellence on Impact Evaluation of International Development. https://supplierportal.dfid.gov.uk/selfservice/pages/public/supplier/publicbulletin/viewPublicNotice.cmd?bm90aWNlSWQ9NTgzNzA%3D] Accessed 1 Nov 2014
US Department of State, (Bureau of Resource Management): Department of State Program Evaluation Policy. 2012, http://www.state.gov/s/d/rm/rls/evaluation/2012/184556.htm#A Accessed 26 Apr 2013
Vasquez EE, Hirsch JS, Giang LM, Parker RG: Rethinking health research capacity strengthening. Global Public Health. 2013, 8 (Supp1): S104-S124.
We thank the members of the Steering Committee of the ESSENCE on Health Research initiative from the Swedish International Development Cooperation Agency (SIDA), Fogarty International Center – National Institutes of Health, and the Wellcome Trust for support during this research. In addition, Ritz Kakuma participated in project formulation, Maniola Sejrani in initial project work, and Quenby Mahood in document selection and tracking.
Funding was provided by the Canadian Institutes of Health Research: IIM-111606. However, neither our study research funder nor these funding agency stakeholders determined the study design, engaged in the collection, analysis, and interpretation of the data, or wrote or decided to submit this paper for publication.
The authors declare that they have no competing interests.
Each author participated in the design, data identification and collection, analysis and interpretation, writing, and revisions. All authors read and approved the final manuscript.
About this article
Cite this article
Cole, D.C., Boyd, A., Aslanyan, G. et al. Indicators for tracking programmes to strengthen health research capacity in lower- and middle-income countries: a qualitative synthesis. Health Res Policy Sys 12, 17 (2014) doi:10.1186/1478-4505-12-17
- Capacity building
- Evaluation studies
- Research personnel
- Research support
- World health