Return to search

Assessing quality in systematic reviews of the effectiveness of health promotion : areas of consensus and dissension

Systematic reviews have played an increasingly important role in health promotion in recent years. Yet there are debates about how they should be conducted, particularly about how the quality of evidence should be assessed. The aim of this research was to assess current approaches to, and general views on, the use of quality assessment in systematic reviews of effectiveness in health promotion, and to identify areas of consensus and dissension around the choice of techniques, methods and criteria employed. There were two stages of data collection. The first was a structured mapping of a random sample of 30 systematic reviews of the effectiveness of health promotion to identify and explain trends and themes in methods and approaches to quality assessment. During the second stage semi-structured interviews were conducted with a purposive sample of 17 systematic reviewers who had conducted at least one review of a health promotion topic, to investigate some of these trends and approaches in greater detail. The mapping found that the majority of systematic reviews had assessed the quality of the included studies, to varying degrees. However, procedures were not always explicitly reported or consistent. There was some degree of consensus over criteria, with experimental evaluation methods commonly favoured. Most frequently used quality assessment criteria included participant attrition, the validity and reliability of data collection and analysis methods, and adequacy of sample sizes. External validity was commonly assessed, primarily in terms of generalisability and replicability, but less so in terms of intervention quality. The interviews revealed some of the barriers to effective systematic reviewing, including: lack of time and resources, complexity of some health promotion interventions, inclusion of observational evaluation designs, and poor reporting of primary studies. Systematic reviewing was commonly done in small teams, mostly comprising academics, sometimes with practitioners. Interviewees learned systematic review skills through a combination of training, support from colleagues and mentors, literature and a strong emphasis on hands-on practical learning. Subjective judgement was often required, contra to the popular belief that systematic reviews are wholly objective. The overall conclusions of this study are that systematic reviewing in health promotion is often challenging due the complexity of interventions and evaluation designs. This places additional demands on reviewers in terms of knowledge and skills required, often exacerbated by finite time scales and limited funding. Initiatives are in place to foster shared ways of working, although the extent to which complete consensus is achievable in a multi- disciplinary area such as health promotion is questionable.

Identiferoai:union.ndltd.org:bl.uk/oai:ethos.bl.uk:503192
Date January 2009
CreatorsShepherd, Jonathan Paul
ContributorsWeare, Katherine
PublisherUniversity of Southampton
Source SetsEthos UK
Detected LanguageEnglish
TypeElectronic Thesis or Dissertation
Sourcehttps://eprints.soton.ac.uk/67475/

Page generated in 0.0017 seconds