Produced in Collaboration
Melissa Sweet and Ray Moynihan
(To see a complete list of Milbank reports, click here.
When ordering, be sure to specify which report you want,
your name, mailing address, and phone number.
To print this report, click here for the pdf version.)
Table of Contents
This report describes the methods, applications, and value for policymakers of systematic reviews that evaluate interventions intended to improve population health. These reviews are the best available scientific guidance for choosing among alternative policies. Systematic reviews apply the methods of research (sometimes called “evidence”) synthesis, a rapidly developing area of inquiry that merges the biomedical and social sciences.
Policymakers and scientists collaborated in planning this report. They also reviewed it in draft, giving particular attention to its accuracy, balance, and accessibility. These colleagues, as well as persons the authors interviewed, are listed in the Acknowledgments.
Considerable evidence suggests that systematic reviews are informing policymakers’ decisions about policy for health, social services, education, and criminal justice. Such evidence can be seen in three examples from one systematic review group based at the Centers for Disease Control and Prevention (CDC) and known as the Guide to Community Preventive Services (Community Guide). In many jurisdictions, policymakers have cited a Community Guide systematic review demonstrating the harmful effects of second-hand cigarette smoke as part of their rationale for enacting laws to ban smoking in public places. Another example is a 2007 Connecticut law ending the automatic assignment of sixteen and seventeen year olds to the adult criminal justice system. This decision took account of a Community Guide systematic review published a few months earlier. The review found that prosecuting youth as adults leads to increased violence. Similarly, in June 2007, a member of the United States House of Representatives told his colleagues that a Community Guide systematic review had “found conclusively that moving the drinking age up to twenty-one decreases alcohol-involved crash fatalities by 16 percent and lowering it increases fatalities by 10 percent.”
The CDC and the Milbank Memorial Fund commissioned this report in order to explain the methods and uses of systematic reviews to a broad audience. Incomplete or erroneous information about systematic reviews continues to be published in the media and distributed by groups trying to influence policy. Some of the misinformation that is circulated by researchers in other fields and health journalists can be attributed to their not understanding the rapidity with which the field is advancing. According to a recent article in Plos Medicine, peer-reviewed journals published approximately 2,500 systematic reviews in 2005up from 87 the first time they were counted in 1987.1 Misinformation can also result from the complex interactions of science and policy, as well as commercial and nonprofit advocacy.
The CDC and the Fund are longtime participants in international activities to advance and apply the science of research synthesis. The CDC is the U.S. federal public health agency whose mission is “to promote health and quality of life by preventing and controlling disease, injury, and disability.” A core function of the CDC is to be a credible source of health information. The Fund is an endowed operating foundation that works to improve health by helping decision makers in the public and private sectors acquire and use the best available evidence to inform policy for health care and population health. This report completes a project begun with the publication in 2004 of a report by Ray Moynihan, Evaluating Health Services: A Reporter Covers the Science of Research Synthesis, available at http://www.milbank.org/reports/2004Moynihan/040330Moynihan.html.
The authors asked us to include the following statement about their qualifications and methods: This is a piece of journalism from two respected health writers, informed by discussions with, and literature from, some of the key figures in the field. The authors have attempted to reflect a diversity of voices in the storytelling, including researchers who conduct systematic reviews, policymakers who use them, and critical thinkers who are imagining ways of enhancing the utility of systematic reviews. The authors use storytelling and anecdote to raise broad issues and case studies to help identify key themes.
Coordinating Center for Environmental Health and Injury Prevention, CDC
Daniel M. Fox
Milbank Memorial Fund
Shawna L. Mercer
The Guide to Community Preventive Services, CDC
Carmen Hooker Odom
Milbank Memorial Fund
The following people planned this report with the authors, staff of the CDC, and the Fund; they also reviewed the report in draft. Their affiliation at the time of review is listed.
Lisa Bero, University of California, San Francisco; Robert Boruch, University of Pennsylvania; Ross Brownson, Saint Louis University; Leah Devlin, North Carolina Department of Health and Human Services; Jonathan Fielding, Los Angeles County Department of Health Services; Richard Gottfried, New York State Assembly; Lee Greenfield, Hennepin County (MN) Department of Human Services and Public Health; John Nilson, Ministry of Environment, Saskatchewan.
The following people were interviewed for this report and/or reviewed it in draft. Their affiliation at the time of their participation is listed.
Alice Ammerman, University of North Carolina at Chapel Hill; Laurie Anderson, Centers for Disease Control and Prevention; David Atkins, Agency for Healthcare Research and Quality; Dileep Bal, Hawaii State Department of Health; Simon Chapman, University of Sydney; Bernard Choi, Public Health Agency of Canada; Karina Davidson, Columbia University Medical Center; Don Des Jarlais, Beth Israel Medical Center; Randy Elder, Centers for Disease Control and Prevention; Brian Flay, Oregon State University; Sandro Galea, University of Michigan; Stanton Glantz, University of California, San Francisco; Lawrence Green, University of California, San Francisco; Jessie Gruman, Center for the Advancement of Health; Trevor Hancock, British Columbia Ministry of Health; Alyson Hazen, Partnership for Prevention; Jim Hedlund, Highway Safety North; Mark Helfand, Oregon Evidence-Based Practice Center; Neal Kohatsu, California Department of Health Services; Kelli Komro, University of Florida; John Lavis, McMaster University; Roxanne Lieb, Washington State Institute for Public Policy; Vivian Lin, La Trobe University; David MacLean, Simon Fraser University; Linda Major, University of NebraskaLincoln; Marc Manley, BlueCross BlueShield of Minnesota; Toben Nelson, Harvard University; Jim Nichols, formerly of the National Highway Traffic Safety Administration; Fran Norris, National Center for Posttraumatic Stress Disorder; Lotrakul Panpimol, Thailand Ministry of Public Health; John Pierce, University of California, San Diego; Marcus Plescia, North Carolina Department of Health and Human Services; Peter Sainsbury, University of Sydney; Jim Sallis, San Diego State University; Michelle Scollo, Cancer Council Victoria; Jan Semenza, European Centre for Disease Prevention and Control; Ruth Shults, Centers for Disease Control and Prevention; Zili Sloboda, University of Akron; Cathy Thomas, North Carolina Department of Health and Human Services; Mark Van Ommeren, World Health Organization; Elizabeth Waters, Deakin University; Tom Workman, University of NebraskaLincoln.
Efforts to improve population health will yield better returns if there is more effective integration of reliable scientific evidence into policymaking. Increasing the use and usefulness of systematic reviews is one powerful mechanism for improving the evidence available to inform population health decision making.
Systematic reviews provide a systematic, transparent means for gathering, synthesizing, and appraising the findings of studies on a particular topic or question. They aim to minimize the bias associated with single studies and nonsystematic reviews. They can include many types of studies from diverse disciplines.
Systematic reviews can help provide information useful to policymakers, practitioners, and researchers, including the extent of a health problem and whether interventions work, at what cost, and for whom. They can help policymakers make the most efficient use of scarce resources and evaluate the relative merits of competing policies or programsthough for the most part current reviews compare an innovative intervention with some sort of “usual care” scenario rather than compare two different interventions or strategies. Systematic reviews can also help policymakers resist pressures from vested or competing interests, and they can help identify gaps in the evidence and priorities for future research.
Population health is affected by policymaking in many areas, including both government and nongovernment enterprises, as well as sectors other than health. Therefore, a broad perspective should be taken in considering the scope of systematic reviews and their potential users. In this report, policymakers are considered to include public and private sector organizations whose decisions influence health, whether they sit in the health sector or elsewhere.
Moves are under way to improve policymakers’ access to and use of relevant, reliable information from systematic reviews. However, scientific evidence is only one of many forces and many types of information that influence policymaking. Systematic reviews face tough competition for policymakers’ attention. They cannot always provide information that is useful or relevant to policymakers’ needs, and policymakers often must make decisions on the basis of incomplete evidence within a very short time frame.
Evidence-based public/population health differs from evidence-based medicine because it bridges complex systems and populations rather than homogenous patient populations. Many methodological issues confront those who produce and use systematic reviews relevant to public/population health, and concerted efforts are under way to improve the quality of systematic reviews in this area. David MacLean, professor and dean of the faculty of health sciences at Simon Fraser University in British Columbia, has expressed concerns that overreliance on systematic reviews or unquestioning use of them might stifle creativity and innovation or lead to useful programs being sidelined because of their inadequate evidence base.
This report includes case studiesranging from tobacco control to binge drinking among college students to the mental health challenges of the Indian Ocean tsunamithat bear many lessons for those seeking to improve population health. First and foremost they suggest there is significant room to enhance the role of evidence in policymaking. They also reveal the importance of
- taking a systems, environmental, or policy approach to changing human behavior, rather than strategies focused solely on individuals,
- the role of social norms in influencing behavior,
- the power of legislative, regulatory, and financial incentives to encourage the implementation of evidence-based policies,
- and the value of a reliable and relevant evidence base to help set political and public agendas and to shape interventions.
These themes are also instructive for those seeking to boost the role of evidence in population health decision making. Improving the use and usefulness of systematic reviews will require individual researchers, policymakers, advocates, and other relevant groups to modify the way they work. Success in this endeavor is most likely to be achieved if systemic, environmental, and cultural changes promote and support them.
As the case studies demonstrate, the real-world interaction between evidence and action is complex, and while each serves to improve the other over time, the ease and speed of this process can be improved. The case studies also highlight the importance of fearless and prominent champions in influencing political and professional agendas. They show the value of collaboration, perseverance, and pragmatism.
Those who seek to improve the use and usefulness of systematic reviews must recognize that policymaking is not a linear process and that a comprehensive array of interventions is more likely to have an impact than any single intervention. They must also be systematic by making deliberate, strategic efforts to disseminate and implement review findings, as well as ready and willing to seize moments of opportunity. Timeliness is often crucial in the policy environment.
This report makes recommendations for a comprehensive range of strategies, involving researchers, policymakers, the media, interest groups, and the broader community, in order to make systematic reviews both more useful and utilized. Recommendations include organizational and cultural changes to improve the translation and dissemination of systematic reviews; collaboration across sectors and disciplines to improve the quality, timeliness, accessibility, and relevance of systematic reviews; and incentives to encourage researchers, policymakers, and interest groups to work towards better integration of evidence into policy.
The importance of the media should not be overlooked. Just as effective media advocacy contributed to many of the public health advances outlined in this report, so can it contribute to efforts to promote both demand for and supply of useful, relevant systematic reviews. Media advocacy has been defined as the “strategic use of mass media to advance a social or public policy initiative.”2
The final message from this report is that efforts to improve the use and usefulness of systematic reviews must be evaluated, and these findings must themselves be disseminated and implemented.
How Systematic Reviews Can Help
- Efforts to improve population health will yield better returns if there is more effective integration into policymaking of reliable scientific evidence, including systematic reviews.
- Increasing the use and usefulness of systematic reviews is one powerful mechanism for improving population health.
- Systematic reviews provide a systematic, transparent means for gathering, synthesizing, and appraising the findings of studies on a particular topic or question.
- Policymakers can be more confident about the findings of systematic reviews because such reviews aim to minimize the bias associated with single studies and nonsystematic reviews.
- Systematic reviews can help provide many types of information useful to policymakers and researchers, including the extent of a health problem and whether interventions work, at what cost, and for whom.
- Systematic reviews can help policymakers to make the most efficient use of scarce resources and to resist pressures from vested interests. They can also help identify gaps in the evidence and priorities for future research.
Improving the Use and Usefulness of Systematic Reviews
- Improving the use and usefulness of systematic reviews will require researchers, policymakers, advocates, and other relevant groups to make changes to the way they work, which are most likely to be achieved if systemic, environmental, and cultural changes promote and support them.
- Those seeking to improve the use and usefulness of systematic reviews must make deliberate, strategic efforts to disseminate and implement review findings as well as be ready and willing to seize the moment when opportunities arise.
- Fearless and prominent champions are needed to influence political, professional, and public agendas around the use of systematic reviews in policymaking.
- Enhancing the “literacy” of decision makers and the public about the strengths and weaknesses of different types of evidence may help inform and improve population health policy.
- Educating journalists about the principles of an evidence-based approach within health may help improve the public’s understanding of evidence-based issues, including systematic reviews.
- Collaboration across sectors and disciplines will help improve the quality, timeliness, accessibility, and relevance of systematic reviews, and thereby promote the acceptability of their findings. The reverse is also true: having a solid evidence base to guide best practice can also facilitate collaboration.
- Incentives are needed to encourage researchers, policymakers, and interest groups to better integrate evidence into policy.
- Efforts to improve the use and usefulness of systematic reviews must be evaluated.
- Key weaknesses of systematic reviews can include limited relevance or applicability and generalizability of the original studies being reviewed. Advocating an unquestioning or inappropriate overreliance on systematic reviews might discourage innovation or promising practices.
- The real-world interaction between evidence and action is complex, and while each serves to improve the other over time, the process can be made more efficient.
Introduction: The Potential of Systematic Reviews . . . and the Limitations
The rise of the evidence-based public/population health policy movement signals a growing awareness of the potential benefits of making better use of reliable scientific evidence in the policymaking process.3 The term evidence-based public health has been defined as “the process of integrating science-based interventions with community preferences to improve the health of populations.”4 The term evidence-informed public health is gaining currency, acknowledging that factors other than evidence influence policymaking. Population health is affected by policymaking in many areas, including both government and nongovernment sectors, as well as fields other than health. Better integration of evidence into the policy process has the potential not only to improve health and well-being, but also to maximize the return on investment in interventions and policies promoting population health.
As the case studies in this report demonstrate, there is significant potential to improve both the use and usefulness of systematic reviews in population health policymaking. The case studies, which cover major population health challenges, including drug and alcohol abuse, tobacco control, traffic safety, and obesity, help illuminate the subtle, complex, and often invisible interactions between research evidence and policymaking. As such, they will be of enormous interest to anyone wanting to improve traffic flow on the bridges between research and action.
Systematic reviews are a methodology using a systematic, transparent process for gathering, synthesizing, and appraising the findings of studies on a particular topic or question. They aim to minimize the bias associated with single studies and nonsystematic reviews. Policymakers can be more confident relying on the results of systematic reviews than those of single studies because they are less likely to be biased or misleading.5 Systematic reviews have been used for some decades in social sciences and other disciplines,6 and in recent years they have become prominent in clinical medicine, fostered by the emergence of the evidence-based medicine movement.
Sometimes the terms systematic review and meta-analysis are used interchangeably, but it is more correct to view the meta-analysis as a subset of a systematic review. A meta-analysis involves mathematically aggregating data from independent studies to produce a more statistically powerful estimate.7 Not all systematic reviews are meta-analyses. Key elements of a systematic review include an explicit question; a comprehensive statement about what types of research evidence were included and excluded in the review; a thorough examination of the quality of the studies included; and a critical and transparent process of interpretation of the findings.8 Those interested in a more detailed understanding of the methods of a systematic review are advised to refer to the Centre for Reviews and Dissemination at the University of York,9 the Cochrane Collaboration,10 the Campbell Collaboration,11 or the Task Force on Community Preventive Services and the Guide to Community Preventive Services (Community Guide).12
Some of the forces encouraging the use of systematic reviews in clinical decision making are also promoting their use in population health. These include the ever increasing volume and complexity of knowledge and the pressures to maximize transparency, accountability, and efficiency. Methodological developments in community research, program evaluation, and data synthesis, as well as an “explosion” of community studies of health promotion and disease prevention programs are also factors.13 One of the most important drivers, however, is widespread evidence that the interventions and policies proven to be effective in promoting public health are not always put into practice. Conversely, interventions and policies are sometimes implemented despite evidence of ineffectiveness or lack of evidence of effectiveness. “In public health, we have discovered much, not yet applied, that holds the promise of important change in health,” leading public health experts wrote recently.14
Systematic reviews are a useful tool for those seeking to promote the translation of knowledge into action. They can help researchers and policymakers to identify gaps in knowledge, as well as areas where further research is not needed.
As we will see in this report’s case studies, systematic reviews can help provide many types of information useful to policymakers, including information about the nature and extent of a problem, and the potential benefits, harms, uncertainties, and costs of interventions and policies. Policymakers may also want to know about the impact on different groups in various settings. Systematic reviews can also help answer questions about how best to disseminate information and innovations; about a particular community’s receptiveness to proposed interventionswhether the interventions are appropriate to local culture and context; and about the factors influencing study outcomes. For example, a systematic review of studies examining the economic impact of tobacco laws revealed the influence of tobacco industry funding on study outcomes: industry-funded studies were more favorable towards industry’s products.15
Furthermore, the transparency of the systematic review process offers policymakers the opportunity to scrutinize the evidence base rather than simply rely on the final conclusions, according to internationally recognized authority on the relationship between evidence and policy John Lavis, the Canada Research Chair in Knowledge Transfer and Exchange and associate professor at McMaster University. “The nice thing about systematic reviews is that they don’t take the next step like a guideline where they put in the values and all kinds of assumptions that experts often bring to the table,” he says. “They just say this is the global pool of knowledge, synthesized in a transparent way. If you disagree with any of the steps that were followed, here are the building blocks so you can go out and build your own.”
Systematic reviews can also help policymakers resist pressures from vested or competing interests. Jonathan Fielding, director of public health and health officer, Los Angeles County Department of Health Services, and professor at the University of California, says this is one of many factors contributing to public health leaders’ growing interest in the quality of evidence available to guide decision making. His department’s decisions are based on evidence, he says. “It doesn’t mean evidence is always going to drive the decisions; there has to be a social context taken into account and also timing is important. But I think that having the evidence is really the most important weapon we have against ideologues on every side.”
In clinical medicine, randomized controlled trials have until recently been the dominant type of study included in systematic reviews. In population health, however, many argue that systematic reviews need to accommodate a wider range of study designs in recognition of the complexity of population health interventions and research. Many of the questions facing policymakers cannot be answered by randomized controlled trials. In population-based interventions, it is often not feasible, for example, to measure the effect of an intervention in a truly controlled environment. At least partly because of these concerns, methodology has been developed to allow for systematic reviews of qualitative and quantitative research, as well as of studies undertaken in disciplines other than the health and medical sciences.16
Common myths and misconceptions about systematic reviews include that they can cover only randomized controlled trials; that they are of value only for assessing effectiveness of health care interventions; that they must adopt a biomedical model; and that they always entail some form of statistical synthesis.17 In reality, systematic reviews frequently include a wide range of study designs and study questions, have no preferred biomedical model, and have methodologies that are more flexible than is sometime realized. “Several arguments are commonly used to reject a wider role for systematic reviews, and these arguments are often based on major misconceptions about the history, purpose, methods, and uses of systematic reviews,” according to an article by Mark Petticrew, associate director of the Social and Public Health Sciences Unit at the University of Glasgow. “Their potential role beyond the world of evidence-based medicine is more wide ranging than is often realized.”18
Key Organizations and Recent Developments
The work of several key organizations and some recent developments, in the United States and elsewhere, reflect growing interest in the role of systematic reviews in policymaking. These include:
The Task Force on Community Preventive Services and the Community Guide
(www.thecommunityguide.org; accessed July 9, 2007)
The Task Force on Community Preventive Services (Task Force) is an independent non-federal body formed in 1996 to oversee the conduct of systematic reviews of population health interventions and to make recommendations for policy and practice. The work of the Task Force is staffed by the Centers for Disease Control and Prevention. While the Task Force is convened by the U.S. Department of Health and Human Services and its members are appointed by the director of the Centers for Disease Control and Prevention, the Task Force is an independent decision making body.
The more than 180 Task Force recommendations made to date and the systematic reviews on which they are based are known collectively as the Guide to Community Preventive Services or Community Guide. The Community Guide summarizes what is known about the effectiveness, economic efficiency, and feasibility of interventions to promote community health and prevent disease, and highlights gaps in the body of evidence across topics and interventions and ways in which the research might be improved. The body of work reveals that some program effectiveness questions are easier to answer than others and that interventions shown to work in specific settings or populations may prove difficult to evaluate when implemented on a broader community scale when more complex processes of social change must be taken into account.19 As a result, the Task Force stresses that a finding of insufficient evidence of effectiveness should not be regarded as evidence of ineffectiveness. The Community Guide has therefore spent considerable effort on methods development and refinement, and its methods and updates are peer reviewed, published, and available online.
The Task Force promotes the Community Guide as relevant to policymakers because it can help them to make the most efficient use of limited resources for public health, to demonstrate accountability for resource allocation choices, and to make stronger arguments in funding requests. It helps researchers identify areas needing further study and ways to improve the quality of research, and helps practitioners, employers, and other health care service purchasers be more efficient in enhancing the health of their employees or members.
The Cochrane Collaboration
(www.cochrane.org; accessed July 9, 2007)
The Cochrane Collaboration, an international not-for-profit organization, was initially established in 1993 to prepare and maintain systematic reviews of the effects of health care interventions but has since expanded its focus to include health promotion and public health. Its major product is the Cochrane Database of Systematic Reviews, published quarterly as part of the Cochrane Library. A history of the Cochrane Health Promotion and Public Health Field is available at: www.vichealth.vic.gov.au/cochrane (accessed July 9, 2007).
Cochrane methodology and reviews have been the basis for many reviews by Evidence-based Practice Centers (EPCs, see www.ahrq.gov/clinic/epc; accessed July 9, 2007) and the Drug Effectiveness Review Project (DERP, see www.ohsu.edu/drugeffectiveness; accessed July 9, 2007). Cochrane reviews are commonly used by the committees that prepare guidelines for specialty societies, and they are also frequently reported in national news media.
Currently far fewer Cochrane reviews relate to public health, as compared with clinical medicine.
The Campbell Collaboration
(www.campbellcollaboration.org; accessed July 9, 2007)
The Campbell Collaboration was established in 2000 to prepare and maintain systematic reviews on the effects of social, behavioral, and educational policies and practices. It collaborates closely with the Cochrane Collaboration. Members of the Methods Groups in Campbell and Cochrane work together with the aim of stimulating the empirical methodological research required to improve the validity, relevance, and precision of systematic reviews and the randomized trials and nonrandomized trials on which they are based.
Robert Boruch, co-chair of the Campbell Steering Group, University Trustee Chair Professor of Education and professor of statistics at the University of Pennsylvania in Philadelphia, says Collaboration participants and partner organizations have substantially assisted the U.S. Department of Education’s Institute of Education Sciences in producing high-quality systematic reviews of evidence on the effectiveness of education interventions. These reviews are designed to help decision makers and are published by the What Works Clearinghouse, established in 2002 to provide educators, policymakers, researchers, and the public with a central, trusted source of scientific evidence of what works in education (www.whatworks.ed.gov; accessed July 9, 2007). Boruch says the reviews have encouraged some publishers of commercial curriculum packages to fund randomized controlled trials comparing their products against those of their competitors.
The Agency for Healthcare Research and Quality
(www.ahrq.gov; accessed July 9, 2007)
The Agency for Healthcare Research and Quality (AHRQ) is the lead federal agency charged with improving the quality, safety, efficiency, and effectiveness of health care. As one of twelve agencies within the U.S. Department of Health and Human Services, the AHRQ supports health services research that will improve the quality of health care and promote evidence-based decision making. The AHRQ works with partners, including thirteen Evidence-based Practice Centers (ten in the United States and three in Canada), to conduct systematic reviews on a wide range of health topics, including both clinical and public health interventions. In addition, the AHRQ supports the Guide to Clinical Preventive Services, the clinical companion to the Community Guide.
Translation and Other Projects
Efforts are under way to translate the evidence from systematic reviews into action plans to assist implementation of effective interventions. For example, the Centers for Disease Control and Prevention (CDC) (www.cdc.gov; accessed July 9, 2007) is funding the Partnership for Prevention (www.prevent.org; accessed July 9, 2007) to develop manuals for public health workers on how to implement evidence-based interventions. These are based both on the results of systematic reviews as well as interviews with key informants. “Systematic reviews are a great starting point in letting you know what works but often don’t provide the level of information that you need to implement interventions,” says Alyson Hazen, a fellow at the Partnership for Prevention. The manuals, called The Community Health Promotion Handbook: Action Guides to Improve Community Health, are expected to be available on the Partnership for Prevention’s website by late 2007. They include steps for implementing an intervention, suggest partners to engage in the process, and offer advice on how to overcome likely barriers and the personnel, financial, and material resources required in order to do so.
Other examples of efforts to better integrate evidence into policy include the European Observatory on Health Systems and Policies (www.euro.who.int/observatory; accessed July 9, 2007), the Canadian Health Services Research Foundation (www.chsrf.ca/home_e.php; accessed July 9, 2007), the National Institute for Health and Clinical Excellence (www.nice.org.uk; accessed September 4, 2007), the Association of Public Health Observatories (www.apho.org.uk/apho/index.htm; accessed July 27, 2007), and the Health Evidence Network (www.euro.who.int/HEN; accessed July 9, 2007) recently set up by the World Health Organization (WHO).20 The Society for Prevention Research (SPR) (www.preventionresearch.org; accessed July 9, 2007) is also working to improve the evidence base of policy and practice. In 2004, the SPR adopted standards of evidence for prevention practices, which cover efficacy, effectiveness, and dissemination.21
Many factors other than scientific evidence influence the policy process, including political and economic imperatives, vested interests and advocacy groups, crises, the media, public opinion, ideology, intuition, emotion, values, and beliefs. Happenstance and timing can also be important. Other influences are policymakers’ experience, expertise, and judgment, as well as the pragmatics of political life, including electoral cycles and timetables, procedures, and unanticipated contingencies.22
Policymakers are influenced by information from many sources, including anecdotal evidence, single studies, commercial interests, disease-specific advocacy groups, and media reports. Reliable scientific evidence such as that from systematic reviews is only one player in the process and often not the most visible or influential. Policymakers are not necessarily familiar with the concept of a hierarchy of scientific evidence, in which different forms of evidence are ranked according to their reliability, with most weight accorded to systematic reviews of well-conducted controlled studies and least to individual studies, expert opinion, or anecdotal evidence.23
According to Bernard Choi, senior research scientist at the Public Health Agency of Canada’s Centre for Chronic Disease Prevention and Control, the hierarchy of scientific evidence is often inverted when policy decisions are made. He cites decision makers’ lack of time, ability, or expertise to access and synthesize the evidence from high-quality studies24 and adds that for policymakers already suffering information overload, systematic reviews can seem more like part of a problem than a solution. Others argue that anecdotal information provided by voters can be a source of valuable information for policymakers, who may be more likely to consider the appropriateness of evidence than where it fits in the scientific hierarchy.25 Some researchers have put it thus: “How and when evidence is used often depends upon the political agenda and ideology of the government of the day, not on the nature of the evidence, however compelling.”26 Many public health experts say that evidence is sometimes used to justify policy decisions rather than being a driving force in itself. As suggested by a recent systematic review of the literature about how policymakers use evidence, research is often used to legitimate decisions after the fact.27
Choi and colleagues note the lack of correlation between the quality of science and the policy derived from it. “Good science does not always guarantee good policy; bad or even no science does not necessarily lead to bad policy,” they say. “It is true good policy does not always depend on waiting for good evidence.”28
Neal Kohatsu, chief of the Cancer Control Branch of the California Department of Health Services, notes that policymaking generally does not happen in a scientific or medical framework. “The political decision making process is not a scientific process but a battle of constituencies who can effectively get their message to decision makers, whether in the executive or legislative branch,” he says.
In such an environment, it is a mistake to see systematic reviews as a final outcome when it is more realistic to view them as the beginning of a process, according to Marc Manley, vice president and medical director of population health at the health insurer BlueCross and BlueShield of Minnesota. “Groups come together and write the review and think this problem is solved but it’s only started to be solved,” he says. In a similar vein, Mark Helfand, director of the Oregon Evidence-Based Practice Center, says the role of systematic reviews is to define the limits and strengths of the evidence but not to tell policymakers what to do.
For Lavis, the most important outcome of strengthening the role of evidence in policymaking is not necessarily the implementation of evidence. “The outcome that matters to me is ensuring that policy debates are informed,” he says, “whereas some of my colleagues want evidence used regardless of how it aligns with the many other influences that policymakers have to consider.”
Whether and how the results of a review are implemented depend on many factors, including political, financial, and human resources.29 As the case studies described in this report illustrate, local context, preferences, culture, history, and other considerations can also come into play.
Systematic reviews cannot always provide the answers policymakers require. This may be because the relevant primary studies or systematic reviews have not been conducted; because those that have been done are not relevant to the particular setting, population, or situation; or because their findings are inconclusive or disputed. Senior policy advisors lament that “policy-free evidence” is common, suggesting that researchers do not always consider the needs of policymakers, and as a result many studies and reviews fail to meet those needs as well.30
There are far fewer systematic reviews to guide decision making in population health than in clinical medicine, and policymakers must often make decisions on the basis of incomplete evidence. Jessie Gruman, president of the Center for the Advancement of Health in Washington DC, which promotes the translation of health research into effective policy and practice, says the population health sector has not focused enough on finding answers. “Their focus is more on describing health problems than it is on finding solutions,” she says, “because that’s easier work to do and there is money to do it.”
Gruman also has concerns about the variable quality of systematic reviews, especially those not based on randomized controlled trials (RCTs). “The RCT people are struggling so hard to get a good solid set of standards together that everyone can agree on,” she says. “It’s a pretty tough row to hoe to put good standards in place for a systematic review that is not based on RCTs.”
Indeed, many methodological challenges arise in systematic reviews of population health interventions. Elizabeth Waters, director of the Cochrane Health Promotion and Public Health Field and professor at Deakin University, and colleagues note that retrieval of studies is more complicated in public health than in clinical medicine because of a more diverse literature, a wider range of bibliographical tools of varying coverage and quality, and terminological difficulties. Many studies may not be well indexed or may be indexed differently among different databases. Research on public health interventions is often difficult to synthesize because of the complexity of the interventions, the study populations, and the outcomes measured, to name just some of the methodological challenges involved.
The effect of context on intervention design, implementation, and effectiveness is another challenge, while the benefits of many public health interventions may not be evenly distributed, and in some cases interventions may unintentionally exacerbate health inequalities.31 For example, broad public health educational interventions can sometimes be most effective in people who are best educated and already least affected by health problems.
The many forms of bias that afflict primary studies, and sometimes whole bodies of evidence, are also a limitation of systematic reviews. Choi characterizes these as publication bias (wherein studies producing positive results are more likely to be published), the false positive research cycle (in which one false positive finding can generate further research and further false positive results), circular epidemiology (or repetitive publications), and the impact of fashions (wherein some important topics are understudied).
Alice Ammerman, professor in the Department of Nutrition and director of the Center for Health Promotion and Disease Prevention, University of North Carolina at Chapel Hill, has learned from her own involvement in systematic reviews of dietary interventions that many judgments and assumptions are involved. “I know from my own experience that many assumptions are made in manipulating the data to synthesize the findings from numerous papers,” she says. “The danger is that the resulting numbers can look so solid and objective when they may be based on a large number of assumptions.” In short, systematic reviews require judgments as do all scientific activities; they are not of themselves perfect; and they cannot compensate for all of the weaknesses or gaps in the primary literature.
Questions about the implementation of systematic reviews trouble many public health professionals. Several persons interviewed noted the difficulty of assessing whether an intervention shown to be effective in a systematic review will also work in other situations and contexts. Others said that there is often a discrepancy between the artificial conditions of a trial and the real-world conditions in which an intervention is implemented. “A critical issue in public health is, can you translate this finding to different settings with different levels of resources?” says David Atkins, chief medical officer at the Center for Outcomes and Evidence at the Agency for Healthcare Research and Quality. “It’s problematic with clinical interventions too but more so with public health.” Hazen agrees: “The studies that go into a systematic review are usually well funded and staffed; they have resources that many communities don’t have,” she says. “When you move that into a community, there is a chance that you won’t have the same impact.”
Local context and culture are often important. Ammerman says it can’t always be assumed that an intervention of proven benefit in one setting will work in another. “Often the intervention staff and study population vary widely, making it impossible to generalize from one setting to another,” she says. Notwithstanding these concerns, it’s important to remember that systematic reviews offer stronger evidence than single studies can that an intervention proven successful over a broad range of contexts is likely to work in other settings.
Systematic reviews and the studies they include have been criticized for not containing enough information about their external validityincluding factors such as how representative was the population involved—to enable judgments to be made about their applicability elsewhere. Lawrence Green, adjunct professor in the Department of Epidemiology and Biostatistics at the University of California, San Francisco, and others say many systematic reviews have been “sterile” in producing studies that are high on internal validityto help maximize their chances of demonstrating effectivenessbut low on external validity, meaning their relevance, generalizability, and applicability to real-world situations are questionable.32 A method of evaluating the extent to which a trial or a review incorporates these factors, called RE-AIM, has been developed.33
Meanwhile, concerns have been raised that supporting only scientifically proven interventions and policies may limit action or stifle innovation. “It is unrealistic to demand that every decision be based on robust scientific evidence from systematic reviews when we know we are far from having all the information we need,” researchers have noted.34
Green is also concerned that the push to entrench systematic reviews in policymaking may be used as an excuse to quash programs. “There are a lot of good programs for which we may not have the systematic reviews required to get special funding,” he says. “So it becomes a very fatal conservative tool for cutting back programs. I don’t want to sound like a scientific nihilist, but I think we’ve painted ourselves into a rather awkward scientific position by trumpeting the importance of a particular tradition of systematic reviews. They’ve served medicine well, but they’re serving population health poorly.”
Many of these pointed concerns about the potential limitations of systematic reviews in public/population health policy have been raised by researchers working to refine and develop the methodology of systematic reviews. Their comments highlight areas for consideration and further work by those individuals and organizations seeking to enhance the usefulness of systematic reviews in policymaking and thus in improving population health outcomes.
Traffic Safety Interventions: Successes and Setbacks—Implementing the Evidence Saves Lives
Summary: The traffic safety field has developed an extensive evidence base, including many systematic reviews, to guide policymakers. The implementation of effective interventions has prevented many thousands of deaths and injuries, but greater gains could be achieved through more widespread and consistent implementation of proven, cost-effective interventions. Advocacy groups have had a powerful impact on policymaking.
When traffic safety experts tally up their policy wins and losses, they inevitably point to the introduction of 0.08 blood alcohol concentration laws as a critical victory. It is also widely cited as a key example of how one systematic review had a marked impact on policy, leading to the prevention of thousands of deaths and injuries.
In the late 1990s when Congress was considering whether to introduce financial incentives to push states towards 0.08 blood alcohol concentration (BAC) laws, the General Accounting Office (GAO) released a report questioning the merits of such laws. “Overall, the evidence does not conclusively establish that .08 BAC laws, by themselves, result in reductions in the number and severity of alcohol-related crashes,” it concluded.35 This report, widely seen as advancing the cause of the alcohol industry, was viewed skeptically by many experts who were then conducting a systematic review, on behalf of the Community Guide, on the impact of 0.08 BAC laws. Their review suggested such laws could be expected to drop alcohol-related traffic fatalities by about 7 percent.36
When the systematic review was completed, some of its authors ensured their findings were made available to federal legislators. It is widely believed that this was one important factor influencing Congress’s subsequent decision to withhold federal highway construction funds from states that did not pass such laws. Lobbying by advocacy groups was also critical to the congressional sanction, which had a dramatic impact. Before it was introduced, seventeen states had 0.08 BAC laws. By the end of 2004, all fifty states, the District of Columbia, and Puerto Rico had them. It has been estimated that the introduction of the laws in the states that did not previously have them should save at least four hundred to six hundred lives each year.37
According to Randy Elder, the scientific director of the Community Guide, the GAO report and the Community Guide’s systematic review were based on much of the same literature, but their differing conclusions reflected their differing processes. “The GAO’s narrative approach meant they were able to critique each study whereas our review looking at all the studies systematically showed the remarkable level of agreement among them,” he says. “When you looked at all of the data, aggregated into the same table, it became very clear that whatever problems the studies had, they were all coming to roughly the same conclusion.”
For Peter Briss, former chief of the Community Guide branch at the Centers for Disease Control and Prevention (CDC), this case study illustrates the value of involving opinion leaders in systematic reviews so that they can help spread awareness about a review’s findings. It also highlights the potential impact of influencing central policy levers. It was far more effective and timely to have Congress exert a single policy lever to achieve consistent laws across the country, Briss says, than to attempt to lobby each state legislature individually. “That would have been a much longer, harder process,” he adds.
By contrast, the history of motorcycle helmet legislation illustrates the public health challenges that can arise in the absence of a national approach to policy. Jim Nichols, who began his career with the National Highway Traffic Safety Administration (NHTSA) in 1970, remembers back to the mid-1970s, a time when national sanctions meant that most states required motorcycle riders to wear a helmet. Now a private consultant, he also remembers the impact of Congress’s subsequent decision, at least partly in response to lobbying by motorcycle riders, to repeal sanctions for states that did not have such laws.
Many state policymakers, when confronted by a group of angry, leather-clad bikers, chose to overlook or ignore the weight of scientific evidence, including a recent systematic review by the Cochrane Collaboration, showing that motorcycle helmets reduce deaths and head injuries.38 Many states repealed their laws requiring helmets and those that still have such legislation are under ongoing pressure to revoke them. “The people who vote against motorcycle helmet laws are not looking primarily at the evidence but at the intensity of the advocacy,” says Nichols.
Jim Hedlund, a traffic safety consultant who has previously worked with the NHTSA, says opponents of the laws base their arguments on civil libertarian principles. “They say that, ‘it’s my head, it’s no business of yours if I crack it up, so get out of my face and let me ride my bike,’” he says. “They are incredibly effective at lobbying state legislatures. You haven’t been lobbied until a group of bikers with long hair and black leather jackets and broad shoulders come and stand in your office.”
Many public health experts are extremely concerned about the policy backdown, and believe it has contributed to an increase in motorcycle deaths and escalating health costs. The most recent figures show that motorcycle fatalities rose 13 percent from 4,028 in 2004 to 4,553 in 2005 and that almost half of the people who died were not wearing a helmet.39
As with the push to ensure motorcyclists wear helmets, the battle for seat belts is a long one. Despite strong systematic review evidence that primary seat belt laws prevent suffering and death,40 many states still do not have them. Primary laws mean police officers can stop a driver for not wearing a belt. Secondary laws, which exist in the majority of the states, mean police officers can only fine someone for not wearing a belt if he or she has been stopped for another reason. One state, New Hampshire, still has no law at all covering adults.
The traffic safety field has developed a solid evidence base, and dozens of systematic reviews are available to help guide policy decisions.41 Investment in research reflects concern about the enormous toll of motor vehicle accidents, which killed more than 43,000 people in the United States in 2005.42
Ruth Shults, who coordinated the traffic safety systematic review43 on behalf of the CDC, says traffic safety interventions face fewer methodological issues than many other population health interventions. This is because many traffic safety interventions can have nearly instantaneous effects on the health outcomes of interest, making it fairly straightforward to attribute changes of clear public health relevance to the interventions. Randomized controlled trials are rarely applicable to the field, but Shults says the Community Guide holds that findings from an observational study with a well-chosen comparison group can be compelling.
Some of the key audiences for traffic safety systematic reviews are outside the traditional public health domain, and therefore great efforts have been made to disseminate findings to traffic safety researchers and practitioners, as well as to engage them in the review process, says Shults.
Shults and other experts believe that scientific evidence has had greatest impact on policy when used to inform the work of advocacy groups and thereby influence public debate. Organizations such as MADD and the Insurance Institute for Highway Safety have referred to the systematic reviews of evidence on sobriety checkpoints in their calls for broader use of sobriety checkpoints throughout the nation.44 Also, the NHTSA, the CDC, and the AAA Foundation for Traffic Safety are incorporating review findings into their funding proposal guidelines for community-based interventions to reduce alcohol-impaired driving and to increase safety belt use.
Elder cautions that third parties such as interest groups will only be interested in the evidence from systematic reviews if it matches their agenda. “There has to be some level of preparedness to accept the information you’re giving and to act on it,” he says.
The dramatic decline in traffic fatalities is often held up as an example of a major public health victory, but traffic safety experts note that the introduction of effective interventions, such as primary seat belt enforcement laws and sobriety checkpoints, has been patchy. Scientific evidence is only one of a “whole raft of considerations,” which influences policymaking, points out Hedlund. Furthermore, even when policies are introduced, the effectiveness of their implementation can vary. The NHTSA notes that the effectiveness of interventions can fluctuate immensely from state to state or community to community. “What is done is often less important than how it is done,” says the NHTSA. “The best countermeasure may have little effect if it is not implemented vigorously, publicized extensively, and funded satisfactorily.”45
In particular, many experts are concerned that little progress has been made in reducing alcohol-related traffic fatalities in the past decade. “We haven’t been able to break through public concern in the way we did in the 1980s and early ’90s,” says Nichols. “The public gets used to issues, and it’s more difficult to generate interest. Plus 9/11 has had an impact—traffic safety used to compete with crime for police officer attention; now it also competes with security.”
The lesson, he says, is that persistence is required. “You have to continue to collect evidence until you have achieved your goal. You can’t let your evidence age. You have to put it in the hands of credible spokespeople and you have to persevere.”
- Evidence-based traffic safety interventions have saved many lives, prevented much suffering, and reduced health care costs. More could be achieved, however, through more widespread, consistent, and effective implementation of such interventions.
- It is crucial to involve recognized leaders of the target audiences in the entire review process and to broadly disseminate the findings to all segments of the target audiences.
- Dissemination of the findings of systematic reviews is particularly important when the target audiences lie outside the traditional public health field.
- The use of financial incentives and sanctions can encourage the implementation of evidence-based policy.
- Complacency is unwise: effective policies can be repealed and public health messages can lose their impact.
- Commercial interests and advocacy groups, whether basing their arguments on evidence or not, can exert a powerful influence on policy.
- Involving opinion leaders in the development and dissemination of reviews can increase their impact.
Tackling Youth Drinking: Collaboration Counts—When the Evidence Engages Communities
Summary: Binge drinking among college students is a major public health problem upon which prevention campaigns have had little impact. At the University of Nebraska–Lincoln, however, binge drinking rates fell after the introduction of a comprehensive program aimed at creating an environment that supports low-risk drinking. The program, which has been running for several years, was developed and implemented by a broad-based coalition of town and gown interests. Ongoing evaluation of the program is contributing to development of the evidence base for alcohol control policies. While the evidence from systematic reviews is important in shaping such programs, the implementation process is also critical.
Binge drinking exacts a costly toll upon the health, well-being, and education of college students and also affects others in their orbit. They are more likely to binge drink than their peers who are not college students and are more likely than other adults to drive while under the influence of alcohol. It has been estimated that seventeen hundred college students die each year from alcohol-related causes, including about thirteen hundred deaths due to alcohol-related motor vehicle crashes. Heavy drinking among college students also contributes to tens of thousands of assaults and sexual assaults each year.46
Efforts to combat the problem of binge drinking have relied mainly upon educational strategies to encourage individual students to change their behavior and have been largely unsuccessful. A three-year investigation by the Task Force on College Drinking, established by the National Advisory Council on Alcohol Abuse and Alcoholism, identified the need for a three-pronged approach to changing the culture of drinking on campus: prevention programs should target individuals including at risk or alcohol-dependent drinkers; the student population as a whole; and the college and its surrounding community.47
The Task Force, reporting its findings in 2002, also urged colleges to base their policies and prevention programs upon scientific evidence. “In general, colleges and universities have not applied the methods, techniques and findings from cutting-edge alcohol prevention research to the problem of college student drinking,” it said. The Task Force noted that college administrators had come to see drinking as an insolvable problem. “When schools have made efforts to reduce drinking among their studentsand many have made considerable effort—they haven’t had significant campus-wide success. With each failed effort, the image of college drinking as an intractable problem is reinforced, administrators are demoralized, and the likelihood that schools will devote resources to prevention programs decreases.”
The Task Force also highlighted the dearth of knowledge about what interventions are effective at preventing students from binge drinking and said evaluations of environmentally focused prevention strategies specific to college populations are “sorely needed.”
A bright light amid this gloom is a program that has had a marked impact on binge drinking and related harms at the University of Nebraska–Lincoln. It is one of ten universities funded under the Robert Wood Johnson Foundation’s Matter of Degree program, administered by the American Medical Association.48 Participants identify the environmental factors such as advertising and marketing, institutional policies and practices, local ordinances, and cultural beliefs that converge to encourage alcohol abuse. Most of the universities funded under the program have high-profile athletics programs and large Greek chapters—characteristics associated with increased rates of harmful drinking. The program supports a collaborative approach between participating universities and the communities where they are located.
Since 1998, the NU Directions Campus-Community Coalition at the University of Nebraska–Lincoln (UNL) has been developing and implementing a raft of strategies to create a culture that supports responsible low-risk drinking, including abstinence. These have included legislative reform to reduce the concentration of liquor outlets, media campaigns, a web-based training program for hospitality and retail workers, and projects focusing on fraternity and sorority chapters. The coalition involves community leaders; city, county, and state officials; police officers; hospitality owners; prevention specialists; educators; students; and parents. Between 1997 and 2003, the UNL binge drinking rate fell from 62 percent to 47 percent, but remains above the national rate.49 At the same time, the student retention rate increased approximately 1 percentage point per year.
The project, which has an onsite evaluator, has been heavily data-driven, according to its director, Linda Major. But the key reason for its success, she adds, is the strong coalition that has been built between the many stakeholders in both the town and gown communities. “Having data available to help guide our strategies and our resource allocation was absolutely critical, but these other process factors were just as critical,” she says.
Tom Workman, the program’s associate director, agrees that the long, laborious process of developing a coalition of interests that understood the problem and supported action was crucial. “You can’t simply present data and assume data cuts across a variety of political considerations,” he says. “Here, there was a willingness across the community to say, we believe the data. It takes years to establish relationships that enable people to see the data for what it is and to be willing to work collaboratively.”
Responding to stakeholders’ concerns was another vital part of the process. For example, when hospitality and liquor outlets opposed a plan for mandatory staff training as expensive and impractical because of their high staff turnover, a free, web-based program was developed. “You will be much more successful in accomplishing the policy once you have worked out the problems of implementation rather than jumping in and fighting resistance,” says Workman.
The project also had to be flexible, for example, when it met resistance from the Greek chapters in developing strategies to tackle their risky drinking culture. The project was much more successful after being reframed as a way of helping the chapters tackle problems such as declining membership, negative press, and image problems. “That allowed the alcohol issues to become part of the discovery when fraternity and sorority members analyzed factors contributing to existing problems,” says Major. “It took much longer this way, but it was much more effective.”
With the project involving dozens of different strategies, Major says it is difficult to know which were the most effective, and she expects it is the “totality” of the approach, with synergistic effects, that counts. But she has no doubt that a critical element was bringing on board a communications professional to engage the media. The five-year evaluation of the project also noted the importance of having a staff member who understands community organizing, uniform messages, and soliciting feedback.50
One of the other notable lessons, Major adds, is that such campaigns take years of consistent work and require patience. The grant was originally for five years, but was extended for another four years. Major says it is unusual to win support for such a long-term project: “Rarely do you have these kinds of resources or a funder who actively supports long-term strategic planning to have the opportunity to create thoughtful, intentional strategies based on literature, experience, and intuition that make an impact.”
For Toben Nelson, a researcher at the Harvard School of Public Health who has been involved in evaluating the Matter of Degree program, the Nebraska project is a rare example of an effective effort to prevent alcohol-related harms on campuses. While an evaluation of the ten universities involved in the program found it did not result in significant improvements overall, a different picture emerged when considering the five sites, including Nebraska, that most closely implemented the environmental model. These universities reported reduced alcohol consumption and related harms, with benefits both for the drinkers and those around them.51
Nelson says there is great potential for policymakers to make better use of the findings of systematic reviews when examining options for reducing alcohol-related harm in the broader community as well as on college campuses. Effective interventions include enforcing the minimum legal drinking age, zoning restrictions to reduce alcohol outlet density, keg registration, mandatory responsible beverage service training, and increased excise taxes.
“My general impression is that the specific policy recommendations tend to be ignored or people seem to think that they are too difficult to implement,” Nelson says. This is largely because policymakers often don’t have the skills, expertise, authority, or power needed for implementation, he adds.
“What you really need,” Nelson continues, “is almost a campaign manager and that’s not who these people areby and large they are people within the bureaucracy and they don’t tend to be the leaders with the authority to make those kinds of changes happen, whether that’s on a college campus or within a country or a city or a state. Another reason is that it takes a long time to get this stuff to happen and there are lots of barriers along the way. The alcohol industry is pretty active in offering prevention alternatives that are much easier to implement but less effective.”
Another hurdle is the misperception that it is the heaviest drinkers causing all the problemsor the “few bad apples” perspective. An understanding of the “prevention paradox” reveals that while extreme drinkers are at highest risk of harm, the majority of harms in a community arise from moderate or low consumption because that involves the largest number of drinkers.52 Nelson says policymakers who fail to grasp this concept do not appreciate that the greatest health gains to a community will flow from incrementally changing the drinking patterns of the majority, rather than focusing on the smaller group of heavy drinkers.
Nelson says the program at the University of Nebraska–Lincoln is a “shining example” of an effective program. Its success has hinged on the program’s strong leadership as well as its high-level support from the university administration and the city. “That’s laid the groundwork for an extremely effective individual with a good team around her that has led a political style campaign,” he says.
- Having a solid evidence base is important but is only one of the elements needed for success in community-based interventions.
- Building broad coalitions can be critical for the success of population-based interventions; stakeholders need to be engaged and their concerns addressed.
- Community-based programs need to be flexible and take a long-term approach.
- Leadership is important, both within and outside the program.
- Effective engagement with media is critical.
- Understanding the prevention paradox is vital when designing population-based health strategies.
- Ongoing evaluation helps inform strategy development and implementation.
Tobacco Control: The Long War—When the Evidence Has to Be Created
Summary: Fearless advocacy and a willingness to experiment with novel approaches drove a ground-breaking tobacco control program in California. It had a dramatic impact in changing community attitudes towards smoking and reducing tobacco consumption in the early 1990s, but has subsequently lost some of its impact. The program, and others that followed elsewhere, contributed to the development of an evidence base, which includes many systematic reviews, on effective tobacco control interventions.
Public health history was made in November 1988 when Californians voted for a citizen-initiated law called Proposition 99. It increased tobacco taxes and devoted 20 percent of the money raised to tobacco control, creating the world’s largest tobacco control program almost overnight. However, this victory marked only the beginning of a long and difficult battle on behalf of what has been described as one of the most intensive, aggressive public health interventions ever undertaken.53 The challenge for Californian policymakers, who were breaking new ground in the early 1990s, was how best to spend their windfallabout $100 million per yearwhile fighting ongoing battles with antagonistic political, professional, and industry interests.
Dileep Bal was one of the key tobacco warriors within the California Department of Health Services for fifteen years, until recently moving to his current position as district health officer at the Hawaii State Department of Health. While at the California Department of Health Services, Bal oversaw the distribution of funding to one of the first statewide anti-tobacco media campaigns, as well as to local health departments, community-based organizations, schools, and other agencies to promote tobacco prevention and cessation. Bal says about 30 percent of tobacco control funds was allocated to media campaigns, about 30 percent to local health departments, about 30 percent to competitive grants, and 10 percent to evaluation. The centerpiece was the innovative media campaign, which, rather than taking the traditional health promotion approach of warning that “smoking will kill you,” attacked the tobacco industry directly and aimed to change social norms around smoking. The goal was to overturn industry marketing portraying smoking as sexy and cool.
“For a government program, it was very, very atypical,” recalls Bal, whose job came under threat many times. “It was very unusual for government to take on what was referred to as a legally constituted industry. I elected to take on the tobacco industry head-on and I have the scars on my back to show for it.”
Bal laughs when asked about the role of scientific evidence in guiding policy decisions. “There was no science on how to do a community intervention on something of this global dimension,” he says. “Where there is no science, you have to go and be venturesomeyou can’t use the paucity of science as an excuse to do nothing. We created the science, we did the interventions and then all the scientists came in behind us and analyzed what we did.”
Bal is frustrated by colleagues who want to wait for high-level evidence before acting. “Public policy must be propped up by science but must not be a handmaiden of the science,” he says. “Most scientists will say you need a randomized controlled trial level of proof to do a community intervention. That’s horse feathers. We tried twenty-five things—twelve worked and we renewed those. Empirical trial and error is the oldest scientific device and we used it to distinction.”
Another critical aspect of the program, says Bal, was having advocates outside the bureaucracy who were fearless in speaking out. These included Stanton Glantz, professor of medicine at the University of California, San Francisco, and his colleague John Pierce, a professor at the University of California, San Diego (UCSD). Both Glantz and Pierce documented the program’s initial dramatic success in dropping smoking rates between 1989 and 1993, before cutbacks in funding and political support stymied further progress.54 More recently, the program has been described as stagnant and facing an uncertain future.55
Despite the setbacks, there is no doubt the program has had a dramatic impactand not only because in its first eight years, it is estimated to have prevented 58,900 deaths between 1989 and 1997.56 Its impact reached beyond California’s borders, providing other jurisdictions with confidence that smoking rates could be dropped and guidance about what interventions might work. California’s contribution to knowledge about tobacco control was particularly important after the unprecedented, multibillion dollar settlement reached between the tobacco industry and most state governments in 1998.
Not all states used these funds for tobacco control, but the settlement contributed to many new anti-tobacco programs. Tobacco control is one of the most common reasons that policymakers turn to the Community Guide,57 which lists more than a dozen systematic reviews in the areas of preventing the initiation of tobacco product use, increasing cessation, and reducing exposure to environmental tobacco smoke. The Community Guide also includes economic evaluations for many recommended interventions.
Systematic reviews have also helped policymakers to appreciate the unreliability of research and review articles supported by the tobacco industry.58 One such investigation found that review articles concluding secondhand smoke was not harmful were about ninety times more likely to be written by authors with financial ties to the tobacco industry rather than those without such ties. Another systematic review found that lower quality studies were much more likely to conclude that smoke-free regulations adversely affect the hospitality industry and were also much more likely to be industry-funded. This review would not have had such an impact on policymakers, according to one of its authors, Michelle Scollo, a researcher with the Cancer Council Victoria in Australia, if the researchers had not been proactive in presenting the findings to politicians and public forums using simple, straightforward language and a graphic PowerPoint presentation. Effective communication of systematic reviews requires “persuasive speech” to help policymakers connect the evidence with their constituency, adds Lisa Bero, a professor and a prominent member of the Cochrane Collaboration, based at the University of California, San Francisco. “It’s not so much about spin but about making the numbers come alive,” she says.
Many experts believe that reports issued by prominent authorities, such as the U.S. surgeon general, have been more influential in driving policy than systematic reviews, which were scarcely cited in an online poll by the journal Tobacco Control of the most important and influential PubMed papers in tobacco control.59 Bero says it is unfortunate many of the landmark reports on tobacco control have been consensus documents, rather than the product of the methodology and rigor of a systematic review. “If there’s a way we can introduce the systematic review methodology more into these reports, not only would they be more influential, they’d be stronger,” she says. There is also room to increase the role of evidence in tobacco control advocacy, she adds.
Bero also points out that systematic reviews played an important role in tobacco control by establishing evidence of the harm caused by secondhand smoke exposure. “Without that evidence of harm, there would have been no need to make policy at all,” she says.
A solid evidence base is only one element of effective advocacy, argues Simon Chapman, professor of public health at the University of Sydney, Australia, and editor of Tobacco Control. He cites a survey of Australian policymakers and advocates, suggesting that economic, ideological, and anecdotal arguments often outweigh scientific evidence in policy decisions about environmental tobacco smoke. Politicians and political advisors told the researchers doing the survey that evidence has limited power in governmental deliberations. One political advisor said “evidence” presented by any lobbyist is viewed skeptically because lobbyists tend to spin their findings to suit their positions. A politician suggested that anecdotes and “folklore” carry more weight in political conversations than scientific studies and other forms of independent evidence.60
In the United States, some of the leading players in tobacco control policy are skeptical about the value of systematic reviews in population health policymaking. Lawrence Green, adjunct professor in the Department of Epidemiology and Biostatistics at the University of California, San Francisco, who was closely involved in evaluating tobacco control programs for the WHO and the CDC, argues that many systematic reviews of population health interventions have been “sterile” and of little help to policymakers. “They’ve produced a lot of studies that are high in internal validity [to help maximize their chances of demonstrating effectiveness] but relatively low in external validity [their relevance, generalizability, and applicability to real-world situations are questionable],” he says. “This means they screen out a lot of studies that might have greatest relevance for policymakers and leave in highly controlled studies, which as a result of all the experimental controls have become somewhat artificial.”
Green says his experience evaluating the California tobacco control program suggests that its comprehensive approach was the reason for its success, rather than the effectiveness of any one intervention. “School-based programs may not have much effect by themselves, but when combined with community efforts, they take on a whole different meaning,” he says. “And when they’re supported further by the mass media, it takes on even more meaning. And the mass media wouldn’t be effective without what’s going on in the home. And then effective enforcement of what kids can buy across the counter is also important. Each of these interventions by themselves may have limited effect if not supported by the others but combined you get a synergistic effect.”
When Green was working for the CDC, helping to promote a manual called Best Practices in Comprehensive Tobacco Control, he says they “took some heat” from headquarters because they broke rank with the use of the term best practices by going beyond controlled trials for evidence. They pushed for the manual to include case studies from California and Massachusetts rather than limiting it to systematic reviews of highly controlled studies. Public health differs from evidence-based medicine because it is intervening across very complex systems and populations rather than across the relatively homogenous biological organisms called patients, he says. “In public health we can’t afford to be too precious about our definitions of evidence,” he says. “It was the comprehensiveness of the approach that we wanted to represent to the other forty-eight states as best practice. And it was the real-world, real-time experience of the two states that made the manual more credible to other states than if it had been limited to carefully controlled studies in more academic settings and circumstances.”
As previously discussed, systematic reviews can include many methodologies other than randomized controlled trials. Pierce, director of the Division of Population Sciences, Moores UCSD Cancer Center, believes randomized controlled trials are best suited to studies of plants and pharmaceuticals and do not provide a useful basis for population health decisions. Their use is particularly limited in communication studies where the intervention cannot be controlled. “In reality, people change their behavior because of social norms, and the rules for systematic reviews frequently leave out studies that address social norms,” he says. “Very few people have done good evaluative studies in which social norms were changed.”
The ideal tobacco control program doesn’t exist and population-based programs are political and unpredictable. “We have some things that work in some places, some things that don’t,” Pierce says. “A population program is more like a rodeoif I’m on a bull, all I can do is tell you where I’ve been, I can’t tell you where I’m going. It’s not like having a map. Things change in front of you, you can’t control the events.”
- Changing social norms through legislative and regulatory changes as well as innovative media campaigns is an effective way of influencing individuals’ behavior.
- Achieving policy change requires courage, persistence, and a thick skin. A reliable evidence base also helps.
- A comprehensive range of interventions working at multiple levels is more likely to have an impact than any single intervention.
- Systematic reviews are more likely to have an impact if efforts are made to disseminate and “sell” their findings.
- Systematic reviews are more likely to have an impact if they address issues of external validity and relevance as well as issues of internal validity and bias.
- Credible evaluations of state programs and policies are needed to continue to add to the evidence base.
Obesity: Dealing with Uncertainty and ComplexityWhen There Are Gaps in the Evidence
Summary: The obesity epidemic presents policymakers with many complex challenges. Multifaceted community-wide interventions are being developed and implemented, but sometimes there is considerable uncertainty about their likely impact and cost-effectiveness. Responses are being guided by common sense as well as scientific evidence from systematic reviews and sources such as the Community Guide.
From Africa to Australasia and from Europe to the Americas, an unprecedented public health problem has arisen on a scale that would have been unimaginable to previous generations. A multitude of factors has contributed to the rapid emergence of the international obesity epidemic, including modern lifestyles, leisure choices, working environments, urban design, transport, food production, and technological development. Clearly, there are no simple solutions to such a multifaceted problem, and policymakers are confronted with many tough dilemmas. When there are so many uncertainties, what courses of action will bring the greatest benefits? How can politicians and the broader community be helped to understand that community-wide strategies are more likely to succeed than those focusing on individuals’ behaviors? How can the push from vested interests to medicalize the response be resisted? How can sectors outside the traditional health domain—such as transport and urban planningbe engaged? What areas of research will yield best returns on investment?
When grappling with such issues, many policymakers have turned to the Task Force on Community Preventive Services and the Community Guide for clues to the way forward.61 The Community Guide has published a number of systematic reviews in the areas of physical activity, nutrition, and obesity prevention, and has many more pending. Its work has highlighted significant gaps in knowledge, often reflecting a lack of primary prevention studies. More is known, for example, about how to promote physical activity than about how to increase healthy eating. Even where interventions are known to be effective, many questions remain about their implementation and cost-effectiveness.62 The Task Force has listed numerous research questions it would like public health practitioners, policymakers, employers/purchasers, and funders to consider when establishing research priorities.
Meanwhile, some policymakers have described their difficulties responding to recommendations from the Task Force, the Institute of Medicine, and other organizations, citing a lack of guidance on how to translate public health and clinical evidence about obesity control into meaningful policies.63 Others have questioned whether “the people, the proficiencies, and the proliferation processes” are available to achieve effective translation and dissemination.64
In the absence of comprehensive evidence about which combination of interventions will be most effective and how best to implement them, policymakers must rely on “common sense and input from a variety of sources including staff and special interests,” according to Ross Brownson, a professor at the Saint Louis University School of Public Health. For example, many states are implementing obesity prevention policy measures without knowing for certain which are most effective.
Policymakers could often also make better use of the evidence that is available, says Neal Kohatsu, chief of the Cancer Control Branch, California Department of Health Services. “There are many gaps and unanswered research questions from both the clinical and population health perspective, but my own feeling is that we aren’t making use of the available evidence,” Kohatsu says. “Overall, most people are not as familiar with the evidence as they should be.”
For the past several years, a broad-based coalition has been working in North Carolina, where obesity rates are above the national average, to increase physical activity and healthy eating wherever “people live, learn, earn, play and pray.”65 The North Carolina Division of Public Health provides staff support and funding for the program’s leadership team, representing more than one hundred public and private partners. Interventions include media and public awareness campaigns, nutrition standards for school meals, mandatory physical activity for grades K8, worksite programs, and grants for communities to implement physical activity and nutrition programs.
The campaign’s materials note that public health victories in areas such as traffic safety and tobacco control offer lessons for obesity control, such as outlining the importance of having highly credible scientific evidence that can persuade policymakers and withstand attack from those whose interests are threatened; advocates who are passionately committed to their cause and who have the inner resources to withstand the pressure applied by the industries whose practices they are criticizing; a partnership with the media; and laws and regulations.66
Leah Devlin, state health director, says that while the Community Guide and recommendations from other expert sources have provided some guidance, policymakers have often had to rely on “what we think makes sense.” “In public health we do make our decisions to the extent that we can on sound science but in obesity we can’t wait for the science to act,” she says.
One of the difficulties, adds Marcus Plescia, chief of the Public Health Division’s Chronic Disease and Injury Section, is that the interventions supported by the strongest evidence—such as making changes to the built environment to promote activityare the most difficult to implement, requiring major societal changes. Interventions that may be more straightforward to implement—such as changing school menusinvolve many uncertainties. “Nobody really knows whether if you start changing menu items in schools and vending machines, how kids will react,” he says. “The big concern is that kids will stop eating school lunches altogether, which we do not want to happen. The question is how hard can you push regarding what nutritional offerings are there, and how quickly or slowly do you need to phase things in? And there’s not a lot of evidence on those issues.”
North Carolina’s plan to prevent overweight, obesity and related chronic diseases from 2007 to 2012 is unlike many health policy documents. Noticeably, it does not carry the logo of the North Carolina Division of Public Health or other groups involved in its preparation. This is part of a deliberate strategy to encourage everyone in North Carolina to feel ownership of the plan and free to disseminate it as his or her own. “One of the great lessons we’re learning is that sometimes it is good to lead but lead in the background and let others carry the wider issue,” says Cathy Thomas, head of the Division’s Physical Activity and Nutrition Branch.
The plan is also notable because it is written in an engaging storytelling style, rather than the dry, academic language and jargon often found in policy documents. “We wanted to write a plan that was readable,” says Thomas. “So many times state plans sit on people’s shelves. We wanted it to be attractive and to tell the story so people understand that while obesity is a public health issue, it is an issue we all have to engage in, whether you’re in a preschool or school or work setting or health care.”
Beverley Perdue, North Carolina’s lieutenant governor, is also cited as having played an important role in championing the need for action on obesity, both among legislators and the broader community. Plescia says the governor’s work, chairing the Health and Wellness Trust Fund Commission and convening a group of experts and politicians to form the Childhood Obesity Study Committee, has led to significant policy change.
In Arkansas, another state noted for its attempts to tackle obesity, the personal support of two prominent local leaders has also been critical in garnering community and political backing for tackling obesity. This is widely seen to reflect their personal experiences with lifestyle-related illnesses: Herschel Cleveland, former House speaker, had a heart attack, while Mike Huckabee, the former governor, went on a high-profile and successful weight loss campaign after being diagnosed with type 2 diabetes.67
When evidence is lacking, policymakers bear an extra responsibility to ensure programs are carefully evaluated so as to contribute to the development of the evidence base. The impact of obesity control efforts in North Carolina, Arkansas, and other states will be closely watched by policymakers who are grappling with similar challenges around the world.
Another important message from the obesity front line is that there will be no quick fix. North Carolina has acknowledged that a long road lies ahead, noting that it took almost a century to move from the discovery that lead paint caused neurological damage in children to legislation and regulation that dropped children’s blood lead levels.68 Devlin hopes progress on obesity will come more quickly than this, however.
- Policymakers often must make decisions on the basis of incomplete evidence.
- A lack of evidence is no excuse for inaction.
- Systematic reviews can help identify gaps in the evidence and priorities for future research.
- Policymakers need more than the findings of systematic reviews: they need advice on how to translate these findings into policy and action.
- Prominent leaders with some personal stake in the issue can be powerful champions.
- Evaluation of programs and policies is vital.
- Complicated health problems require complex solutions and persistence.
The Mental Health Aftermath of the Tsunami: Levels of ChaosWhen the Evidence Is Not Relevant or Applicable
Summary: While grappling with the chaos that follows a disaster, policymakers can struggle to find systematic reviews or other scientific evidence that is relevant or helpful to their needs. Most disaster research has been conducted in developed countries, although disasters are most frequent and damaging in poorer countries. Population-based mental health responses to disasters should focus on minimizing stressors and providing social support, as well as ensuring medical services are available to those who need them.
In late 2004 and early 2005, chaos reigned in many countries bordering the Indian Ocean after a tsunami wrought destruction on an unimaginable scale. Confusion also resulted when a plethora of governments, charities, emergency workers, and volunteers raced to respond to the massive humanitarian crisis.
In some senses, the chaos that inevitably follows in the wake of disasters is also a metaphor for the complexities confronting policymakers in such situations. They must respond quickly to an emergency despite confusion and controversy about which interventions might be most beneficial and growing awareness of the potential for some interventions to do more harm than good. Much debate also surrounds whether some Western responses to disaster, including a focus on post-traumatic stress disorder and other mental health problems, are appropriate or useful for other cultures.69
Within days of the tsunami swamping international news headlines, Fran Norris sat in her office in Vermont and methodically worked through a systematic review she’d published with colleagues in 2002, analyzing studies of the mental health of 160 samples involving more than sixty thousand disaster survivors.70 Among other things, the review highlighted how skewed the literature was towards developed countriesfar fewer studies had been done in developing countries, despite the fact that they tend to suffer more disasters and that these tend to have a worse impact. The finding illustrates both a limitation of systematic reviews, that they are the constrained by the nature of the primary studies, and one of their key values: their ability to identify gaps in the literature.
After the review’s publication, Norris, a research professor in the Department of Psychiatry at Dartmouth Medical School and a psychologist with the National Center for Posttraumatic Stress Disorder, prepared a short executive summary to help disseminate its findings. She has made a conscious effort to learn to write in a nonacademic style although she finds this difficult. “Academics don’t always know how to write interestingly and concisely, to skip the details and qualifications,” she says. “It’s hard but if you don’t do it based on research, someone else is going to do it based on something else. It’s challenging to go out on a limb; others may say you should never go beyond the data.”
Norris had feedback suggesting the short executive summary generated more interest than the review itself, although she was not sure what impact, if any, it had on policy. “It got attention but I’d be hard-pressed to say it changed policy,” she says. “But in some ways, who am I to tell any government what their policy should be? I can try to give some information; here is our best understanding, but I am not even sure I can tell you what it says you should do,” she says.
After the tsunami, Norris extracted from the review those studies involving hurricanes, floods, tornadoes, and dam bursts, which might be relevant to the tsunami response. She prepared and posted on the Internet a summary providing the following tips for policymakers: that most people would recover over time but some groups (women, children, people with pre-existing psychiatric conditions, minority ethnic groups, and the poor) were at increased risk of persisting problems; that secondary stressors were an important predictor of chronic problems; that naturally occurring social resources were vital; and that interventions should address both psychological and social/community needs.71
Coincidentally, at about the same time as this summary hit cyberspace, WHO officials published an article suggesting a way to move beyond the controversies that have surrounded mental health responses to disasters.72 It noted that disasters in poor countries are often followed by the arrival of foreign clinicians promoting trauma-focused treatment in the absence of “a system-wide public health approach that considers pre-existing human and community resources, social interventions, and care for people with pre-existing mental disorders.”
Based on a nonsystematic literature review, a survey of experts, and the findings of consensus statements and guidelines, the article proposed the following set of principles and strategies to guide responses to populations exposed to extreme stressors: contingency planning before the acute emergency, assessment before intervention, use of a long-term development perspective, collaboration with other agencies, provision of treatment in primary health care settings, access to services for all, training and supervision, and monitoring indicators. Most of these WHO recommendations had been included in a “mental and social aspects of health” standard in a revised edition of the disaster response “bible,” the 2004 Sphere Handbook. It was the first time this widely used handbook included such a standard.
After the article was published, officials at the WHO commissioned systematic reviews, which included anthropological, qualitative, and social science literature, to examine how their recommendations equated with the evidence base. The findings of these reviews were used to inform the development of a more detailed set of guidelines.73 According to Mark Van Ommeren, a public health scientist at the WHO in Geneva responsible for mental health in emergencies, this example illustrates the reality of how policymaking often occurs. “A lot of people in academe think you do a systematic review and then there’s policy,” he says. “I don’t think it happens at all like that. My experience is that people make policy and then they look for the evidence, and if they are objective, they adjust the policy. For better or worse, the policy process runs way ahead of the evidence process.”
In the aftermath of the tsunami, officials at the Thailand Ministry of Public Health turned to the WHO and UNICEF for information to guide their policy responses for strengthening community networks and public education. They were apparently not aware of the Norris review or the related summary that she produced.
Van Ommeren says it is impossible to know how widely WHO and UNICEF recommendations were followed on the ground after the tsunami. “I went to rural Sri Lanka and found our guidance being written on the blackboard and that was somewhat reassuring,” he says, “but tons of groups respond to such emergencies and everyone seems to have their own approach. The aid workers at the scene act according to what they think is the best at the time and it’s usually not fantastic; for example, people try to use a clinical model, thinking that anything is better than nothing, which is a very dangerous assumption. The assumption that any care is better than no care often leads to harm.”
Van Ommeren says much evidence developed in the West is irrelevant when responding to disasters in developing countries. For example, he says, the Cochrane Collaboration circulated systematic reviews to help the tsunami response efforts, but these were rarely useful because most involved interventions that were not applicable or were relevant only to professionals with advanced training. In addition, these reviews tended to address the effectiveness of individual interventions rather than how services and responses should be organized.
Even if applicable and relevant, health and medical literature alone cannot provide all the information required in the aftermath of a disaster. Sandro Galea, associate professor of epidemiology at the University of Michigan School of Public Health and co-director of the Disaster Research Education and Mentoring Center, says it is unfortunate there are not better linkages between all areas of research relevant to disaster responses, including engineering, political sciences, and health.
An understudied area, he adds, is how disasters affect the behavior of populations. “If I were in charge of policy, I would be much more interested in how people behave after a disaster than their mental health or health,” he says. “Companies want to know, for example, if they will still have workers or consumers. The GAO report on the impact of pandemic flu says a large proportion of the costs will be due to behavioral changes.”
Galea also stresses the importance of ensuring that the mental and physical health responses are closely integrated. Overall, however, the most critical message from the disaster response research for policymakers is to minimize ongoing stressors and to provide social support. “Rather than devising interventions based on cognitive behavioral therapy or treatments, our biggest population impact is going to be doing something about stressors and social support,” he says. “You have to make sure people are housed, fed, and know where their kids are, and have the formal and informal resources to help them cope with the stressors.”
Galea says there is a gulf between policymakers and the disaster response evidence base, and he does not expect it to be bridged quickly. “There is always a gap between science and policy, but this is a relatively new field,” he says. “Given that this field has coalesced in the last ten years, it will be another ten years until it makes it into policy.”
- Findings can be extracted from systematic reviews to provide information relevant to particular situations or contexts.
- Dissemination of systematic reviews requires particular skills, which do not always come naturally to the producers of such reviews.
- The policy process frequently runs ahead of the evidence base, which is often developed after policies have been implemented.
- Systematic reviews are not always relevant or applicable to all cultures or contexts.
- Policymakers require better integration of evidence from a range of fields beyond health and medicine.
Drug Abuse Resistance Education (DARE): Lessons LearnedWhen the Evidence Is Lacking
Summary: The Drug Abuse Resistance Education (DARE) program has been widely implemented in schools in the United States and elsewhere since its development in 1983, despite good evidence from a meta-analysis that it has negligible or no long-term benefits. The program has been significantly revised in response to widespread criticism and following the advent of federal education funding policies that encourage prevention programs to be based upon reliable evidence such as that from systematic reviews. Results of further evaluations of the revised program are awaited. DARE’s successful diffusion may hold lessons for those seeking to disseminate effective programs.
In 1983, a former police chief and a health education specialist in Los Angeles combined forces to create a drug education program that soon swept across the nation, becoming the most widely implemented drug prevention program in the United States.74 In its first year, ten police officers were trained to take the Drug Abuse Resistance Education (DARE) program into fifty elementary schools.75 In 2006, the DARE organization expected to reach thirty-six million school children around the world, including twenty-six million in the United States, where it has been implemented in 75 percent of school districts.76
The program has evolved since its inception to include a variety of activities, including substance abuse prevention curricula for primary and secondary school students, parent education, and after-school and summer programming. Despite its popularity, DARE has attracted widespread concern among health and education policymakers and researchers because the weight of evidence suggests it is not effective. A 1994 meta-analysis, aggregating the findings of several evaluations of DARE, is widely cited by those concerned about its opportunity cost in diverting resources from programs more likely to be effective.77 Brian Flay, professor of public health at Oregon State University, argues that the meta-analysis findings should inform policy: “The policy implications are clear. Don’t do DARE. Rather, spend your money on a program of proven effectiveness.”
In 1998, following widespread controversy about DARE programs, the U.S. Department of Education introduced measures to promote the use of evidence-based prevention programs by state and local education agencies that receive funding under the Safe and Drug-Free Schools and Communities Act. Nevertheless, DARE continued to be extremely popular with many schools and communities, and one study found it was often implemented more widely than interventions of proven benefit.78
In response to concerns about the effectiveness of its programs, DARE significantly modified its approach and in 1999 collaborated with the Robert Wood Johnson Foundation on a five-year national study evaluating a new curricula for seventh and ninth grade students developed by a national team of prevention experts. The results of this study are in preparation and should be available in 2007. Zili Sloboda, the study’s principal investigator and a senior research associate at the Institute for Health and Social Policy at the University of Akron, does not agree with calls to abolish DARE, and instead proposes that experts work with the organization to improve the program. Many critics fail, she says, to make a distinction between the program’s content and the effective delivery network created by DARE, which includes impressive quality assurance measures.
“DARE has been extremely effective in creating and sustaining the network,” she says. “DARE has the ability that no other unit in the U.S. really has to introduce a new program and disseminate it within a few months to thousands of communities. This is an incredible network, and it’s underestimated.”
Sloboda believes that some of the opposition to DARE is driven by ideology, including animosity towards police officers among researchers belonging to the “hippie generation” and beliefs that teachers are best at delivering prevention programming. She says prevention programs that serve to alter attitudes and misconceptions regarding the use of drugs require skills that may be quite different from those needed to teach math or science.
Others argue, however, that DARE’s widespread implementation illustrates the victory of ideology over evidence, the power of the “feel good” effect and the intuitive appeal of connecting police officers and students. DARE’s successful dissemination has also been attributed to the impact of successful marketing.79
Critics of DARE have sometimes felt a backlash. Kelli Komro, an associate professor at the Institute for Child Health Policy at the University of Florida, who has been involved in evaluating DARE programs, says DARE’s supporters are often vocal and passionate. “I recently gave a presentation in Washington DC on alcohol and drug prevention programs and said, get rid of DARE,” she says. “The audience reacted. It was like an uproar.”
While DARE is sometimes held up as an example of the failure of evidence to influence policy, Jonathan Fielding, director of public health and health officer for the Los Angeles County Department of Health Services and professor at the University of California, Los Angeles, notes that the program was revised in response to negative evidence. The DARE story is also a reminder that public health interventions are less likely to be effective if they focus on changing individual behavior rather than the environment. “We’ve learned increasingly that health education alone is going to be insufficient to change behavior,” says Fielding. “It may affect attitudes and knowledge but is unlikely alone to affect behavior.”
Don Des Jarlais, director of research at the Beth Israel Medical Center Chemical Dependency Institute, contrasts the rapid spread of DARE with the much slower diffusion of another drug-related program, one backed by compelling evidence of benefit—the syringe exchange program. These contrasting case studies illustrate the impact of how issues are framed, he says. The concept of framing was developed within psychological decision theory and refers to the context within which a particular problem is posed. Framing can determine whether something is considered a problem, the range of acceptable solutions, and the most preferred solutions.
DARE was framed as a law enforcement response to a criminal problem whereas the syringe exchange program was framed as a response to the public health problem of HIV in injecting drug users. “The critical thing is that he who gets to frame the question usually gets to decide the answers,” says Des Jarlais.
Another lesson from these case studies is that evidence eventually will influence policy, Des Jarlais says. “DARE is being cut back or reformed, and we do have more state and local governments supporting syringe exchange programs,” he says. “For big public health problems, over the long run evidence will work, but it can be a very long run.”
- Programs can diffuse rapidly despite evidence of lack of effectiveness.
- Programs entrenched within communities and defended by powerful interests can successfully resist evidence-based criticisms.
- Even when a systematic review provides clear answers about the effectiveness or otherwise of an intervention, the implications of these findings for policymaking can be interpreted in different ways.
- It may take many years before evidence influences policy.
- Evidence can inform policy by leading to improvements in programs, as well as replacement of programs.
- How problems are framed can have a powerful impact on the type of solutions adopted.
- Reviews should address both individual and environmental interventions and how such interventions should be organized.
The Washington State Institute for Public Policy: Connecting Policy and ScienceWhen the Infrastructure Supports Evidence
Summary: In 1983, the Washington State Legislature established the Washington State Institute for Public Policy to provide evidence-based advice on policy questions in diverse fields, including education, health, and criminal justice. The Institute’s systematic reviews help answer questions not only about which interventions work or do not work, but also about cost-effectiveness. The Institute is highly regarded by policymakers as a source of credible advice, but its findings do not always influence policy decisions.
In 2005, legislators in Washington State were facing a dilemma. They had been warned that the state would need to build two new prisons by 2020 to cope with increasing numbers of prisoners. With each prison costing about $250 million to build and $45 million a year to operate, the politicians were looking for advice, based on the best available scientific evidence, about how to reduce the impending burden on taxpayers while maintaining community safety.
The Washington State Legislature has a long history of interest in evidence-based policymaking. In 1983, it established the Washington State Institute for Public Policy to carry out practical, nonpartisan research to help answer policy questions nominated by legislators.
The Institute conducts research using its own policy analysts and economists, specialists from universities, and consultants, and its staff works closely with legislators, legislative and state agency staff, and experts in the field. It has developed expertise in conducting systematic reviews that not only identify interventions that work or do not work, but also answer questions about cost-effectiveness.
In 2005, the Legislature allocated funding for the Institute to conduct a systematic review of cost-effective, evidence-based options to reduce the projected future demand for prison beds, to save taxpayers money, and to contribute to lower crime rates.
The reviewers identified 560 controlled evaluations of adult corrections, juvenile corrections, and prevention programs. To be included, an evaluation had to have a nontreatment or treatment-as-usual comparison group well matched to the program group. The findings of evaluations with nonrandomized designs were discounted by a uniform percentage.
The reviewers preferred evaluation studies using real-world samples from programs in the field. The findings from programs that had only been implemented in a university setting were discounted by a fixed percentage. The reviewers also discounted findings from evaluations where the researcher was also the program’s developer to account for potential conflicts of interest and the difficulties of replicating the efforts of exceptionally motivated program originators.
The review identified economically attractive, evidence-based programs in the areas of adult and juvenile corrections and prevention.80 “We find that if Washington can successfully implement a moderate-to-aggressive portfolio of evidence-based options, then a significant level of future prison construction can be avoided, state and local taxpayers can save about two billion dollars, and net crime rates can be lowered slightly,” the reviewers concluded.
These estimates were cautious, they said, to reflect the difficulty often encountered when taking programs to larger scale, and the reviewers suggested the Legislature consider establishing an ongoing oversight process to ensure quality control. They noted that a previous review of a juvenile justice program found it produced outstanding returns on investment when delivered as designed, but failed to reduce crime when not implemented competently.81 “Ensuring competent delivery of programs while maintaining fidelity to the program model appears to be essential,” they said. “Thus safeguarding the state’s investment in evidence-based programs requires ongoing efforts to assess program delivery and, when necessary, taking the required steps to make corrective changes.”
It is too early to judge the review’s impact, but the state’s criminal justice policies already have had a marked effect, with incarceration rates 56 percent below the national rate. The review notes that the rate began to diverge from national trends in the early 1980s after the state enacted sentencing reform laws.
The Institute’s director, Roxanne Lieb, says policymakers have become far more aware of the value of systematic reviews in recent years, partly due to the efforts of the Cochrane and Campbell Collaborations. “Our state’s policymakers are not directly aware of these groups’ efforts; however, this work has influenced researchers and others who advise legislators and has contributed to a climate where meta-analysis is valued,” she says. Policymakers particularly value the use of systematic reviews in cost-benefit analysis. “It’s one thing to report on effect sizes,” says Lieb. “It’s another thing for policymakers to understand what the consequences of that effect size might be in terms of the state budget.”
Another advantage is that systematic reviews tend to identify a number of effective options, she says. “Local governments can pick which of the effective programs to implement so you get a bit of that local control that people value.”
Lieb has learned to be pragmatic during her sixteen years with the Institute. Often their findings meet resistance from interest groups, are contrary to the “current wisdom” of the time, do not answer the questions confronting policymakers, or are not politically welcome or timely. “While the influence of systematic reviews is growing, I would say a majority of decisions on funding that could be potentially influenced by such reviews are still based on many other factors,” she says.
The Institute has, for example, done extensive analysis of programs preventing youth problems. “But our influence on funding and programmatic decisions is still relatively modest,” Lieb says. “So although in general people are becoming more comfortable with the notion of reviews and acknowledge their value, decision making around funding is often driven by continuation of the status quo, the interests of constituents, and lobbying by service providers. This is not a job that you do with a high expectation of seeing immediate applications. One must have patience. Often, the effects of our findings are relatively invisible as they change the discussion rather than directly influencing decisions that can be singled out.”
Lieb stresses that the Institute does not engage in advocacy or lobbying: “We are very softly spoken in terms of presenting information. We work for the Legislature. If they are interested in our results, they will take advantage of them.”
If systematic reviews are to gain more influence in policy circles, Lieb says it is critical to improve the consistent quality and to achieve transparency. “Two sets of researchers can produce a meta-analysis on the same topic and come up with different results depending on their selection criteria and measurement decisions,” she says. “And while reviews by the Cochrane and Campbell Collaborations are seen as highly credible, they often take a long time to produce because they are conducted by volunteers.” Lieb adds that if the selection criteria for systematic reviews prioritized issues involving significant government investment, their influence on policy would likely be increased.
Lieb would also like to see undergraduate training improved for the wide range of disciplines involved in producing and using systematic reviews, including psychologists and social workers: “It would be helpful if social science graduates were universally knowledgeable about the term effect size and knew how to recognize a good meta-analysis.” Lieb also believes the policy process would benefit if charities and other advocacy groups became more familiar with the benefits of systematic reviews. The experience of Washington State suggests that many benefits can accrue from forging closer connections between policy and science.
- Institutionalizing an evidence-based approach in the policy process has the potential to pay significant dividends.
- Systematic reviews are being used to help answer important and complex policy questions in diverse fields, including education, health, and criminal justice.
- Assessing program delivery and making any needed changes are vital when implementing programs.
- There is a need to improve the quality of systematic reviews and the expertise of those who prepare and use them.
- Even when reliable scientific evidence is available, it may not be reflected immediately in policy decisions.
- Systematic reviews would more likely be used if policymakers contributed to selecting the review topics most important to them.
- Providing economic as well as effectiveness data can enhance the relevance and utility of reviews.
Conclusion and Recommendations of the Authors of This Report
The case studies in this report highlight some of the great gains that have occurred in population and public health, as well as some of the frustrations and disappointments. They make it clear that there is a long way to go to maximize the effective use of systematic reviews in population health. These case studies show the multiplicity of forces affecting population health as well as the complexity of issues facing researchers and policymakers within the many sectors that influence population health. They also suggest there is significant potential to enhance the contribution of research evidence to policymaking, while acknowledging that it is only one of many factors influencing policy.
The case studies also reveal the importance of taking a systems or environmental approach to behavior change rather than individual-focused strategies; the role of social norms in influencing behavior; the power of legislative, regulatory, and financial incentives; and the value of a reliable and relevant evidence base to help set political and public agendas and to shape interventions. Improvements to a population’s health are rarely achieved quickly or easily: perseverance and pragmatism are often required.
These themes are also instructive for those seeking to boost the role of evidence in population health decision making. Improving the use and usefulness of systematic reviews will require individual researchers, policymakers, advocates, and other relevant groups to modify the way they work. Success in this endeavor is most likely to be achieved if systemic, environmental, and cultural changes promote and support them.
Those who seek to improve the use and usefulness of systematic reviews must recognize that policymaking is not a linear process and that a comprehensive array of interventions is more likely to have an impact than any single intervention. They must also be systematic by making deliberate, strategic efforts to disseminate and implement review findings as well as ready and willing to seize the moment when opportunities arise. Timeliness is often crucial in the policy environment.
At the same time, it is helpful to take a long-term perspective, acknowledging that significant change is rarely achieved in the typical bureaucratic or political time frame of two to three years, advises Trevor Hancock, public health consultant at the British Columbia Ministry of Health. “It’s a useful lesson generally for people trying to promote the use of evidence,” he says. “If you take tobacco as an example, look back on how long it’s taken us to make progress on something as clear-cut as that. That’s the bad news. The good news is if you look back twenty or thirty years, we’ve come a huge way. So huge changes are possible but they take time.”
Recommendations to Improve the Use and Usefulness of Systematic Reviews
The following recommendations are categorized in general groupings. This does not mean to imply that the research community, for example, alone bears responsibility for implementing recommendations around research or that only policymakers can effect change in the policymaking domain. The case studies in this report highlight the value of genuine, cross-sectoral collaboration. Fostering collaboration among researchers, policymakers, and other stakeholders will also be important for efforts to improve the use and usefulness of systematic reviews. Prominent champions have also been vital in many public health victories. Identifying, mentoring, and supporting the champions of evidence-informed policy may help achieve the following goals.
Improving Systematic Reviews
There is significant room to improve the quality, timeliness, accessibility, and relevance of systematic reviews, so that they better answer the key questions of policymakers: can it work, how well will it work, and is it worth it? Strategies include:
- Ensuring undergraduate and postgraduate curricula in a wide range of disciplines—including public health, psychology, social sciences, political sciences, social work, and journalism—give a solid grounding in evidence-related concepts, including systematic reviews.
- Providing continuing education and mid-career training in evidence-related concepts.
- Providing funding and infrastructure to better support the production, updating, and local adaptation of systematic reviews.
- Involving policymakers and other users of systematic reviews in their topic selection, planning, and conduct. This may help make the reviews more relevant to users’ needs, as well as develop networks to aid dissemination and implementation. Higher quality reviews are more likely to result when people with a broad range of experience, in both topic and methodology, are involved.82
- Improving the quality and reporting of primary research upon which systematic reviews are based. Participatory research has the potential to increase the relevance of research findings and their subsequent use.
- Ensuring better reporting of costing issues, including separation of research from implementation costs, start-up costs from ongoing operational costs, and actual expenditures from opportunity costs.83
- Identifying the harms as well as benefits of interventions both to highlight uncertainty associated with estimates and to describe differential effects by subgroups.
- Determining whether findings of ineffectiveness within primary studies are simply due to failure of implementation or to failure of intervention concept or theory.
- Specifying, where possible, the critical aspects of an intervention that cannot be modified, as well as those that can be modified to suit local conditions without jeopardizing the intervention’s effectiveness.
- Ensuring maximum transparency of the review process and full disclosure of the interests of those involved.
- Focusing more review topics towards identifying gaps and limitations in primary studies.
- Better reporting of information on external validity.
- Better reporting of information on factors that seem to predict the success of interventions being reviewed.
Improving the Translation of Systematic Reviews
Systematic reviews themselves are often of little use for policymakers already overwhelmed by information. They need concise, meaningful summaries of relevant findings from relevant systematic reviews and their implications. It is unrealistic to expect that researchers will have the skills required for this work. The task may require information brokers or translational scientists—specialist writers or communicators who can speak the language of both policymakers and researchers, and negotiate these different worlds. Strategies include:
- Establishing infrastructure and systems, whether in research organizations, interest groups, or government agencies, to support effective translation of systematic review findings.
- Explaining, where possible, the implications of findings for the local context. Policymakers usually want to know what an intervention will mean for their constituency, or their budget.
- Ensuring an engaging and accessible format and providing contact details for content experts.84
- Using anecdotes and storytelling as powerful methods of transmitting evidence-based information; more research is needed on how to combine empirical data with storytelling.
- Engaging a variety of formats, including videos and multimedia productions.
- Translating systematic review findings into manuals and other formats to guide those designing and implementing interventions.
Improving the Dissemination of Systematic Reviews
Strategies to help ensure relevant systematic reviews reach potential users include:
- Implementing systems to identify systematic reviews with the potential to improve population health and to ensure their findings reach relevant audiences, both within and outside the traditional health domain.
- Ensuring active dissemination strategies that move beyond the traditional academic approach of publication in journals, and exploiting the power of the media, Internet, and other communication channels. Information brokers could, for example, provide policymakers with timely access to content experts, whether by phone, video conference, or in person.
- Providing in public/population health education and training the skills needed to use and adapt evidence-based approaches, and the development of communication skills, including media advocacy. This is especially important in many public health agencies where workforce turnover is high.
- Developing networks, relationships, and collaborations among researchers, policymakers, knowledge brokers, interest groups, media, and other stakeholders. These networks can be formal or informal and can include such mechanisms as conferences, task forces, e-networks, or briefings. Note that personal contact can have more impact than an impersonal briefing document.
- Developing infrastructures, programs, and processes to facilitate exchange between the policy and research communities. The American Association for the Advancement of Science, for example, has a program encouraging scientists to enter the policymaking arena.85
- Including elected officials in dissemination efforts.
- Developing incentives to encourage collaboration and dissemination. Performance reviews for academics could ask for evidence of their work’s societal impact or their role in assisting policy development. Funding agencies could also reward translation and dissemination efforts.
- Making existing databases of systematic reviews, such as the Cochrane Collaboration, more policy user-friendly. Gaps in current databases could be filled; for example, some researchers have proposed that a global registry of anticipated public health studies be established.86
- Evaluating systematic review dissemination strategies and ensuring these findings are themselves disseminated and implemented.
Policymaking environments create many barriers to the better integration of evidence. Strategies to overcome some of these barriers include:
- Implementing incentives for policymaking groups to incorporate systematic reviews into their decision making processes and to commission systematic reviews where they are needed. Such incentives could include:
- Policy briefs or reports and other publications could be required to describe relevant systematic reviews.
- Research grant applications could be required to detail where the proposed study fits into knowledge gaps identified by systematic reviews.
- Program funding applications could be required to state where the proposed program fits with evidence from systematic reviews.
- Performance review criteria for policymakers and advocates could incorporate assessment of the implementation of evidence-informed practices.
- Implementing feedback mechanisms to track the quality of the evidence base being used to develop policies, programs, or advocacy initiatives.
- Implementing incentives to encourage training in evidence-related concepts for policymakers and those who influence policy, including legislators, the legal professions, advocacy groups, and the media.
- Fostering and supporting champions and leaders in evidence-informed policy, at both the individual and organizational level.
- Supporting policy-related research, including studies investigating the determinants of policy and assessing policy outcomes.
- Supporting existing programs and initiatives to strengthen the role of evidence in population health. For example, Health Impact Assessment encourages intersectoral analysis of population health issues,87 and the Robert Wood Johnson Foundation Health & Society Scholars program aims to build the nation’s capacity for research, leadership, and action to address the broad range of factors affecting the health of populations.88
- Ensuring evaluation is built-in—those seeking to improve the use and usefulness of systematic reviews should ensure their work is evaluated and the findings disseminated, where appropriate and possible.
Engaging with the Media
The media has a powerful impact on public health by helping to set the political, policy, and public agendas and disseminating information. Strategies to promote evidence-informed reporting include:
- Incorporating evidence-related concepts into journalism education and training.
- Targeting relevant professional groups, such as organizations for health care writers or education writers to raise their awareness of the use of evidence in decision making.
- Encouraging active engagement with the media by both organizations and individuals. This could involve both traditional media strategies as well as approaches such as that of the Center for the Advancement of Health, which has established an electronic news service to promote the translation of health research into action.
1Moher et al. 2007.
2cited in Wallack 1994.
3Waters et al. 2006.
4Kohatsu, Robinson, and Torner 2004.
5Lavis et al. 2004.
8Lavis et al. 2005.
9www.york.ac.uk/inst/crd/report4.htm (accessed July 9, 2007).
10www.cochrane.org (accessed July 9, 2007).
11www.campbellcollaboration.org (accessed July 9, 2007).
12www.thecommunityguide.org (accessed July 9, 2007).
13Green and Kreuter 2000.
14Brownson, Kreuter et al. 2006.
15Scollo et al. 2003.
16Popay, Rogers, and Williams 1998.
19Anderson et al. 2005.
20Choi, Pang et al. 2005.
21Flay et al. 2004.
23For more on the hierarchy of evidence see Moynihan 2004.
25Whitehead et al. 2004.
26Bowen and Zwi 2005.
27Innvær et al. 2002.
28Choi, Pang et al. 2005.
29Lavis et al. 2004.
30Petticrew et al. 2004.
31Waters et al. 2006; Jackson and Waters, for the Guidelines . . . 2005.
32Green and Glasgow 2006.
33Dzewaltowski et al. 2004.
34Anderson et al. 2005.
35United States General Accounting Office 1999.
36Shults et al. 2001.
38Liu et al. 2003.
39National Highway Traffic Safety Administration 2006a.
40Dinh-Zarr et al. 2001.
41Morrison, Petticrew, and Thomson 2003.
42National Highway Traffic Safety Administration 2006a.
43Shults et al. 2001.
44Elder et al. 2002.
45National Highway Traffic Safety Administration 2006b.
46Cited in Nelson, Weitzman, and Wechsler 2005; Task Force of the National Advisory Council on Alcohol Abuse and Alcoholism 2002.
47Task Force of the National Advisory Council on Alcohol Abuse and Alcoholism 2002.
48A Matter of Degree program websites: www.ama-assn.org/ama/pub/category/3558.html (accessed July 9, 2007); www.hsph.harvard.edu/amod/resources.html (accessed July 9, 2007); www.rwjf.org (accessed July 9, 2007); www.alcoholpolicymd.com (accessed July 9, 2007).
49Wechsler et al. 2002.
50NU Directions Campus-Community Coalition 2004.
51Weitzman et al. 2004.
52Weitzman and Nelson 2004.
53Brownson and Bright 2004; Glantz and Balbach 2000.
54Glantz and Balbach 2000; Pierce et al. 1998.
55Tobacco Education and Research Oversight Committee 2003.
56Fichtenberg and Glantz 2000; 2001.
57Burrus et al. 2006.
58Scollo et al. 2003; Barnes and Bero 1998.
60Bryan-Jones and Chapman 2006.
61Burrus et al. 2006.
62Ball, Timperio, and Crawford 2006; Brownson et al. 2001; Kahn et al. 2001.
63Ryan et al. 2006.
64Dunn and Blair 2002.
65Caldwell et al. 2006.
66Ammerman et al. 2002.
67Ryan et al. 2006.
68Ammerman et al. 2002.
69Miller 2005; Silove 2005; Van Ommeren, Saxena, and Saraceno 2005.
70Norris et al. 2002; Norris, Friedman, and Watson 2002.
71Norris 2005a; 2005b; 2005c.
72Van Ommeren, Saxena, and Saraceno 2005.
73Inter-Agency Standing Committee 2007.
74Perry et al. 2000.
75Des Jarlais et al. 2006.
77Des Jarlais et al. 2006; Hallfors and Godette 2002; Komro et al. 2004; Perry et al. 2000; Perry et al. 2003; Ennett et al. 1994.
78Hallfors and Godette 2002.
80Aos, Miller, and Drake 2006.
82Jackson and Waters, for the Guidelines . . . 2005.
84Lavis et al. 2005.
85Choi, Pang et al. 2005.
86Author’s interview with Bernard Choi, senior research scientist, the Public Health Agency of Canada’s Centre for Chronic Disease Prevention and Control, 2006.
87www.ph.ucla.edu/hs/health-impact (accessed July 12, 2007).
88www.healthandsocietyscholars.org (accessed July, 12, 2007).
Ammerman, A., C. Cooke, C. Dunn, J. Longenecker, B. Matthews, E. Ngui, C. Samuel-Hodge, R. Schwartz, and D. Ward. 2002. Moving Our Children toward a Healthy Weight: Finding the Will and the Way. Edited by D. Caldwell and J. Lebeuf. Raleigh, NC: North Carolina DHHS, Division of Public Health. Available at http://www.eatsmartmovemorenc.com/stateplan/docs/healthy_weight_initiative .pdf (accessed June 27, 2007).
Anderson, L.M., R.C. Brownson, M.T. Fullilove, S.M. Teutsch, L.F. Novkic, J. Fielding, and G.H. Land. 2005. Evidence-Based Public Health Policy and Practice: Promises and Limits. American Journal of Preventive Medicine 28(5, suppl. 1):22630. doi:10.1016/j.amepre.2005.02.014. Available at http://dx.doi.org/10.1016/j.amepre.2005.02.014 (accessed June 28, 2007).
Aos, S., M. Miller, and E. Drake. 2006. Evidence-Based Public Policy Options to Reduce Future Prison Construction, Criminal Justice Costs, and Crime Rates. Olympia: Washington State Institute for Public Policy. Available at http://www.wsipp.wa.gov/rptfiles/06-10-1201.pdf (accessed June 27, 2007).
Ball, K., A.F. Timperio, and D.A. Crawford. 2006. Understanding Environmental Influences on Nutrition and Physical Activity Behaviors: Where Should We Look and What Should We Count? International Journal of Behavioral Nutrition and Physical Activity 3:33. doi:10.1186/1479-5868-3-33. Available at http://dx.doi.org/10.1186/1479-5868-3-33 (accessed June 28, 2007).
Barnes, D.E., and L.A. Bero. 1998. Why Review Articles on the Health Effects of Passive Smoking Reach Different Conclusions. JAMA 279(19):1566–70. doi:10.1001/jama.279.19.1566. Available at http://dx.doi.org/10.1001/jama.279.19.1566 (accessed June 28, 2007).
Bowen, S., and A.B. Zwi. 2005. Pathways to “Evidence-Informed” Policy and Practice: A Framework for Action. PLoS Medicine 2(7):e166. doi:10.1371/journal.pmed.0020166. Available at http://dx.doi .org/10.1371/journal.pmed.0020166 (accessed June 28, 2007).
Brownson, R.C., E.A. Baker, R.A. Housemann, L.K. Brennan, and S.J. Bacak. 2001. Environmental and Policy Determinants of Physical Activity in the United States. American Journal of Public Health 91(12):1995–2003. Available at http://www.ajph.org/cgi/content/full/91/12/1995 (accessed June 28, 2007).
Brownson, R.C., and F.S. Bright. 2004. Chronic Disease Control in Public Health Practice: Looking Back and Moving Forward (Viewpoint). Public Health Reports 119:230–38. doi:10.1016/j.phr.2004.04.001. Available at http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=1497626 (accessed June 28, 2007).
Brownson, R.C., M.W. Kreuter, B.A. Arrington, and W.R. True. 2006. From the Schools of Public Health—Translating Scientific Discoveries into Public Health Action: How Can Schools of Public Health Move Us Forward? Public Health Reports 121(1):97121.
Bryan-Jones, K., and S. Chapman. 2006. Political Dynamics Promoting the Incremental Regulation of Secondhand Smoke: A Case Study of New South Wales, Australia. BMC Public Health 6:192. doi:10.1186/1471-2458-6-192. Available at http://dx.doi.org/10.1186/1471-2458-6-192 (accessed June 28, 2007).
Burrus, B., E. Dean, L. Flicker, D. Aiken, T. Heinrich, and D. Weizenkamp. 2006. Surveillance Survey for the Guide to Community Services. Research Triangle Park, NC: RTI International.
Caldwell, D., C. Dunn, A. Hardison, A. Hardison, A. Keene, K. Kolasa, B. Lafontant, et al. 2006. Eat Smart, Move More: North Carolina’s Plan to Prevent Overweight, Obesity and Related Chronic Diseases—2007–2012. Raleigh, NC: Eat Smart, Move More . . . North Carolina. Available at www.eatsmartmovemorenc.com/stateplan/docs/esmm_state_plan_prof_printer_ver.pdf (accessed June 28, 2007).
Chapman, S. 2005. The Most Important and Influential Papers in Tobacco Control: Results of an Online Poll. Tobacco Control 14:e1. doi:10.1136/tc.2005.013177. Available at http://dx.doi.org/10.1136/tc.2005.013177 (accessed June 28, 2007).
Choi, B.C.K. 2005. Twelve Essentials of Science-Based Policy. Preventing Chronic Disease: Public Health Research, Practice, and Policy. 2(4):111. Available at http://www.cdc.gov/pcd/issues/2005/oct/05_0005.htm (accessed June 28, 2007).
Choi, B.C.K., T. Pang, V. Lin, P. Puska, G. Sherman, M. Goddard, M.J. Ackland, et al. 2005. Can Scientists and Policy Makers Work Together? Journal of Epidemiology and Community Health 59:632–37. doi:10.1136/jech.2004.031765. Available at http://dx.doi.org/10.1136/jech.2004.031765 (accessed June 28, 2007).
Davies, P. 2004. “Is Evidence-Based Government Possible? Jerry Lee Lecture 2004.” Paper presented at the 4th Annual Campbell Collaboration Colloquium, Washington DC, February 19. Available at http://www.policyhub.gov.uk/downloads/JerryLeeLecture1202041.pdf (accessed June 28, 2007).
Des Jarlais, D.C., Z. Sloboda, S.R. Friedman, B. Tempalski, C. McKnight, and N. Braine. 2006. Diffusion of the D.A.R.E. and Syringe Exchange Programs. American Journal of Public Health 96(8):135458. doi:10.2105/AJPH.2004.060152. Available at http://dx.doi.org/10.2105/AJPH.2004.060152 (accessed June 28, 2007).
Dinh-Zarr, T.B., D.A. Sleet, R.A. Shults, S. Zaza, R.W. Elder, J.L. Nichols, R.S. Thompson, D.M. Sosin, and the Task Force on Community Preventive Services. 2001. Reviews of Evidence Regarding Interventions to Increase the Use of Safety Belts. American Journal of Preventive Medicine 21(4, suppl.1):4865. doi:10.1016/S0749-3797(01)00378-6. Available at http://dx.doi.org/10.1016/S0749-3797(01)00378-6 (accessed June 28, 2007).
Dunn, A.L., and S.N. Blair. 2002. Translating Evidenced-Based Physical Activity Interventions into Practice—The 2010 Challenge. American Journal of Preventive Medicine 22(4, suppl. 1):89. doi:10.1016/S0749-3797(02)00432-4. Available at http://dx.doi.org/10.1016/S0749-3797(02)00432-4 (accessed June 28, 2007).
Dzewaltowski, D.A., P.A. Estabrooks, L.M. Klesges, S. Bull, and R.E. Glasgow. 2004. Behavior Change Intervention Research in Community Settings: How Generalizable Are the Results? Health Promotion International 19(2):23545. doi:10.1093/heapro/dah211. Available at http://dx.doi.org/10.1093/heapro/dah211 (accessed June 28, 2007).
Elder, R.W., R.A. Shults, D.A. Sleet, J.L. Nichols, S. Zaza, and R.S. Thompson. 2002. Effectiveness of Sobriety Checkpoints for Reducing Alcohol-Involved Crashes. Traffic Injury Prevention 3(4):26674. doi:10.1080/15389580214623. Available at http://dx.doi.org/10.1080/15389580214623 (accessed June 28, 2007).
Ennett, S.T., N.S. Tobler, C.L. Ringwalt, and R.L. Flewelling. 1994. How Effective Is Drug Abuse Resistance Education? A Meta-Analysis of Project DARE Outcome Evaluations. American Journal of Public Health 84(9):13941401. Available at http://www.ajph.org/cgi/content/abstract/84/9/1394 (accessed July 27, 2007).
Fichtenberg, C.M., and S.A. Glantz. 2000. Association of the California Tobacco Control Program with Declines in Cigarette Consumption and Mortality from Heart Disease. The New England Journal of Medicine 343(24):177277. doi:10.1056/NEJM200012143432406. Available at http://dx.doi.org/10.1056/NEJM200012143432406 (accessed July 5, 2007).
Fichtenberg, C.M., and S.A. Glantz. 2001. Controlling Tobacco Use: The Authors Reply (Correspondence). The New England Journal of Medicine 344(23):179799. Extract available at http://content.nejm.org/cgi/content/short/344/23/1797?query=prevarrow (accessed July 5, 2007).
Flay, B.R., A. Biglan, R.F. Boruch, F. González Castro, D. Gottfredson, S.Kellam, E.K. Moscicki, S. Schinke, J.Valentine, and P. Ji. 2004. Standards of Evidence: Criteria for Efficacy, Effectiveness and Dissemination. Falls Church, VA: Society for Prevention Research. Available at http://www.preventionresearch.org/StandardsofEvidencebook.pdf (accessed July 2, 2007).
Glantz, S.A., and E.D. Balbach. 2000. Tobacco War: Inside the California Battles. Berkeley: University of California Press. Available at http://ark.cdlib.org/ark:/13030/ft167nb0vq/ (accessed June 28, 2007).
Green, L.W., and R.E. Glasgow. 2006. Evaluating the Relevance, Generalization, and Applicability of Research: Issues in External Validation and Translation Methodology. Evaluation and the Health Professions 29(1):12653. doi:10.1177/0163278705284445. Available at http://dx.doi.org/10.1177/0163278705284445 (accessed June 28, 2007).
Green, L.W., and M.W. Kreuter. 2000. Commentary on the Emerging Guide to Community Preventive Services from a Health Promotion Perspective. American Journal of Preventive Medicine 18(1 suppl. 1):79. doi:10.1016/S0749-3797(99)00131-2. Available at http://dx.doi.org/10.1016/S0749-3797(99)00131-2 (accessed June 28, 2007).
Hallfors, D., and D. Godette. 2002. Will the “Principles of Effectiveness” Improve Prevention Practice? Early Findings from a Diffusion Study. Health Education Research 17(4):46170. doi:10.1093/her/17.4.461. Available at http://dx.doi.org/10.1093/her/17.4.461 (accessed June 28, 2007).
Innvær, S., G. Vist, M. Trommald, and A. Oxman. 2002. Health Policy-Makers’ Perceptions of Their Use of Evidence: A Systematic Review. Journal of Health Services Research and Policy 7(4):23944. Available at http://www.ingentaconnect.com/content/rsm/jhsrp/2002/00000007/00000004/art00008 (accessed September 20, 2007).
Inter-Agency Standing Committee (IASC). 2007. IASC Guidelines on Mental Health and Psychosocial Support in Emergency Settings. Geneva: IASC. Available at http://www.humanitarianinfo.org/iasc/content/documents/subsidi/tf_mhps/Guidelines%20IASC%20Mental%20Health%20Psychosocial.pdf (accessed June 28, 2007).
Jackson, N., and E. Waters, for the Guidelines for Systematic Reviews in Health Promotion and Public Health Taskforce. 2005. Criteria for the Systematic Review of Health Promotion and Public Health Interventions. Health Promotion International 20(4):36774. doi:10.1093/her/17.4.461. Available at http://dx.doi.org/10.1093/heapro/dai022 (accessed June 29, 2007).
Kahn, E.B., G.W. Heath, K.E. Powell, E.J. Stone, and R.C. Brownson. 2001. Increasing Physical Activity: A Report on Recommendations of the Task Force on Community Preventive Services. Morbidity and Mortality Weekly Report 50(RR18):116. Available at http://www.cdc.gov/mmwr/preview/mmwrhtml/rr5018a1.htm (accessed June 29, 2007).
Kohatsu, N.D., J.G. Robinson, and J.C. Torner. 2004. Evidence-Based Public Health: An Evolving Concept. American Journal of Preventive Medicine 27(5):41721. doi:10.1016/S0749-3797(04)00196-5. Available at http://dx.doi.org/10.1016/S0749-3797(04)00196-5 (accessed June 29, 2007).
Komro, K.A., C.L. Perry, S. Veblen-Mortenson, M.H. Stigler, L.M. Bosma, K.A. Munson, and K. Farbakhsh. 2004. Violence-Related Outcomes of the D.A.R.E Plus Project. Health Education and Behavior 31(3):33554. doi:10.1177/1090198104263337. Available at http://dx.doi.org/10.1177/1090198104263337 (accessed June 29, 2007).
Lavis, J., H. Davies, A. Oxman, J-L Denis, K. Golden-Biddle, and E. Ferlie. 2005. Towards Systematic Reviews That Inform Health Care Management and Policy-Making. Journal of Health Services Research and Policy 10(3, suppl.1):3548. doi:10.1258/1355819054308549. Available at http://dx.doi.org/10.1258/1355819054308549 (accessed June 29, 2007).
Lavis, J.N., F.B. Posada, A. Haines, and E. Osei. 2004. Use of Research to Inform Public Policymaking. The Lancet 364:161521. doi:10.1016/S0140-6736(04)17317-0. Available at http://dx.doi.org/10.1016/S0140-6736(04)17317-0 (accessed June 29, 2007).
Liu B., R. Ivers, R. Norton, S. Blows, and S.K. Lo. 2003. Helmets for Preventing Injury in Motorcycle Riders. Cochrane Database of Systematic Reviews, Issue 4. Art. No.: CD004333. doi:10.1002/14651858.CD004333.pub2. Available at http://dx.doi.org/10.1002/14651858.CD004333.pub2 (accessed June 29, 2007).
Miller, G. 2005. The Tsunami’s Psychological Aftermath. Science 309:10301033. doi:10.1126/science.309.5737.1030. Available at http://dx.doi.org/10.1126/science.309.5737.1030 (accessed June 29, 2007).
Miller, M. 2006. Intensive Family Preservation Services: Program Fidelity Influences Effectiveness—Revised. Olympia: Washington State Institute for Public Policy. Available at http://www.wsipp.wa.gov/pub.asp?docid=06-02-3901 (accessed June 29, 2007).
Moher, D., J. Tetzlaff, A. Tricco , M. Sampson, and D.G. Altman. 2007. Epidemiology and Reporting Characteristics of Systematic Reviews. PLoS Medicine 4(3):e78. doi:10.1371/journal.pmed.0040078. Available at http://dx.doi.org/10.1371/journal.pmed.0040078 (accessed September 6, 2007).
Morrison, D.S., M. Petticrew, and H. Thomson. 2003. What Are the Most Effective Ways of Improving Population Health through Transport Interventions? Evidence from Systematic Reviews. Journal of Epidemiology and Community Health 57:32733. doi:10.1136/jech.57.5.327. Available at http://dx.doi.org/10.1136/jech.57.5.327 (accessed June 29, 2007).
Moynihan, R. 2004. Evaluating Health Services: A Reporter Covers the Science of Research Synthesis. NY: Milbank Memorial Fund. Available at http://www.milbank.org/reports/2004Moynihan/040330Moynihan.html (accessed July 9, 2007).
National Highway Traffic Safety Administration. 2006a. “Rise in Motorcycle and Pedestrian Deaths Led to Increase in Overall Highway Fatality Rate in 2005.” Press release issued August 22. Available at http://www.nhtsa.dot.gov/portal/site/nhtsa/template.MAXIMIZE/menuitem.f2217bee37fb302f6d7c121046108a0c/?javax.portlet.tpst=1e51531b2220b0f8ea14201046108a0c_ws _MX&javax.portlet.prp_1e51531b2220b0f8ea14201046108a0c_viewID=detail_view&javax.portlet .begCacheTok=com.vignette.cachetoken&javax.portlet.endCacheTok=com.vignette .cachetoken&itemID=cfd8aeeb8212d010VgnVCM1000002c567798RCRD&viewType= standard&pressReleaseYearSelect=2006 (accessed June 29, 2007).
National Highway Traffic Safety Administration. 2006b. Countermeasures That Work: A Highway Safety Countermeasure Guide for State Highway Safety Offices. Washington DC: U.S. Department of Transportation. Available at http://www.nhtsa.dot.gov/people/injury/airbags/Countermeasures/images/Countermeasures.pdf (accessed June 29, 2007).
Nelson, T.F., E.R. Weitzman, and H. Wechsler. 2005. The Effect of a Campus-Community Environmental Alcohol Prevention Initiative on Student Drinking and Driving: Results from the “Matter of Degree” Program Evaluation. Traffic Injury Prevention 6:(4)323–30. doi:10.1080/15389580500253778. Available at http://dx.doi.org/10.1080/15389580500253778 (accessed June 29, 2007).
Norris, F.H. 2005a. Range, Magnitude, and Duration of the Effects of Disasters on Mental Health: Review Update 2005. White River Junction, VT: Research Education in Disaster Mental Health. Available at http://www.redmh.org/research/general/REDMH_effects.pdf (accessed July 30, 2007).
Norris, F.H. 2005b. Psychosocial Consequences of Major Hurricanes and Floods: Range, Duration, and Magnitude of Effects and Risk Factors for Adverse Outcomes. White River Junction, VT: National Center for Posttraumatic Stress Disorder. Available at http://www.ncptsd.va.gov/ncmain/ncdocs/fact_shts/fs_range_hurricane.html?opm=1&rr=rr141&srt=d&echorr=true (accessed July 30, 2007).
Norris, F.H. 2005c. Psychosocial Consequences of Natural Disasters in Developing Countries: What Does Past Research Tell Us about the Potential Effects of the 2004 Tsunami? White River Junction, VT: Research Education in Disaster Mental Health. Available at http://www.redmh.org/research/specialized/tsunami.html (accessed July 30, 2007).
Norris, F.H., M.J. Friedman, and P.J. Watson. 2002. 60,000 Disaster Victims Speak: Part II. Summary and Implications of the Disaster Mental Health Research. Psychiatry: Interpersonal and Biological Processes 65(3):24060. doi:10.1521/psyc.184.108.40.20669. Available at http://dx.doi.org/10.1521/psyc.220.127.116.1169 (accessed June 29, 2007).
Norris, F.H., M.J. Friedman, P.J. Watson, C.M. Byrne, E. Diaz, and K. Kaniasty. 2002. 60,000 Disaster Victims Speak: Part I. An Empirical Review of the Empirical Literature, 19812001. Psychiatry: Interpersonal and Biological Processes 65(3):20739. doi:10.1521/psyc.18.104.22.16873. Available at http://dx.doi.org/10.1521/psyc.22.214.171.12473 (accessed June 29, 2007).
NU Directions Campus-Community Coalition. 2004. Five Year Report 19982003. Lincoln: University of Nebraska. Available at http://www.nudirections.org/5yrreport2003.pdf (accessed June 29, 2007).
Perry, C.L., K.A. Komro, S. Veblen-Mortenson, L.M. Bosma, K. Farbakhsh, K.A. Munson, M.H. Stigler, L.A. Lytle. 2003. A Randomized Controlled Trial of the Middle and Junior High School D.A.R.E. and D.A.R.E. Plus Programs. Archives of Pediatrics and Adolescent Medicine 157(2):17884. Available at http://archpedi.ama-assn.org/cgi/reprint/157/2/178 (accessed July 2, 2007).
Perry, C.L., K.A. Komro, S. Veblen-Mortenson, L. Bosma, K. Munson, M. Stigler, L.A. Lytle, J.L. Forster, and S.L. Welles. 2000. The Minnesota DARE PLUS Project: Creating Community Partnerships to Prevent Drug Use and Violence. Journal of School Health 70(3):8488. Abstract available at http://www.ncbi.nlm.nih.gov/sites/entrez?cmd=Retrieve&db=PubMed&list_uids= 10763475&dopt=Abstract (accessed July 2, 2007).
Petticrew, M. 2001. Systematic Reviews from Astronomy to Zoology: Myths and Misconceptions. BMJ 322:98101. doi:10.1136/bmj.322.7278.98. Available at http://dx.doi.org/10.1136/bmj.322.7278.98 (accessed July 2, 2007).
Petticrew, M., M. Whitehead, S.J. Macintyre, H. Graham, and M. Egan. 2004. Evidence for Public Health Policy on Inequalities1: The Reality According to Policymakers. Journal of Epidemiology and Community Health 58:81116. doi:10.1136/jech.2003.015289. Available at http://dx.doi.org/10.1136/jech.2003.015289 (accessed July 2, 2007).
Pierce, J.P., E.A. Gilpin, S.L. Emery, M.M. White, B. Rosbrook, and C.C. Berry. 1998. Has the California Tobacco Control Program Reduced Smoking? JAMA 280(10):89399. doi:10.1001/jama.280.10.893. Available at http://dx.doi.org/10.1001/jama.280.10.893 (accessed July 2, 2007).
Popay, J., A. Rogers, and G. Williams. 1998. Rationale and Standards for the Systematic Review of Qualitative Literature in Health Services Research. Qualitative Health Research 8(3):34151. doi:10.1177/104973239800800305. Available at http://dx.doi.org/10.1177/104973239800800305 (accessed July 12, 2007).
RE-AIM. 2006. “Final Summary and Recommendations from Meeting of Health Journal Editors on External Validity Reporting Issues, Chapel Hill, North Carolina, April 1718, 2006.” Available at http://www.re-aim.org/Documents/Editors%20Mtg%20synposis%20-%20July%202006.pdf (accessed June 28, 2007).
Ryan, K.W., P. Card-Higginson, S.G. McCarthy, M.B. Justus, and J.W. Thompson. 2006. Arkansas Fights Fat: Translating Research into Policy to Combat Childhood and Adolescent Obesity. Health Affairs 25(4):992–1004. doi:10.1377/hlthaff.25.4.992. Available at http://dx.doi.org/10.1377/hlthaff.25.4.992 (accessed July 2, 2007).
Scollo, M., A. Lal, A. Hyland, and S. Glantz. 2003. Review of the Quality of Studies on the Economic Effects of Smoke-Free Policies on the Hospitality Industry. Tobacco Control 12:1320. doi:10.1136/tc.12.1.13. Available at http://dx.doi.org/10.1136/tc.12.1.13 (accessed July 2, 2007).
Shults, R.A., R.W. Elder, D.A. Sleet, J.L. Nichols, M.O. Alao, V.G. Carande-Kulis, S. Zaza, D.M. Sosin, R.S. Thompson, and the Task Force on Community Preventive Services. 2001. Reviews of Evidence Regarding Interventions to Reduce Alcohol-Impaired Driving. American Journal of Preventive Medicine 21(4, suppl. 1):6688. doi:10.1016/S0749-3797(01)00381-6. Available at http://dx.doi.org/10.1016/S0749-3797(01)00381-6 (accessed July 2, 2007).
Silove, D. 2005. The Best Immediate Therapy for Acute Stress Is Social. Bulletin of the World Health Organization 83(1):7576.
Task Force of the National Advisory Council on Alcohol Abuse and Alcoholism. 2002. A Call to Action: Changing the Culture of Drinking at U.S. Colleges. Bethesda, MD: National Institute on Alcohol Abuse and Alcoholism, National Institutes of Health. Available at http://www.collegedrinkingprevention.gov/media/TaskForceReport.pdf (accessed July 2, 2007).
Tobacco Education and Research Oversight Committee (TEROC). 2003. Toward a Tobacco-Free California 2003–2005: The Myth of Victory. State of California: TEROC. Available at http://www.dhs.ca.gov/tobacco/documents/pubs/TobaccoMasterPlan2003.pdf (accessed July 2, 2007).
United States General Accounting Office (GAO). 1999. Highway Safety: Effectiveness of State .08 Blood Alcohol Laws. Report to Congressional Committees: GAO/RCED-99-179. Washington, DC: GAO. Available at http://www.gao.gov/archive/1999/rc99179.pdf (accessed July 2, 2007).
Van Ommeren, M., S. Saxena, and B. Saraceno. 2005. Mental and Social Health during and after Acute Emergencies: Emerging Consensus? Bulletin of the World Health Organization 83(1):7175. Available at http://www.who.int/bulletin/volumes/83/1/71arabic.pdf (accessed July 2, 2007).
Wallack, L. 1994. Media Advocacy: A Strategy for Empowering People and Communities. Journal of Public Health Policy 15(4):42036.
Waters, E., J. Doyle, N. Jackson, F. Howes, G. Brunton, and A. Oakley. 2006. Evaluating the Effectiveness of Public Health Interventions: The Role and Activities of the Cochrane Collaboration. Journal of Epidemiology and Community Health 60:28589. doi:10.1136/jech.2003.015354. Available at http://dx.doi.org/10.1136/jech.2003.015354 (accessed July 2, 2007).
Wechsler, H., J.E. Lee, M. Kuo, M. Seibring, T.F. Nelson, and H. Lee. 2002. Trends in College Binge Drinking during a Period of Increased Prevention Efforts: Findings From 4 Harvard School of Public Health College Alcohol Study Surveys: 19932001. Journal of American College Health. 50(5):203217. Abstract available at http://www.ncbi.nlm.nih.gov/sites/entrez?Db=pubmed&Cmd=ShowDetailView&TermToSearch=11990979&ordinalpos=1&itool=EntrezSystem2.PEntrez.Pubmed.Pubmed_ResultsPanel.Pubmed_RVAbstractPlus (accessed July 12, 2007).
Weitzman, E.R., and T.F. Nelson. 2004. College Student Binge Drinking and the “Prevention Paradox”: Implications for Prevention and Harm Reduction. Journal of Drug Education 34(3):24765. doi:10.2190/W6L6-G171-M4FT-TWAP. Available at http://dx.doi.org/10.2190/W6L6-G171-M4FT-TWAP (accessed July 2, 2007).
Weitzman, E.R., T.F. Nelson, H. Lee, and H. Wechsler. 2004. Reducing Drinking and Related Harms in College: Evaluation of the “A Matter of Degree” Program. American Journal of Preventive Medicine 27(3):18796. doi:10.1016/S0749-3797(04)00140-0. Available at http://dx.doi.org/10.1016/S0749-3797(04)00140-0 (accessed July 2, 2007).
Whitehead, M., M. Petticrew, H. Graham, S.J. Macintyre, C. Bambra, and M. Egan. 2004. Evidence for Public Health Policy on Inequalities—2: Assembling the Evidence Jigsaw. Journal of Epidemiology and Community Health 58:81721. doi:10.1136/jech.2003.015297. Available at http://dx.doi.org/10.1136/jech.2003.015297 (accessed July 2, 2007).
www.ahrq.gov (accessed July 9, 2007).
www.ahrq.gov/clinic/epc (accessed July 9, 2007).
www.alcoholpolicymd.com (accessed July 9, 2007).
www.ama-assn.org/ama/pub/category/3558.html (accessed July 9, 2007).
www.apho.org.uk/apho/index.htm (accessed July 27, 2007).
www.campbellcollaboration.org (accessed July 9, 2007).
www.cdc.gov (accessed July 9, 2007).
www.chsrf.ca/home_e.php (accessed July 9, 2007).
www.cochrane.org (accessed July 9, 2007).
www.dare.org/home/default.asp (accessed July 2, 2007).
www.euro.who.int/HEN (accessed July 9, 2007).
www.euro.who.int/observatory (accessed July 9, 2007).
www.healthandsocietyscholars.org (accessed July, 12, 2007).
www.hsph.harvard.edu/amod/resources.html (accessed July 9, 2007).
www.ohsu.edu/drugeffectiveness (accessed July 9, 2007).
www.ph.ucla.edu/hs/health-impact (accessed July 12, 2007).
www.preventionresearch.org (accessed July 9, 2007).
www.prevent.org (accessed July 9, 2007).
www.rwjf.org (accessed July 9, 2007).
www.thecommunityguide.org (accessed July 9, 2007).
www.vichealth.vic.gov.au/cochrane (accessed July 9, 2007).
www.whatworks.ed.gov (accessed July 9, 2007).
www.york.ac.uk/inst/crd/report4.htm (accessed July 9, 2007).
For Further Reading
Aos, S., J. Mayfield, M. Miller, and W. Yen. 2006. Evidence-Based Treatment of Alcohol, Drug, and Mental Health Disorders: Potential Benefits, Costs, and Fiscal Impacts for Washington State. Olympia: Washington State Institute for Public Policy. Available at http://www.wsipp.wa.gov/rptfiles/06-06-3901.pdf (accessed June 27, 2007).
Aos, S., M. Miller, and E. Drake. 2006. Evidence-Based Adult Corrections Programs: What Works and What Does Not. Olympia: Washington State Institute for Public Policy. Available at http://www.wsipp.wa.gov/rptfiles/06-01-1201.pdf (accessed June 27, 2007).
Bero, L.A., and A.R. Jadad. 1997. How Consumers and Policymakers Can Use Systematic Reviews for Decision Making. Annals of Internal Medicine 127(1):3742. Available at http://www.annals.org/cgi/content/full/127/1/37 (accessed June 28, 2007).
Bero, L.A., T. Montini, K. Bryan-Jones, and C. Mangurian. 2001. Science in Regulatory Policy Making: Case Studies in the Development of Workplace Smoking Restrictions. Tobacco Control 10:32936. doi:10.1136/tc.10.4.329. Available at http://dx.doi.org/10.1136/tc.10.4.329 (accessed June 28, 2007).
Brownson, R.C., D. Haire-Joshu, and D.A. Luke. 2006. Shaping the Context of Health: A Review of Environmental and Policy Approaches in the Prevention of Chronic Diseases. Annual Review of Public Health 27:34170. doi:10.1146/annurev.publhealth.27.021405.102137. Available at http://dx.doi.org/10.1146/annurev.publhealth.27.021405.102137 (accessed June 28, 2007).
Brownson, R.C., C. Royer, R. Ewing, T.D. McBride. 2006. Researchers and Policymakers: Travelers in Parallel Universes. American Journal of Preventive Medicine 30(2):16472. doi:10.1016/j.amepre.2005.10.004. Available at http://dx.doi.org/10.1016/j.amepre.2005.10.004 (accessed June 28, 2007).
Carnevale Associates. 2006. A Longitudinal Evaluation of the New Curricula for the D.A.R.E. Middle (7th Grade) and High School (9th Grade) Programs: Take Charge of Your Life. Year Four Progress Report. Gaithersburg, MD: Carnevale Associates. Available at http://www.dare.org/home/Resources/documents/DAREMarch06ProgressReport.pdf (accessed June 27, 2007).
Choi, B.C.K., D.J. Hunter, W. Tsou, and P. Sainsbury. 2005. Diseases of Comfort: Primary Cause of Death in the 22nd Century. Journal of Epidemiology and Community Health 59:103034. doi:10.1136/jech.2005.032805. Available at http://dx.doi.org/10.1136/jech.2005.032805 (accessed June 28, 2007).
Devlin, L., and M. Plescia. 2006. The Public Health Challenge of Obesity in North Carolina. (Commentary). North Carolina Medical Journal. 67(4):27882. Available at http://www.ncmedicaljournal.com/jul-aug-06/Devlin.pdf (accessed June 28, 2007).
Dobbins, M., R. Cockerill, and J. Barnsley. 2001. Factors Affecting the Utilization of Systematic Reviews. International Journal of Technology Assessment in Health Care 17(2):20314. doi:10.1017/S0266462300105069. Available at http://dx.doi.org/10.1017/S0266462300105069 (accessed June 28, 2007).
Dobbins, M., H. Thomas, M.A. O’Brien, and M. Duggan. 2004. Use of Systematic Reviews in the Development of New Provincial Public Health Policies in Ontario. International Journal of Technology Assessment in Health Care 20(4):399404. Available at http://journals.cambridge.org/action/displayIssue?jid=THC&volumeId=20&issueId=04 (accessed June 28, 2007).
Eat Smart, Move More . . . North Carolina. 2002. North Carolina Blueprint for Changing Policies and Environments in Support of Increased Physical Activity. Raleigh, NC: Eat Smart, Move More . . . North Carolina. Available at http://www.eatsmartmovemorenc.com/stateplan/docs/esmm_blueprint_moving.pdf (accessed June 28, 2007).
Elder, J.P., C.C. Edwards, T.L. Conway, E. Kenney, C.A. Johnson, and E.D. Bennett. 1996. Independent Evaluation of the California Tobacco Education Program. Public Health Reports 111(4):35358. Available at http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=1381881 (accessed June 28, 2007).
Elder, R.W., J.L. Nichols, R.A. Shults, D.A. Sleet, L.C. Barrios, R. Compton, and the Task Force on Community Preventive Services. 2005. Effectiveness of School-Based Programs for Reducing Drinking and Driving and Riding with Drinking Drivers—A Systematic Review. American Journal of Preventive Medicine 28(5, suppl. 1):288304. doi:10.1016/j.amepre.2005.02.015. Available at http://dx.doi.org/10.1016/j.amepre.2005.02.015 (accessed June 28, 2007).
Elliott, H., and J. Popay. 2000. How Are Policy Makers Using Evidence? Models of Research Utilisation and Local NHS Policy Making. Journal of Epidemiology and Community Health 54:46168. doi:10.1136/jech.54.6.461. Available at http://dx.doi.org/10.1136/jech.54.6.461 (accessed June 28, 2007).
Enkin, M.W., S. Glouberman, P. Groff, A.R. Jadad, and A. Stern. 2006. Beyond Evidence: The Complexity of Maternity Care. Birth 33(4):26569. doi:10.1111/j.1523-536X.2006.00117.x. Available at http://dx.doi.org/10.1111/j.1523-536X.2006.00117.x (accessed June 28, 2007).
Farquhar, J.W. 1996. The Case for Dissemination Research in Health Promotion and Disease Prevention. Canadian Journal of Public Health 87(S2):S449. Abstract available at http://www.ncbi.nlm.nih.gov/sites/entrez?cmd=Retrieve&db=PubMed&list_uids=9002343&dopt=Abstract (accessed June 29, 2007).
Fichtenberg, C.M., and S.A. Glantz. 2002. Effect of Smoke-Free Workplaces on Smoking Behaviour: Systematic Review. BMJ 325:188. doi:10.1136/bmj.325.7357.188. Available at http://dx.doi.org/10.1136/bmj.325.7357.188 (accessed June 28, 2007).
Fielding, J.E., and P.A. Briss. 2006. Promoting Evidence-Based Public Health Policy: Can We Have Better Evidence and More Action? Health Affairs 25(4):96978. doi:10.1377/hlthaff.25.4.969. Available at http://dx.doi.org/10.1377/hlthaff.25.4.969 (accessed June 28, 2007).
Galea, S., A. Nandi, and D. Vlahov. 2005. The Epidemiology of Post-Traumatic Stress Disorder after Disasters. Epidemiologic Reviews 27(1):7891. doi:10.1093/epirev/mxi003. Available at http://dx.doi.org/10.1093/epirev/mxi003 (accessed June 28, 2007).
Gibson, L.E. 2006. Executive Summary: A Review of the Published Empirical Literature Regarding Early- and Later-Stage Interventions for Individuals Exposed to Traumatic Stress. White River Junction, VT: Research Education in Disaster Mental Health (REDMH) Available at http://www.redmh.org/research/general/ExecSum_intervention.pdf (accessed June 28, 2007).
Glasgow, R.E. 2003. Translating Research to Practice—Lessons Learned, Areas for Improvement, and Future Directions. Diabetes Care 26:245156. doi:10.2337/diacare.26.8.2451. Available at http://dx.doi.org/10.2337/diacare.26.8.2451 (accessed June 28, 2007).
Goodwin, A., R. Foss, J. Hedlund, J. Sohn, R. Pfefer, T.R. Neuman, K.L. Slack, and K.K. Hardy. 2005. NCHRP Report 500, Guidance for Implementation of the AASHTO Strategic Highway Safety Plan, Volume 16: A Guide for Reducing Alcohol-Related Collisions. Washington DC: Transportation Research Board. Available at http://onlinepubs.trb.org/onlinepubs/nchrp/nchrp_rpt_500v16.pdf (accessed June 28, 2007).
Gorman, D.M. 2003. Alcohol and Drug Abuse—The Best of Practices, the Worst of Practices: The Making of Science-Based Primary Prevention Programs. Psychiatric Services 54(8):108789. doi:10.1176/appi.ps.54.8.1087. Available at http://dx.doi.org/10.1176/appi.ps.54.8.1087 (accessed June 28, 2007).
Johnson, J.L., L.W. Green. C.J. Frankish, D.R. MacLean, and S. Stachenko. 1996. A Dissemination Research Agenda to Strengthen Health Promotion and Disease Prevention. Canadian Journal of Public Health 87(S2):S510. Abstract available at http://www.ncbi.nlm.nih.gov/sites/entrez?cmd=Retrieve&db=PubMed&list_uids=9002336&dopt=Abstract (accessed June 29, 2007).
Kerner, J.F. 2006. Knowledge Translation versus Knowledge Integration: A “Funder’s” Perspective. Journal of Continuing Education in the Health Professions 26(1):7280. doi:10.1002/chp.53. Available at http://dx.doi.org/10.1002/chp.53 (accessed June 29, 2007).
Lamberg, L. 2005. As Tsunami Recovery Proceeds, Experts Ponder Lessons for Future Disasters. JAMA 294(8):88990. doi:10.1001/jama.294.8.889. Available at http://dx.doi.org/10.1001/jama.294.8.889 (accessed June 29, 2007).
Lavis, J.N. 2006. Moving Forward on Both Systematic Reviews and Deliberative Processes. Healthcare Policy 1(2):5963. Available at http://www.longwoods.com/product.php?productid=17877&cat=412&page=1 (accessed June 29, 2007).
Lavis, J.N., H.T.O. Davies, R.L. Gruen, K. Walshe, C.M. Farquhar. 2006. Working within and beyond the Cochrane Collaboration to Make Systematic Reviews More Useful to Healthcare Managers and Policy Makers. Healthcare Policy 1(2):2133. Available at http://www.longwoods.com/home.php?cat=412 (accessed June 29, 2007).
Lavis, J.N., J. Lomas, M. Hamid, and N.K. Sewankambo. 2006. Assessing Country-Level Efforts to Link Research to Action. Bulletin of the World Health Organization 84(8):62028 doi:10.2471/BLT.06.030312. Available at http://www.who.int/bulletin/volumes/84/8/06-030312.pdf (accessed June 29, 2007).
Lin, V. 2004. From Public Health Research to Health Promotion Policy: On the 10 Major Contradictions. Social and Preventive Medicine 49(3):17984. doi:10.1007/s00038-004-3083-1. Available at http://dx.doi.org/10.1007/s00038-004-3083-1 (accessed June 29, 2007).
Lipsey, M.W. 2005. The Challenges of Interpreting Research for Use by Practitioners: Comments on the Latest Products from the Task Force on Community Preventive Services (Commentary). American Journal of Preventive Medicine 28(2, suppl. 1)13. doi:/10.1016/j.amepre.2004.09.026. Available at http://dx.doi.org/10.1016/j.amepre.2004.09.026 (accessed June 29, 2007).
MacLean, D.R. 1996. Positioning Dissemination in Public Health Policy. Canadian Journal of Public Health. 87(S2):S403. Abstract available at http://www.ncbi.nlm.nih.gov/sites/entrez?cmd=Retrieve&db=PubMed&list_uids=9002342&dopt=Abstract (accessed June 29, 2007).
MADD Public Policy Department. Protect, Serve and Prevent: Successful Law Enforcement Strategies to Stop Drunk Driving. MADD 2004 Law Enforcement Leadership Summit Report. Irving, TX: MADD Public Policy Department. Available at http://www.dps.state.la.us/tiger/Law%20Enforcement%20Summit,%202005.pdf (accessed June 29, 2007).
Mangurian, C.V., and L.A. Bero. 2000. Lessons Learned from the Tobacco Industry’s Efforts to Prevent the Passage of a Workplace Smoking Regulation. American Journal of Public Health 90(12):192630. Available at http://www.ajph.org/cgi/reprint/90/12/1926 (accessed June 29, 2007).
Mays, N., C. Pope, and J. Popay. 2005. Systematically Reviewing Qualitative and Quantitative Evidence to Inform Management and Policy-Making in the Health Field. Journal of Health Services Research and Policy 10(3, suppl.1):620. doi:10.1258/1355819054308576. Available at http://dx.doi.org/10.1258/1355819054308576 (accessed June 29, 2003).
Montini, T., and L.A. Bero. 2001. Policy Makers’ Perspectives on Tobacco Control Advocates’ Roles in Regulation Development. Tobacco Control 10:21824. doi:10.1136/tc.10.3.218. Available at http://dx.doi.org/10.1136/tc.10.3.218 (accessed June 29, 2007).
Morris, J., M. Van Ommeren, M. Belfer, S. Saxena, and B. Saraceno. 2007. Children and the Sphere Standard on Mental and Social Aspects of Health. Disasters 31(1)7190. doi:10.1111/ j.1467-7717.2007.00341.x. Available at http://dx.doi.org/10.1111/j.1467-7717.2007.00341.x (accessed September 17, 2007).
Mullen, P.D., and G. Ramírez. 2006. The Promise and Pitfalls of Systematic Reviews. Annual Review of Public Health 27:81102. doi:10.1146/annurev.publhealth.27.021405.102239. Available at http://dx.doi.org/10.1146/annurev.publhealth.27.021405.102239 (accessed June 29, 2007).
Nelson, T.F., T.S. Naimi, R.D. Brewer, and H. Wechsler. 2005. The State Sets the Rate: The Relationship among State-Specific College Binge Drinking, State Binge Drinking Rates, and Selected State Alcohol Control Policies. American Journal of Public Health 95(3):44146. doi:10.2105/AJPH.2004.043810. Available at http://dx.doi.org/10.2105/AJPH.2004.043810 (accessed June 29, 2007).
Quinlan, K.P., R.D. Brewer, P. Siegel, D.A. Sleet, A.H. Mokdad, R.A. Shults, and N. Flowers. 2005. Alcohol-Impaired Driving among U.S. Adults, 19932002. American Journal of Preventive Medicine 28(4):34650. doi:10.1016/j.amepre.2005.01.006. Available at http://dx.doi.org/10.1016/j.amepre.2005.01.006 (accessed July 2, 2007).
Robert Wood Johnson Foundation. 2005. Addressing Tobacco in Managed Care: The Path Ahead, September 23, 2005, Conference Report. Princeton, NJ: The Robert Wood Johnson Foundation. Available at http://www.rwjf.org/files/research/Path%20Ahead%20Final%20Report.pdf (accessed July 2, 2007).
Saltz, R., A. Biglan, L.M. Brotman, F.G. Castro, and D. Gorman-Smith. Advocacy for Prevention Science. Fairfax, VA: Society for Prevention Research. Available at http://www.preventionresearch.org/Advocacy_principles_051205RFSa2.pdf (accessed July 2, 2007).
Saltz, R.F., and W. DeJong. 2002. Reducing Alcohol Problems on Campus: A Guide to Planning and Evaluation. Bethesda, MD: National Institute on Alcohol Abuse and Alcoholism, National Institutes of Health. Available at http://www.collegedrinkingprevention.gov/media/FINALHandbook.pdf (accessed July 2, 2007).
Schotland, M.S., and L.A. Bero. 2002. Evaluating Public Commentary and Scientific Evidence Submitted in the Development of a Risk Assessment. Risk Analysis 22(1):13140. doi:10.1111/0272 -4332.t01-1-00011. Available at http://dx.doi.org/10.1111/0272-4332.t01-1-00011 (accessed July 2, 2007).
Shults, R.A., R.W. Elder, D.A. Sleet, R.S. Thompson, and J.L. Nichols. 2004. Primary Enforcement Seat Belt Laws Are Effective Even in the Face of Rising Belt Use Rates. Accident Analysis and Prevention 36(3):49193. doi:10.1016/S0001-4575(03)00038-1. Available at http://dx.doi.org/10.1016/S0001-4575(03)00038-1 (accessed July 2, 2007).
Shults, R.A., D.A. Sleet, R.W. Elder, G.W. Ryan, and M. Sehgal. 2002. Association between State Level Drinking and Driving Countermeasures and Self Reported Alcohol Impaired Driving. Injury Prevention 8:10610. doi:10.1136/ip.8.2.106. Available at http://dx.doi.org/10.1136/ip.8.2.106 (accessed July, 2, 2007).
Summerfield, D. 2005. What Exactly Is Emergency or Disaster “Mental Health”? Bulletin of the World Health Organization 83(1):76.
Task Force on Community Preventive Services. 2001. Recommendations to Reduce Injuries to Motor Vehicle Occupants. American Journal of Preventive Medicine 21(4, suppl. 1):1622. doi:10.1016/S0749-3797(01)00380-4. Available at http://dx.doi.org/10.1016/S0749-3797(01)00380-4 (accessed July 2, 2007).
World Health Organization (WHO). 2003. Mental Health in Emergencies: Mental and Social Aspects of Health of Populations Exposed to Extreme Stressors. WHO/MSD/MER/03.01. Geneva: Department of Mental Health and Substance Dependence, World Health Organization. Available at http://www.who.int/mental_health/media/en/640.pdf (accessed July 2, 2007).
Zaza, S., V.G. Carande-Kulis, D.A. Sleet, D.M. Sosin, R.W. Elder, R.A. Shults, T.B. Dinh-Zarr, J.L. Nichols, R.S. Thompson, and the Task Force on Community Preventive Services. 2001. Methods for Conducting Systematic Reviews of the Evidence of Effectiveness and Economic Efficiency of Interventions to Reduce Injuries to Motor Vehicle Occupants. American Journal of Preventive Medicine 21(4, suppl. 1):2330. doi:10.1016/S0749-3797(01)00379-8. Available at http://dx.doi.org/10.1016/S0749-3797(01)00379-8 (accessed July 2, 2007).
Melissa Sweet is a freelance health writer and journalist in Australia. She holds honorary positions as an adjunct senior lecturer in the School of Public Health at the University of Sydney and as an adjunct senior lecturer in the School of Medicine of the Sydney Campus of the University of Notre Dame Australia. Sweet has been writing about the promises and pitfalls of evidence-informed health care for more than a decade. Together with Judy and Les Irwig, she published Smart Health Choices: How to Make Informed Health Decisions (Allen and Unwin, 1999), which gives readers some tools for critically assessing health information. Sweet recently published Inside Madness (Pan Macmillan, 2006), which combines a number of nonfiction stories, including that of murdered psychiatrist Margaret Tobin, the difficulties of achieving change in complex health systems, and a history of mental health in Australia. Her latest book is The Big Fat Conspiracy: How to Protect Your Family’s Health (ABC Books, 2007), which attempts to reframe obesity as an environmental health issue and gives families, communities, and other groups some tips for promoting healthier environments for children.
Award-winning journalist, author, and documentary maker Ray Moynihan is an internationally recognized health writer, a visiting editor with the British Medical Journal, and a conjoint lecturer at the University of Newcastle. He has a global reputation for writing critically about the dangers of modern medicine and how to reform it.
His 2005 book Selling Sickness: How the World’s Biggest Pharmaceutical Companies Are Turning Us All into Patients (co-written with Alan Cassels and published by Nation Books) has been successfully sold in many nations, including the United States, the United Kingdom, and Canada, and across Asia and Europe. It is being translated into Japanese, Chinese, Italian, Greek, and Spanish, among other languages; it has inspired several TV and radio documentaries; and it was shortlisted for a Walkley and Queensland Premier Award.
Selected Publications of the Centers for Disease Control and Prevention
Useful General Web References with Which to Start
www.CDC.govYour online source for credible health information.
www.thecommunityguide.orgA complete compendium of Community Guide information and references.
Additional Specific Community Guide References
Overviews and General References
Zaza, S., P.A. Briss, and K.W. Harris. 2005. The Guide to Community Preventive Services. Oxford: Oxford University Press.
Briss, P.A., R.C. Brownson, J.E. Fielding, and S. Zaza. 2004. Developing and Using the Guide to Community Preventive Services: Lessons Learned about Evidence-Based Public Health. Annual Review of Public Health 25:281302. doi:10.1146/annurev.publhealth.25.050503.153933. Available at http://dx.doi.org/10.1146/annurev.publhealth.25.050503.153933 (accessed October 5, 2007).
McGinnis, J.M., and W. Foege. 2000. Guide to Community Preventive Services: Harnessing the Science. American Journal of Preventive Medicine 18(1, suppl. 1):12. doi:10.1016/S0749-377(99)00125-7. Available at http://dx.doi.org/10.1016/S0749-3797(99)00125-7 (accessed October 5, 2007).
How Can Reviews (and Dissemination) Better Inform Policy and Practice?
Fielding, J.E., and P.A. Briss. 2006. Promoting Evidence-Based Public Health Policy: Can We Have Better Evidence and More Action? Health Affairs 25(4):96978. doi:10.1377/hlthaff.25.4.969. Available at http://dx.doi.org/10.1377/hlthaff.25.4.969 (accessed October 5, 2007).
Importance of Integrating Evidence-Based Health Care and Public Health
Ockene, J.K., E.A. Edgerton, S.M. Teutsch, L.N. Marion, T. Miller, J.L. Genevro, C.J. Loveland-Cherry, J.E. Fielding, and P.A. Briss. 2007. Integrating Evidence-Based Clinical and Community Strategies to Improve Health. American Journal of Preventive Medicine 32(3):24452. doi:10.1016/j.amepre.2006.11.007. Available at http://dx.doi.org/10.1016/j.amepre.2006.11.007 (accessed October 5, 2007).
McGowan A., R. Hahn, A. Liberman, A. Crosby, M. Fullilove, R. Johnson, E. Moscicki, L. Price, S. Snyder, F. Tuma, J. Lowy, P. Briss, S. Cory, G. Stone, and the Task Force on Community Preventive Services. 2007. Effects on Violence of Laws and Policies Facilitating the Transfer of Juveniles from the Juvenile Justice System to the Adult Justice System: A Systematic Review. American Journal of Preventive Medicine 32(4, suppl. 1):728. doi:10.1016/j.amepre.2006.12.003. Available at http://dx.doi.org/10.1016/j.amepre.2006.12.003 (accessed October 5, 2007).
Shults, R.A., R.W. Elder, D.A. Sleet, J.L. Nichols, M.O. Alao, V.G. Carande-Kulis, S. Zaza, D.M. Sosin, R.S. Thompson, and the Task Force on Community Preventive Services. 2001. Reviews of Evidence Regarding Interventions to Reduce Alcohol-Impaired Driving. American Journal of Preventive Medicine 21(4, suppl. 1):6688. doi:10.1016/S0749-3797(01)00381-6. Available at http://dx.doi.org/10.1016/S0749-3797(01)00381-6 (accessed October 5, 2007).
McGinnis, J.M., and W. Foege. 2000. Guide to Community Preventive Services: Harnessing the Science. American Journal of Preventive Medicine 18(1, suppl. 1):12. doi:10.1016/S0749-3797(99)00125-7. Available at http://dx.doi.org/10.1016/S0749-3797(99)00125-7 (accessed October 5, 2007).
Methods and Process
Briss, P.A., S. Zaza, M. Pappaioanou, J. Fielding, L. Wright-De Agüero, B.I. Truman, D.P. Hopkins, P.D. Mullen, R.S. Thompson, S.H. Woolf, V.G. Carande-Kulis, L. Anderson, A.R. Hinman, D.V. McQueen, S.M. Teutsch, J.R. Harris, and the Task Force on Community Preventive Services. 2000. Developing an Evidence-Based Guide to Community Preventive Services—Methods. American Journal of Preventive Medicine 18(1, suppl. 1):3543. doi:10.1016/S0749-3797(99)00119-1. Available at http://dx.doi.org/10.1016/S0749-3797(99)00119-1 (accessed October 5, 2007).
Data Collection Instrument and Procedure
Zaza, S., L.K. Wright-De Agüero, P.A. Briss, B.I. Truman, D.P. Hopkins, M.H. Hennessy, D.M. Sosin, L. Anderson, V.G. Carande-Kulis, S.M. Teutsch, M. Pappaioanou, and the Task Force on Community Preventive Services. 2000. Data Collection Instrument and Procedure for Systematic Reviews in the Guide to Community Preventive Services. American Journal of Preventive Medicine 18(1, suppl. 1):4474. doi:10.1016/S0749-3797(99)00122-1. Available at http://dx.doi.org/10.1016/S0749-3797(99)00122-1 (accessed October 5, 2007).
Commentary on Methods Challenges of Evidence-Based Public Health
Briss, P.A. 2005. Evidence-Based: US Road and Public-Health Side of the Street. The Lancet 365:82830. doi:10.1016/S0140-6736(05)71019-9. Available at http://dx.doi.org/10.1016/S0140-6736(05)71019-9 (accessed October 5, 2007).
(To request a bound copy of this report, click here.
To see a complete list of Milbank reports, click here.
Be sure to specify which report you want,
your name, mailing address, and phone number.)
Milbank Memorial Fund
645 Madison Avenue
New York, NY 10022
The Milbank Memorial Fund is an endowed operating foundation that engages in nonpartisan analysis, study, research, and communication on significant issues in health policy. In the Fund's own publications, in reports, films, or books it publishes with other organizations, and in articles it commissions for publication by other organizations, the Fund endeavors to maintain the highest standards for accuracy and fairness. Statements by individual authors, however, do not necessarily reflect opinions or factual determinations of the Fund.
©2007 Milbank Memorial Fund. All rights reserved. This publication may be redistributed electronically, digitally, or in print for noncommercial purposes only as long as it remains wholly intact, including this copyright notice and disclaimer.
Centers for Disease Control and Prevention
1600 Clifton Road, N.E.
Atlanta, GA 30333
The Centers for Disease Control and Prevention (CDC) is the U.S. federal public health agency whose mission is “to promote health and quality of life by preventing and controlling disease, injury, and disability.” A core function of the CDC is to be a credible source of health information. The findings and conclusions in this report are those of the authors and do not necessarily represent the views or endorsement of the CDC.
Printed in the United States of America.
Online producer: Stephanie Moe-Quiggle