Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2015 Oct 18.
Published in final edited form as: Am J Prev Med. 2014 Jun 18;47(2):212–219. doi: 10.1016/j.amepre.2014.03.007

Conducting Rapid, Relevant Research

Lessons Learned from the My Own Health Report Project

Russell E Glasgow 1, Rodger S Kessler 2, Marcia G Ory 3, Dylan Roby 4, Sherri Sheinfeld Gorin 5,6, Alex Krist 7
PMCID: PMC4609529  NIHMSID: NIHMS607465  PMID: 24953520

Abstract

The lengthy and uncertain translation of research into clinical practice is well documented. Much of the current “gold standard” clinical research is slow, expensive, and lacks perceived relevance for practitioners and decision makers. In contrast, we summarize experiences conducting the My Own Health Report (MOHR) project to collect and address patient reported measures using principles of rapid, relevant pragmatic research. The methods used for rapid design and fielding of the MOHR project to improve attention to health behaviors and mental health are detailed. Within the multisite, pragmatic, implementation–focused MOHR study, we describe the four phases of the research and the key decisions made and actions taken within each. We provide concrete examples of how relevant research can be conducted transparently to rapidly provide information to practitioners. Data were collected and analyzed in 2013.

The multisite (seven research centers partnered with 18 clinics) cluster randomized pragmatic delayed intervention trial was conducted in less than 18 months from receipt of funding applications to completion of data collection. Phases that were especially accelerated included funding and review, and recruitment and implementation. Conducting complex studies rapidly and efficiently is a realistic goal. Key lessons learned for prevention research include: use of existing research networks; use of web-based assessment/feedback tools that are tailored to fit local needs; engaging relevant stakeholders early on and throughout the process to minimize need for redesign; and making pragmatic decisions that balance internal and external validity concerns rather than waiting for perfect solutions.

Introduction

Most research evidence based on “gold standard” trials has little immediate impact on practice or policy.13 Such trials are often so focused on being rigorously internally valid that the results do not disseminate well into practice.4,5 Traditional trials are also slow and expensive.68 Most importantly, research often fails to respond to the questions of stakeholders who make decisions about health care, and who need rapid, actionable data to make decisions.911

Riley et al.7 recently presented a conceptual model and set of key decisions to improve the speed and utility of health research. This model calls for ongoing stakeholder involvement, streamlining the grant announcement and review process, and planning for rapid dissemination, implementation, and analyses, resulting in faster availability of data for decision-making. Despite the need for alternative research administration, designs, and strategies,12 as yet there are few examples of such rapid research.

This paper reviews the actions, challenges, and lessons learned in developing and conducting the My Own Health Report (MOHR) pragmatic trial, providing an example of a trial designed with the observations of Riley and colleagues7 as a guide. We report key decisions in the research phases of funding, collaborative planning, implementation, and analysis undertaken to balance research rigor with speed and relevance to practice. The primary study outcomes, which include the intervention reach, effectiveness, context, and costs will be reported separately. To our knowledge, this is the first report systematically describing the steps undertaken to design, implement, and analyze data from a rapid, complex pragmatic trial for healthcare change.

Methods

MOHR is a practice–level, cluster randomized pragmatic implementation study. The general methods, measures, and intervention components are described elsewhere.13 It is designed to develop rapid, actionable evidence around the use of patient-reported measures in patient care.13 The intervention uses the MOHR web–based patient assessment and feedback instrument that addresses ten domains of health behaviors and psychosocial issues. MOHR assesses a range of prevention and care issues including health behaviors, mental health issues, and substance abuse, as well as health–related quality of life and key demographic factors (Table 1). This automated assessment and feedback tool encourages collaborative goal setting between primary care providers and patients presenting for usual care, wellness, or chronic care visits. Appendices 1 and 2 present sample feedback reports for a patient and the practice team, respectively. Diverse patients and practice staff were involved in selecting and pilot testing the assessment items and feedback system.14 The primary outcome of the study was the establishment of collaboratively set action plans. The randomized delayed intervention sites provided a control for temporal trends and local context. The delayed intervention sites were paired with intervention sites and did not receive either the MOHR assessment and feedback tool or explicit goal setting support, but served as usual care controls for what was happening in these contexts given the rapidly changing healthcare environment. Diverse stakeholders (e.g., clinicians, behavioral scientists, public health researchers, and dissemination and implementation scientists) contributed to the measures, study design, and implementation.15

Table 1.

My Own Health Report measures for adult primary care

Domain Number of items and final measure (source)
F Eating patterns 3 items: modified from Starting the Conversation (Adapted from Paxton AE, et al. Am J Prev Med 2011;40(1):67–71)
F Physical activity 2 items: the Exercise Vital Sign (Sallis R. Br J Sports Med 2011;45(6):473–4)
F Sleep 2 items: a. Adapted from BRFSS b. Neuro-QOL (Item PQSLP04)
F Smoking/tobacco use 2 items: Tobacco Use Screener (Adapted from YRBSS Questionnaire)
SA Risky drinking 1 item: Alcohol Use Screener (Smith PC, et al. J Gen Int Med 2009;24(7):783–8)
SA Substance abuse 1 item: NIDA Quick Screen (Smith PC, et al. Arch Int Med 2010;170(13):1155–60)
MH Stress 1 item: Distress Thermometer (Roth AJ, et al. Cancer 1998;15(82):1904–8)
MH Anxiety and depression 4 items: Patient Health Questionnaire—Depression and Anxiety (Kroenke K, et al. Psychosomatics 2009;50(6):613–21)
G Overall health status 1 item: BRFSS Questionnaire
G Demographics

BRFSS, Behavioral Risk Factor Surveillance System; F, Framingham health behaviors; G, general; MH, mental health; NIDA, National Institute on Drug Abuse; QOL, quality of life; SA, substance abuse; YRBSS, Youth Risk Behavior Surveillance System

Setting and Participant Characteristics

MOHR collected data from over 2,700 patients in 18 primary care clinics associated with seven research centers (Table 2). Six clinic sites were urban, three suburban, and nine rural. The majority of clinic sites had patient–centered medical home (PCMH) status or were in the process of applying. The vast majority of clinics that we approached for participation took part in the project, although one research network that was initially interested declined participation owing to timing issues.

Table 2.

Participating practice characteristics

State Practice type Setting Patients seen annually Provider FTEs Rooming staff FTE Patient ethnicity and race Insurance type PCMH status Behavior change expertise

Clinician MLP Nurse MA Non-clinical Latino Black Medicare Medicaid Uninsured
Site1 VA PBRN Suburban 5,000 2 0 2 0 2 15% 10% 5% 0% 30% No None
Site2 VA PRBN Suburban 1,500 1 0 0 1 1 20% 10% 9% 0% 1% No None
Site3 VA PRBN Rural 2,500 1.2 0.4 0 2.6 4.4 1% 49% 12% 1% 49% No Some
Site4 VA PRBN Suburban 5,200 2 2 0 5 6 2% 18% 15% 2% 18% No None
Site5 VA PRBN Urban 4,440 4.9 0.5 7.3 0 5 2% 19% 24% 42% 17% Yes Some
Site6 VA PRBN Urban 4,770 5.3 0.5 7.3 0 5 1% 17% 26% 42% 17% Yes Some
Site7 CA PRBN Rural 3,500 5.1 0.4 3.1 6.7 5.2 3% 1% 13% 3% 1% Applying None
Site8 CA PRBN Rural 5,400 3.9 3.1 5.6 13.8 6.1 13% 2% 12% 13% 2% Applying A lot
Site9 VT PRBN Rural 9,500 3.3 1.7 2.5 2 9 1% 5% 13% 1% 5% Yes Some
Site10 VT PRBN Rural 10,000 5 0 3 5 6 1% 2% 15% 1% 2% Yes Some
Site11 NC CPCRN Rural 1,100 1.5 3 2 6 4 2% 60% 49% 2% 60% Yes Some
Site12 NC CPCRN Rural 7,500 2.5 1 2 4 4 40% 60% 10% 10% 70% Yes A lot
Site13 CA CPCRN Urban 2,040 1 0 0 7 5 75% 25% 5% 45% 50% Applying Some
Site14 CA CPCRN Urban 2,180 1 1 0 6 5 75% 25% 5% 45% 50% Applying Some
Site15 TX CPCRN Rural 4,800 1 1 0 4 2 48% 23% 2% 48% 23% No None
Site16 TX CPCRN Rural 3,800 1 1 0 3 3 23% 32% 2% 23% 32% No None
Site17 TX CPCRN Urban 2,800 1.7 1.3 0 5 14 82% 6% 1% 82% 6% Yes None
Site18 TX CPCRN Urban 2,800 1.8 1.8 0 5 7 80% 5% 1% 80% 5% Yes Some

FTE, full-time equivalent; PCMH, patient–centered medical home

There was variability in panel ethnicity/racial background with a few sites predominantly serving Latinos (approximately 80%) and most remaining clinics serving a small percentage of Latinos (≤23%). Two clinics served more than 50% African Americans. In one clinic, almost 50% of its clients were covered by Medicare, in contrast to most clinics where less than 15% received Medicare benefits. Although most sites (n=16) had electronic health record (EHR) capability, the diversity of EHR systems, limited time, and restricted funding did not allow for full MOHR integration with the diverse EHRs. Comparison of the race and ethnicity of actual MOHR participants to the general characteristics of the overall involved clinics indicated that there was good representation of ethnic and racial groups. The proportion of African Americans was higher among MOHR intervention participants than the overall clinic in four sites, equivalent (± 3%) in two sites, and lower in three sites. Latino ethnicity was higher in two sites, equivalent in four, and lower in three sites.

Study Phases

Funding and review of applications

MOHR was funded using supplemental funds from the National Cancer Institute (NCI), Agency for Healthcare Research and Quality (AHRQ), and NIH Office of Behavioral and Social Science Research. To accelerate the review and start-up process, project eligibility was restricted to currently funded research organizations that were part of one of two existing networks: the AHRQ–funded Practice Based Research Network (PBRN) program16 or the NCI Cancer Prevention and Control Research Network (CPCRN).19

Applicants completed a brief, structured application (three pages) that summarized their expertise, proposed key study issues and outcomes, and identified a matched set of proposed primary care practices. Internal review was conducted by staff from the funding organizations, supplemented by outside experts. Priority was given to applications proposing well-matched pairs of sites that varied across dimensions including geographical location, clinic type and size, patient population, urban versus rural setting, and level of EHR integration.

Collaborative planning and refinement

Study design was accelerated by agreement to use a set of common patient–report items and an automated assessment/feedback tool, which was eventually called the MOHR patient assessment and feedback system. The items in MOHR were based on prior work by the NIH to identify brief practical measures feasible for use in primary care.15 The resultant set of 17 items, covering ten different health behavior, mental health, and substance abuse domains (Table 1) are described by Estabrooks et al.15 and Glasgow and colleagues.17

A decentralized decision–making process was used in the collaborative trial planning process. Work groups varying in size from four to eight members were established for: assessment and feedback tool creation; patient-reported outcomes and assessment procedures; context assessment; cost and resources collection; and papers, publications, and public relations. These groups worked efficiently and made the major decisions and recommendations that were brought to overall project conference calls bimonthly.

Patient recruitment and implementation

Efforts were made to recruit a diverse set of primary care practices and patients. There were few exclusion criteria (e.g., we did not exclude patients based on health behavior profile or presence of existing conditions) and practices were encouraged to invite the widest possible variety of adult patients regardless of disease status who were coming for usual care, wellness, or chronic disease visits. Primary care practices had to agree to randomization to either early or delayed intervention, and to recruit 150–200 adult primary care patients within a 1-year period.

Analyses and reporting

The automated tool generated real-time reports on patient enrollment for each clinic. An initial summary data report summarized recruitment experiences and prevalence of the ten different health risks for practices.

Results

Time to Accomplish Key MOHR Activities

Table 3 summarizes the approximate time to complete each of the MOHR phases. The review and funding process took approximately 3 months and was conducted in two waves: one for CPCRN sites (four research sites and eight federally qualified health center clinics) and one for PBRN sites (three research sites and ten PBRN practices). The practices were not randomly selected, but rather resulted from a purposive, pragmatic selection of clinics from two federally funded stakeholder groups designed to be collectively diverse in terms of geography, practice characteristics, and patient profiles.

Table 3.

Key phases, decisions, and actions in the My Own Health Report study

My Own Health Report phases Decisions and actions to increase speed and enhance relevance
Funding and review of applications (2–3 months) • Restricted eligibility to existing research networks
• Short response time and rapid review process
• Internal review and use of supplement funding
• Stipulated agreement to use common measures
Collaborative planning and refinement phases (3–4 months) • Decision to use intermediate implementation outcomes rather than ultimate outcomes
• Utilized existing measures where possible and built on prior electronic health record measures work
• Conducted small scale, rapid tests of automated tool and items
• Engaged stakeholders (clinicians and patients) at multiple points
• Gave authority to subgroups to make decisions (e.g., context assessment, automated tool components, patient surveys, and cost collection)
• Worked with each clinic to adapt My Own Health Report to their setting, clinic flow, and patients
Patient recruitment and implementation phases (9–10 months) • Webinar rapid training and biweekly collaborative calls
• Responsive “core steering committee”
• Adaptations made based on real-time data including weekly feedback on reach and survey completion
• Modified recruitment and follow-up procedures
Plans for analyses and reporting (Ongoing) • Initially outlined table shells and prioritized analyses
• First analyses focused on feedback to participating clinics on reach and patient health behaviors
• Commitment to transparent reporting on adaptations, variations across sites, and time

Planning and refinement

Project planning, testing, and training on use of the MOHR assessment/feedback tool was accomplished in another 3–4 months. Previous experience with a paper–based pilot test of the MOHR items14 allowed for efficient creation of the MOHR web– based assessment and feedback tool.13 Programming of items, including follow-up questions if the initial screening item(s) were positive for four items (i.e., the Patient Health Questionnaire-2 for depression), were accomplished relatively quickly. It took somewhat longer to create and obtain feedback from both patients and practice staff on the prototype summary reports.

Work groups for the patient experience survey (to develop the patient experience items used as primary outcomes), context assessment (to develop mixed methods to assess key contextual factors), and cost and resources completed their tasks by the time the MOHR assessment/feedback tool was developed and piloted. In general, the larger steering committee endorsed the recommendations of these work groups with only minor refinements.

Recruitment and implementation

These activities were completed in approximately 10 months. Most sites experienced challenges to patient recruitment, consistent implementation of the feedback protocol, follow–up survey completion, or some combination. The general approach used to enhance quality was tracking and providing rapid, continuous feedback on progress, using collaborative semi–weekly learning calls to share successes and brainstorm responses to challenges, and if an initial implementation strategy still did not work, substituting or replacing it.

An example was one site that experienced particularly low response rates to mailed invitations to complete the MOHR assessment, despite using identical procedures as other sites. After consultation regarding options, it was possible to change to an existing call service that was part of the delivery organization. This change resulted in a ten-fold increase in participation. Most sites experienced challenges with obtaining high response rates to the post-visit survey sent 2–4 weeks following the visit and designed to assess patient experiences concerning health risk assessment and shared decision-making regarding identified risks. After reviewing initial data and consulting Dillman et al.,18 to increase survey return rates, each site supplemented their initial survey return procedures with an added modality (e.g., if initial survey was mailed, following up with a phone call for non-respondents), so that all sites made multiple contact attempts using different modalities.

Challenges Encountered

There were challenges in study implementation, both expected1920 and unexpected. We encouraged practices to recruit a large number of representative patients, not just a convenience sample or those they felt would most benefit. This created challenges for busy practices with multiple priorities, more so than just attempting to recruit an occasional, non-complex patient.

Some of the lengthiest delays were due to IRB review. Having seven different IRBs needing to review the protocol, and any changes, resulted in considerable delays. We attempted to address this challenge by having the coordinating center obtain approval early in the process, as a precedent for other IRBs. This was only partially successful, as no IRBs decided to cede review authority, and different decisions were made by different IRBs about which aspects of the study were exempt, expedited, or constituted research as opposed to quality improvement. Although most IRBs considered MOHR assessment and feedback as part of clinical care or a quality improvement process, one IRB required patient consent before patients could be approached. All IRBs except one required informed consent for the follow–up patient experience survey. Also, the IRBs had different requirements about what they would allow versus consider coercive in terms of follow-up contact and incentives for completing surveys.

The heterogeneity across sites and adoption of pragmatic principles that adapted the intervention to fit each setting required different approaches to the details of patient recruitment, MOHR implementation, and feedback procedures. Having to develop strategies that retained common elements, but were tailored to settings (e.g., if the assessment was administered over the web versus in-person, before versus during the visit, in English versus Spanish), proved time consuming. Part of the diversity resulted from the decision to include two care networks (i.e., PBRNs and FQHCs). This decision created greater diversity in the settings, samples, and contexts and increased confidence that the MOHR procedures can work in diverse settings, but it increased the complexity and length of the planning and implementation process.

Finally, it became necessary to replace the initial contractor engaged to create the web–based MOHR assessment/feedback tool when they could not meet the specifications required in a timely manner. Fortunately, all parties were willing to work together to make the transition as smooth as possible, making the delays and added costs as minimal as possible.

Discussion

This paper demonstrates that it is feasible to rapidly conduct complex, high–quality, pragmatic research that is relevant to stakeholders. Despite the complexity of issues addressed, MOHR was conducted in less than 18 months from concept to final data collection. Below, we summarize the primary reasons this was accomplished in a short time relative to traditional multisite trials. Two key influences were the rapid review process for funding and use of existing research networks. Compared to traditional review and funding mechanisms, these two decisions likely saved at least 12–18 months. Paired with a management team providing leadership in the implementation of pragmatic trials, selection of sites that had experience in collaborative multisite research, partnership research principles,2123 and working in real–world, non-academic primary care and community settings expedited the process.

MOHR might have proceeded even more quickly if we had restricted sites to only those using a common EHR platform, with shared data infrastructures, or who were at a common stage of PCMH implementation. Adapting the intervention and training, and automating a tool to fit varied clinic flows and implementation plans were challenging. Not attempting to integrate the MOHR into the various EHRs can be seen as a major limitation. Although we will pursue such integration in future research, we are not convinced that it is necessary to have all tools and guidance within an EHR as the most provider or patient-centered way to achieve high-quality implementation.

The use of patient experience outcomes (e.g., surveys about receipt of collaborative action planning) for our primary endpoint was another key factor in accomplishing the project efficiently. If we had required a biological outcome, this would have considerably increased time and dramatically increased sample size and expense, as discussed by Proctor et al.24 and Glasgow and colleagues.25 The decision to focus on many health risks rather than one or two in isolation, such as smoking cessation or depression, as is typically done substantially increased the relevance for primary care settings faced with all of these issues concurrently.26 Further, only a subset of patients had any one risk factor.

MOHR was designed to see if a brief, low-cost procedure to address the complex array of health behavior, psychosocial, and substance abuse issues encountered in primary care could be implemented consistently in diverse real–world settings. Additional research is needed to clarify linkages among key implementation steps such as establishing an action plan27 and follow-up contact28 with behavior change or biological outcomes.

Finally, we used rapid learning principles2931 and stakeholder engagement32 processes to simultaneously address speed and relevance. MOHR adapted procedures based upon initial data and evolving issues—in particular, we had to adjust recruitment processes and survey procedures33 to enhance participation rates. We think that such adaptations are important and that transparency is critical. Although MOHR could have been conducted even more quickly, we made several decisions to enhance relevance or rigor that resulted in modest delays. We recruited clinics interested in pragmatic trials and evaluating processes for ongoing use rather than just asking permission “to let us recruit 150 patients and not bother you.” The decision to translate the MOHR assessment and feedback tool and all measures and procedures into Spanish also took longer, but we judged this important for future use in an increasingly Hispanic patient caseload.

There have been other recent funding mechanisms to support rapid research.7 For example, the AHRQ Developing Evidence to Inform Decisions about Effectiveness network provides a quick review process by restricting applicants to a set of previously vetted research groups with demonstrated capacity; the National Institute of Diabetes Digestive and Kidney Diseases has recently used rapid review procedures to fund ongoing natural experiments and policy interventions related to obesity; and Patient–Centered Outcomes Research Institute funding mechanisms are experimenting with ways to enhance both the relevance (e.g., including patients and stakeholders in the review) and speed of review.34 Although limiting funding to sites that are part of research networks would have been restrictive and possibly unrepresentative 10 years ago, today a vast array of PBRNs, community health centers, and rural networks exist, and many more are being established with the advent of accountable care organizations and other real– world learning healthcare systems.

Limitations of this project include that the pragmatic trial is admittedly not definitive. By focusing on multiple issues facing primary care, it is less likely to impact any one of the ten targets than if we had chosen to focus on only one. MOHR was not intended to produce practice transformation,35 but rather to see if diverse practices could find a way to integrate screening and tailored action planning into their clinic flow and procedures. Finally, although we purposely included a wide range of clinics in different geographic areas, in rural, urban, and suburban settings and at different levels of EHR integration36 and PCMH status, the study did not include integrated care plans such as Veterans Affairs or HMO settings, and thus is not generalizable to all primary care settings. Even with these limitations, we think it instructive to present a transparent real–world example of how valid multisite research can be conducted on complex issues in diverse settings relatively rapidly.

Recommendations for future research include: transparent reporting, especially about settings, inclusion and exclusion criteria, and adaptations made; differences across sites; and attempts to balance rigor with relevance, speed, and research efficiency. Remaining challenges related to addressing ten health behaviors, mental health, and substance abuse concerns concurrently include identification of optimal ways to help patients and staff prioritize goals. Long-term sustainability of enhanced patient assessment and counseling remains a question for future research.

Supplementary Material

01
02

Acknowledgments

Financial support provided by National Cancer Institute, Agency for Healthcare Research and Quality, and NIH Office of Behavioral and Social Science Research.

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

No other financial disclosures were reported by the authors of this paper.

Contributor Information

Russell E. Glasgow, Colorado Health Outcomes Program, University of Colorado School of Medicine, Denver, Colorado.

Rodger S. Kessler, Department of Family Medicine, University of Vermont, Burlington, Vermont.

Marcia G. Ory, Department of Health Promotion and Community Health Sciences, Texas A&M Health Science School of Public Health, College Station, Texas.

Dylan Roby, Department of Health Policy and Management, University of California, Los Angeles, California.

Sherri Sheinfeld Gorin, Senior Scientific Consultant, Division of Cancer Control and Population Sciences, National Cancer Institute, Rockville, Maryland; Director of New York Physicians against Cancer, Herbert Irving Comprehensive Cancer Center, Columbia University, New York.

Alex Krist, Department of Family Medicine and Population Health, Virginia Commonwealth University, Richmond, Virginia.

References

  • 1.Greenhalgh T, Robert G, Macfarlane F, Bate P, Kyriakidou O. Diffusion of innovations in service organizations: systematic review and recommendations. Milbank Q. 2004;82(4):581–629. doi: 10.1111/j.0887-378X.2004.00325.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Green LW, Fielding J. The U.S. healthy people initiative: its genesis and its sustainability. Annu Rev Public Health. 2011;32:451–70. doi: 10.1146/annurev-publhealth-031210-101148. [DOI] [PubMed] [Google Scholar]
  • 3.Kessler R, Glasgow RE. A proposal to speed translation of healthcare research into practice: dramatic change is needed. Am J Prev Med. 2011;40(6):637–44. doi: 10.1016/j.amepre.2011.02.023. [DOI] [PubMed] [Google Scholar]
  • 4.Green LW, Glasgow RE, Atkins D, Stange K. Making evidence from research more relevant, useful, and actionable in policy, program planning, and practice slips “twixt cup and lip.”. Am J Prev Med. 2009;37(6S):S187–S191. doi: 10.1016/j.amepre.2009.08.017. [DOI] [PubMed] [Google Scholar]
  • 5.Westfall JM, Mold J, Fagnan L. Practice-based research—“Blue highways” on the NIH roadmap. JAMA. 2007;24297(4):403–6. doi: 10.1001/jama.297.4.403. [DOI] [PubMed] [Google Scholar]
  • 6.Rothwell PM. Factors that can affect the external validity of randomised controlled trials. PLoS Clin Trials. 2006;1(1):e9. doi: 10.1371/journal.pctr.0010009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Riley WT, Glasgow RE, Etheredge L, Abernethy AP. Rapid, responsive, relevant (R3) research: a call for a rapid learning health research enterprise. Clin Transl Med. 2013;2(1):10. doi: 10.1186/2001-1326-2-10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Kumar S, Nilsen WJ, Abernethy A, et al. Mobile health technology evaluation: the mHealth evidence workshop. Am J Prev Med. 2013;45(2):228–36. doi: 10.1016/j.amepre.2013.03.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Kessler R. The patient centered medical home: an oppportunity to move past brilliant and irrelevant research and practice. Transl Behav Med. 2012;2(3):311–2. doi: 10.1007/s13142-012-0151-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Thorpe KE, Zwarenstein M, Oxman AD, et al. A pragmatic-explanatory continuum indicator summary (PRECIS): a tool to help trial designers. CMAJ. 2009;180(10):E47–57. doi: 10.1503/cmaj.090523. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Tunis SR, Stryer DB, Clancey CM. Practical clinical trials: increasing the value of clinical research for decision making in clinical and health policy. JAMA. 2003;290(12):1624–32. doi: 10.1001/jama.290.12.1624. [DOI] [PubMed] [Google Scholar]
  • 12.Glasgow RE. What does it mean to be pragmatic? Pragmatic methods, measures, and models to facilitate research translation. Health Educ Behav. 2013;40(3):257–65. doi: 10.1177/1090198113486805. [DOI] [PubMed] [Google Scholar]
  • 13.Krist AH, Glenn BA, Glasgow RE, et al. Designing a valid randomized pragmatic primary care implementation trial: the my own health report (MOHR) project. Implement Sci. 2013;8:73. doi: 10.1186/1748-5908-8-73. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Rodgriguez HP, Glenn BA, Olmos T, et al. The impact of implementing point-of-care behavioral health assessment on clinical discussion across diverse primary care practices. J Am Board Fam Med. 2014 In press. [Google Scholar]
  • 15.Estabrooks PA, Boyle M, Emmons KM, et al. Harmonized patient-reported data elements in the electronic health record: supporting meaningful use by primary care action on health behaviors and key psychosocial factors. J Am Med Inform Assoc. 2012;19(4):575–82. doi: 10.1136/amiajnl-2011-000576. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Agency for Health Care Research and Quality Practice-Based Research Networks. http://pbrn.ahrq.gov.
  • 17.Glasgow RE, Kaplan RM, Ockene JK, Fisher EB, Emmons KM. Patient-reported measures of psychosocial issues and health behavior should be added to electronic health records. Health Aff (Millwood) 2012;31(3):497–504. doi: 10.1377/hlthaff.2010.1295. [DOI] [PubMed] [Google Scholar]
  • 18.Dillman DA. Mail and Internet surveys: the tailored design method. 2nd ed. John Wiley & Sons; New York: 2000. [Google Scholar]
  • 19.Solberg LI, Glasgow RE, Unutzer J, et al. Partnership research: a practical trial design for evaluation of a natural experiment to improve depression care. Med Care. 2010;48(7):576–82. doi: 10.1097/MLR.0b013e3181dbea62. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Zwarenstein M, Treweek S. What kind of randomised trials do patients and clinicians need? Evid Based Med. 2009;14(4):101–3. doi: 10.1136/ebm.14.4.101. [DOI] [PubMed] [Google Scholar]
  • 21.Green LA, Cifuentes M, Glasgow RE, Stange KC. Redesigning primary care practice to incorporate health behavior change: prescription for health round 2 results. Am J Prev Med. 2008;35(5S):S347–S349. doi: 10.1016/j.amepre.2008.08.013. [DOI] [PubMed] [Google Scholar]
  • 22.Stange KC, Glasgow RE. Considering and reporting important contextual factors in research on the patient-centered medical home. Agency for Healthcare Research and Quality; Rockville MD: May, 2013. AHRQ Publication No.: 13-0045-EF. [Google Scholar]
  • 23.Harris JR, Brown PK, Coughlin S, et al. The cancer prevention and control research network. Prev Chronic Dis. 2005;2(1):A21. [PMC free article] [PubMed] [Google Scholar]
  • 24.Proctor E, Silmere H, Raghavan R, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health. 2011;38(2):65–76. doi: 10.1007/s10488-010-0319-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Glasgow RE, Brownson RC, Kessler RS. Thinking about health-related outcomes: what do we need evidence about? Clin Transl Sci. 2013;6(4):286–91. doi: 10.1111/cts.12080. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Yarnell KS, Pollack KI, Ostbye T, Krause KM, Michener JL. Primary care: is there enough time for prevention? Am J Public Health. 2003;93(4):635–41. doi: 10.2105/ajph.93.4.635. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Lorig KR, Holman HR. Self-management education: history, definition, outcomes, and mechanisms. Ann Behav Med. 2003;26(1):1–7. doi: 10.1207/S15324796ABM2601_01. [DOI] [PubMed] [Google Scholar]
  • 28.Weinberger M, Kirkman MS, Samsa GP, et al. A nurse-coordinated intervention for primary care patients with non-insulin-dependent mellitus: impact on glycemic control and health-related quality of life. J Gen Intern Med. 1995;10(2):59–66. doi: 10.1007/BF02600227. [DOI] [PubMed] [Google Scholar]
  • 29.Berwick DM. The science of improvement. JAMA. 2008;299(10):1182–4. doi: 10.1001/jama.299.10.1182. [DOI] [PubMed] [Google Scholar]
  • 30.Etheredge LM. A rapid-learning health system. Health Aff (Millwood) 2007;26(2):w107–18. doi: 10.1377/hlthaff.26.2.w107. [DOI] [PubMed] [Google Scholar]
  • 31.IOM The Learning Healthcare System in America. www8.nationalacademies.org/cp/projectview.aspx?key=IOM-EO-10-06.
  • 32.Westfall JM, Ingram B, Navarro D, et al. Engaging communities in education and research: PBRNs, AHEC, and CTSA. Clin Transl Sci. 2012;5(3):250–8. doi: 10.1111/j.1752-8062.2011.00389.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Dillman D. Design effects in the transition to web-based surveys. Am J Prev Med. 2007;32(5S):S90–S96. doi: 10.1016/j.amepre.2007.03.008. [DOI] [PubMed] [Google Scholar]
  • 34.Selby JV, Beal AC, Frank L. The Patient-Centered Outcomes Research Institute (PCORI) national priorities for research and initial research agenda. JAMA. 2012;307(15):1583–4. doi: 10.1001/jama.2012.500. [DOI] [PubMed] [Google Scholar]
  • 35.Patient-Centered Primary Care Collaborative 2013 www.pcpcc.org/
  • 36.Krist AH, Woolf SH, Rothemich SF, et al. Interactive preventive health record to enhance delivery of recommended care: a randomized trial. Ann Fam Med. 2012;10(4):312–9. doi: 10.1370/afm.1383. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

01
02

RESOURCES