Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Aug 1.
Published in final edited form as: Curr HIV/AIDS Rep. 2019 Aug;16(4):335–348. doi: 10.1007/s11904-019-00455-4

Design Considerations for Implementing eHealth Behavioral Interventions for HIV Prevention in Evolving Sociotechnical Landscapes

Dennis H Li 1,2,*, C Hendricks Brown 3,4, Carlos Gallo 3,4, Ethan Morgan 1, Patrick S Sullivan 5, Sean D Young 6,7, Brian Mustanski 1,2
PMCID: PMC6837839  NIHMSID: NIHMS1055264  PMID: 31250195

Abstract

Background:

Despite tremendous potential for public health impact and continued investments in development and evaluation, it is rare for eHealth behavioral interventions to be implemented broadly in practice. Intervention developers may not be planning for implementation when designing technology-enabled interventions, thus creating greater challenges for real-world deployment following a research trial. To facilitate faster translation to practice, we aimed to provide researchers and developers with an implementation-focused approach and set of design considerations as they develop new eHealth programs.

Methods:

Using the Accelerated Creation-to-Sustainment model as a lens, we examined challenges and successes experienced during the development and evaluation of four diverse eHealth HIV prevention programs for young men who have sex with men: Keep It Up!, Harnessing Online Peer Education, Guy2Guy, and HealthMindr. HIV is useful for studying eHealth implementation because of the substantial proliferation of diverse eHealth interventions with strong evidence of reach and efficacy and the responsiveness to rapid and radical disruptions in the field.

Results:

Rather than locked-down products to be disseminated, eHealth interventions are complex sociotechnical systems that require continual optimization, vigilance to monitor and troubleshoot technological issues, and decision rules to refresh content and functionality while maintaining fidelity to core intervention principles. Platform choice and sociotechnical relationships (among end users, implementers, and the technology) heavily influence implementation needs and challenges. We present a checklist of critical implementation questions to address during intervention development.

Conclusion:

In the absence of a clear path forward for eHealth implementation, deliberate design of an eHealth intervention’s service and technological components in tandem with their implementation plans is critical to mitigating barriers to widespread use. The design considerations presented can be used by developers, evaluators, reviewers, and funders to prioritize the pragmatic scalability of eHealth interventions in research.

Keywords: eHealth, mHealth, implementation, intervention development, scalability, sustainability, HIV, young men who have sex with men

Background

Over the last two decades, there has been increasing development of patient-facing behavior change interventions that incorporate digital technologies (e.g., web-based multimedia, telephone/video conferencing, mobile and sensor technology, gaming, virtual reality, social media) across health domains to augment and/or fill the gaps of traditional prevention and treatment programs [14]. One reason behind this growth is the promise of effectiveness: eHealth interventions afford design elements not previously available or feasible with in-person, human-delivered interventions. Features such as personalization, privacy, variable workflow, timecasting, and integration into daily routines and environments are hypothesized to improve engagement and learning, thereby leading to better outcomes [2, 59]. More importantly, eHealth interventions offer the promise of reach: Digital technologies not only allow interventions to circumvent geographic, social, and economic barriers to access [7] but also provide strong opportunities for cost-efficient scalability with fidelity [2, 7]. With capacity to improve both reach and effectiveness, eHealth interventions have tremendous potential for public health impact [10].

Despite continued investments in the development and testing of new eHealth programs [3, 4, 1113], few interventions with demonstrated evidence of effectiveness have achieved widespread use [24, 14]. This gap between research and implementation is not unique to eHealth [15], but the model of establishing effectiveness in incrementally less-controlled settings creates challenges for eventual implementation that are more pronounced for technology-based interventions. First, research-based eHealth interventions are typically developed within highly controlled trials, often with a focus on the technology but limited input from end users, particularly future implementers [16]. Upon moving into practice, an intervention may not be accepted by its intended targets or those delivering it because it does not fit their needs, contexts, or capabilities. Second, although the greatest cost of eHealth interventions is usually borne during development, building out components to make them pragmatic also incurs considerable resources that are rarely budgeted for at the end of a proof-of-concept study [15]. Third, the speed at which clinical innovations move through the research pipeline to the field—about 17 years by one estimate [17]—is far outpaced by the rate of technological advancements and changes in consumer expectations [6, 16]. By the time an eHealth program’s effectiveness has been demonstrated, the technology has almost certainly become outdated [18]. Fourth, limited theoretical guidance exists on how to implement eHealth interventions. Most implementation theories and frameworks begin with the premise that the innovation is a discrete, stable product to be disseminated [1923]. For eHealth, however, there is no option to “freeze” or even “stiffen” interventions in this way because failing to update software, hardware, form, and/or functionality would relegate them to increasing dysfunction and obsolescence [6]. Fifth, eHealth evidenced-based interventions (EBIs) that are ready for implementation must compete in a market with home-grown and commercial programs that have not been rigorously evaluated, making it difficult for consumers to know which to use [20].

To maximize return on research investments in eHealth and realize their potential for greater public health impact, guidance on overcoming these technology-related implementation challenges is needed. Specifically, we believe researchers can prepare for and mitigate many of the aforementioned barriers to implementation via strategic choices during intervention development. Such planning can then be extended as intervention content evolves based on new research findings and technology adapts to changing user preferences [6, 24]. By prioritizing implementation, program developers can minimize stakeholder rejection, reduce build-out costs, keep pace with shifting social and technological (i.e., sociotechnical) factors during and after research trials, and build supports for sustained delivery of their interventions. The aim of this manuscript is thus to articulate a pragmatic approach and set of guiding questions to consider when designing or adapting eHealth behavioral interventions. To achieve this, we describe four rigorously evaluated eHealth HIV programs as exemplars of such implementation considerations during development and evaluation, and we examine challenges and successes experienced in each case.

eHealth in HIV

HIV prevention is a useful domain for studying eHealth implementation because of the substantial proliferation of diverse eHealth interventions [1113, 25], limited resources for prevention, and a nationally coordinated infrastructure sensitive to dynamic changes in the field. Regarding end users, the epidemic in the US disproportionately affects young men who have sex with men (YMSM) [2629], who, because of factors such as stigma against sexual minorities, are difficult to reach through traditional youth settings (e.g., schools, families) [3032]. Targeted face-to-face HIV EBIs have had some success in high-density areas but are insufficient to meet goals for reducing HIV incidence due to economic and structural barriers to implementation [21, 23, 3336]. Even in urban centers, only 28% of HIV-negative YMSM report participating in HIV prevention programs [37]. Simultaneously, YMSM tend to be early adopters and frequent users of digital technologies because they provide opportunities for learning, connection, and expression free from stigma [3845]. Trials of eHealth HIV interventions have accordingly shown exceptionally high acceptability, interest, and actual use, particularly among subgroups at greater risk (e.g., YMSM of color) [4648]. Available evidence also demonstrates comparability to in-person programs in effectiveness at changing HIV risk and protective behaviors [4, 13, 4951].

Regarding implementers, the science of preventing HIV is rapidly and repeatedly transformed by biomedical, behavioral, and structural innovations [28, 52, 53], resulting in relationships among research institutions, government agencies, healthcare providers, community-based organizations, pharmaceutical companies, and YMSM communities that are synergistic, adaptive, and responsive to sociotechnical volatility [54]. As funding for HIV prevention, differing from treatment, becomes more scarce [55], the collective zeal from stakeholders to scale up and scale out [56] eHealth HIV EBIs grows.

Accelerated Creation-to-Sustainment (ACTS) Model

Our analysis is informed by the ACTS model, a “framework for accelerating research and integrating design, evaluation, and sustainable implementation into a unified effort” [16]. As noted, one of the challenges for eHealth implementation research is that most implementation theories focus on delivering and replicating a complete, “locked-down” product. The ACTS model refutes this characterization by explicitly separating eHealth interventions into a service component, representing what end users receive through the intervention from implementers, and a technology component, representing how technology supports delivery of the service. This reconceptualization of eHealth as technology-enabled services instead of human-supported technologies allows for both components to evolve over time and incorporates the role of implementers into design and evaluation [57].

Building on a paradigm shift in implementation science toward a recognition of ongoing change as important for long-term implementation (e.g., dynamic sustainability framework [24]), the three-phase ACTS model proposes that eHealth interventions are initially designed only insofar as to have service protocols, technology prototypes, and implementation plans for both service and technology that are safe and free of significant usability problems (Create phase). The protocol, prototype, and plans are optimized and evaluated on effectiveness and implementation outcomes in a hybrid optimization–effectiveness–implementation trial, combining what would traditionally be multiple distinct studies (OEI Hybrid Trial phase). Finally, research support is removed, leaving in place a fully functioning, integrated, and independently sustained program (Sustainment phase). Across the phases, a process of iterative evaluation and redesign continually fits the intervention to real-world contexts to avoid over-optimization for non-pragmatic conditions, expedite use, and adapt to unpredicted sociotechnical disruptions.

Methods

We applied the ACTS model’s reconceptualization of eHealth interventions to four previously tested HIV prevention programs for YMSM—Keep It Up!, Harnessing Online Peer Education, Guy2Guy, and HealthMindr—to identify where implementation was or should have been considered and draw comparisons across different types of digital technologies. Although the interventions were developed and evaluated using traditional trial methodology, the ACTS framework nonetheless provides a lens through which to examine how each program addressed the need for upkeep and revolutions in the field, most notably the arrival of pre-exposure prophylaxis (PrEP), a medication that prevents HIV transmission [28, 53]. Our goal was to use the lessons learned from these experiences to help accelerate the transition of other interventions from studies to scalable solutions.

Results

Table 1 presents the service, technology, and implementation plans of the eHealth interventions and summaries of the adaptations that have been made over time. The following narratives further describe the interventions along with challenges encountered during each program’s operation and subsequent considerations around intervention design.

Table 1.

Four eHealth HIV prevention interventions for MSM, framed using the Accelerated Creation-to-Sustainment model

Keep It Up! Harnessing Online Peer Education Guy2Guy HealthMindr
Year(s) developed 2007–2009 (1.0); 2012 (1.5); 2012–2013 (2.0); 2017–2018 (2.5); 2018–present (3.0) 2011 (Los Angeles); 2012–2013 (Peru); 2017–present (Los Angeles) 2012–2014 2013–2015
Service As designed • Individual education, motivation, and skills practice
• 7 modules over 3 sessions and 1 booster (later 2)
• Based on IMB model
• For MSM ages 18–29 recruited by CBOs (later, also online)
• Peer-led group education and motivation paired with optional free HIV testing
• 12 weeks
• Based on C-POL model
• For MSM ages 18+ recruited on and offline (mean age = 31, 29)
• Individual and paired education, motivation, and skills practice
• 5–10 daily messages over 5 weeks and a 1-week booster
• Based on IMB model
• For MSM ages 14–18 recruited online
• Individual HIV service coordination and free condom and at-home HIV test ordering
• 4 months (length of pilot study)
• Based on social cognitive theory
• For MSM ages 18+ recruited online (median age = 28)
Adaptations • Continual updates to incorporate new HIV research, particularly around PrEP
• Added second booster starting in 2.0
• Updates to incorporate new HIV research (e.g., PrEP) • No major adaptations to service during trial
• Planned updates to messaging content (e.g., PrEP) and tailoring by linguistic style and context
• Added content and services related to PrEP and STI testing
Technology As designed • Web-based platform
• Includes videos, interactive activities, provider locators, and virtual simulations
• Utilizes Closed Group function on MSM’s personal social media accounts • Utilizes text messaging service on MSM’s personal phones
• Interactive platform that matches AMSM into dyads
• Automated texting from library of standardized messages
• Smartphone application
• Includes assessments, customizable reminders, provider locators, and commodity ordering
Adaptations • Multiple platform changes
• Mobile compatibility
• Updated form, function, and feel to match evolving user expectations
• Attempted to move to different platform to address decreased usability due to changes in Group functionality and resistance to using Facebook • No major adaptations to technology during trial
• Planned automated evaluation of and tailoring to MSM’s linguistic profiles
• Added video content supported by improved cellular bandwidth
• Updated data security and privacy measures to conform to user expectations
• Updated look and feel
Implementation plan As designed • CBOs recruit eligible MSM during HIV testing and counseling and provide link to program for MSM to complete
• Technology hosted by developer
• Peer leaders recruited from CBOs and trained by developer
• Peer leaders deliver content via social media Groups
• Test kits mailed to MSM
• Platform coordinates texts
• Automated and manual monitoring for disclosure of sensitive information (e.g., contact info, mental health)
• Health departments and CBOs deploy and operate the app as a public health strategy
• Developer provides technical assistance
Adaptations • Varied integration of program with other services by CBOs
• Added online recruitment option (direct-to-consumer delivery) with at-home testing in 2.0 and 3.0
• Changed posting protocol because of software behavior
• Partnered with CBO to provide in-person HIV testing option
• Modified incentive payment structure for CBOs
• No major adaptations to implementation plan during trial • Integrated with Amazon.com to streamline commodity shipping
• Continued challenges around recognition as a public health strategy and limited technical capacity among implementers

Note. MSM = men who have sex with men. IMB = information–motivation–behavioral skills. C-POL = community popular opinion leader. CBO = community-based organization. PrEP = pre-exposure prophylaxis.

Keep It Up! (KIU!)

KIU! is an online HIV prevention program designed to get YMSM ages 18–29 who test HIV-negative to “keep it up,” or maintain their negative status, by reducing risk and enacting protections [47]. Through accounts registered to their emails, participants can access KIU! at their own convenience from any web browser (originally not mobile devices). They move linearly through seven modules (approximately one hour total) across three sessions, with a forced day-long break between sessions. There are also two (originally one) booster sessions 3 and 6 months later. Based on the information–motivation–behavioral skills (IMB) model [58], each module focuses on a setting/situation relevant to YMSM (e.g., gay bars, dating) and uses diverse multimedia (e.g., soap opera videos, testimonials, animations, games, testing/clinic locator) to address gaps in HIV knowledge, motivate safer behaviors, teach behavioral skills, and instill self-efficacy for preventive behaviors via active learning, role modeling, dramatic relief, goal setting, and self-reevaluation [59, 60].

Through two randomized controlled trials (RCTs), labeled 1.0 [47] and 2.0 [61, 62], KIU! was shown to be highly acceptable and efficacious in reducing condomless anal sex and incidence of rectal/urethral STIs. KIU! has also been implemented as service projects (KIU! 1.5 and 2.5) by community-based organizations (CBOs) in Chicago, IL [48], and Jackson, MI. A national implementation RCT (KIU! 3.0) comparing two delivery approaches is in progress.

Service Challenges

Originally focused on condom use and testing, the educational components of KIU! have had to respond to changes in HIV prevention. The Food and Drug Administration approved PrEP for high-risk YMSM in 2012, posing a methodological issue for KIU! 2.0 because the intervention content was frozen in the context of the RCT. However, leaving out this information would be unethical and immediately antiquate the intervention. PrEP content was therefore added to a booster session that no participants had yet completed, which allowed for rapid inclusion of information with consistent delivery to everyone in the trial. KIU! 2.5 and 3.0 have been further refreshed to include PrEP and other recent scientific advances (e.g., viral suppression [63]).

Anticipating more such changes, the developers have been deliberate about how new content integrates with technology. Media that are more costly to update (e.g., filmed video, interactive applications) are reserved for relatively stable information (e.g., communication skills) whereas more easily edited media (e.g., digital pamphlets) are used to deliver facts that could change over time (e.g., forms of biomedical prevention). Real and animated characters and scenes (e.g., clothing, hairstyles, music) are kept simple but diverse to maximize shelf lives.

Technology Challenges

The KIU! platform has also had to evolve within a changing technology landscape. KIU! 1.0 was developed for web-based delivery on desktop and laptop computers. In preparation for implementation, the software was transferred in KIU! 1.5 to a university-supported platform that, though designed for collecting patient-reported outcomes rather than delivering interventions [64], was purported to be stable and scalable. However, a sociotechnical shift toward mobile devices saw young people accessing the Internet more through smartphones [40]. Despite requests by KIU! 2.0 participants to view KIU! on their phones, the Adobe-Flash-based platform was not compatible with all mobile devices, thus requiring another resource-intensive switch in KIU! 3.0 to stay current. The latest version is a mobile-responsive website built on an open-source system that has ongoing university support and investment. The developers kept the web platform instead of migrating to a smartphone application so that KIU! would be accessible across a range of devices and more easily updated [65].

The multimedia content has faced similar challenges, as exemplified by a virtual club simulation activity: Created in KIU! 1.0 using Second Life and exported to Flash for easier access and usability, it suffered from the aforementioned mobile-compatibility issues as well as aging graphics and functionality. Updates have been cost-prohibitive until KIU! 3.0, where the simulation is being built anew; however, the developers recognize it will likely require redevelopment within a few years.

Implementation Considerations

Weighing the substantial technical requirements of KIU! against the limited technical capacity of most CBOs, KIU! was designed to be centrally maintained by the developers but integrated into community-based HIV testing as an opportunity to reach diverse YMSM and supplement standard counseling [37, 66]. Across versions, the developers have continually acquired feedback from CBOs and other stakeholders to ensure KIU!’s appropriateness and acceptability upon deployment. One recurring theme was localized tailoring. KIU! 2.0 was conducted in 3 metropolitan areas, so Module 1 videos were filmed in each city, with the idea that future versions could include such kinds of adaptations generally favored by CBOs. Meanwhile, rating and feedback pages were incorporated within the program to provide ongoing monitoring of end-user acceptability.

In practice, there has been unanticipated variation in how implementers deliver KIU!. Contrary to the purported advantage of independence from in-person delivery, CBOs consistently describe value in using KIU! to engage YMSM with services. The KIU! 1.5 CBO flipped the order and used the program as an incentive to bring YMSM in for testing. The KIU! 2.5 CBO is delivering the program on-site, bundled with other services and group discussion. These deviations from the original implementation plan raise questions about how CBOs view eHealth in general. Furthermore, with the advent of low-cost at-home HIV testing, a new approach of delivering the intervention directly to YMSM arose in KIU! 2.0. This direct-to-consumer strategy is now being evaluated in the KIU! 3.0 comparative implementation trial against the CBO approach on outcomes that include reach, responsiveness, and cost-effectiveness. The design of this head-to-head implementation trial [67] provides for distinct strategies within each arm for engaging YMSM, creating sustainable delivery systems as recognized in the ACTS model.

Harnessing Online Peer Education (HOPE)

HOPE is a stigma reduction and behavior change intervention that uses social media (so far Facebook and HealthCheckins) to increase HIV testing among MSM, but it is also tailorable for different populations and areas of need. Following a modified community popular opinion leader model [68], influential members of the target communities are trained to communicate about HIV (in HIV testing studies) and motivate change in their peers’ behaviors over a 12-week period [69]. MSM participants are added to a closed (private) group and instructed to use their social media as they normally do. Peer leaders attempt to engage with their assigned participants around HIV prevention and testing knowledge and attitudes via direct messages, chats, and wall posts, though participants are not required to respond to peer leaders or engage with other participants.

HOPE has been shown to increase HIV self-testing behavior among primarily African American and Latino MSM in Los Angeles, CA [70], and in-person HIV testing among MSM in Lima, Peru [71]. Acceptability and retention in both RCTs were high [72]. A third trial in Los Angeles with modifications for delivery by CBOs is underway. HOPE has been adapted to increase HIV testing among women in jails, increase retention in care among minority MSM living with HIV, reduce addiction and overdose among chronic pain patients on opioid therapy [73, 74], and reduce substance use and underage drinking among youth [75].

Service Challenges

In the HIV trials, peer-led group education was paired with optional free HIV testing, which doubled as an outcome for the studies. The major service challenge has been incorporating updated HIV research findings and clinical tools (e.g., PrEP) into the intervention via peer leader training.

Technology Challenges

Facebook was initially selected as the platform for HOPE because of its high traffic and acceptability among MSM [69, 76]. After conducting the first two efficacious HIV trials using Facebook, a third trial was planned. Before the trial launched, though, Facebook altered its user interface, which included moving Facebook Groups from visibly front and center to a side menu, making it more difficult for people to find and use the feature. Additionally, some MSM expressed resistance to having the intervention on Facebook. Anticipating implementation problems, the developers planned and created another platform (i.e., HealthCheckins) that could mimic the original functionality of and replace Facebook. However, a university health system compliance officer ruled the new third-party software to be unacceptable because it could not be owned by the university or stored on its servers, so the technology reverted to the original platform despite the less-ideal Group functionality.

Implementation Considerations

Using a popular platform like Facebook for HOPE saves on development and maintenance costs and includes additional benefits like a way to verify participant identities via Facebook Connect, an integrated single sign-on feature. However, there are challenges to developing an implementation plan for technology one does not control. The intervention becomes subject to the behaviors of that particular platform, often with unanticipated consequences. For example, the order that posts are displayed in Facebook Groups is determined by Facebook’s proprietary algorithm, which changed over the course of the three HIV trials but most recently appeared stacked in reverse chronological order. Because MSM randomly assigned to the intervention arm were more actively engaged than those assigned to the control (as intended), during the third trial, the testing invitation posts within the intervention Groups kept getting buried underneath other posts, whereas the testing posts in the control Groups remained at the top due to the limited number of other posts. Thus, the developers modified the research protocol to repost the testing invitation multiple times in the intervention Groups so that it could be seen as frequently as in the control.

Another adaptation to the implementation plan occurred in the second trial. As an alternative to requesting an at-home HIV test, participants could visit a local CBO for testing, of which 17% of the intervention arm and almost 7% of the control arm did. This was an encouraging finding for considering alternative dissemination approaches for HOPE. Familiarity with Facebook may also mean CBOs could more easily adopt the intervention. So, while it is unfortunate that the developers’ attempt to better control the implementation of their technology-enabled service by switching to HealthCheckins met with unanticipated sociotechnical barriers (data security policies), the tradeoff could be a lower burden for eventual deployment. However, modifying HOPE for other researchers or service organizations who lack the resources to pay peer leaders has been an ongoing challenge.

Guy2Guy (G2G)

G2G was a comprehensive text-messaging-based HIV prevention program for adolescent MSM ages 14–18 [77]. Every day for 5 weeks, G2G delivered 6 to 8 short, standardized messages to AMSM participants’ phones. Booster text messages were sent approximately 6 weeks after intervention completion. Based on the IMB model [58], message content primarily focused on HIV knowledge, attitudes (e.g., reasons AMSM use condoms), and skills (e.g., correct condom use) and was tailored to participants’ sexual experience. Messages could also include other topics, such as healthy and unhealthy relationships, coming out to parents, and bullying; links to online resources; and interactive quiz questions to which participants could respond. Furthermore, AMSM could text questions to an automated intervention number, which would use text mining to select and return an answer from a library of scripted messages. Finally, AMSM were paired based on demographics, sexual experience, and geographic location, and automated messages would encourage participants to text their partner to discuss HIV-related topics and provide peer support.

During an iterative development process, G2G was deemed acceptable by AMSM [77]. In the pilot RCT, G2G had a positive effect on rates of HIV testing among a national sample, but it did not significantly reduce condomless anal sex compared to the active control [78].

Service Challenges

No major changes to the service were made to G2G during the pilot study, but several adaptations would be required in future iterations. First, none of the original messaging contained information about PrEP because it was not indicated for adolescents at the time of the trial. As these recommendations have since changed [79], new content will need to be integrated into the library of scripted messages while outdated material is refreshed or purged. Second, the original library comprised messages that were tailored to the participants’ sexual experience (i.e., initiated sexual activity versus abstinent) but untailored to individual linguistic style and the linguistic context of the peer-to-peer conversation thread. Current work exploring the optimization of pairing AMSM based on linguistic compatibility suggests that G2G dyads who were closer in style were more engaged [80]. To strengthen the program’s effectiveness, the developers could incorporate this information into the dyad matching algorithm and deliver scripted messages from the library that would be sensitive to each pair’s style and/or context.

Technology Challenges

No major changes to the technology were needed during evaluation. In planning for the aforementioned service adaptations, the platform that pairs AMSM and coordinates messages would need to be updated with algorithms that automatically measure and respond to AMSM’s linguistic profiles.

Implementation Considerations

Because text messaging is an existing feature of phones, reliable and consistent across different mobile platforms, and widely adopted by youth [40], G2G is purportedly scalable without need for graphic or user interface redesign. However, the lack of multimedia content may have contributed to the limited behavioral outcomes, so additional features (e.g., language tailoring) may prove critical to improving effectiveness. Another consideration is the potential cost to participants; whereas the trial included only individuals with unlimited texting plans, this will not be the case for all AMSM.

G2G incorporated some automation into its implementation plan, including algorithms for detecting keywords and patterns that signal problems (e.g., personal contact information, suicidal ideation). However, some participants tried to circumvent the terms of service prohibiting attempts to exchange contact information, so substantial manual monitoring of peer-to-peer conversations was also required. Future scale out would need to plan how to maintain safety and confidentiality, especially because the users are minors.

HealthMindr (HM)

HM is a smartphone native application (i.e., installs onto the phone) that provides a comprehensive package of HIV prevention services for MSM, including monthly HIV risk assessments with tailored feedback; PrEP and PEP eligibility screeners; testing decision support (e.g., recommended frequency based on risk behaviors); a customizable testing plan and results tracker; a testing/clinic locator; customizable reminders; answers to frequently asked questions; and a mechanism to order free condoms, lubricant, and at-home test kits [81]. Based on social cognitive theory [82], these features promote goal setting, self-efficacy, outcome expectations, and self-regulation around multiple protective health behaviors (e.g., testing, condom use, PrEP self-screening), but no components are required; MSM participants who download the app can use it however they see fit.

HM was developed after an extensive formative research process that included several cycles of input from potential app users, HIV prevention counselors, health department officials, and federal funders of HIV prevention programs [83, 84]. An uncontrolled pilot study of HM showed high (>50%) utilization of planning for a regular HIV testing schedule, ordering of at-home HIV test kits, and ordering of condoms and condom-compatible lubricant for home delivery [81]. Additionally, 9% of PrEP-eligible men initiated PrEP during the 4-month evaluation period. HM was found to be both usable and acceptable by MSM. A multicity effectiveness RCT is underway.

Service Challenges

Over the period of HM development and evaluation, there were important changes in the HIV/STI prevention environment. Awareness of, willingness to use, and use of PrEP increased substantially [85]; state and local programs to offset costs of PrEP became more common; and a national directory of PrEP providers was developed [86]. In response, content about available PrEP navigation programs and a geolocation-based locator of PrEP providers (via an automated interface to a live database) were added to the app. Furthermore, mail-out kits for specimen self-collection for uretheral, pharyngeal, and rectal STI testing were found to be very acceptable to MSM in research settings [87, 88], so the commodity ordering function in HM was modified to include mail-out STI kits.

Technology Challenges

Considerable changes in the sociotechnical landscape occurred during the same period, which impacted the transition of HM from evaluation to implementation. For example, as cellular bandwidth has increased, so have opportunities to include video content in mobile interventions and demand for such from users, who may find static content less engaging over time. Consequently, HM was modified to present more video in place of text. In another example, public disclosures, data breaches, and hacking of social media platforms continue to legitimate concerns about data security and privacy in sexual health apps [89]. It was necessary to modify HM to incorporate additional security measures (drawing, for instance, on the advent of biometric identification tools in operating systems) and more explicit and visible mechanisms to secure the most sensitive health and sexual risk behavior data. More generally, the look and feel of smartphone apps has evolved, requiring maintenance to keep the app contemporary and functioning across different devices.

Implementation Considerations

Regarding the service implementation plan, because of the increased volume of prevention supplies available for ordering, HM integrated ordering and shipping with Amazon.com to streamline workflow for fulfilling test kit and condom orders and to provide a consistent level of service in delivery times. More broadly, however, two challenges that were not fully anticipated, despite extensive formative work to lay out the path for eventual implementation, have delayed rollout of HM. First, most funding for HIV prevention programs in the US is provided by the CDC to health departments and CBOs, and only EBIs and recognized public health strategies can be supported by these funds. There is a question about whether HM is a standalone intervention, which requires more stringent evidence of efficacy, or a strategy to promote the efficient provision of prevention tools known to be impactful. If the latter, then distribution of HIV test kits and referrals to care through HM would be eligible for immediate uptake by CDC grantees, but the process to evaluate eHealth tools as public health strategies is not as well described as the process to evaluate interventions [90]. Second, most health departments and CBOs have minimal technical capacity to manage the deployment and operations of an eHealth intervention. To support widespread use of HM in community settings, the CDC must delineate a more explicit process for evaluating eHealth programs as public health strategies and establish capacity for technical assistance for eHealth.

Discussion

Examining our experiences developing KIU!, HOPE, G2G, and HM and addressing emergent challenges throughout their evaluations and implementations, we have identified several lessons that have practical implications for planning future eHealth programs for HIV and other health domains.

  1. Continual adjustment to the ACTS model targets (service, technology, implementation plan) should be expected. Whereas the iterative optimization process outlined in the ACTS model suggests these adjustments are small and incremental, though, we encountered major and radical interruptions during implementation (e.g., move toward smartphones, introduction of home-based testing). We posit that such sociotechnical disruptions are common, especially in a field with frequent innovations. Therefore, contingencies including time, resources, and processes should be incorporated into the implementation plan and, if possible, built into the technology. This is particularly true of multimedia interventions (e.g., KIU!, HM), as some formats are more costly to update than others. Such planning may limit developers’ technology options, but to ignore these inevitabilities would likely reduce the life expectancy of the intervention. However, that is not to say that changes should be viewed negatively. Dynamic models of implementation [24], including the ACTS model, reject the notion that deviations from the original intervention will inherently produce a sub-optimal effect. Conversely, adaptability to evolving contexts, whether in terms of technology (e.g., more video content in HM), service (e.g., new PrEP content in KIU!), or implementation (e.g., more testing options in HOPE), can potentially improve intervention quality and is key to maintaining long-term relevance and viability for an eHealth program.

  2. Because adjustments for optimization and disruptions are expected, one cannot just “set and forget” an eHealth intervention. Implementation requires both vigilance to catch issues and capacity to troubleshoot/fix them. Under traditional models of HIV EBI dissemination [91], few community settings would have the resources to solely manage a technology-heavy intervention, constraining scalability of the program. Thus, intervention design should consider the capabilities not only of the technology but also of its end users. Implementation plans that share the cost burden of eHealth intervention delivery, such as KIU!’s coordinating center and direct-to-consumer models, can help alleviated these challenges.

  3. It is important to distinguish between minor updates (e.g., bug fixes, system upgrades, additional device support) and major content or functionality changes that alter or improve users’ experience and benefits. Though ongoing optimization is desirable, to what extent does adding features around STI testing, for example, make HM a different intervention than its prototype? To respond to this reality while maintaining scientific rigor in studying eHealth programs, we must move beyond traditional RCTs of frozen or locked-down products. One approach is to shift the paradigm to evaluating core intervention principles, which may take on different forms but retain the same fundamental function [6, 16]. Operationally, this includes (a) identifying methods (e.g., peer norms) and instantiation strategies (e.g., role model stories) that underlie behavior change, which are part of the service component that is constant during a trial; (b) partnering with stakeholders to identify new technology features or scientific content to be considered for adoption; (c) matching proposed changes to the behavior change methods and strategies and deciding if the changes interfere with the core principles being tested; and (d) monitoring and testing usability and acceptability to ensure functionality is maintained.

    When the content and/or delivery systems of an eHealth EBI evolve more substantially, researchers may choose to establish new evidence of efficacy. In cases where the implementation plan generally remains the same but the service and/or technology change, new participants can be randomized to older versus newer versions of the intervention. In cases where there is a change in implementation plan, researchers can pinpoint which system components need testing and which components can “borrow strength” from previous studies [56]. Novel study designs, such as factorial experiments [92] and sequential multiple-assignment randomized trials [93], may be used to isolate the effects of specific combinations of components, whereas optimization frameworks, like the ACTS model and the multiphase optimization strategy [94], can guide that process.

  4. The choice of platform for an eHealth intervention has far-reaching ramifications for its implementation needs, and different technologies have their own unique challenges. A complex intervention with multiple components built on a custom or proprietary platform requires high technical expertise to manage and update (e.g., KIU!, HM). Conversely, a platform that is maintained by a widely used third party delegates upkeep responsibilities but also relinquishes control over changes in function (e.g., HOPE). Other factors like adaptability, data and security issues, cost to participants (e.g., texting in G2G), and funding streams (e.g., prevention grants for HM) are tied to platform choice, all of which influence pragmatic scalability: Platforms that are not widely used, do not have extensive and ongoing external support, or are esoteric to one institution are more likely to encounter challenges. Intervention developers should carefully weigh the pros and cons of selected technologies for scalability early in the design process.

  5. Understanding all aspects of user engagement is critical. The ACTS model distinguishes three relationships in a sociotechnical system: between the participant and the technology, the implementer and the technology, and the participant and the implementer [16]. Of the three, the first receives the most attention because acceptability and usability of an intervention by its target population are major factors in effectiveness and successful implementation [3, 95]. Within HIV prevention, eHealth interventions have thus far had high acceptability among YMSM in trials, but retention drops outside the research context [48]. Local adaptations; better technology, tailoring, interactivity, and graphics; and more human involvement have been shown to increase engagement and utilization, but the added complexity may come at the cost of greater resource demand and upkeep across the life of the program [2].

    The implementer–technology interaction encompasses implementers’ attitudes about the intervention, an important implementation outcome [95], but it also represents behind-the-scenes tasks required to administer the program that are often neglected in non-pragmatic studies. For example, eHealth technology tends to be built or calibrated to meet research needs like efficiently tracking participants, monitoring fidelity, and delivering assessments. Though some features may be useful for future implementers as well, the software is not typically deployment-ready, requiring modifications later. These additional costs could be mitigated by constructing pragmatic interfaces for implementers from the start. Data management is another example. Tremendous amounts of data and meta-data, such as the number of clicks on a feature or the content of typed messages, can be captured from eHealth interventions but will overwhelm organizations that do not have the capacity to store, handle, analyze, and interpret those data [16]. Data visualizations and dashboards with monitoring algorithms like in G2G can help implementers filter through large volumes of information to triage cases that need attention.

    Finally, the participant–implementer connection is often overlooked in eHealth, but reframing eHealth interventions as technology-enabled services identifies important human interactions necessary for implementation. Participant recruitment is often considered a function of research, so enrollment drops during real-world implementation [48, 96]. However, the ability to reach and retain participants is key to scaling up an intervention. Those processes must be planned as much as the technology itself, but who provides these critical roles of recruitment, support, and engagement in practice is an unanswered question.

Design Considerations for eHealth Intervention Implementation

From these lessons, we distilled a practical checklist of design decisions for eHealth intervention developers to think through in order to prevent or moderate issues that may hinder implementation of their programs (Table 2). Several guidelines for eHealth already exist in the literature, but they primarily focus on categorizing an intervention’s applications and functions (e.g., mHealth and ICT Framework [97]) or reporting on its core features and evaluation methods (e.g., CONSORT-EHEALTH [98], mHealth Evidence Reporting and Assessment [99]). Although referring to these guides a priori may help developers plan out their study designs, with maybe a nod toward future adaptability and scalability, the guides are intended to apply after an intervention has been created and tested, as a way to comprehensively codify the research evidence. In contrast, our checklist is meant to be used during intervention development. Rather than describe a finished product, we hope that it stimulates thinking around how service, technology, and implementation plans are intricately linked and must be designed in tandem so that eHealth intervention developers, along with reviewers and funders evaluating the merits of their proposals, can prioritize the pragmatic scalability of their programs. The questions outlined therein are also not meant to be an exhaustive solution to all implementation challenges for eHealth interventions, but they do represent at minimum what must be asked and answered for all technology-enabled behavior change services. In the absence of a clear path forward for eHealth implementation, addressing these considerations early on will set up program developers to better traverse the gap between research trials and widespread use.

Table 2.

Critical implementation questions for eHealth intervention design

Question
1. What is the service being provided?
    a. What are the participant–implementer interactions?

2. What is the technology being used?
    a. What are the participant–technology interactions?
    b. What are the implementer–technology interactions?

3. How is acceptability by end users established and monitored?
    a. For participants?
    b. For implementers?

4. How pragmatic/scalable is the technology?
    a. How widely is it used?
    b. How difficult is it to maintain?
    c. How adaptable is it to local contexts?
    d. How much does it cost end users?

5. How are data collected, handled, stored, and presented?
    a. What are the legal ramifications outside of research?

6. How does the intervention monitor for adjustments and disruptions?

7. How does the intervention respond to adjustments and disruptions?
    a. Content (e.g., new information)?
    b. Technical (e.g., incompatible software update)?
    c. Participant–technology relationship (e.g., shift away from platform)?
    d. Implementer–technology relationship (e.g., delivery not as intended)?

8. How does the intervention integrate with other services in the service setting?

9. What human requirements/skills are needed to implement the intervention?

10. What technical requirements/skills are needed to implement the intervention?

11. What dissemination model is needed to support implementation of the intervention?
    a. Implementer training?
    b. Technical assistance?
    c. Funding stream?

As a caveat, if researchers try to address only implementation issues up front, they could succeed at implementation but accidentally fail at proof of concept. Our checklist should be applied alongside the ACTS model and established protocols for designing behavioral change interventions [100]. It is also impossible to predict and optimize for all possible futures and sociotechnical disruptions, so developers and their technology teams must balance the tension between adaptability and parsimony. The tradeoffs are many, and there are no right answers; the only wrong answer is to not consider implementation issues before embarking on the costly endeavor of creating a new program.

Conclusions

As eHealth interventions continue to propagate in HIV and across other health domains, it is critical that we establish new paradigms in program design, evaluation, and implementation that keep pace with the rapidly shifting dynamics of modern sociotechnical landscapes. Just as we upgrade technology, so too must we employ new frameworks, like the ACTS model [16], as well as new methods (e.g., trials of intervention principles [6], multiphase optimization strategy [94]); paradigms (e.g., dynamic sustainability framework [24]); and research agendas (e.g., disruptive innovations [20]). We propose our design considerations for eHealth behavioral intervention implementation to contribute to the burgeoning science in this area and to aid developers, evaluators, reviewers, and funders in achieving a future of eHealth intervention scalability.

Acknowledgements

We thank all of our Ce-PIM colleagues and community advisors for their contributions and are grateful for current and prior funding from the National Institutes of Health and other sponsors.

Funding

This study is funded by the National Institutes of Health and other sponsors. Development, evaluation, and implementation of Keep It Up! was previously sponsored by the National Institute of Mental Health (R34MH079714, Mustanski), National Institute on Drug Abuse (R01DA035145, Mustanski), Chicago Department of Public Health, and ViiV Healthcare. Harnessing Online Peer Education was previously sponsored by the National Institute of Mental Health (K01MH090884, Young) and UCLA Center for AIDS Research (P30AI028697, Zack), with ongoing sponsorship from the former (R01MH106415, Young). Guy2Guy was previously sponsored by the National Institute of Mental Health (R01MH096660, Mustanski & Ybarra), with ongoing work supported by the Third Coast Center for AIDS Research (P30AI117943, D’Aquila). HealthMindr was previously sponsored by the MAC AIDS Fund, Emory Center for AIDS Research (P30AI050409, Del Rio), and National Institute on Drug Abuse (R01DA045612, Sullivan).

This manuscript was a product of a workgroup on eHealth interventions formed through the Center for Prevention Implementation Methodology for Drug Abuse and HIV (Ce-PIM), which is supported by the National Institute on Drug Abuse (P30DA027828, Brown & Mustanski). This work was also supported by the Keep It Up! 3.0 grant from the National Institute of Mental Health, National Institute on Drug Abuse, Office of Behavioral and Social Sciences Research, and Office of Disease Prevention at the National Institutes of Health (R01MH118213, Mustanski), as well as an individual postdoctoral fellowship grant from the National Institute on Drug Abuse (F32DA046313, Morgan). The sponsors had no involvement in the conduct of the research or the preparation of the manuscript. The content is the sole responsibility of the authors and does not necessarily represent the official views of the funders.

Dr. Sullivan reports grants and personal fees from NIH, grants and personal fees from CDC, grants and personal fees from Gilead Sciences outside the submitted work.

Dr. Li reports grants from National Institutes of Health, during the conduct of the study.

Dr. Brown reports grants from National Institute on Drug Abuse/NIH, during the conduct of the study.

Dr. Mustanski reports grants from National Institutes of Health, during the conduct of the study.

Dr. Morgan reports grants from National Institutes of Health, during the conduct of the study.

List of abbreviations

(A/Y)MSM

(adolescent/young) men who have sex with men

ACTS

Accelerated Creation-to-Sustainment (model)

CBO

community-based organization

C-POL

Community Popular Opinion Leader (model)

EBI

evidence-based intervention

G2G

Guy2Guy

HIV

human immunodeficiency virus

HM

HealthMindr

HOPE

Harnessing Online Peer Education

IMB

Information–Motivation–Behavioral Skills (model)

KIU!

Keep It Up!

PrEP

pre-exposure prophylaxis

RCT

randomized controlled trial

STI

sexually transmitted infection

Footnotes

Conflict of Interest: Drs. Gallo and Young have nothing to disclose.

Human and Animal rights and Informed Consent: This article does not contain any studies with human or animal subjects performed by any of the authors.

Availability of Data and Material: Data sharing is not applicable to this article as no datasets were generated or analyzed during the current study.

Competing Interests: The authors declare that they have no competing interests.

References

  • 1.Oh H, Rizo C, Enkin M, Jadad A. What is eHealth (3): a systematic review of published definitions. J Med Internet Res. 2005;7(1):e1. doi: 10.2196/jmir.7.1.e1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Bennett GG, Glasgow RE. The delivery of public health interventions via the Internet: actualizing their potential. Annu Rev Public Health. 2009;30:273–92. doi: 10.1146/annurev.publhealth.031308.100235. [DOI] [PubMed] [Google Scholar]
  • 3.Mohr DC, Burns MN, Schueller SM, Clarke G, Klinkman M. Behavioral intervention technologies: evidence review and recommendations for future research in mental health. Gen Hosp Psychiatry. 2013;35(4):332–8. doi: 10.1016/j.genhosppsych.2013.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Bailey J, Mann S, Wayal S, Hunter R, Free C, Abraham C, Murray E. Sexual health promotion for young people delivered via digital media: A scoping review. Public Health Research. 2015;3(13). doi: 10.3310/phr03130. [DOI] [PubMed] [Google Scholar]
  • 5.Ybarra ML, Bull SS. Current trends in Internet- and cell phone-based HIV prevention and intervention programs. Curr HIV/AIDS Rep. 2007;4(4):201–7. [DOI] [PubMed] [Google Scholar]
  • 6.Mohr DC, Schueller SM, Riley WT, Brown CH, Cuijpers P, Duan N, Kwasny MJ, Stiles-Shields C, Cheung K. Trials of intervention principles: Evaluation methods for evolving behavioral intervention technologies. J Med Internet Res. 2015;17(7):e166. doi: 10.2196/jmir.4391. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Schueller SM, Munoz RF, Mohr DC. Realizing the potential of behavioral intervention technologies. Curr Dir Psychol Sci. 2013;22(6):478–83. [Google Scholar]
  • 8.Brown CH, Mohr DC, Gallo CG, Mader C, Palinkas L, Wingood G, Prado G, Kellam SG, Pantin H, Poduska J, Gibbons R, McManus J, Ogihara M, Valente T, Wulczyn F, Czaja S, Sutcliffe G, Villamar J, Jacobs C. A computational future for preventing HIV in minority communities: how advanced technology can improve implementation of effective programs. J Acquir Immune Defic Syndr. 2013;63 Suppl 1:S72–84. doi: 10.1097/QAI.0b013e31829372bd. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Noar SM, Pierce LB, Black HG. Can computer-mediated interventions change theoretical mediators of safer sex? A meta-analysis. Human Communication Research. 2010;36(3):261–97. doi: 10.1111/j.1468-2958.2010.01376.x. [DOI] [Google Scholar]
  • 10.Glasgow RE, Vogt TM, Boles SM. Evaluating the public health impact of health promotion interventions: the RE-AIM framework. Am J Public Health. 1999;89(9):1322–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Muessig KE, Nekkanti M, Bauermeister J, Bull S, Hightow-Weidman LB. A systematic review of recent smartphone, Internet and Web 2.0 interventions to address the HIV continuum of care. Curr HIV/AIDS Rep. 2015;12(1):173–90. doi: 10.1007/s11904-014-0239-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Hightow-Weidman LB, Muessig KE, Bauermeister J, Zhang C, LeGrand S. Youth, Technology, and HIV: Recent Advances and Future Directions. Curr HIV/AIDS Rep. 2015;12(4):500–15. doi: 10.1007/s11904-015-0280-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Noar SM. Computer technology-based interventions in HIV prevention: state of the evidence and future directions for research. AIDS Care. 2011;23(5):525–33. doi: 10.1080/09540121.2010.516349. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Kemp CG, Velloza J. Implementation of eHealth Interventions Across the HIV Care Cascade: a Review of Recent Research. Curr HIV/AIDS Rep. 2018. doi: 10.1007/s11904-018-0415-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Brownson RC, Jacobs JA, Tabak RG, Hoehner CM, Stamatakis KA. Designing for dissemination among public health researchers: findings from a national survey in the United States. Am J Public Health. 2013;103(9):1693–9. doi: 10.2105/AJPH.2012.301165. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Mohr DC, Lyon AR, Lattie EG, Reddy M, Schueller SM. Accelerating Digital Mental Health Research From Early Design and Creation to Successful Implementation and Sustainment. J Med Internet Res. 2017;19(5):e153. doi: 10.2196/jmir.7725. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Balas EA, Boren SA. Managing Clinical Knowledge for Health Care Improvement. Yearb Med Inform. 2000(1):65–70. [PubMed] [Google Scholar]
  • 18.Allison S, Bauermeister JA, Bull S, Lightfoot M, Mustanski B, Shegog R, Levine D. The intersection of youth, technology, and new media with sexual health: moving the research agenda forward. J Adolesc Health. 2012;51(3):207–12. doi: 10.1016/j.jadohealth.2012.06.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Tabak RG, Khoong EC, Chambers DA, Brownson RC. Bridging research and practice: models for dissemination and implementation research. Am J Prev Med. 2012;43(3):337–50. doi: 10.1016/j.amepre.2012.05.024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Rotheram-Borus MJ, Swendeman D, Chorpita BF. Disruptive innovations for designing and diffusing evidence-based interventions. Am Psychol. 2012;67(6):463–76. doi: 10.1037/a0028180. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Bell SG, Newcomer SF, Bachrach C, Borawski E, Jemmott JB 3rd, Morrison D, Stanton B, Tortolero S, Zimmerman R. Challenges in replicating interventions. J Adolesc Health. 2007;40(6):514–20. doi: 10.1016/j.jadohealth.2006.09.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Kilbourne AM, Neumann MS, Pincus HA, Bauer MS, Stall R. Implementing evidence-based interventions in health care: application of the replicating effective programs framework. Implement Sci. 2007;2:42. doi: 10.1186/1748-5908-2-42. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Neumann MS, Sogolow ED. Replicating effective programs: HIV/AIDS prevention technology transfer. AIDS Educ Prev. 2000;12(5 Suppl):35–48. [PubMed] [Google Scholar]
  • 24.Chambers DA, Glasgow RE, Stange KC. The dynamic sustainability framework: addressing the paradox of sustainment amid ongoing change. Implement Sci. 2013;8:117. doi: 10.1186/1748-5908-8-117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Henny KD, Wilkes AL, McDonald CM, Denson DJ, Neumann MS. A Rapid Review of eHealth Interventions Addressing the Continuum of HIV Care (2007–2017). AIDS Behav. 2018;22(1):43–63. doi: 10.1007/s10461-017-1923-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Centers for Disease Control and Prevention. HIV among youth. 2016. http://www.cdc.gov/hiv/group/age/youth/. Accessed 18 Nov 2018. [PubMed]
  • 27.Centers for Disease Control and Prevention. CDC Fact Sheet: HIV among Gay and Bisexual Men. 2017. https://www.cdc.gov/nchhstp/newsroom/docs/factsheets/cdc-msm-508.pdf. Accessed 18 Nov 2018.
  • 28.Centers for Disease Control and Prevention. HIV prevention in the United States: New opportunities, new expectations. Atlanta, GA: CDC, 2015. [Google Scholar]
  • 29.Mustanski B, Newcomb M, Du Bois SN, Garcia SC, Grov C. HIV in young men who have sex with men: A review of epidemiology, risk and protective factors, and interventions. J Sex Res. 2011;48(2–3):218–53. doi: 10.1080/00224499.2011.558645. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Barrett B, Bound AM. A Critical Discourse Analysis of No Promo Homo Policies in US Schools. Educational Studies. 2015;51(4):267–83. doi: 10.1080/00131946.2015.1052445. [DOI] [Google Scholar]
  • 31.D’Amico E, Julien D. Disclosure of sexual orientation and gay, lesbian, and bisexual youths’ adjustment: Associations with past and current parental acceptance and rejection. Journal of GLBT Family Studies. 2012;8(3):215–42. doi: 10.1080/1550428X.2012.677232. [DOI] [Google Scholar]
  • 32.Kuper LE, Coleman BR, Mustanski BS. Coping With LGBT and Racial-Ethnic-Related Stressors: A Mixed-Methods Study of LGBT Youth of Color. J Res Adolesc. 2014;24(4):703–19. doi: 10.1111/jora.12079. [DOI] [Google Scholar]
  • 33.Swendeman D, Rotheram-Borus MJ. Innovation in sexually transmitted disease and HIV prevention: internet and mobile phone delivery vehicles for global diffusion. Curr Opin Psychiatry. 2010;23(2):139–44. doi: 10.1097/YCO.0b013e328336656a. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Kelly JA, Spielberg F, McAuliffe TL. Defining, designing, implementing, and evaluating phase 4 HIV prevention effectiveness trials for vulnerable populations. J Acquir Immune Defic Syndr. 2008;47 Suppl 1:S28–33. doi: 10.1097/QAI.0b013e3181605c77. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Rietmeijer CA. Risk reduction counselling for prevention of sexually transmitted infections: how it works and how to make it work. Sex Transm Infect. 2007;83(1):2–9. doi: 10.1136/sti.2006.017319. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Solomon J, Card JJ, Malow RM. Adapting efficacious interventions: Advancing translational research in HIV prevention. Eval Health Prof. 2006;29(2):162–94. [DOI] [PubMed] [Google Scholar]
  • 37.Centers for Disease Control and Prevention. HIV infection risk, prevention, and testing behaviors among men who have sex with men - National HIV Behavioral Surveillance, 20 U.S. cities, 2014. 2016.
  • 38.Harper GW, Bruce D, Serrano P, Jamil OB. The role of the Internet in the sexual identity development of gay and bisexual male adolescents. In: Hammack PL, Cohler BJ, editors. The story of sexual identity: Narrative perspectives on the gay and lesbian life course. New York, NY: Oxford University Press; 2009. p. 297–326. [Google Scholar]
  • 39.Macapagal K, Moskowitz DA, Li DH, Carrion A, Bettin E, Fisher CB, Mustanski B. Hookup app use, sexual behavior, and sexual health among adolescent men who have sex with men in the United States. J Adolesc Health. 2018;62(6):708–15. doi: doi: 10.1016/j.jadohealth.2018.01.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Lenhart A. Teen, Social Media and Technology Overview 2015. Pew Research Center, 2015.
  • 41.Garofalo R, Herrick A, Mustanski BS, Donenberg GR. Tip of the iceberg: young men who have sex with men, the Internet, and HIV risk. Am J Public Health. 2007;97(6):1113–7. doi: 10.2105/AJPH.2005.075630. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Bolding G, Davis M, Hart G, Sherr L, Elford J. Where young MSM meet their first sexual partner: the role of the Internet. AIDS Behav. 2007;11(4):522–6. [DOI] [PubMed] [Google Scholar]
  • 43.Paul JP, Ayala G, Choi KH. Internet sex ads for MSM and partner selection criteria: the potency of race/ethnicity online. J Sex Res. 2010;47(6):528–38. doi: 10.1080/00224490903244575. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Chiasson MA, Hirshfield S, Rietmeijer C. HIV prevention and care in the digital age. J Acquir Immune Defic Syndr. 2010;55 Suppl 2:S94–7. doi: 10.1097/QAI.0b013e3181fcb878. [DOI] [PubMed] [Google Scholar]
  • 45.Mustanski B, Lyons T, Garcia SC. Internet use and sexual health of young men who have sex with men: a mixed-methods study. Arch Sex Behav. 2011;40(2):289–300. doi: 10.1007/s10508-009-9596-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Bauermeister JA, Pingel ES, Jadwin-Cakmak L, Harper GW, Horvath K, Weiss G, Dittus P. Acceptability and preliminary efficacy of a tailored online HIV/STI testing intervention for young men who have sex with men: the Get Connected! program. AIDS Behav. 2015;19(10):1860–74. doi: 10.1007/s10461-015-1009-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Mustanski B, Garofalo R, Monahan C, Gratzer B, Andrews R. Feasibility, acceptability, and preliminary efficacy of an online HIV prevention program for diverse young men who have sex with men: the Keep It Up! intervention. AIDS Behav. 2013;17(9):2999–3012. doi: 10.1007/s10461-013-0507-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Greene GJ, Madkins K, Andrews K, Dispenza J, Mustanski B. Implementation and Evaluation of the Keep It Up! Online HIV Prevention Intervention in a Community-Based Setting. AIDS Educ Prev. 2016;28(3):231–45. doi: 10.1521/aeap.2016.28.3.231. [DOI] [PubMed] [Google Scholar]
  • 49.Guse K, Levine D, Martins S, Lira A, Gaarde J, Westmorland W, Gilliam M. Interventions using new digital media to improve adolescent sexual health: a systematic review. J Adolesc Health. 2012;51(6):535–43. doi: 10.1016/j.jadohealth.2012.03.014. [DOI] [PubMed] [Google Scholar]
  • 50.Noar SM, Willoughby JF. eHealth interventions for HIV prevention. AIDS Care. 2012;24(8):945–52. doi: 10.1080/09540121.2012.668167. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Gabarron E, Wynn R. Use of social media for sexual health promotion: a scoping review. Global Health Action. 2016;9. doi: 10.3402/gha.v9.32193. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Rotheram-Borus MJ, Swendeman D, Chovnick G. The past, present, and future of HIV prevention: integrating behavioral, biomedical, and structural intervention strategies for the next generation of HIV prevention. Annu Rev Clin Psychol. 2009;5:143–67. doi: 10.1146/annurev.clinpsy.032408.153530. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Rausch DM, Grossman CI, Erbelding EJ. Integrating behavioral and biomedical research in HIV interventions: challenges and opportunities. J Acquir Immune Defic Syndr. 2013;63 Suppl 1:S6–11. doi: 10.1097/QAI.0b013e318292153b. [DOI] [PubMed] [Google Scholar]
  • 54.Centers for Disease Control and Prevention. Supported activities: Prioritizing high impact HIV prevention. 2015. https://www.cdc.gov/hiv/programresources/healthdepartments/supportedactivities.html. Accessed 18 Nov 2018.
  • 55.Kaiser Family Foundation. U.S. Federal Funding for HIV/AIDS: Trends Over Time. 2017. https://www.kff.org/global-health-policy/fact-sheet/u-s-federal-funding-for-hivaids-trends-over-time/. Accessed 18 Nov 2018.
  • 56.Aarons GA, Sklar M, Mustanski B, Benbow N, Brown CH. “Scaling-out” evidence-based interventions to new populations or new health care delivery systems. Implement Sci. 2017;12(1):111. doi: 10.1186/s13012-017-0640-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Mohr DC, Weingardt KR, Reddy M, Schueller SM. Three Problems With Current Digital Mental Health Research … and Three Things We Can Do About Them. Psychiatr Serv. 2017;68(5):427–9. doi: 10.1176/appi.ps.201600541. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Fisher JD, Fisher WA, Williams SS, Malloy TE. Empirical tests of an information-motivation-behavioral skills model of AIDS-preventive behavior with gay men and heterosexual university students. Health Psychology. 1994;13(3):238–50. [DOI] [PubMed] [Google Scholar]
  • 59.Fisher CM. Adapting the information-motivation-behavioral skills model: Predicting HIV-related sexual risk among sexual minority youth. Health Educ Behav. 2012;39(3):290–302. doi: 10.1177/1090198111406537. [DOI] [PubMed] [Google Scholar]
  • 60.Fisher JD, Fisher WA. The information-motivation-behavioral skills model. In: DiClemente RJ, Crosby RA, Kegler MC, editors. Emerging theories in health promotion practice and research: Strategies for improving public health. San Francisco: Jossey-Bass; 2002. p. 40–70. [Google Scholar]
  • 61.Mustanski B, Madkins K, Greene GJ, Parsons JT, Johnson BA, Sullivan P, Bass M, Abel R. Internet-based HIV prevention with at-home sexually transmitted infection testing for young men having sex with men: study protocol of a randomized controlled trial of Keep It Up! 2.0. JMIR Res Protoc. 2017;6(1):e1. doi: 10.2196/resprot.5740. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Mustanski B, Parsons JT, Sullivan PS, Madkins K, Rosenberg E, Swann G. Biomedical and Behavioral Outcomes of Keep It Up!: An eHealth HIV Prevention Program RCT. Am J Prev Med. 2018;55(2):151–8. doi: 10.1016/j.amepre.2018.04.026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Centers for Disease Control and Prevention. Evidence of HIV Treatment and Viral Suppression in Preventing the Sexual Transmission of HIV. 2018. https://www.cdc.gov/hiv/pdf/risk/art/cdc-hiv-art-viral-suppression.pdf. Accessed 18 Nov 2018.
  • 64.Gershon R, Rothrock NE, Hanrahan RT, Jansky LJ, Harniss M, Riley W. The development of a clinical outcomes survey research application: Assessment Center. Qual Life Res. 2010;19(5):677–85. doi: 10.1007/s11136-010-9634-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Turner-McGrievy GM, Hales SB, Schoffman DE, Valafar H, Brazendale K, Weaver RG, Beets MW, Wirth MD, Shivappa N, Mandes T, Hebert JR, Wilcox S, Hester A, McGrievy MJ. Choosing between responsive-design websites versus mobile apps for your mobile behavioral intervention: presenting four case studies. Transl Behav Med. 2017;7(2):224–32. doi: 10.1007/s13142-016-0448-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Du Bois SN, Johnson SE, Mustanski B. Examining racial and ethnic minority differences among YMSM during recruitment for an online HIV prevention intervention study. AIDS Behav. 2012;16(6):1430–5. doi: 10.1007/s10461-011-0058-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Brown CH, Curran G, Palinkas LA, Aarons GA, Wells KB, Jones L, Collins LM, Duan N, Mittman BS, Wallace A, Tabak RG, Ducharme L, Chambers DA, Neta G, Wiley T, Landsverk J, Cheung K, Cruden G. An Overview of Research and Evaluation Designs for Dissemination and Implementation. Annual Review of Public Health, Vol 38. 2017;38:1–22. doi: 10.1146/annurev-publhealth-031816-044215. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Rogers EM. Diffusion of Innovations. 5th ed. New York, NY: Free Press; 2003. [Google Scholar]
  • 69.Jaganath D, Gill HK, Cohen AC, Young SD. Harnessing Online Peer Education (HOPE): integrating C-POL and social media to train peer leaders in HIV prevention. AIDS Care. 2012;24(5):593–600. doi: 10.1080/09540121.2011.630355. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Young SD, Cumberland WG, Lee S-J, Jaganath D, Szekeres G, Coates T. Social Networking Technologies as an Emerging Tool for HIV PreventionA Cluster Randomized Trial. Ann Intern Med. 2013;159(5):318–24. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Young SD, Cumberland WG, Nianogo R, Menacho LA, Galea JT, Coates T. The HOPE social media intervention for global HIV prevention in Peru: a cluster randomised controlled trial. Lancet HIV. 2015;2(1):e27–32. doi: 10.1016/S2352-3018(14)00006-X. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Young SD. Social media technologies for HIV prevention study retention among minority men who have sex with men (MSM). AIDS Behav. 2014;18(9):1625–9. doi: 10.1007/s10461-013-0604-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Young SD, Koussa M, Lee J, Gill K, Perez H, Gelberg L, Heinzerling K. Feasibility of a Social Media/Online Community Support Group Intervention among Chronic Pain Patients on Opioid Therapy. J Addict Dis. in press. [DOI] [PMC free article] [PubMed]
  • 74.Young SD, Heinzerling K. The Harnessing Online Peer Education (HOPE) Intervention for Reducing Prescription Drug Abuse: A Qualitative Study. J Subst Use. 2017;22(6):592–6. doi: 10.1080/14659891.2016.1271039. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Young SD. Stick with It: A Scientifically Proven Process for Changing Your Life--for Good. New York, NY: Harper; 2017. [Google Scholar]
  • 76.Young SD, Jaganath D. Feasibility of using social networking technologies for health research among men who have sex with men: a mixed methods study. Am J Mens Health. 2014;8(1):6–14. doi: 10.1177/1557988313476878. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Ybarra ML, Prescott TL, Philips GL 2nd, Bull SS, Parsons JT, Mustanski B Iteratively developing an mHealth HIV orevention program for sexual minority adolescent men. AIDS Behav. 2016;20(6):1157–72. doi: 10.1007/s10461-015-1146-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Ybarra ML, Prescott TL, Phillips GL 2nd, Bull SS, Parsons JT, Mustanski B. Pilot RCT Results of an mHealth HIV Prevention Program for Sexual Minority Male Adolescents. Pediatrics. 2017;140(1). doi: 10.1542/peds.2016-2999. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Gilead Sciences Inc. U.S. Food and Drug Administration Approves Expanded Indication for Truvada® (Emtricitabine and Tenofovir Disoproxil Fumarate) for Reducing the Risk of Acquiring HIV-1 in Adolescents. Business Wire; 2018.
  • 80.Gallo C, Moran K, Brown CH, Mustanski B. Optimizing HIV prevention mHealth interventions by measuring linguistic alignment. 10th Annual Conference on the Science of Dissemination and Implementation; Arlington, VA 2017. [Google Scholar]
  • 81.Sullivan PS, Driggers R, Stekler JD, Siegler A, Goldenberg T, McDougal SJ, Caucutt J, Jones J, Stephenson R. Usability and Acceptability of a Mobile Comprehensive HIV Prevention App for Men Who Have Sex With Men: A Pilot Study. JMIR Mhealth Uhealth. 2017;5(3):e26. doi: 10.2196/mhealth.7199. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Bandura A Social foundations of thought and action: A social cognitive theory. Engelwood Cliffs, NJ: Prentice Hall; 1986. [Google Scholar]
  • 83.Goldenberg T, McDougal SJ, Sullivan PS, Stekler JD, Stephenson R. Building a Mobile HIV Prevention App for Men Who Have Sex With Men: An Iterative and Community-Driven Process. JMIR Public Health Surveill. 2015;1(2):e18. doi: 10.2196/publichealth.4449. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Goldenberg T, McDougal SJ, Sullivan PS, Stekler JD, Stephenson R. Preferences for a Mobile HIV Prevention App for Men Who Have Sex With Men. JMIR Mhealth Uhealth. 2014;2(4):e47. doi: 10.2196/mhealth.3745. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Sullivan PS, Sineath C, Kahle ES, T. Awareness, willingness, and use of oral pre-exposure prophylaxis (PrEP) among a national sample of US men who have sex with men. AIDS Impact; Amsterdam, Netherlands: 2015. [Google Scholar]
  • 86.Siegler AJ, Wirtz S, Weber S, Sullivan PS. Developing a Web-Based Geolocated Directory of HIV Pre-Exposure Prophylaxis-Providing Clinics: The PrEP Locator Protocol and Operating Procedures. JMIR Public Health Surveill. 2017;3(3):e58. doi: 10.2196/publichealth.7902. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Mustanski B, Feinstein BA, Madkins K, Sullivan P, Swann G. Prevalence and Risk Factors for Rectal and Urethral Sexually Transmitted Infections From Self-Collected Samples Among Young Men Who Have Sex With Men Participating in the Keep It Up! 2.0 Randomized Controlled Trial. Sex Transm Dis. 2017;44(8):483–8. doi: 10.1097/OLQ.0000000000000636. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Siegler AJ, Mayer KH, Liu AY, Patel RR, Ahlschlager LM, Kraft CS, Fish R, Wiatrek SE, Sullivan PS. Developing and assessing the feasibility of a home-based PrEP monitoring and support program. Clin Infect Dis. 2018. doi: 10.1093/cid/ciy529. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Whittaker R Key issues in mobile health and implications for New Zealand. Health Care and Informatics Review Online. 2012;16(2):2–7. [Google Scholar]
  • 90.Centers for Disease Control and Prevention. Compendium of evidence-based interventions and best practices for HIV prevention. 2019. http://www.cdc.gov/hiv/prevention/research/compendium/rr/complete.html. Accessed 12 Mar 2019.
  • 91.Collins CB Jr., Wilson KM. CDC’s dissemination of evidence-based behavioral HIV prevention interventions. Transl Behav Med. 2011;1(2):203–4. doi: 10.1007/s13142-011-0048-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92.Collins LM, Dziak JJ, Kugler KC, Trail JB. Factorial experiments: efficient tools for evaluation of intervention components. Am J Prev Med. 2014;47(4):498–504. doi: 10.1016/j.amepre.2014.06.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93.Murphy SA. An experimental design for the development of adaptive treatment strategies. Stat Med. 2005;24(10):1455–81. doi: 10.1002/sim.2022. [DOI] [PubMed] [Google Scholar]
  • 94.Collins LM, Murphy SA, Strecher V. The multiphase optimization strategy (MOST) and the sequential multiple assignment randomized trial (SMART): new methods for more potent eHealth interventions. Am J Prev Med. 2007;32(5 Suppl):S112–8. doi: 10.1016/j.amepre.2007.01.022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95.Proctor E, Silmere H, Raghavan R, Hovmand P, Aarons G, Bunger A, Griffey R, Hensley M. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health. 2011;38(2):65–76. doi: 10.1007/s10488-010-0319-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 96.Kegeles SM, Rebchook G, Tebbetts S, Arnold E, Team T. Facilitators and barriers to effective scale-up of an evidence-based multilevel HIV prevention intervention. Implement Sci. 2015;10:50. doi: 10.1186/s13012-015-0216-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 97.Labrique AB, Vasudevan L, Kochi E, Fabricant R, Mehl G. mHealth innovations as health system strengthening tools: 12 common applications and a visual framework. Glob Health Sci Pract. 2013;1(2):160–71. doi: 10.9745/GHSP-D-13-00031.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98.Eysenbach G, Group C-E. CONSORT-EHEALTH: improving and standardizing evaluation reports of Web-based and mobile health interventions. J Med Internet Res. 2011;13(4):e126. doi: 10.2196/jmir.1923. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99.Agarwal S, Lefevre AE, Labrique AB. A call to digital health practitioners: new guidelines can help improve the quality of digital health evidence. JMIR Mhealth Uhealth. 2017;5(10):e136. doi: 10.2196/mhealth.6640. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 100.Bartholomew Eldredge LK, Parcel GS, Kok G, Gottlieb NH, Fernandez ME. Planning health promotion programs: An intervention mapping approach. Fourth edition. ed. San Francisco, CA: Jossey-Bass & Pfeiffer Imprints, Wiley; 2016. [Google Scholar]

RESOURCES