Abstract
Objective
The increased availability of public data and accessible visualization technologies enhanced the popularity of public health data dashboards and broadened their audience from professionals to the general public. However, many dashboards have not achieved their full potential due to design complexities that are not optimized to users’ needs.
Material and Methods
We used a 4-step human-centered design approach to develop a data dashboard of sexually transmitted infections for the New York State Department of Health: (1) stakeholder requirements gathering, (2) an expert review of existing data dashboards, (3) a user evaluation of existing data dashboards, and (4) an usability evaluation of the prototype dashboard with an embedded experiment about visualizing missing race and ethnicity data.
Results
Step 1 uncovered data limitations and software requirements that informed the platform choice and measures included. Step 2 yielded a checklist of general principles for dashboard design. Step 3 revealed user preferences that influenced the chart types and interactive features. Step 4 uncovered usability problems resulting in features such as prompts, data notes, and displaying imputed values for missing race and ethnicity data.
Discussion
Our final design was accepted by program stakeholders. Our modifications to traditional human-centered design methodologies to minimize stakeholders’ time burden and collect data virtually enabled project success despite barriers to meeting participants in-person and limited public health agency staff capacity during the COVID-19 pandemic.
Conclusion
Our human-centered design approach and the final data dashboard architecture could serve as a template for designing public health data dashboards elsewhere.
Keywords: human-centered design, public health informatics, data visualization, sexually transmitted infections
INTRODUCTION
Although public health data dashboards have become more common, many have not reached their full potential due to design complexities that are not optimized to their users’ needs.1–4 For example, a review of 2 dashboards for displaying influenza surveillance data shows that although these tools provide useful information, some of their design features (eg, use of acronyms and jargon and lack of granular data limit the usefulness of these applications).4 Data dashboards have become more popular due to the increased availability of publicly available data and accessible visualization technologies such as Tableau.5 The adoption of dashboards in different domains broadened both the form and content of the original dashboards. Traditionally, data dashboards comprised single-view representations of key performance measures for executive decision-makers in organizational contexts.6 In contrast, new dashboards usually include multiple data views with interactive interfaces and are used to facilitate learning, communication, and motivation in different areas, including public health.5 Moreover, using data dashboards for communication and education expanded their audiences from executives to decision-makers at all levels and the general public.5 Public health data dashboards allow these diverse users to interact with public health data, thereby promoting data-driven decision-making, encouraging new research, and supporting advocacy work.7,8
The human-centered design has been useful in developing data dashboards and other information visualization tools. Taking this approach, previous studies have included users in the design process to ensure products’ ease of use and usefulness for the intended audience.9–11 It is critical that new information visualization tools and technologies be evaluated by users because tool development is mostly conducted by computer scientists and engineers who are trained in statistical techniques and data analytics methods and therefore have a different perception of visualizations than the general population.12,13 Despite the usefulness of human-centered design, its application in developing public health data dashboards has been limited. One reason is the diversity of target users and use cases14; consequently dashboards cannot meet the needs of all users and it is critical to identify a clear focus and purpose. For example, while certain graphical features might improve quantitative reasoning, other features may more effectively encourage behavior changes or perceptions in the target audience such as willingness to be vaccinated or satisfaction with healthcare decisions related to breast cancer treatment.15 Moreover, dashboard usability might differ among user groups with heterogeneous computer and visualization literacy. For example, in a usability study of a “Treatment as Prevention” intervention related to HIV prevention in South Africa, task completion and accuracy varied between scientific and operational staff versus community advisory board members and nursing staff.16 Employing human-centered design requires that dashboard designers have explicit goals and recruit participants who are similar to target users. Another challenge of integrating human-centered design in public health data dashboards is that traditional methodologies, such as observation of users’ activities, are time-consuming and require heavy investment from users.10,17
We collaborated with the New York State Department of Health AIDS Institute (hereafter, AIDS Institute) to develop a prototype data dashboard for sexually transmitted infections (STIs) using a human-centered design approach. New York State was an early leader in publicizing open health data18 and integrating stakeholder feedback into dashboards such as the New York State Prevention Agenda dashboard19 and the Ending the (HIV) Epidemic Dashboard.20,21 The focus on STIs was a useful case for US public health practice because of the high morbidity (estimated STI prevalence, 67.6 million cases in 2018), high health-related costs (estimated at $15.9 billion in 2018), and increasing rates of gonorrhea, syphilis, and congenital syphilis.22–24 The new National STI Strategic Plan set a goal to increase accessibility and use of STI data to inform decisions,25 and data dashboards could provide an opportunity for education, increasing community awareness, monitoring trends, and evaluating programs. Moreover, the general knowledge about STIs is low in the US population26 and there is a gap in provider communication about STIs,27 making it especially important to make reliable sources of information available in an accessible format.
METHODS
Overview
This project comprised 4 steps, which are described in detail in Table 1. The University at Albany Institutional Review Board determined the study was exempt from human subjects review.
Table 1.
Steps involved in the development and evaluation of the data dashboard
Steps | Detailed procedure |
---|---|
Step 1. Stakeholder requirements gathering and establishing the partnership |
|
Step 2. Expert review of current US states’ STI data dashboards to produce a usability checklist |
|
Step 3. Pairwise user study with domain experts to understand needs and preferences of target users |
|
Step 4. Development and usability evaluation of the prototype data dashboard for STIs in New York State |
|
This checklist is now published28 and can be used in addition to other published work for evaluating data dashboards’ usability29 to avoid usability issues without the effort of conducting usability testing.
For the expert review (Step 2), one of the reviewers was an AIDS Institute Office of Sexual Health staff member. All participants in the pairwise user study (Step 3) were from the community, and represented a mix of direct service providers, healthcare administrators, and volunteers. The Step 4 participants were similar to Step 3 participants, with no AIDS Institute staff included.
Step 1. Stakeholder requirements gathering and establishing the partnership
Three preliminary meetings with the data owners (AIDS Institute Office of Sexual Health and Epidemiology) and an in-person site visit were used to establish relationships, secure buy-in, and discuss topics such as data dashboard goals, the intended audience, available resources, data limitations, software requirements, and other relevant organizational contexts. Additionally, 2 meetings with staff from the AIDS Institute Office of Planning and Community Affairs, which is responsible for policy development, strategic planning, budgetary and legislative analysis, and community were used to expand the organization’s engagement with the project and finalize the participant recruitment plan. A meeting series was established to keep staff engaged throughout the project lifecycle.
Step 2. Expert review of current US states’ STI data dashboards to produce a usability checklist
An expert review of all US state health department STI data dashboards (N = 13, as of June 1, 2021) was used to develop a usability checklist for data dashboard designers to conform to usability principles.28 The selection of dashboards was limited to state (rather than local) health departments because, although exceptions exist, local departments usually have limited resources and staff to prepare complex data dashboards, making their websites less comparable to state departments. The included states were Arizona, Colorado, District of Columbia, Florida, Georgia, Hawaii, Iowa, Kansas, Maryland, New York, North Carolina, Tennessee, and Texas. Six expert evaluators evaluated the STI data dashboards, representing domain experts and nondomain experts. Per the requirements of expert review,30 all evaluators had knowledge of usability and data visualization. The domain expert evaluators comprised a state health department employee, a faculty member with public health expertise, and a doctoral student with health informatics expertise. The nondomain expert evaluators comprised a faculty member with expertise in digital government, a faculty member with expertise in human–computer interaction, and a doctoral student with expertise in human–computer interaction. The usability checklist was derived from a qualitative synthesis of evaluators’ comments regarding major and minor usability problems. The usability evaluation was accompanied by an analysis of the data dashboards’ characteristics such as available visualizations, interactive features, and use of filtering or faceting to show the breakdown of the represented data.
Step 3. Pairwise user study with domain experts to understand needs and preferences of target users
A pairwise user study was conducted with practitioners who primarily worked in community-based organizations that serve populations at high risk for STIs. These community champions were ideal participants because they could represent the needs of target users (community-based organizations and patients) and had established relationships with the AIDS Institute, making them relatively easy to access. Participants were paired to minimize researcher intrusion, and increase the chance of revealing usability problems through the process of helping each other work on the specified activity.31 During virtual meetings, pairs of participants collaborated to complete several tasks on Arizona’s STI dashboard, which was selected from the data dashboards that were evaluated in Step 2. The Arizona STI data dashboard had different visualizations with interactive features and was scored highly in Step 2, making it an ideal choice for a user study to determine if users and expert opinions diverged. The tasks included exploration, pattern identification, association, comparison, verification, policy development, and dissemination, and were designed based on common tasks in visual analytics of public health data.32 All meetings were held online and the participants were given 30–45 min to complete the tasks. Details of the data collection meetings are available in Supplementary Appendix S1. Usability problems and user preferences were identified from the moderator’s direct observations of their task completion, a thematic analysis of audio transcripts, and a brief survey of their user experiences.33
Step 4. Development and usability evaluation of the prototype data dashboard for STIs in New York State
A prototype dashboard was developed using the NYS STI surveillance data. Multiple data preparations steps were conducted in R software, including merging with Census population data to produce STI rates among different populations and redistributing missing values. The data dashboard was created in Tableau using a dashboard layout and interactive charts. All findings from the first 3 steps were integrated in the prototype data dashboard’s design. Each of the prior 3 steps informed the design. For example, the data limitations and the software requirements from Step 1 were used in choosing the platform and selecting appropriate measures for presentation. The user preferences from Step 3 were considered when selecting the chart types and adding interactive features. The checklist produced in Step 2 was used as a general guide for designing the layout, text, and charts.
Participants from Step 3 were recruited for a Zoom-based usability evaluation of the prototype, in which participants were asked to freely explore each page, use interactive features, and narrate what they saw. In addition, the participants were asked to use the data dashboard to complete some tasks so potential usability problems would emerge. Details of the data collection meetings are available in Supplementary Appendix S2. When working with the data dashboard, participants used their pointing device to point to where they were looking on their screen and verbalized their thoughts in a sentence such as, “The red button is the first thing that caught my eyes. I am wondering what this button will do, and I’m going to click to find out.” The moderator monitored users’ activities and specific usability problems encountered and took notes of usability issues as well as the way participants interacted with the data dashboard (eg, the first thing to which they pointed on each page, and if they read the guides or used the interactive features at any time during their interaction). The moderator synthesized her notes to identify common usability issues and generalizable insights for other data dashboards.
One specific issue that emerged in Step 3 was how to present missing race and ethnicity data, which are common in chlamydia and gonorrhea surveillance data.34 Two versions of a visualization of cases stratified by race and ethnicity, using different methods to present missing demographic information, were included in the prototype to assess users’ understanding and preferences. The findings of the experiment and the usability evaluation were incorporated in the final prototype data dashboard.
RESULTS
Key findings from usability evaluations
Table 2 summarizes key findings that influenced the design. The stakeholder requirements gathering (Step 1) revealed important considerations about the data structure and limitations, organizational context, compatible software, and resources. For example, the AIDS Institute preferred Tableau for consistency with other projects and easier maintenance. Another finding was that the final data dashboard would need to be hosted on an external organization’s website due to state government agency “executive clearance” requirements that each website update be reviewed and approved before release to the public.
Table 2.
Key findings from evaluations that influenced the data dashboard design
Key findings | |
---|---|
Step 1 findings |
|
Step 2 findings |
|
Step 3 findings |
|
Step 4 findings |
|
The evaluation of existing STI data dashboards (Step 2) uncovered a tension between complexity and usability. Complex data dashboards, with multiple visualizations and interactive features, were more suitable for public health experts interested in specific populations but received lower usability scores because of their limited usefulness for novice users. In contrast, simple data dashboards received higher usability scores but were less useful for expert users. When we examined filtering or faceting, 2 common methods to deal with complexity in data visualization,35 data dashboards that used both methods simultaneously achieved a good balance of usability and complexity that would be useful for expert and novice users.
The user study (Step 3) identified important preferences among target users. First, participants were more attentive to the content with the use of simple visualization techniques such as bar charts and line charts. Second, some participants were concerned about the potential misinterpretation of disease maps. They shared that showing highly impacted regions in dark colors could be stigmatizing. Third, participants found charts showing the number of STI cases stratified by race and ethnicity with a “missing” category (including cases for whom the race and ethnicity information was not available) to be confusing and raised concerns about the trustworthiness of the data dashboard. Fourth, participants raised concerns about the use of outdated terminology for population groups (eg, transgender male to female). Fifth, navigation and interaction possibilities were not immediately clear to all participants, and many needed guidance.
The usability evaluation and the embedded experiment in Step 4 had several findings. First, participants repeatedly discussed how consistency in layouts and interaction possibilities helped them learn and use that knowledge throughout the dashboard. Second, participants exhibited different reading patterns, with some reading all guides and notes before interacting with the charts and others ignoring the text and using the interactive charts exclusively, which led to later confusion. Third, participants frequently asked about the excluded diseases, regions, and populations. For example, human papillomavirus was not included because data were not available, but participants found it confusing that some STIs were omitted from the data dashboard. Fourth, following the data visualization principle of “removing extraneous ink,”28 some items were removed (eg, axis titles that were repetitive of page titles). However, these choices reduced chart comprehension among some participants. Fifth, the initial terminology used to explain interactivity options (eg, filter, cluster, aggregate) was unclear to users, and they requested more common terms, such as drop-down menus. Finally, in the embedded experiment, the visualization with the added imputed values improved users’ accuracy in interpretation and their confidence in the data.
Final data dashboard prototype
Figure 1 shows one page of the final prototype data dashboard, which incorporates changes made in response to the Step 4 feedback. A multipage format was selected to help users digest each piece of information and encourage them to assess information from different perspectives. Arrow “a” points to the use of contrast in the navigation bar to help users see their current location and other pages available for their exploration. Arrow “b” points to the reading and interaction guides for each page to support novice users. Arrow “c” points to the filters for expert users to drill down into specific population groups. Each filter had a drop-down menu and a prompt to encourage users to read the footnotes. Arrow “d” points to the faceting technique to produce simpler charts for nonexperts. Arrow “e” shows the extensive data footnotes.
Figure 1.
Screenshot of one data dashboard page demonstrating the features included based on the usability evaluations. Note: Data are unvalidated and should not be used for program planning or other purposes. Arrows indicate features included in response to issues identified during the usability evaluations.
Figure 2 shows another data dashboard page that displays geo-faceting as a solution for a concern raised in the usability evaluation about inadvertently stigmatizing highly impacted counties in standard maps with higher-morbidity counties shaded in a darker color. Moreover, equal-sized boxes allowed each county to have the same prominence regardless of their relative population size or land area. The layout and available interactions were consistent with prior pages for improved usability. The final user study found that the geographical resemblance between the layout of the county charts and New York State was not clear to all users. The final prototype included bold and underlined font in the reading guide to communicate the positioning to users, county names on each chart, and a zooming function (zooming each box when hovered over) to provide additional guidance.
Figure 2.
Screenshot of one data dashboard page demonstrating the use of geo-faceting to map infection trends in the counties of the New York State. Note: Data are unvalidated and should not be used for program planning or other purposes.
DISCUSSION
The prototype design was accepted by the AIDS Institute and is currently in the process of being implemented as a public-facing data dashboard. Several research methodology choices enabled the successful use of the human-centered design approach. First, traditional human-centered design methods were modified to be more pragmatic, which was critical during the COVID-19 pandemic (a highly contagious respiratory disease caused by the SARS-CoV-2 virus) when public health staff were incredibly busy. The pragmatic modifications included learning from existing data dashboards instead of starting from scratch and conducting virtual meetings instead of observing user behavior in their workplace. Also, we recruited participants from practitioners who already had close relationships with the AIDS Institute as community-based public health champions. These participants were knowledgeable about STIs but most of them had little previous experience with data dashboards. Therefore, we did not expect them to understand data dashboards better than a layperson. Moreover, they work directly with healthcare providers and patients, and could share their own professional experiences with the dashboard and represent the needs of community members.
A second critical factor contributing to project success was the engagement of the AIDS Institute throughout the project lifecycle to increase the likelihood that research findings would be implemented. In Step 1, multiple meetings with the AIDS Institute were conducted to understand the organizational context, their goals for the data dashboard, and the data presented on the data dashboard. In Step 2, one AIDS Institute staff member participated as an expert evaluator. In Step 3, AIDS Institute staff identified a participant pool from its existing workgroups, and community-based listservs assisted with recruitment. In Step 4, early findings from the usability evaluations were discussed with AIDS Institute staff, and they were invited to contribute ideas for solutions for the found usability problems. Collaborating with practitioners enhanced the usefulness of findings for stakeholders and ensured buy-in and collective ownership of the final design.
Our methods have several contributions to usability evaluation practices. First, traditional human-centered methods suggest starting with the users’ current technology and improving areas in which the current technology does not meet the users’ needs.36 In this study, we initially considered New York State dashboards such as the influenza tracker but decided to go beyond the state health department to review all STI data dashboards on other comparable health department websites. The reviewed data dashboards had different levels of complexity, which we found was related to the number of identified usability problems. These findings led us to explicitly consider an appropriate balance between complexity and usability throughout our data dashboard design process. Second, for understanding users’ needs and preferences, 2 traditional human-centered methods are contextual inquiry (observation of user activities in their work environment) and domain knowledge workshops (conducting workshops with users to discuss their work practices in detail). These methods need heavy investment from domain experts, which is not realistic in many situations.10,11 Contextual inquiry requires the acceptance of researchers in users’ work environments and their interruptions for inquiries Domain knowledge workshops require users to attend lengthy workshops to discuss their work practices. Both were impractical in our context, between staff being diverted to the COVID-19 response (thereby severely constraining their time) and staff working in confidential spaces due to the sensitive data (thereby making it difficult for a researcher team to gain security clearances to observe workflows directly). Our virtual meetings in Steps 3 and 4 may have been limited compared to the traditional methods because the data collection was not in the users’ work environment, and we could not collect some potentially useful data such as eye movements or facial expressions. However, our meetings required 2 h of participants’ time (outside their working hours), was easier to schedule as virtual meetings with 1 or 2 participants (vs intensive long workshops with many participants), and did not require us to obtain special approvals to observe their interactions with the technologies.
Using the human-centered design approach helped us prevent some of the usability problems that were found in previous usability evaluations of public health data dashboards. For example, one of the usability problems found in previous dashboard evaluations is that data were not available at a granular level.4 We provided multiple filters (with dropdown menus) on each page of the dashboard to enable access to granular data by county, by sex, by age, and by race and ethnicity. Moreover, we discussed the levels of these variables with participants to ensure the output would be useful for them. For example, STI clinics sometimes focus on a specific age group (eg, adolescents) or a specific population group (eg, Black women). We used our participants’ comments about target audiences to ensure these groups were included in our dropdown menus. Another usability problem found in previous dashboard evaluations is that the terminology (eg, surveillance, prevalence, or antiviral resistance) was not suitable for those unfamiliar with health-related information.4 Our participants had experience directly working with general consumers and helped us avoid terminology that would not resonate with laypersons and develop footnotes to facilitate data interpretation.
This project had several limitations. First, focusing on one disease area and state might limit the generalizability of findings. Second, although the evaluators and participants comprised diverse groups, their opinions may not reflect all stakeholders’ opinions. Third, the data collection of this project was conducted virtually because of the concerns regarding COVID-19 transmission and the difficulties of scheduling time with busy public health experts during the pandemic. Virtual meetings limited the ability to collect additional user data such as body expressions and eye movements. Finally, based on the human-centered design framework, a long-term and consistent evaluation of the data dashboard is necessary, which was not yet possible because the dashboard is currently being implemented. However, this is a promising area for future research to investigate the emerging usability problems as technology changes and the user base grows.
Supplementary Material
ACKNOWLEDGMENTS
First and foremost, we like to thank the experts from New York State who participated in the virtual data collection sessions. Without their generously sharing their time and expertise, this study would not be possible. We are also immensely grateful to Rachel Hart-Malloy, Karen Bovell, Claire Simon, Julie Harris, Wilson Miranda, and other amazing staff at the New York State Department of Health AIDS Institute, who provided data access, administrative support, assistance with participant recruitment, and ongoing feedback. Finally, we are grateful to Xiaojun Yuan and Luis Luna-Reyes (University at Albany) for their helpful feedback throughout the project.
Contributor Information
Bahareh Ansari, Organization, Work, and Leadership Department, Queen’s Management School, Queen’s University Belfast, Belfast, UK.
Erika G Martin, Department of Public Administration and Policy, Rockefeller College of Public Affairs and Policy, University at Albany, Albany, New York, USA.
FUNDING
This research received no specific grant from any funding agency in the public, commercial or not-for-profit sectors.
AUTHOR CONTRIBUTIONS
Both authors conceived the study. Ansari directed the research, collected data, conducted the analysis, built the dashboard, and wrote the initial draft. Martin reviewed the methods and results and revised the manuscript for intellectual content.
SUPPLEMENTARY MATERIAL
Supplementary material is available at Journal of the American Medical Informatics Association online.
CONFLICT OF INTEREST STATEMENT
The authors have no competing interests to declare.
DATA AVAILABILITY
The data underlying this article will be shared on reasonable request to the corresponding author.
REFERENCES
- 1. Alcaraz-Martinez R, Ribera-Turró M.. An evaluation of accessibility of Covid-19 statistical charts of governments and health organisations for people with low vision. El Profes Inform 2020; 29: 1. [Google Scholar]
- 2. Bhowmick T, Robinson AC, Gruver A, et al. Distributed usability evaluation of the Pennsylvania Cancer Atlas. Int J Health Geogr 2008; 7: 36. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Blacklow SO, Lisker S, Ng MY, et al. Usability, inclusivity, and content evaluation of COVID-19 contact tracing apps in the United States. J Am Med Inform Assoc 2021; 28 (9): 1982–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Charbonneau DH, James LN.. FluView and FluNet: tools for influenza activity and surveillance. Med Ref Serv Q 2019; 38 (4): 358–68. [DOI] [PubMed] [Google Scholar]
- 5. Sarikaya A, Correll M, Bartram L, et al. What do we talk about when we talk about dashboards? IEEE Trans Visual Comput Graphics 2019; 25 (1): 682–92. [DOI] [PubMed] [Google Scholar]
- 6. Few S. Information Dashboard Design: The Effective Visual Communication of Data. Sebastopol, CA: Oreilly Media;2006. [Google Scholar]
- 7. Valdiserri R, Sullivan PS.. Data visualization in public health promotes Sound public health practices: The AIDSVU example. AIDS Educ Prev 2018; 30 (1): 26–34. [DOI] [PubMed] [Google Scholar]
- 8. Zakkar M, Sedig K.. Interactive visualization of public health indicators to support policymaking: An exploratory study. Online J Public Health Inform 2017; 9 (2): e190. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Tory M, Moller T.. Human factors in visualization research. IEEE Trans Vis Comput Graph 2004; 10 (1): 72–84. [DOI] [PubMed] [Google Scholar]
- 10. Lloyd D, Dykes J.. Human-centered approaches in geovisualization design: investigating multiple methods through a long-term case study. IEEE Trans Vis Comput Graph 2011; 17 (12): 2498–507. [DOI] [PubMed] [Google Scholar]
- 11. McKenna S, Staheli D, Fulcher C, et al. Bubblenet: A cyber security dashboard for visualizing patterns. Comput Graphics Forum 2016; 35 (3): 281–90. [Google Scholar]
- 12. Conati C, Maclaren H. Exploring the role of individual differences in information visualization. In: Proceedings of the Working Conference on Advanced Visual Interfaces. ACM; 2008, 199–206; Napoli, Italy.
- 13. Colin W. Visual Thinking for Design. Burlington, MA: Morgan Kaufmann; 2008. [Google Scholar]
- 14.Visualizing Health. Vizhealth.org; 2014. http://vizhealth-assets.s3.amazonaws.com/static/Visualizing%20Health%20Report.pdf. Accessed July 1, 2021.
- 15. Ancker JS, Senathirajah Y, Kukafka R, et al. Design features of graphs in health risk communication: a systematic review. J Am Med Inform Assoc 2006; 13 (6): 608–18. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Concannon D, Herbst K, Manley E.. Developing a data dashboard framework for population health surveillance: widening access to clinical trial findings. JMIR Form Res 2019; 3 (2): e11342. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Roth R, Ross K, Finch W, et al. A usercentered approach for designing and developing spatiotemporal crime analysis tools. In: GIScience 2010. Zurich, Switzerland; 2010.
- 18. Martin EG, Helbig N, Shah NR.. Liberating data to transform health care: New York’s open data experience. JAMA 2014; 311 (24): 2481–2. [DOI] [PubMed] [Google Scholar]
- 19. New York State Department of Health. New York State prevention agenda dashboard; 2022. https://webbi1.health.ny.gov/SASStoredProcess/guest?_program=/EBI/PHIG/apps/dashboard/pa_dashboard&p=sh. Accessed February 19, 2022.
- 20. New York State Department of Health. ETE dashboard: ending the AIDS epidemic; 2022. http://etedashboardny.org. Accessed February 19, 2022.
- 21. Joshi A, Amadi C, Katz B, et al. A human-centered platform for HIV infection reduction in New York: development and usage analysis of the ending the epidemic (ETE) dashboard. JMIR Public Health Surveill 2017; 3 (4): e95. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Kreisel KM, Spicknall IH, Gargano JW, et al. Sexually transmitted infections among US women and men: prevalence and incidence estimates, 2018. Sex Transm Dis 2021; 48 (4): 208–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Chesson HW, Spicknall IH, Bingham A, et al. The estimated direct lifetime medical costs of sexually transmitted infections acquired in the United States in 2018. Sex Transm Dis 2021; 48 (4): 215–21. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Centers for Disease Control and Prevention. Sexually Transmitted Disease Surveillance 2020; 2022. https://www.cdc.gov/std/statistics/2020/overview.htm. Accessed February 11, 2022.
- 25. U.S. Department of Health and Human Services. Sexually Transmitted Infections National Strategic Plan for the United States: 2021–2025. Washington, DC; 2020. https://www.hhs.gov/programs/topic-sites/sexually-transmitted-infections/plan-overview/index.html. Accessed February 19, 2022.
- 26. Ansari B. Information seeking behaviour differences indicate which US population groups lack information about human papilloma virus and associated cancer risks. Health Info Libraries J 2023; 40 (1): 42–53. [DOI] [PubMed] [Google Scholar]
- 27. Gilkey MB, McRee A-L.. Provider communication about HPV vaccination: A systematic review. Hum Vac Immunother 2016; 12 (6): 1454–68. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Ansari B, Martin EG.. Development of a usability checklist for public health dashboards to identify violations of usability principles. J Am Med Inform Assoc 2022; 29 (11): 1847–58. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Dowding D, Merrill JA.. The development of heuristics for evaluation of dashboard visualizations. Appl Clin Inform 2018; 9 (3): 511–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Nielsen J. Finding usability problems through heuristic evaluation. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. New York, NY: Association for Computing Machinery; 1992, 373–80.
- 31. Wildman D. Getting the most from paired-user testing. Interactions 1995; 2 (3): 21–7. [Google Scholar]
- 32. Preim B, Lawonn K.. A survey of visual analytics for public health. Comput Graph Forum 2020; 39 (1): 543–80. [Google Scholar]
- 33. Braun V, Clarke V.. Using thematic analysis in psychology. Qual Res Psychol 2006; 3 (2): 77–101. [Google Scholar]
- 34. Ansari B, Hart-Malloy R, Rosenberg ES, Trigg M, Martin EG.. Modeling the potential impact of missing race and ethnicity data in infectious disease surveillance systems on disparity measures: scenario analysis of different imputation strategies. JMIR Public Health Surveill 2022; 8 (11): e38037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Munzner T. Visualization Analysis and Design. Natick, MA: A K Peters/CRC Press; 2014. [Google Scholar]
- 36. Sharp H, Preece J, Rogers Y.. Interaction Design: Beyond Human–Computer Interaction. 5th ed. Hoboken, NJ: Wiley; 2019. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The data underlying this article will be shared on reasonable request to the corresponding author.