Skip to main content
eGEMs logoLink to eGEMs
. 2016 Mar 31;4(2):1207. doi: 10.13063/2327-9214.1207

Governance Through Privacy, Fairness, and Respect for Individuals

Dixie B Baker i, Jane Kaye ii, Sharon F Terry iii
PMCID: PMC4827784  PMID: 27141520

Abstract

Introduction:

Individuals have a moral claim to be involved in the governance of their personal data. Individuals’ rights include privacy, autonomy, and the ability to choose for themselves how they want to manage risk, consistent with their own personal values and life situations. The Fair Information Practices principles (FIPPs) offer a framework for governance. Privacy-enhancing technology that complies with applicable law and FIPPs offers a dynamic governance tool for enabling the fair and open use of individual’s personal data.

Perceptions of Risk:

Any governance model must protect against the risks posed by data misuse. Individual perceptions of risks are a subjective function involving individuals’ values toward self, family, and society, their perceptions of trust, and their cognitive decision-making skills.

The HIPAA Privacy Rule Puts Some Governance in the Hands of Individuals:

Individual privacy protections and individuals’ right to choose are codified in the HIPAA Privacy Rule, which attempts to strike a balance between the dual goals of information flow and privacy protection. The choices most commonly given individuals regarding the use of their health information are binary (“yes” or “no”) and immutable. Recent federal recommendations and law recognize the need for granular, dynamic choices.

Building a Governance Framework Based in Trust: Avoiding Surprises:

Individuals expect that they will govern the use of their own health and genomic data. Failure to build and maintain individuals’ trust increases the likelihood that they will refuse to grant permission to access or use their data. The “no surprises principle” asserts that an individual’s personal information should never be collected, used, transmitted, or disclosed in a way that would surprise the individual were she to learn about it.

Fair Information Practices Principles:

The FIPPs provide a powerful framework for enabling data sharing and use, while maintaining trust. We introduce the eight FIPPs adopted by the Department of Health and Human Services, and provide examples of their interpretation and implementation.

Reducing Risk through Consumer Engagement:

Privacy risk and health risk can be reduced by giving consumers control, autonomy, and transparency, and by engaging them in managing their own health. Explicit “consent” may not always be necessary – the FIPPs offer multiple ways to engender trust and avoid surprises.

Keywords: Participant governance, health information technology, patient involvement, privacy

Introduction

The provision of safe, high-quality health care and the advancement of biomedical science are highly dependent on the availability of a large volume of personal health information, longitudinally collected, from large populations. The use of health information for these different purposes raises questions about the governance of such data. Governance refers to the collective set of “shared principles, norms, rules, decision-making procedures”1 that are associated with the collection, storage, use, and dissemination of health information and biospecimens. Individuals contributing the data have a moral claim to be involved in the governance of their personal data based on their autonomy and as a demonstration of respect for them as human beings.2 This claim is enforced through data protection law and human rights. How to implement this fundamental aspect of governance raises complex decision-making issues for individuals, families, and medical professionals as to what are their rights, duties, and responsibilities in regard to health information in general, and genomic information in particular.

Within the United States, the Health Insurance Portability and Accountability Act (HIPAA)3 Privacy Rule4 provides a framework for navigating these requirements. However, the Privacy Rule was largely designed to protect privacy while enabling access for providing and paying for health care services, operating health care enterprises, and protecting public health. The Privacy Rule recognizes the value of health data for other purposes, including research, and defines de identification requirements for enabling these additional uses. However, the Privacy Rule did not consider the special issue of the genome as the ultimate identifier, since few health records contained exome or genome sequences at the time the Privacy Rule was originally crafted. When these legal requirements are implemented in practice, individual choices are often limited and very little information is provided to individuals about their right to govern their information, or about the relevant risks and benefits of sharing it. Thus governance of these data is sometimes mistakenly limited by misapplication of the Privacy Rule.

In any system of governance for health data, short of a property ownership model, trust is a key element. Governance of genomic and genetic data must be built on transparency and accountability to engender trust. Maintaining individuals’ trust in an environment of transparency is a core attribute of this governance, and is essential to assuring continuing access to these data. Trust is engendered by respecting individuals’ rights and values; treating them fairly; and giving them the information and tools they need to make contextually informed decisions about the use and sharing of their own health information—based on their personal and contextual perception of risk, and at the level of specificity they require. Respecting individuals implies respecting not only their privacy, but also their individual autonomy and right to choose for themselves how they want to manage risk, consistent with their personal values and life situations.

In this paper, we argue that the application of the Fair Information Practices principles offers a framework for governance, and that privacy-enhancing technology that complies with the Privacy Rule and situations when the Privacy Rule does not apply, can offer a dynamic governance tool. In the final section of this paper, we describe a digital platform that enables this model of governance by giving individuals the tools to control how their health information is used, embedded in a trust context, while at the same time respecting individual choice.

Perceptions of Risk

Any governance model must protect against the risks posed by data misuse: infringements upon individuals’ and families’ rights to privacy; decisions and processes that fail to respond to societal values regarding privacy and data sharing; exposure of individuals to harms, such as social and insurance discrimination based on genetic predisposition; social stratification leading to class disparities; and decisions and processes that weaken societal trust in health care providers and governments.5 Individual perceptions of these risks are sometimes not an objective or quantifiable quality, but a subjective function involving individuals’ values toward self, family, and society; their perceptions of trust; and their cognitive decision-making skills. What seems reasonable and appropriate for one individual may be considered wildly reckless for another. In regard to genomic information, individuals may hold considerably divergent views. Individuals at increased risk for late onset conditions such as Parkinson’s or Alzheimer’s disease may or may not want to be informed about their susceptibility. Conversely, individuals at increased risk for breast cancer may want to know, in order to take steps to decrease that risk. Across their life spans, individuals are free to exercise their right to be informed (or not to be informed) of their health risks, and to choose a risk-management strategy that is consistent with each individual’s life situation and personal values, in the context of their culture and community.

Consciously and subconsciously, individuals continually assess risk and decide how to handle it, based on their individual values and perceived benefits, within the context of their personal experience and the information available to them. A report developed by the President’s Council of Advisors on Science and Technology (PCAST)6 illustrates some of the risks patients consider in deciding whether and how to share their health information:

Patients are concerned that the storage of their health information in electronic form will make it easier for employers, insurers, government, or malicious electronic intruders to improperly access their records. This concern may make them unwilling to participate in health [information technology] systems or [to] grant consent for their information to be used in research, even though the aggregation of patient data to compare treatments and providers is a major benefit of health [information technology]. Data can be anonymized by removing all personal identifiers from the data. But patients also may want to be re-contacted if analysis of their data reveals a problem with a medication they are taking or a treatment that could benefit them.

In reality, very few individuals consider all of their health-related information equally sensitive in all contexts. They may feel quite comfortable allowing their physician to use unencrypted email to remind them of an appointment or to tell them their prescription order has been sent to the pharmacy, but less comfortable receiving the result of a biopsy or genetic test via unencrypted email. Individuals typically desire more granular privacy controls over their health information than is afforded them by yes-or-no consent forms.7

The HIPAA Privacy Rule Puts Some Governance into the Hands of Individuals

Within the health care arena, privacy protections and individuals’ right to choose are codified in the HIPAA Privacy Rule,4 which requires that health care providers and health plans obtain individual consent or authorization for various types of uses and disclosures. The specific requirements range from “may obtain consent” (i.e., optional) for uses and disclosures for the purposes of treatment, payment, and health care operations, to written authorization for uses and disclosures of psychotherapy notes. The Privacy Rule attempts to strike a balance between the dual goals of information flow and privacy protection.

A major goal of the Privacy Rule is to assure that individuals’ health information is properly protected while allowing the flow of health information needed to provide and promote high quality health care and to protect the public’s health and well-being. The Rule strikes a balance that permits important uses of information, while protecting the privacy of people who seek care and healing.8

[italics added]

However, the HIPAA Privacy Rule is a complex law whose interpretation and implementation challenge both the organizations seeking to comply and the individuals seeking to understand their rights and protections. As a result, organizations’ approaches to compliance with the Privacy Rule sometimes have resulted in practices that comply with the letter of the law, while enabling access and use of individuals’ health information in ways these individuals might find surprising or possibly alarming. Transparency is key to avoiding such surprises and reactions. For example, for convenience, some organizations ask individuals to authorize specific uses of their data at the same time they are asked to acknowledge receipt of an already complex Notice of Privacy Practices (NPP). Such a practice may make it difficult for an individual to discriminate between the uses and rights he is being notified about, and uses which he is being asked to authorize.

The complexity of the HIPAA Privacy Rule also leads to organizations’ placing a greater emphasis on the procedural governance issues—such as collection, storage, use, and dissemination—than on the governing concept of individuals’ rights and protections, which are less well understood. Recent clarifications of these rights from the United States Department of Health and Human Services (HHS) Office of Civil Rights may bolster individuals’ ability to better govern access to their own information.9 Current policies and practices that govern the use and disclosure of individually identifiable health information are designed to enable access for purposes such as treatment, payment, health care operations, activities preparatory to research, and certain legally required disclosures, while requiring individual, binary (“yes” or “no” to “all” or “none”) authorization for other uses. Health care providers and researchers are keenly aware of the uses and disclosures allowed by the Privacy Rule, and can be very creative in devising practices that enable access and use within the bounds of the law. Unfortunately, when making decisions about how their health information is used and disclosed, individuals’ choices often are limited to dichotomous choices: “yes” or “no” to use and disclose “all” or “none” of their information for any purpose not permitted or required by the Privacy Rule—and their decisions usually are made with very little information about the relevant risks and benefits of sharing. Furthermore, these choices are usually made once, with no opportunity to change one’s choices as circumstances change. The PCAST set forth a proposal for achieving such granular controls, and the Health Information Technology for Economic and Clinical Health (HITECH) Act10 called for investigation into technologies for segmenting and protecting specific and sensitive individually identifiable information. However, except for psychotherapy notes, the Privacy Rule offers individuals no options for granting permission to use or disclose specific subsets of their health information.

Some uses of health information that the HIPAA Privacy Rule allows without individual authorization may be confusing, surprising, or even suspicious to individuals lacking an understanding of how health care operates. For example, the average patient may not fully understand the full range of activities included under “health care operations” and “preparatory to research.” Nor is the average patient likely to know what business associates her provider is using, the services these business associates are providing, or their obligations to protect health information. In fact, even within health care, the term “health care operations” has been inconsistently interpreted, prompting the HITECH Act to enact restrictions on what can be considered “health care operations.”

Today’s health portals, mobile apps, and online health advisers present new opportunities to engage consumers in improving their own health and in advancing biomedical knowledge. At the same time, these tools present new opportunities to surreptitiously collect and use personal information for purposes relating to “treatment, payment, and health care operations” that the individual may not expect. To increase transparency and possibly avoid surprises, HITECH also added a requirement to maintain an accounting of all disclosures, even for treatment, payment, and health care operations, and a provision that makes business associates directly regulated under HIPAA.

Building a Governance Framework Based in Trust: Avoiding Surprises

Too often health care providers and biomedical researchers view privacy as a barrier to getting access to the data they need; they see their challenge as finding creative ways to obtain the data, while complying with applicable privacy laws and regulations. While legal compliance obviously is necessary, it is not sufficient to build and maintain the perception of responsible stewardship that is essential to gaining an individual’s trust that their health information will be collected, accessed and used fairly, respectfully, and responsibly.

Failing to build and maintain the trust of the individuals whose information is needed increases the likelihood that those individuals will refuse to grant permission to access or use their data, if they are asked, or that they will push back mightily if their information is used without their knowledge or consent. A clear example is a federal lawsuit filed by the Texas Civil Rights Project against the Texas Department of State Health Services and the Texas A&M University System. The lawsuit alleged that the state’s failure to ask parents for permission to store and possibly use blood samples collected during newborn screening for birth defects violated constitutional protections against unlawful search and seizure. A settlement reached in 2010 resulted in the destruction of more than 5 million blood samples that, in the typical course of newborn screening for that state, had been “taken from babies without parental consent and stored indefinitely for scientific research.”11 It also undermined the public trust in newborn screening.12 Newborn screening is arguably the most successful public health program in America, with more than 98 percent of the more than 4 million babies born in the United States annually being screened at birth for several dozen treatable conditions.13 Nonetheless, clashes between privacy advocates and state public health departments continue to cause the public unease and mistrust.

A study published in 2009 could have predicted such a response from parents. Using an Internet-based survey of a nationally representative sample of parents, the study examined parents’ willingness to permit use of their children’s newborn-screening blood samples for research with and without the parents’ permission. The study found that if the parents’ permission was obtained before using the samples, 76 percent were either very willing or somewhat willing to allow their children’s bloodspots to be used for research. If the bloodspots were used without the parents’ permission, 73 percent were somewhat unwilling or very unwilling to allow their children’s bloodspots to be used.15

In 1951, Henrietta Lacks’s cells were collected without her knowledge or consent; the prolific cells subsequently were used in more than 74,000 studies, many leading to profound insights into cell biology, vaccines, in vitro fertilization, and cancer. The family was never consulted, so they were quite surprised to learn about the marvelous Henrietta Lacks (HeLa) cell line and questioned why no one ever approached them for their consent. Rebecca Skloot, in her book The Immortal Life of Henrietta Lacks, very effectively portrayed the family’s anguish and confusion over learning of the proliferation of her cells.16 In 2013, the family was even more surprised to learn that two researchers had used the HeLa cells to sequence her genome and had made it publically available for downloading, again without the family’s knowledge or permission.17 Further, public exposure of the genome presents risk to all of the family members, and they felt even more vulnerable in this very public sharing without their permission. When the Lacks family raised objections, the National Institutes of Health (NIH) acknowledged that they should have sought the family’s permission before funding research to sequence the HeLa cell, and NIH then negotiated an agreement with the Lacks family. Terms of the agreement included storage of the genomic data from the two studies in the NIH’s database of genotypes and phenotypes, with controlled access to the data and annual reporting of the use of the data. A HeLa Genome Data Access working group, which would include two Lacks family members, would be responsible for reviewing data-access applications, and the cells.18

At a workshop convened by NIH to explore scientific and ethical issues related to open access to HeLa genomic data, several Lacks family members clearly articulated their thoughts and requests. They repeatedly expressed their extreme pride in their family’s contributions to medical science, and their strong support of the continuing use of the HeLa cell line and genome in biomedical research. They asked only for respect, fairness, and to be kept informed of how the cells and genome are used. As David Lacks explained, the family came up with the idea of governance in the form of controlled access so that they “wouldn’t be surprised,” while still allowing his grandmother’s genome to be used to advance medical science. As one participant pointed out, the family’s requests are highly consistent with the Fair Information Practices Principles that for decades have been used throughout the world to guide the fair use of personal information.19 The use of the cells is now governed by a family advisory board in collaboration with NIH.20

Trust is built through experience. When consumers are surprised to learn that their personal information has been collected, used, or disclosed in ways they were unaware of, did not approve of, and did not expect, their trust is eroded. Individuals expect that they will govern the disclosure of their health information. This concept is sometimes referred to as the “no surprises principle”—an individual’s personal information should never be collected, used, transmitted, or disclosed in a way that would surprise the individual were she to learn about it. Every time an individual is surprised to learn that a trusted caregiver has shared her health information with someone whom she did not expect to see it, or used her information in research without her permission, trust is eroded. At the same time, the risk that the unauthorized use will be exposed in ways that will affect the perceptions and behaviors of the broader public is elevated. Every caregiver and researcher whose success depends on the availability of high-quality health information should assiduously act in accordance with the “no surprises principle.” If an individual is likely to be surprised to learn that his information is being collected, used, or disclosed in the way contemplated, then his permission should first be obtained—even if the law does not require the individual’s consent.

England’s National Health Service (NHS) learned the importance of the “no surprises principle” the hard way. The NHS had high aspirations when it decided to build a national database of individuals’ health information. Having access to such an extensive repository would allow researchers to investigate drug side effects and patient outcomes, thus facilitating medical advances and ultimately saving lives. Building such a database was legal; the data were pseudonymized, and all individuals were given the opportunity to opt out. Unfortunately, the NHS’s plans were not effectively communicated to the public—resulting in a “surprised” public and attendant mistrust of the NHS. As a result of public outcry, the NHS was compelled to delay the program to permit “more time to build understanding of the benefits of using the information, what safeguards are in place, and how people can opt out if they choose to.”21

Researchers fear that if health information is not freely accessible, it cannot be used to advance biomedical knowledge. A general perception in the research community is that if researchers ask individuals for permission to use their information, and allow them to establish highly granular, context-specific access rules that they can change at any time, biomedical research will at best be skewed by selection bias and at worst grind to a halt. Our hypothesis is that an engaged public will participate in research at a much higher rate than the current 4 percent or 5 percent enrollment in clinical trials, and that any improvement in this enrollment rate will result in less selection bias.

Fair Information Practices Principles

Several decades ago, recognition of the futility of anticipating every questionable or evasive method that might be used, combined with an appreciation of the multidimensionality of trust, led to the development of a set of principles for responsible information stewardship, which is essential to establishing and maintaining public trust when collecting, using, disclosing, and sharing personal information. The Fair Information Practices Principles (FIPPs) were first published in 1973 as the Code of Fair Information Practices22 and became the basis for the United States federal Privacy Act of 1974.23 In more recent years, these principles were adopted by the HHS as the Nationwide Privacy and Security Framework for Electronic Exchange of Individually Identifiable Health Information24 and by the White House as a Consumer Privacy Bill of Rights to serve as a code of conduct for companies conducting business on the Internet.25 The American Institute of Certified Public Accountants (AICPA) and Canadian Institute of Chartered Accountants (CICA) also have published similar principles as Generally Accepted Privacy Principles.26

These principles provide a powerful framework for enabling data sharing and use, while maintaining trust. The FIPPs included in the HHS guidance comprise the following eight principles, which have been implemented in the HIPAA Privacy Rule or related statutes and have led to the development of governance practices to implement the principle. Each principle has been adopted by the HHS as guidance for health care entities who electronically exchange individually identifiable health information,24 and is followed by interpretation, examples, and where applicable, examples of how the principle has been translated, and possibly expanded, in law.

  1. Individual Access: Individuals should be provided with a simple and timely means to access and obtain their individually identifiable health information in a readable form and format.

    The HIPAA Privacy Rule gave individuals the right to inspect and obtain a copy of their own health information; the HITECH Act further provided that, at the individual’s request, the information must be provided electronically to either the individual or to a third party named by that individual in the format requested by the individual.9 The Clinical Laboratory Improvement Amendments (CLIA) rule enacted in February 2014 expanded this provision further to include the right for a patient to obtain laboratory test results directly from the lab.27 Any entity that holds health information about an individual should provide the capability for that individual to access and obtain a copy of his or her own information in either print or electronic form, in the format requested.

  2. Correction: Individuals should be provided with a timely means for disputing the accuracy or integrity of their individually identifiable health information, and to have erroneous information corrected or to have a dispute documented if their requests are denied.

    Any entity that holds health information about an individual should provide the capability for that individual to point out errors and to request that the errors be corrected. If the entity concludes that no changes are warranted, the individual’s dispute should be recorded. The Consumer Privacy Bill of Rights25 states that consumers have a right to access and correct personal data in accordance with the sensitivity of the data and the risk of adverse consequences to consumers if the data are inaccurate.

  3. Openness and Transparency: There should be openness and transparency about policies, procedures, and technologies that directly affect individuals and their individually identifiable health information.

    Consumers should be able to trust that their health information will be collected, used, and disclosed only in ways that are consistent with their expectations for a given context. This principle embodies the “no surprises principle” discussed above—an individual’s health information should never be collected, used, transmitted, or disclosed in a way that would surprise the individual were he to learn about it. If an individual would likely be surprised, then permission should first be obtained. For example, after a nurse has drawn a blood sample, patients are unlikely to be surprised to learn that their name and birthdate have been sent to a laboratory along with the sample. However, they might be very surprised to learn that the laboratory uses the blood sample for purposes other than running the test the doctor has ordered. This principle is particularly relevant with respect to Internet-based services and social networking applications, which often surreptitiously collect information about the user’s actions as a means of “customizing the user experience”—which may be to the good or detriment of the individual. Consumers have a right to easily accessible and understandable information about privacy- and security policies and practices, including the use of technologies that collect information about users and their actions, outside the user experience.

  4. Individual Choice: Individuals should be provided with a reasonable opportunity and capability to make informed decisions about the collection, use, and disclosure of their individually identifiable health information.

    Key concepts here are “choice” and “informed.” Not only do individuals have the right to decide what personal information is collected and how that information may be used and disclosed, they also have the right to be informed of both the risks and potential benefits associated with their decision. Whether an entity uses an opt in, opt out, or open choice, the individual must be given the opportunity to make that choice before the action in question has been taken—in sufficient time for it to be meaningful.

    Today, “individual choice” is generally presented to the individual at a given point in time, in the form of a printed “consent” document that the consenter signs by hand that is then filed away for safekeeping. Individual choice needs to be an ongoing activity through which an individual expresses her wishes and preferences with respect to the use of her personal information. A single consent document cannot bear the burden of what must be a much fuller, ongoing engagement. A one-time consent form allows only a brief snapshot of an individual’s health status and values, and not the longitudinal engagement of an individual over time, health conditions, and changing priorities and values. “Individual choice” must be viewed as being a dynamic, ongoing process that occurs within the context of the full set of fair information practices.

  5. Collection, Use, and Disclosure Limitation: Individually identifiable health information should be collected, used, and disclosed only to the extent necessary to accomplish a specified purpose and never to discriminate inappropriately.

    In security engineering, this is known as the “principle of least privilege,”28 which means that each user, software program, and process should be able to authorize use of only the information, resources, and privileges necessary to fulfill its assigned responsibilities. For the HIPAA Privacy Rule, it is the “minimum necessary” standard, which includes minimizing the access authorized for each person within an organization, minimizing the amount of information that is disclosed when disclosure is necessary, and minimizing the amount of information that is requested. The introduction of genomic and genetic information into health care and biomedical research raises the stakes for those seeking to limit the use of identifiable information since the individual genome is inherently the ultimate “biometric identifier”—one of the 18 data elements the HIPAA Privacy Rule enumerates as individual identifiers.

  6. Data Quality and Integrity: Persons and entities should take reasonable steps to ensure that individually identifiable health information is complete, accurate, and up-to-date to the extent necessary for the person’s or entity’s intended purposes, and that it has not been altered or destroyed in an unauthorized manner. Technical security safeguards are available to protect information from accidental and malicious alteration and to detect when data have been corrupted. This principle goes a step further to say that an organization has a responsibility to help ensure that each individual’s health information is complete, accurate, and current. This means, for example, that all of the information an entity holds for a given individual can be associated with the same identity.

  7. Safeguards: Individually identifiable health information should be protected with reasonable administrative, technical, and physical safeguards to ensure its confidentiality, integrity, and availability and to prevent unauthorized or inappropriate access, use, or disclosure. The HIPAA Security Rule29 defines administrative, technical, and physical safeguards designed to protect the confidentiality of information, the integrity of data, and the availability of resources. The Security Rule is strongly grounded in risk assessment and risk management. Entities should assess risks on an ongoing basis to identify new vulnerabilities and to assure that their security safeguards can adequately protect against new and emerging threats.

  8. Accountability: These principles should be implemented, and adherence assured, through appropriate monitoring, and other means and methods should be in place to report and mitigate nonadherence and breaches. Entities should make these principles part of their governance process so that adherence is continuously monitored and any policy breaches are detected and corrected. The HIPAA Security and Privacy Rules call for two types of accountability: the Security Rule requires the recording and review of a system audit trail, and the Privacy Rule requires maintaining an accounting of all disclosures of protected health information. An audit trail records security-relevant events, such as logon attempts, launching a software program, and creating a record in a database, while an “accounting of disclosures” records information about the release of health information from one organization to another. Both system audit trails and disclosure logs will be useful in monitoring adherence with these principles.

Reducing Risk Through Consumer Engagement

Privacy risk and health risk can be reduced by giving consumers control, autonomy, and transparency, and by engaging them in managing their own health. Adhering to the FIPPs helps reduce individuals’ privacy risk by giving them greater visibility into how their information is accessed, used and shared—reducing the likelihood of surprises. Consistent with the “no surprises principle,” individuals should not be surprised to learn that their health information is being disclosed or used in any particular way. Explicit consent may not always be necessary—the FIPPs offer multiple ways to engender trust and avoid surprises. One might speculate how the outcomes of some of the scenarios cited in this paper might have been different had the affected parties been engaged in the decision-making and kept informed of actions thereafter—specifically, the Texas A&M newborn screening program, NHS national health database, and the use of HeLa cells. Authentic engagement in the health care setting, such as peer-to-peer outreach or social networking among patients involved in clinical trials, or personal notification of the use of health information in observational studies, can help avoid surprises.

Health care professionals, biomedical researchers, and consumers are broadly recognizing the value of engaging individuals in improving their own health and the health of their families, and in helping to advance biomedical science. Today’s mobile technology, social networking, consumer health technology, ubiquitous connectivity, and powerful search capabilities provide consumers with the tools they need to engage in health advancement to whatever extent, using whatever means, they wish. However, consumers’ willingness to participate in advancing biomedical science at the highest level can be thwarted by the research enterprise’s almost exclusive reliance on consent as the single mechanism for engagement. Confining the interaction between the individual and biomedical research to a single transaction—signing a consent form—denigrates what could be a robust engagement, and limits the opportunity to create and nurture a more scientifically and medically literate populace. Longitudinal engagement, within the context of fair information practices, enables the collection of genomic, clinical, environmental, and lifestyle data critical to making headway in prevention, diagnoses, and interventions.

Platform for Engaging Everyone Responsibly (PEER)

The Platform for Engaging Everyone Responsibly (PEER)30 is a technology solution that enables individuals to govern the access to and use of their health information. Using FIPPs as a basis, Genetic Alliance and Private Access Inc. are collaborating on an initiative that enables individuals to participate in advancing medical science, while protecting individuals’ privacy consistent with their own perceptions of risks and benefits. PEER includes a repository of health information that is made discoverable and accessible to authorized researchers only in accordance with permissions established by the individual participants who contribute their information.31 At present 40 communities (e.g., disease advocacy organizations, health provider professional societies, special interest groups) use PEER to build registries, cohorts, and campaigns. Approximately 10,000 individuals are sharing their health information through PEER. This information includes self-reported health information, electronic health record (EHR) data, genetic test reports, and (soon) genomes and exomes. Permissions can be very granular and specific (e.g., “only my de-identified data may be included in searches, and only for clinical trials for drugs to treat type 1 diabetes”) or very liberal (e.g., “all of my identifiable data may be used for any research”) or something in between (e.g., “my linked data may be discovered by diabetes researchers who may then ask PEER to contact me”).

Participants make their health information available and establish their privacy preferences through a PEER interface that uses a simple, interactive, gamified survey to collect answers to questions using common data instruments provided by Genetic Alliance and many instruments created by the sponsoring community. PEER was created on the premise that each FIPP is critical, and that the experience must be local, centered in a trusted community. Thus, participants can view short videos of members of their community, human guides, each offering suggestions for privacy settings corresponding to the guide’s own perceptions of high, medium, and low risk-benefit ratios (Figure 1). An individual may change her privacy permissions at any time, and as often as she wishes. Regardless of how a participant may choose to set her privacy permissions, the system provides defense-in-depth security protection for all of the data entrusted to PEER.

Figure 1.

Figure 1.

Guides Offer Three Levels of Suggested Privacy Settings

Contact information, privacy and sharing preferences, and health information are held in three separate databases, with Genetic Alliance responsible for health information, and Private Access holding responsibility for contact information and privacy settings (Figure 2). All data are encrypted for both storage and transmission, and a participant may view his information and an audit trail of accesses at any time. Genetic Alliance’s Ethics Team32 provides ethical, legal, and policy oversight. Western Institutional Review Board (IRB) has approved the PEER system itself,33 and various IRBs, including the Genetic Alliance IRB, have approved specific projects that use PEER.

Figure 2.

Figure 2.

PEER Stores Individual Contact Information, Privacy and Sharing Preferences, and Health Data in Three Separate Data Bases

PEER exemplifies a system that empowers individuals to manage their own health information consistent with their personal values and those of their community, their tolerance of perceived risks, and their own desire for specificity and autonomy. The online tools, information, and guidance that PEER offers convey respect for each individual’s personal values while enabling everyone to participate in advancing medical research. As shown in Figure 3, PEER applies the principles of fair use, no surprises, and consumer engagement to help advance biomedical science. PEER’s strict adherence to the FIPPs creates an environment designed to engender trust for all participants. PEER’s privacy policy also details this and can be viewed at peerplatform.org/privacy. This adherence is largely managed through the system and does not rely on the data seekers, the researchers, and investigators to manage the governance. Interested researchers and clinicians apply for an account, and after their credentials are examined, they are given access to any data that individuals have given them permission to see or use, including contact information. The governance is in the hands of the individuals. At present, we do not have universal return of results and notification.

Figure 3.

Figure 3.

Figure 3.

PEER Conforms to the Fipps to Engage Consumers in Medical Research

Conclusion

While many discussions of governance are concerned primarily with health information after it has been collected from the individuals, governance must begin with the individual’s overseeing the contribution and use of this information. Current policies and practices that govern the use and disclosure of individually identifiable health information are designed to enable access for purposes such as treatment, payment, health care operations, activities preparatory to research, and certain legally required disclosures, while requiring individual, binary (“yes” or “no” to “all” or “none”) authorization for other uses. Health care providers and researchers are keenly aware of the uses and disclosures allowed by the HIPAA Privacy Rule without the individual’s authorization, and can be very creative in devising practices that enable access and use within the bounds of the law. Both providing safe, high-quality health care and advancing biomedical science are highly dependent on the availability of large volumes of health information, longitudinally collected, from large populations of individuals. Maintaining individuals’ trust in an environment of transparency is essential to assuring continuing access to these data. Trust is engendered by respecting individuals’ rights and values, treating them fairly, and giving them the information and tools they need to make dynamic, contextual, informed decisions about the use and sharing of their own health and genomic information, based on their personal perceptions of risk, and at the level of specificity they require. Maintaining trust requires assiduous adherence to the Fair Information Practices Principles.

Acknowledgments

We acknowledge the work of Robert Shelton and Private Access. In addition, we also recognize the Genetic Alliance Staff and the Genetic Alliance Think Tank for their contributions. JK is funded by Wellcome Trust Award 096599/2/11/Z.

Footnotes

Disciplines

Health Information Technology

References


Articles from eGEMs are provided here courtesy of Ubiquity Press

RESOURCES