Abstract
With the advancement of genome-sequencing technologies, new genomes are being sequenced daily. Although these sequences are deposited in publicly available data warehouses, their functional and genomic annotations (beyond genes which are predicted automatically) mostly reside in the text of primary publications. Professional curators are hard at work extracting those annotations from the literature for the most studied organisms and depositing them in structured databases. However, the resources don’t exist to fund the comprehensive curation of the thousands of newly sequenced organisms in this manner. Here, we describe WikiGenomes (wikigenomes.org), a web application that facilitates the consumption and curation of genomic data by the entire scientific community. WikiGenomes is based on Wikidata, an openly editable knowledge graph with the goal of aggregating published knowledge into a free and open database. WikiGenomes empowers the individual genomic researcher to contribute their expertise to the curation effort and integrates the knowledge into Wikidata, enabling it to be accessed by anyone without restriction.
Database URL: www.wikigenomes.org
Introduction
Sequencing an organism’s genome has become a routine procedure in the life sciences, but until that genome is annotated, it tells us very little about the organism. The knowledge captured in genomic and functional annotations provides the ‘blueprint’ of the biology of the organism that can be leveraged to drive all manner of scientific inquiry. The knowledge represented by annotations is mostly published in the free text of scientific journal articles. In order to make this type of knowledge computable and accessible as structured annotations, we typically rely on centralized curation efforts like those supported by model organism databases such as the Zebrafish Model Organism database (1) and Mouse Genome Informatics (2). Unfortunately, the vast majority of sequenced genomes do not have a dedicated curation team, hence new curation models need to be explored.
Wikidata is a recently developed project of the Wikimedia Foundation that enables the collaborative construction of a centralized graph database (3). It was initially created as a central hub for structured data across all the language-specific Wikipedias, but its potential applications are much broader. Wikidata follows the Resource Description Framework (RDF) (https://www.w3.org/RDF/ (4)) model, the W3C standard for data interchange. RDF links related entities together into ‘triples’, defined by a subject concept, an object concept and a predicate that describes the relationship between them. This semantic structure allows the modeling of complex systems in a queryable knowledge graph (3, 5). The contents of Wikidata currently represent 28 million concepts, from all domains of knowledge, spanning over 1.3 billion triples (grafana.wikimedia.org/dashboard/db/wikidata-query-service).
Although other resources have also mapped their data onto an RDF data model (6–8), Wikidata is unique in that it allows both read and write access. Wikidata content can be both queried and edited programmatically via an application programming interface (API) (www.wikidata.org/w/api.php) and can be queried using the Wikidata Query Service (WDQS) (query.wikidata.org) via the SPARQL Protocol and RDF Query Language(4, 9, 10). SPARQL is a query language for graph databases similar in syntax to SQL for relational databases.
Notably, all Wikidata content is in the public domain, eliminating the potential for problematic restrictions on downstream reuse and redistribution (11). It is populated through various programs (‘bots’) (12) en masse, and through the discrete contributions of individual users through Wikidata’s own web interface or through a handful of tools developed by Wikimedia Foundation’s WMF Labs (tools.wmflabs.org/).
There is growing appreciation within the bioinformatics community for the potential of Wikidata as an open resource that can be populated, queried and edited by anyone. However, Wikidata has not yet widely reached the biological domain experts who have much to gain and contribute. Taking advantage of the querying capabilities of Wikidata can be a challenge for biologists as structured query languages like SPARQL are not commonly part of a researcher’s toolkit. Further, the domain-specific data models (e.g. the patterns of connections between genes, proteins and their annotations (www.wikidata.org/wiki/Wikidata:WikiProject_Molecular_biology)) underlying the content in the Wikidata graph is not readily apparent or automatically enforced by the generic Wikidata web interface. This makes providing meaningful contributions to the graph a challenge for the researcher and a roadblock for community curation in general. Although Wikidata offers a powerful technical platform for constructing an open community knowledge base of unprecedented scope, it cannot reach this potential without the large-scale participation of experts in the domains of knowledge that it seeks to represent. To attract specialists, e.g. life scientists, its content must be presented in compelling applications that are more useful than the applications that these specialists currently use to do their work.
Here, we describe WikiGenomes (www.wikigenomes.org), the first domain-specific application built on Wikidata, tailored to the needs of biomedical researchers. WikiGenomes is a Web application that is designed and built to allow the genomic researcher to contribute to curating the knowledge they are discovering. It facilitates both the consumption and community curation of genomic data in Wikidata, extending the reach of the biocuration effort deeper into the long tail of sequenced genomes.
Results
Wikidata
Our group and others have been working to populate Wikidata with a comprehensive and centralized knowledge graph that represents biomedical knowledge in a structured, queryable and computable format (5, 12). Previously, we loaded genomic data from a variety of organisms including Homo sapiens ‘human’ (www.wikidata.org/wiki/Q15978631), Mus musculus ‘mouse’ (www.wikidata.org/wiki/Q83310), Rattus norvegicus ‘brown rat’ (www.wikidata.org/wiki/Q184224), Saccharomyces cerevisiae S288c ‘baker’s yeast’ (www.wikidata.org/wiki/Q27510868) and Macaca nemestrina ‘Southern pig-tailed macaque’ (www.wikidata.org/wiki/Q618026). In this work, we systematically expanded this list to include all genes (390 719) and proteins (372 178) from 120 NCBI prokaryotic reference genomes (https://www.ncbi.nlm.nih.gov/genome/browse/reference/) (Figure 1). These organisms are all reasonably well-studied, but in general did not have community-maintained genome databases. In addition to organism/genetic data, we contribute to and maintain data for proteins, chemical compounds and diseases (5, 12, 13).
To ensure that the work we are doing in Wikidata is trusted as high quality by the scientific community, our group references the source and provenance of all claims in a consistent and structured way. Direct links to the data sources allow the user to see for themselves the source of an annotation and read further in the referenced publication to come to their own conclusions. Standards for representing the evidence and provenance of any claim made using our infrastructure are outlined in a detailed project wiki (https://www.wikidata.org/wiki/User:ProteinBoxBot/evidence). References can be accessed either manually by the web interface or programmatically via the Wikidata API (https://www.wikidata.org/w/api.php).
In addition to recording data provenance, Wikidata claims can be qualified by how the initial claim was made. For example, a Gene Ontology (GO) annotation can be qualified using the GO Evidence Codes to indicate the determination method for this claim. The combination of references and qualifiers provides the necessary context to the stated claim, making it more trustworthy and useful. This model allows the researcher to, for example, treat electronically curated annotations differently from those that were manually curated.
Data aggregation and maintenance in Wikidata
As a foundation, we loaded key information for each genome about each gene and gene product into Wikidata. Wikidata is not simply an open database into which data can be dumped. It is a structured knowledge graph that requires careful consideration and community consensus in its design. Currently, these entity models in Wikidata are not comprehensive mirrors of those models in their sources (e.g. NCBI, UniProt etc…). We have chosen to aggregate key data from a variety of sources, creating a minimal viable model that can be expanded on by us or others. The basic data in Wikidata for all the organisms we have loaded includes genomic position and orientation of genomic features (For representation of Wikidata genome feature data as compared with the GFF3, see property mappings in online supplementary Table S1), GO annotations, Enzyme Commission Numbers, InterPro domains and external database identifiers. The data model that we have created in Wikidata, and the sources of the annotations and entities are illustrated and reported in Figure 1.
Data retrieval, standardization, loading and maintenance are done routinely (currently monthly) via our publically available WikidataIntegrator Python package (https://github.com/SuLab/WikidataIntegrator and https://pypi.python.org/pypi/wikidataintegrator) and we are working towards custom update schedules that would follow each source’s update schedule. WikidataIntegrator interacts with the Wikidata API to create new or edit existing Wikidata items when appropriate. The data points are mapped to the appropriate ‘properties’ in Wikidata (see online supplementary Table S2). Mapping these properties to a new semantic model in Wikidata is key to aggregation of data from different sources into a single cohesive graph that represents our collective knowledge of each concept and the relationships between those concepts.
An example of other data we have loaded and maintain is the GO (http://www.geneontology.org/), which we have loaded to Wikidata to provide a semantic framework for gene product annotations. Because Wikidata requires assimilating data into its own semantic model, and because we include links directly to the GO Consortium entries, we did not use all the fields of the GO Consortium RDF model. For the ontology, we included the name, description, synonym and accession for each term, while the structure was represented as a hierarchical tree using the Wikidata properties ‘Subclass of’ (P279) (used in Wikidata to indicate entity type and the child of a parent in an ontology) and part of (P361) (see online supplementary Table S3). For annotations we programmatically gathered curated GO annotations from the European Bioinformatics Institute’s QuickGo API (https://www.ebi.ac.uk/QuickGO/), and NCBI’s Gene Database FTP resources (ftp.ncbi.nlm.nih.gov/gene/DATA/gene2go.gz) for all of the annotated gene products in Wikidata (5, 12, 13). Like the GO ontology, the GO annotations were partially represented in Wikidata to include only the fields necessary for creating a meaningful knowledge graph (see online supplementary Table S4). Any gene product item in Wikidata that is annotated with a specific GO term will be linked to that GO term’s ontology item in Wikidata via the appropriate Wikidata GO property (e.g. molecular function P680) (Figure 1). Our methods for loading and maintaining other domains, such as disease and chemical data, are described in our previous works (12, 13).
WikiGenomes
To allow non-developer biologists to interface with the genomic data knowledge graph that we developed in Wikidata, we built WikiGenomes. WikiGenomes currently supports 120 NCBI Bacterial reference genomes (https://www.ncbi.nlm.nih.gov/genome/browse/reference/) and their genes and gene products. We designed WikiGenomes to allow a user to consume genomic data from any organism in Wikidata in a single web application. The user begins by selecting an organism using a type-ahead search feature on the application’s landing page. Once an organism has been selected, the application is directed to the primary interface page where the data for the selected organism is loaded and displayed. In addition to providing access to genetic data for these organisms, WikiGenomes also allows users to directly contribute genomic and protein annotations to the Wikidata knowledge graph (Figure 2).
The primary page of the interface is divided into four windows that render the data for the selected organism: its genome, genes/proteins, and annotations for the currently displayed gene/protein. To select and load data, there are two forms that allow organism and gene/protein selection (Figure 3). Each of the data elements are drawn dynamically from the Wikidata knowledge graph using the WDQS SPARQL query system. The genome data is rendered in the open source JBrowse genome browser (14, 15). The WebApollo and JBrowse development teams expanded on already existing code in the JBrowse framework to allow JBrowse to load annotations gathered from Wikidata SPARQL queries to the WDQS endpoint (16). The DNA sequences themselves are curated by the RefSeq Project (17) and because bulk primary data is not suitable for storing in Wikidata, are collected monthly from GenBank’s FTP web service (see online supplementary materials for all Wikidata queries that drive the system and see online supplementary Table S2 for the FTP paths of data collected from NCBI). The open source code for WikiGenomes is available at https://github.com/putmantime/CMOD.Django.
Although bioinformaticians can programmatically gather data and upload it to Wikidata, WikiGenomes provides a domain specific editing interface for non-programmers. WikiGenomes distills editing tasks into straightforward forms that guide the user through the process of building annotations. Internally, the server translates the data entered through these forms into the structures needed to populate the Wikidata knowledge graph and submits them via the Wikidata API (Figure 2). WikiGenomes uses the WikiMedia OAuth extension to allow users to edit Wikidata using their own Wikidata account (www.mediawiki.org/wiki/Extension:OAuth). This makes it possible to utilize the Wikidata edit history to track the contributions of individual editors, potentially offering ways to reward good editors as well as mechanisms for detecting vandalism automatically (18).
Community curation forms
There are currently two editing forms in WikiGenomes, one for adding functional annotations and one for adding operon annotations. The forms are designed to allow a user to curate data from a publication by prompting the user to provide the necessary information and references to make a useful contribution to Wikidata (Figure 4).
GO form
Molecular function, cellular component and biological process annotations types are added to Wikidata (and subsequently WikiGenomes) through the GO Annotation Form (Figure 4A). The WikiGenomes GO form allows users to make structured GO annotations to gene products by identifying an annotation from a primary publication, mapping the annotation to the proper GO term, and then pushing it to Wikidata referencing the publication. We have chosen to adhere to the GO annotation model, rather than user generated free-text, to guide the annotation process in a consistent manner. This is accomplished by using the form to search Wikidata by type-ahead for the proper GO term (as mentioned we maintain the GO ontology in Wikidata), selecting the proper GO evidence code (http://geneontology.org/page/guide-go-evidence-codes) to provide a determination method (i.e. EXP: Inferred from Experiment http://geneontology.org/page/exp-inferred-experiment), and select the publication as a reference using the PubMed Identifier (PMID) (Figure 4A). When the annotation is submitted, the annotation data is pushed to Wikidata using the Wikidata API and the annotation will be viewable in WikiGenomes shortly after (there is ∼5 min delay as the Wikidata SPARQL endpoint updates).
Operon form
Regulatory annotations are rare in large data warehouses like NCBI and regulatory annotation curation would greatly benefit from community involvement. To this end we created a form to allow user curation of prokaryotic operons (Figure 4B). The WikiGenomes operon form works similarly to the GO form in using type-ahead search boxes to retrieve relevant data from Wikidata, as well as input boxes to upload novel data. The user can create a new operon or add genes to an existing operon, input the determination method and reference the PMID, then submit that annotation and (as in the GO Form) our backend framework will do the work of creating the structured and standardized annotation in Wikidata.
Discussion
WikiGenomes was built as a proof a concept that Wikidata can serve as a universal database backend for domain specific applications. WikiGenomes is in the stage of functional prototype and usage stats are thus not yet available. WikiGenomes provides a working demonstration of a new technical pattern that, we posit, offers the potential to fundamentally change how biocuration is organized in ways that, over time, will result in a more efficient movement of knowledge throughout the scientific community. Although this is not yet a tool in common usage, we believe that it shows the potential of centralized, community owned knowledge aggregation projects like Wikidata, and hope it will ultimately help stimulate community curation at both the level of individual researchers and bioinformatics laboratories.
Distributed curation and access, centralized and integrated content
The incorporation of operon information for Chlamydia pneumoniae into WikiGenomes illustrates the pattern of centralizing content while decentralizing control. Despite prokaryotic operons being heavily studied, the actual annotations are rarely included in genomic assemblies and often reside in online supplementary tables of primary publications. Chlamydia pneumoniae is a pathogen affecting animals and humans causing lung infections (www.cdc.gov/pneumonia/atypical/cpneumoniae/index.html). The operons of the C. pneumoniae strain CWL029 genome were annotated in a 2011 study (19) but access to this information was only provided in the form of an online supplementary Excel file. To our knowledge, none of those ∼200 experimentally derived annotations have been curated in GenBank or in operon-specific databases such as DOOR (20) or ODB3 (21). Illustrating the ‘small data to big data’ approach of WikiGenomes, we loaded this content programmatically via the Wikidata API. Individual domain experts could also have loaded these data via the operon form (Figure 4B). Using either method, these annotations are now fully linked data, viewable as a track in the WikiGenomes browser and in any other application drawing content from the open Wikidata knowledge base. Integrating this knowledge into the central Wikidata knowledge graph creates queryable connections to other organisms, diseases, drugs and small molecules. The more extensive the knowledge graph becomes, the more it can be used to explore biological research questions (5, 12, 13).
Stepping towards community curation
So far, the great majority of content accessible through WikiGenomes has been added by our research group. Although this content is valuable, it is only a starting point on the path towards an open, community-curated knowledge graph for biology. The fundamental logic of the community curation pattern has been well-described many times (22–24). By incorporating the scientific community into the process of curating the knowledge that they are generating and consuming, we could, in theory, dramatically increase the scale at which biocuration can collectively operate. Here, as usual, the devil is in the details. With certain notable exceptions, most attempts to stimulate community curation fail to attract enough editors to achieve their goals. With the WikiGenomes application, we make no claim that we have solved this challenge of incentives and recruitment. To realize its vision, WikiGenomes will need to transition to a model where the content consumers become the content producers. To make this transition, a number of problems must be addressed. In facing them, some inspiration can be taken from one of WikiGenomes’ most important and successful predecessors, the Gene Wiki effort to organize functional gene information in the context of Wikipedia articles (25).
The key distinction between the Gene Wiki and other projects with very similar goals but less success in recruiting contributors (22, 26), was that it was embedded directly in Wikipedia. The Wikipedia context provided the Gene Wiki with immediate discoverability (e.g. Gene Wiki articles rank highly in Google search results), connection to a large community of editors, and a proven social and technical framework for supporting large-scale community content creation. Wikidata now offers most of these same characteristics for structured information. Its content is highly discoverable, with hyperlinks coming in from nearly every Wikipedia article and a high performance query engine that provides a single point of entry into all of its data. It has already attracted thousands of editors from the Wikipedia communities as well as external groups more focused on structured data. It offers a user interface for human editing of content as well as an effective API for programmatic updates. Finally, communities are forming around it in similar ways to Wikipedia. There are social structures in place for administration of the higher-level data structures it utilizes (e.g. the allowed properties) and emerging domain-focused communities such as WikiProject Medicine and WikiProject Molecular and Cellular Biology that work to build and maintain domain-specific content within the larger knowledge graph.
With all of those similarities, one key distinction is that structured information is much more difficult to edit and to present effectively than the hypertext of Wikipedia. Although the Wikidata interface (www.wikidata.org) provides a useful, general-purpose foundation for navigating and editing its content, it is not nearly sufficient as a way to connect end-users with the information they need. WikiGenomes provides a genome-focused, editable view of biological knowledge in Wikidata; it offers the large community of scientists an easy way to both access and share knowledge in a structured way, thus opening a door towards open community curation.
Conclusions and next steps
Although tools like WikiGenomes are a step in the right direction for converting ‘small data’ to ‘big data’ several key problems remain to be solved in the community curation model. Interfaces like WikiGenomes that provide access to content drawn from openly editable public sources should consider ways to help users assess the trustworthiness of content based on automatic inspection of the provenance information provided in associated references as well as in the histories of the editors who generated the content. In addition to making trust computable, such provenance information will also be of value in attributing the work of the curators who devote time to these open resources—potentially providing an aspect of the solution to the incentive problem. Aside from presenting trust-related information, the forms for adding content should be designed in ways that help domain experts generate high-quality annotations in the first place. The addition of documentation, logical constraints based on underlying ontologies, and helper interfaces are areas where the experience of the existing biocuration resources would be very valuable.
WikiGenomes is a first foray into the world of applications that could be built with the content that is growing in Wikidata. As that content expands and diversifies, many new applications will need to be created to bring that information most effectively to the people that need it and who can contribute to it. Wikidata is unlikely to be the end-all and be-all of open biological knowledge bases. Although it has a number of advantages that make it an ideal platform for initiating the endeavor of linking all biological knowledge, it has limitations in terms of content scope that will eventually need to be overcome through the provision of equally open systems that are tailored more specifically to biological content.
Immediate further development of WikiGenomes is focused on teaming up with biocuration groups to help refine the user experience and identify important areas where the data model needs to be expanded. One example would be to work further with the WebApollo team to integrate WebApollo (27) into WikiGenomes, taking advantage of WebApollo’s annotation features. Another useful feature we will work towards implementing is the ability to export data from Wikidata via WikiGenomes in standard formats (e.g. GFF3). We hope to eventually create a tool in WikiGenomes that brings the biocuration and basic research communities together, creating a more efficient system that will allow for more comprehensive reference genome curation of biological knowledge than is currently possible. Eventually we would make WikiGenomes a deployable toolkit that smaller research communities can customize and deploy while easily sharing their work with the global community.
Supplementary data
Supplementary data are available at Database Online.
Supplementary Material
Acknowledgements
We would like to thank the Wikimedia foundation for supporting the Wikidata project, and especially the Wikidata community for their efforts in populating it.
Funding
This project was supported by the National Institutes of General Medical Sciences (R01GM089820) and National Institutes for Health Common Fund programs for Big Data to Knowledge (U54GM114833) and Extracellular RNA Communication (U54DA036134). Nathan Dunn s and Monica Munoz-Torres s work with Apollo is supported by National Institutes for Health grants 5R01GM080203 from National Institute of General Medical Sciences, and 5R01HG004483 from National Human Genome Research Institute. Nathan Dunn and Monica Munoz-Torres are also supported by the Director, Office of Science, Office of Basic Energy Sciences, of the U.S. Department of Energy under Contract No. DE-AC02-05CH11231.
Conflict of interest. None declared.
References
- 1. Howe D.G., Bradford Y.M., Conlin T.. et al. (2013) ZFIN, the Zebrafish Model Organism Database: increased support for mutants and transgenics. Nucleic Acids Res., 41, D854–D860. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Bult C.J., Eppig J.T., Blake J.A.. et al. (2016) Mouse genome database 2016. Nucleic Acids Res., 44, D840–D847. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Vrandečić D., Krötzsch M. (2014) Wikidata: a free collaborative knowledgebase. Commun. ACM, 57, 78–85. [Google Scholar]
- 4. Quilitz B., Leser U. (2008) The semantic web: research and applications In: Bechhofer S., Hauswirth M., Hoffmann J., Koubarakis M. (eds). Lecture Notes in Computer Science. Springer, Berlin Heidelberg, pp. 524–538. [Google Scholar]
- 5. Putman T.E., Burgstaller-Muehlbacher S., Waagmeester A.. et al. (2016) Centralizing content and distributing labor: a community model for curating the very long tail of microbial genomes. Database, 2016, doi:10.1093/database/baw028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Quest D.J., Land M.L., Brettin T.S.. et al. (2010) Next generation models for storage and representation of microbial biological annotation. BMC Bioinformatics, 11, S15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Miles A., Zhao J., Klyne G.. et al. (2010) OpenFlyData: an exemplar data web integrating gene expression data on the fruit fly Drosophila melanogaster. J. Biomed. Inform., 43, 752–761. [DOI] [PubMed] [Google Scholar]
- 8. Cheung K.-H., Yip K.Y., Smith A.. et al. (2005) YeastHub: a semantic web use case for integrating data in the life sciences domain. Bioinformatics, 21 Suppl 1, i85–i96. [DOI] [PubMed] [Google Scholar]
- 9. Prud’Hommeaux E., Seaborne A.. et al. (2008) SPARQL query language for RDF. W3C Recommendation, 15, https://www.w3.org/TR/rdf-sparql-query/ (accessed Sep 1, 2016). [Google Scholar]
- 10. Pérez J., Arenas M., Gutierrez C. (2006) The Semantic Web - ISWC 2006 In: Cruz I., Decker S., Allemang D., Preist C., Schwabe D., Mika P., Uschold M., Aroyo L.M. (eds). Lecture Notes in Computer Science. Springer, Berlin Heidelberg, pp. 30–43. [Google Scholar]
- 11. Himmelstein D. Integrating resources with disparate licensing into an open network. https://thinklab.com/discussion/integrating-resources-with-disparate-licensing-into-an-open-network/107#1 (accessed Sep 16, 2016).
- 12. Burgstaller-Muehlbacher S., Waagmeester A., Mitraka E.. et al. (2016) Wikidata as a semantic framework for the Gene Wiki initiative. Database, 2016, doi:10.1093/database/baw015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Elvira M., Waagmeester A., Burgstaller S., Schriml L.M., Good B.M., Su A.I. Wikidata: a platform for data integration and dissemination for the life sciences and beyond. In: Proceedings of the 2015 Swat4LS International Conference in Cambridge England.
- 14. Skinner M.E., Uzilov A.V., Stein L.D.. et al. (2009) JBrowse: a next-generation genome browser. Genome Res., 19, 1630–1638. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Buels R., Yao E., Diesh C.M.. et al. (2016) JBrowse: a dynamic web platform for genome visualization and analysis. Genome Biol., 17, 66. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Bolleman J.T., Mungall C.J., Strozzi F.. et al. (2016) FALDO: a semantic standard for describing the location of nucleotide and protein feature annotation. Journal of Biomedical Semantics, 7, 39. [DOI] [PMC free article] [PubMed]
- 17. O’Leary N.A., Wright M.W., Brister J.R.. et al. (2016) Reference sequence (RefSeq) database at NCBI: current status, taxonomic expansion, and functional annotation. Nucleic Acids Res., 44, D733–D745. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Good B.M., Clarke E.L., de Alfaro L.. et al. (2012) The Gene Wiki in 2011: community intelligence applied to human gene annotation. Nucleic Acids Res., 40, D1255–D1261. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Albrecht M., Sharma C.M., Dittrich M.T.. et al. (2011) The transcriptional landscape of Chlamydia pneumoniae. Genome Biol, 12, R98. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Mao F., Dam P., Chou J.. et al. (2009) DOOR: a database for prokaryotic operons. Nucleic Acids Res., 37, D459–D463. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Okuda S., Yoshizawa A.C. (2011) ODB: a database for operon organizations, 2011 update. Nucleic Acids Res., 39, D552–D555. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Mons B., Ashburner M., Chichester C.. et al. (2008) Calling on a million minds for community annotation in WikiProteins. Genome Bio., 9, R89. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Howe D., Costanzo M., Fey P.. et al. (2008) Big data: the future of biocuration. Nature, 455, 47–50. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Pico A.R., Kelder T., van Iersel M.P.. et al. (2008) WikiPathways: pathway editing for the people. PLoS Biol., 6, e184. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Huss J.W. 3rd, Orozco C., Goodale J.. et al. (2008) A gene wiki for community annotation of gene function. PLoS Biol., 6, e175. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Hoffmann R. (2008) A wiki for the life sciences where authorship matters. Nat. Genet., 40, 1047–1051. [DOI] [PubMed] [Google Scholar]
- 27. Lee E., Helt G.A., Reese J.T.. et al. (2013) Web Apollo: a web-based genomic annotation editing platform. Genome Biol., 14, R93. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.