Biodiversity Data Journal : Editorial / Correspondence
Beyond dead trees: integrating the scientific process in the Biodiversity Data Journal
Corresponding author: Lyubomir Penev (firstname.lastname@example.org)
Academic editor: Lyubomir Penev
Received: 14 Sep 2013 | Accepted: 14 Sep 2013 | Published: 16 Sep 2013
© 2013 Vincent Smith, Teodor Georgiev, Pavel Stoev, Jordan Biserkov, Jeremy Miller, Laurence Livermore, Edward Baker, Daniel Mietchen, Thomas L.P. Couvreur, Gregory Mueller, Torsten Dikow, Kristofer M. Helgen, Jiři Frank, Donat Agosti, David Roberts, Lyubomir Penev.
This is an open access article distributed under the terms of the Creative Commons Attribution 3.0 (CC-BY) which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Citation: Smith V, Georgiev T, Stoev P, Biserkov J, Miller J, Livermore L, Baker E, Mietchen D, Couvreur T, Mueller G, Dikow T, Helgen K, Frank J, Agosti D, Roberts D, Penev L (2013) Beyond dead trees: integrating the scientific process in the Biodiversity Data Journal. Biodiversity Data Journal 1: e995. doi: 10.3897/BDJ.1.e995
Driven by changes to policies of governments and funding agencies, Open Access to content and data is quickly becoming the prevailing model in academic publishing. Open Access benefits scientists with greater dissemination and citation of their work, and provides society as a whole with access to the latest research. Open Access is, however, only one facet of scholarly communication. Core scientific statements or assertions are intertwined and hidden in the scholarly narratives, and the data underlying these statements are often obscured to the point that replication of results is impossible (
An often cited reason for the lack of published data is the absence of a reward mechanism for the individuals involved in creating and managing information (
If we are going to incentivise the mass publication of data, we also need mechanisms to ensure quality. Traditional peer review is one of the bottlenecks in standard publication practice (
An abundance of small isolated datasets does not, however, allow us to address the fundamental problems within the biodiversity science community. These islands of data are only of value if connected and interlinked. The task of interlinking is performed by biodiversity data aggregators like the Global Biodiversity Information Facility (GBIF) and Encylopedia of Life (EOL) which form the backbone of data-driven biodiversity research. By automating the submission of data to these aggregators, we can increase their value to more than the sum of their parts, making small data big. A renewed appreciation of the value of small data will help to reduce the vast amount of research data that exists only on laptops and memory sticks - data that is often lost when people change roles or retire.
Works of potentially very limited length can hold intrinsic value to the community, but are almost impossible to publish in traditional journals chasing impact factors. Examples include single species descriptions, local checklists and software descriptions, or ecological surveys and plot data. An infrastructure that allows datasets of any size to be important means we can publish them at any time. There is no need to wait for datasets to reach a critical mass suitable for publication in a traditional journal.
Today, we are pleased to announce the official release of the first series of papers published in Biodiversity Data Journal (BDJ). After years of hard work in analyzing, planning and programming the Pensoft Writing Tool (PWT), we now have a publishing platform that addresses the key concerns raised above. This provides the first workflow to support the full life cycle of a manuscript - from writing through submission, community peer-review, publication and dissemination, all within a single online collaborative environment.
Most journals nowadays clearly separate data from narrative (text). Moreover, data publishing through data centres and repositories has almost become a separate sector within the scholarly publishing landscape. BDJ is not a conventional journal, nor is it a conventional “data journal”. It aims to integrate data and text in a single publication by converting several kinds of biodiversity data (e.g., species occurrences, checklists, or data tables) into the text for human-readable use, while simultaneously making data units from the same article harvestable and downloadable. The text itself is marked up and presented in a highly structured and machine readable form.
BDJ aims to integrate small data into the text whenever possible. Supplementary data files that underpin graphs, hypotheses and results can also be uploaded on the journal’s website and published with the article.
Nonetheless, this is usually not possible for large or complex data, for which we recommend deposition in an established open international repository (for details, see
All external data used in a BDJ paper must be cited in the reference list, and links to these data (as deposited in external repositories) must be included in a separate data resources section of the article.
All datasets, images or multimedia are freely downloadable from the text under the Open Data Commons Attribution License or a Creative Commons CC-Zero waiver / Public Domain Dedication. The article text is available under a Creative Commons (CC-BY) 3.0 license. Primary biodiversity data within an article can be exported in Darwin Core Archive format, which makes them interoperable with biodiversity tools based on the Darwin Core standard.
By facilitating open access to the data that underlie every publication, BDJ is setting a new standard in transparency and repeatability in biodiversity science. Perpetual and universal access to primary data stimulates scientific progress by helping authors build upon existing datasets. BDJ’s commitment to supporting automated data aggregation and interlinking is happening alongside multiple advances in biodiversity informatics infrastructure that herald the dawning of an era of collaborative, big-data biodiversity science (
The online, collaborative, article-authoring platform (Pensoft Writing Tool, PWT) is the principal way to write and submit a manuscript to BDJ. It provides a set of pre-defined, but flexible article templates (
Three-step editorial workflow in the Biodiversity Data Journal: manuscript authoring, peer-review and publication.
A special feature of PWT is that the authors can see at any time an editable preview of their manuscript in a format that is very close to the final published version. On completion of the manuscript, it can be submitted to the journal with a simple click of a button that will initiate the review process. The tool also allows automated import of manuscripts from data management platforms such as Scratchpads. Several tools in PWT facilitate import of data, references, images and other data.
A major advantage of the PWT is that it handles much of the semantic enhancement of a manuscript automatically during validation, eliminating the need for the authors or editors to manually markup portions of text. Examples of this include taxonomic names and georeferenced localities. The validation tool checks for compliance with the relevant biological code, for example checking that a holotype designation has been made for a new species description and that a new genus has a designated type species. In the near future, the PWT will also automatically register nomenclatural acts in the appropriate registry (International Plant Names Index, Index Fungorum, MycoBank or ZooBank).
The technology used by the PWT largely eliminates the conventional layout stage, just as the validation tool saves work for the copyeditors. Our goal is to greatly reduce the publication costs for all. This is particularly important because many authors working within biodiversity science are not backed by large institutions who can cover large page charges.
A novel community-based peer review of the manuscripts submitted to BDJ provides the opportunity for many specialists in the field to review a manuscript. The purpose of community peer review is to distribute effort, increase speed and transparency, engage the broader community of experts, and enhance the quality of the science we publish.
There are three groups of reviewers that may participate in the community peer review process: nominated, panel, and public reviewers. Nominated reviewers are expected to agree to provide a formal review by a deadline, and in this sense, they operate in the same way as conventional referees in most other journals. Panel reviewers are also invited to evaluate the manuscript, but without the formal acceptance of the deadline. They can submit their review, if they wish, at any time before the editorial process is finalised. Both nominated and panel reviewers can propose changes and corrections, make comments in the manuscript online and submit a concise reviewers’ evaluation form. Reviewers may opt to be anonymous but we encourage them to disclose their names. In the near future, authors will be able to opt for an entirely public peer-review process. Finally, comments can be posted after publication, so as to extend the review process even further and to enrich it with new insights, corrections or follow-up work.
The editor’s work is reduced by a tool that collates reviewers’ comments and corrections into a single document. Upon receipt of this consolidated review and editorial evaluation (
Consolidated review version. The editor sees all reviewer's corrections and comments in one place and can filter them out. The editor can also insert his/her own corrections and comments before submitting the editorial evaluation.
Accepted articles are published in semantically enhanced HTML, PDF and XML versions, compliant with the TaxPub schema, an extension of the NLM/NCBI Journal Article Tag Suite (JATS) used by the PubMedCentral archive (
In the Internet era, dissemination of published information is at least as important as the act of publishing. The highly structured text, domain-specific markup and underlying data can be used not only for effective reading but also to provide users direct access to the precise data they need (
Literature references are exported to the community-owned Bibliography of Life (based on the RefBank database and the ReFinder bibliographic search tool) as well as to several other bibliographic databases. This allows for their further re-use and import into new publications, saving authors a great deal of time locating historical literature.
Images are exported to Encyclopedia of Life, which increases their visibility and re-use.
The workflow has been funded in part by the European Union’s Seventh Framework Program (FP7) project ViBRANT (Contract no. RI-261532). Our thanks are due to all who supported the concept and contributed to its realisation over the duration of the project. We particularly thank all of our colleagues that participated in testing the PWT and BDJ workflow during the hot August of 2013.