Do we need to “fail fast” to achieve open access?
Progress to open access has stalled. After two decades of trying, the proportion of born-free articles is stuck at 20%. Kicking off the Impact Blog’s Open Access Week coverage, Toby Green suggests the solution to our financially unsustainable scholarly publishing system may lie in rethinking traditional processes using internet-era norms. Embracing the principle of “fail fast”, all papers should first be published as freely available preprints to test whether they “succeed” or “fail”, with journals then competing to invite authors to publish. This would reduce the costs of the expensive, straining peer review system while ensuring all papers are available to all readers.
Let’s face it, progress to open access has stalled. No progress has been made over the past year – roughly 80% of all new articles published this year will be paywalled – same as last year. As Open Access Week dawns, let’s take a closer look at why.
No one has been idle these past 12 months. Librarians have been getting tougher with publishers, most notably in Germany and Sweden; publishers have innovated with Read and Publish offers; and, with the EU’s blessing, 13 funders are peddling Plan S. My feeling is that these efforts are the final throes of the tired “Green-Gold-Diamond” approach to open access which seeks a flip to a supply-side funding model from the traditional consumption-side model. A flip that’s flawed because all it does is transfer inequity of access to inequity of authoring; i.e. previously those without funds couldn’t read, now they won’t be able to publish. A flip that’s failed because after two decades of trying, we’re stuck at 20%.
In thinking about this problem, I have come to the conclusion that open access is the wrong target, it’s beside the point. The crux of the matter is that scholarly publishing is unsustainable both financially and in terms of human effort. Let me count the ways.
1.The funds available to pay for publishing research are not growing fast enough to keep pace with the growth in research budgets and, consequently, the number of articles that emerge. The number of articles submitted for publication is growing ~6% per annum; the library and funder budgets that pay for publishing are not.
2.Publons’ report on peer review shows a system under severe strain: it’s taking longer to find reviewers and they are less likely to complete a review quickly. Peer review costs around US$1,500 per paper; that’s a lot of money if the result is rejection.
3.It’s kind of ironic that the weakest papers are costing the most to publish. Authors are encouraged to re-submit rejected papers to another journal, sometimes only to once more be rejected, before, finally, the paper finds a home in a title further down the foodchain. Every submission and rejection costs money. Elsevier alone rejects over 4,000 papers every working day – that’s an estimated daily cost of US$100,000.
4.Authors are “double-dipping”: they increasingly post their articles as preprints to share their findings with peers fast, then submit to impact-factored journals to boost their career and grant-winning prospects. With changes between the former and the latter versions being small, we’re paying to publish the same content twice. This is not to cast blame. Authors need the internet-era speed of preprints to counter the analogue-era timescale of formal publishing. They need traditional, impact-factored journals to counter the exclusion of preprints from the reputation economy on which their careers depend.
Until the scholarly publications ecosystem is transformed in line with the digital age, I argue that open access can’t be afforded. So, how to transform it?
I think the answer lies in “digital transformation”, the rethinking of traditional processes using internet-era norms. An example is the process to apply for a British passport. Previously, application involved lots of form-filling, “peer review” in the form of a signature from another passport holder, and other user-unfriendly, bureaucratic, pen-pushing practices. Today’s online system is user-centric and would make any internet start-up proud. It’s undoubtedly a lot less costly for the UK authorities too.
So, inspired by this example, how could we rethink the process of scholarly publishing? One internet-era principle is “fail fast” – if your project fails, you stop and move on in another direction. What if all papers were first published on preprint servers to test whether it “succeeds” or “fails”? If it succeeds, journal editors would compete to invite authors to publish in their journal, flipping the submission process. If it failed to garner interest, no matter, the paper remains on the preprint server (perhaps to gain attention later as a slow-burner) and the author moves on in another direction.
Let’s assume that half of the preprints succeed in gaining the attention of a journal editor and, of these, half survive peer review – the total saving on the current publishing system would be significant. Cutting 15% off today’s cost of publishing journal articles is US$1.5bn. Yet, in terms of getting papers in front of readers, nothing would have changed: all would be available online just as they are today.
One catalyst is needed: the “reputation economy” (comprising tenure, promotion, and grant-giving committees) must value preprints just as it does articles published in impact-factored journals today. To help this process along, preprint servers need to have comment fields like those in TripAdvisor and Airbnb. Just as consumers trust consumers in making choices about where they eat and stay the night, readers will trust readers in making choices about what they read next. Perhaps reader comments could be codified and included in altmetric scores?
Nothing comes for free and this proposal implies another change: authors will have to do more to promote their papers. Funders are increasingly looking to measure the impact of the research they fund so this is something authors will have to do more of in any case. There is a danger that those who are already well-known will do better than newcomers (the Matthew Effect) but I would argue that a preprint system open to all offers newcomers a greater chance of breaking through than today’s closed world of peer-reviewed journals.
Once significant costs have been stripped out of the system, it should be possible for libraries and funders to fund both open preprint repositories and open access journals without the need for paywalls or play-walls. But until the costs come down, I fear we’ll remain stuck with the same frustrations we have today, only things will become more heated. Worst of all, I bet I’ll be writing that the number of articles born-free is still stuck at ~20% in 12 months’ time.
This blog post is based on the author’s preprint article, “We’re still failing to deliver open access and solve the serials crisis: to succeed we need a digital transformation of scholarly communication using internet-era principles” (6 September, 2018), available via Zenodo (DOI: 10.5281/zenodo.1410000).
Featured image credit: Open innovation: The new bright idea, by opensource.com (licensed under a CC BY-SA 2.0 license).
Note: This article gives the views of the author, and not the position of the LSE Impact Blog, nor of the London School of Economics. Please review our comments policy if you have any concerns on posting a comment below.
About the author
Toby Green has spent 35 years in scholarly publishing working with commercial, society, IGO and NGO organisations on all types of content: books, journals, databases, A&I services and encyclopedias – always with an eye on the reader experience. He writes this piece in a personal capacity and in the hope that it contributes to thinking about how to find a sustainable and effective way to make scholarship available to all. His ORCID iD is 0000-0002-9601-9130.
Posted 24/01/2019 12:33Back