Open Access Progress: Anecdotes from close to home

nanoparticles
Solution extinction measurements of (A) faceted (B) and Au@MNPs and (C) photos of the particles. From Silva et al Chem. Commun., 2016 DOI:10.1039/C6CC03225G

It has become rather fashionable in some circles to decry the complain about the lack of progress on Open Access. Particularly to decry the apparent failure of UK policies to move things forward. I’ve been guilty of frustration at various stages in the past and one thing I’ve always found useful is thinking back to where things were. So with that in mind here’s an anecdote or two that suggests not just progress but a substantial shift in the underlying practice.

I live with a chemist, a group not known for their engagement with Open Access. More than most other disciplines in my experience there is a rigid hierarchy of journals, a mechanistic view of productivity, and – particularly in those areas not awash with pharmaceutical funding – not a huge amount of money. Combine that with a tendency to think everything is – or at least should be – patentable (which tends to rule out preprints) and this is not fertile ground for OA advocacy.

Over the years we’ve had our fair share of disagreements. A less than ideal wording on the local institutional mandate meant that archiving was off the menu for a while (the agreement to deposit required all staff to deposit but also required the depositor to take personal responsibility for any copyright breaches) and a lack of funds (and an institutional decision to concentrate RCUK funds and RSC vouchers on only the journals at the top of that rigid hierarchy) meant that OA publication in the journals of choice was not feasible either. That argument about whether you choose to pay an APC or buy reagents for the student was not a hypothetical in our household.

But over the past year things have shifted. A few weeks ago: “You know, I just realised my last two papers were published Open Access”. The systems and the funds are starting to work, are starting to reach even into those corners  of resistance, yes even into chemistry. Yes it’s still the natural sciences, and yes it’s only two articles out of who knows how many (I’m not the successful scientist in the house), buts its a quite substantial shift from it being out totally out of the question.

But around about the same time something that I found even more interesting. Glimpsed over a shoulder I saw something I found odd…searching on a publisher website, which is strange enough, and also searching only for Open Access content. A query raised the response: “Yeah, these CC BY articles is great, I can use the images directly in my lectures without having to worry; I just cite the article, which after all I would have obviously done anyway”. It turns out that with lecture video capture now becoming standard universities are getting steadily more worried about copyright. The Attribution licensed content meant there was no need to worry.

Sure these are just anecdotes but they’re indicative to me of a shift in the narrative. A shift from “this is expensive and irrelevant to me” to “the system takes care of it and I’m seeing benefits”. Of course we can complain that its costing too much, that much of the system is flakey at best and absent at worst, or that the world could be so much better. We can and should point to all the things that are sub-optimal. But just as the road may stretch out some distance ahead, and there may be roadblocks and barriers in front of us, there is also a long stretch of road behind, with the barriers cleared or overcome.

As much as anything it was the sense of “that’s just how things are now” that made me feel like real progress has been made. If that is spreading, even if slowly, then the shift towards a new normal may finally be underway.

First thoughts on the Finch Report: Good steps but missed opportunities

The Finch Report was commissioned by the UK Minister for Universities and Science to investigate possible routes for the UK to adopt Open Access for publicly funded research. The report was released last night and I have had just the chance to skim it over breakfast. These are just some first observations. Overall my impression is that the overall direction of travel is very positive but the detail shows some important missed opportunities.

The Good

The report comes out strongly in favour of Open Access to publicly funded research. Perhaps the core of this is found in the introduction [p5].

The principle that the results of research that has been publicly funded should be freely accessible in the public domain is a compelling one, and fundamentally unanswerable.

What follows this is a clear listing of other potential returns. On the cost side the report makes clear that in achieving open access through journal it is necessary that the first copy costs of publication be paid in some form and that appropriate mechanisms are in place to make that happen. This focus on Gold OA is a result in large part of the terms of reference for the report that placed retention of peer review at its heart. The other excellent aspect of the report is the detailed cost and economic modelling for multiple scenarios of UK Open Access adoption. These will be a valuable basis for discussion of managing the transition and how cost flows will change.

The bad

The report is maddeningly vague on the potential of repositories to play a major role in the transition to full open access. Throughout there is a focus on hybrid journals, a route which – with a few exceptions – appears to me to have failed to deliver any appreciable gains and simply allowed publishers to charge unjustified fee for very limited services. By comparison the repository offers an existing infrastructure that can deliver at relatively low marginal cost and will enable a dispassionate view of the additional value that publishers add. Because the value of peer review was baked into the report as an assumption this important issue gets lost but as I have noted before if publishers are adding value then repositories should pose no threat to them whatsoever.

The second issue I have with the report is that it fails to address the question of what Open Access is. The report does not seek to define open access. This is a difficult issue and I can appreciate a strict definition may be best avoided but the report does not raise the issues that such a definition would require and in this it misses an opportunity to lay out clearly the discussions required to make decisions on the critical issues of what is functionally required to realise the benefits laid out in the introduction. Thus in the end it is a report on increasing access but with no clear statement of what level of access is desirable or what the end target for this might look like.

This is most serious on the issue of licences for open access content which has been seriously fudged. Four key pieces of text from the report:

“…support for open access publication should be accompanied by policies to minimise restrictions on the rights of use and re-use, especially for non-commercial purposes, and on the ability to use the latest tools and services to organise and manipulate text and other content” [recommendations, p7]

“…[in a section on instituional and subject repositories]…But for subscription-based publishers, re-use rights may pose problems. Any requirement for them to use a Creative Commons ‘CC-BY’ licence, for example, would allow users to modify, build upon and distribute the licensed work, for commercial as well as non-commercial purposes, so long as the original authors were credited178. Publishers – and some researchers – are especially concerned about allowing commercial re-use. Medical journal publishers, who derive a considerable part of their revenues from the sale of reprints to pharmaceutical companies, could face significant loss of income. But more generally, commercial re-use would allow third parties to harvest published content from repositories and present them on new platforms that would compete with the original publisher.” [p87]

“…[from the summary on OA journals]…A particular advantage of open access journals is that publishers can afford to be more relaxed about rights of use and re-use.” [p92]

“…[from the summary on repositories]…But publishers have strong concerns about the possibility that funders might introduce further limits on the restrictions on access that they allow in their terms and conditions of grant. They believe that a reduction in the allowable embargo period to six months, especially if it were to be combined with a Creative Commons CC-BY licence that would allow commercial as well as non-commercial re-use, would represent a fundamental threat to the viability of their subscription-based journals.” [p96]

As far as I can tell the comment on page 92 is the only one that even suggests a requirement for CC-BY for open access through journals where the costs are paid. As a critical portion of the whole business model for full OA publishers it worried me that this is given almost a brief throw away line, when it is at the centre of the debate. But more widely a concern over a requirement for liberal licensing in the context of repositories appears to colour the whole discussion of licences in the report. There is, as far as I have been able to tell, no strong statement that where a fee is paid CC-BY should be required – and much that will enable incumbent subscription publishers to continue making claims that they provide “Open Access” under a variety of non-commercial licences satisfying no community definition of either “Open” nor “Open Access”.

But more critically this fudge risks failing to deliver on the minister’s brief, to support innovation and exploitation of UK research. This whole report is embedded in a government innovation strategy that places publicly funded knowledge creation at the heart of an effort to kick start the UK economy. Non-commercial licences can not deliver on this and we should avoid them at all costs. This whole discussion seems to revolve around protecting publishers rights to sell reprints, as though it made sense to legislate to protect candle makers from innovators threatening to put in an electric grid.

Much of this report is positive – and taken in the context of the RCUK draft policy there is a real opportunity to get this right. If we both make a concerted effort to utilise the potential of repositories as a transitional infrastructure, and if we get the licensing right, then the report maps out a credible route with the financial guidelines to make it through a transition. It also sends a strong signal to the White House and the European Commission, both currently considering policy statements on open access, that the UK is ready to move which will strengthen the hands of those arguing for strong policy.

This is a big step – and it heads in the right direction. The devil is in the details of implementation. But then it always is.

More will follow – particularly on the financial modelling – when I have a chance to digest more fully. This is a first pass draft based on a quick skim and I may modify this post if I discover I have made errors in my reading.

Enhanced by Zemanta

They. Just. Don’t. Get. It…

English: Traffic Jam in Delhi Français : Un em...
Image via Wikipedia

…although some are perhaps starting to see the problems that are going to arise.

Last week I spoke at a Question Time style event held at Oxford University and organised by Simon Benjamin and Victoria Watson called “The Scientific Evolution: Open Science and the Future of Publishing” featuring Tim Gowers (Cambridge), Victor Henning (Mendeley), Alison Mitchell (Nature Publishing Group), Alicia Wise (Elsevier), and Robert Winston (mainly in his role as TV talking head on science issues). You can get a feel for the proceedings from Lucy Pratt’s summary but I want to focus on one specific issue.

As is common for me recently I emphasised the fact that networked research communication needs to be different to what we are used to. I made a comparison to the fact that when the printing press was developed one of the first things that happened was that people created facsimiles of hand written manuscripts. It took hundreds of years for someone to come up with the idea of a newspaper and to some extent our current use of the network is exactly that – digital facsimiles of paper objects, not truly networked communication.

It’s difficult to predict exactly what form a real networked communication system will take, in much the same way that asking a 16th century printer how newspaper advertising would work would not provide a detailed and accurate answer, but there are some principles of successful network systems that we can see emerging. Effective network systems distribute control and avoid centralisation, they are loosely coupled, and distributed. Very different to the centralised systems for control of access and control we have today.

This is a difficult concept and one that scholarly publishers simply don’t get for the most part. This is not particularly suprising because truly disruptive innovation rarely comes from incumbent players. Large and entrenched organisations don’t generally enable the kind of thinking that is required to see the new possibilities. This is seen in publishers statements that they are providing “more access than ever before” via “more routes”, but all routes that are under tight centralised control, with control systems that don’t scale. By insisting on centralised control over access publishers are setting themselves up to fail.

Nowhere is this going to play out more starkly than in the area of text mining. Bob Campbell from Wiley-Blackwell walked into this – but few noticed it – with the now familiar claim that “text mining is not a problem because people can ask permission”. Centralised control, failure to appreciate scale, and failure to understand the necessity of distribution and distributed systems. I have with me a device capable of holding the text of perhaps 100,000 papers It also has the processor power to mine that text. It is my phone. In 2-3 years our phones, hell our watches, will have the capacity to not only hold the world’s literature but also to mine it, in context for what I want right now. Is Bob Campbell ready for every researcher, indeed every interested person in the world, to come into his office and discuss an agreement for text mining? Because the mining I want to do and the mining that Peter Murray-Rust wants to do will be different, and what I will want to do tomorrow is different to what I want to do today. This kind of personalised mining is going to be the accepted norm of handling information online very soon and will be at the very centre of how we discover the information we need. Google will provide a high quality service for free, subscription based scholarly publishers will charge an arm and a leg for a deeply inferior one – because Google is built to exploit network scale.

The problem of scale has also just played out in fact. Heather Piwowar writing yesterday describes a call with six Elsevier staffers to discuss her project and needs for text mining. Heather of course now has to have this same conversation with Wiley, NPG, ACS, and all the other subscription based publishers, who will no doubt demand different conditions, creating a nightmare patchwork of different levels of access on different parts of the corpus. But the bit I want to draw out is at the bottom of the post where Heather describes the concerns of Alicia Wise:

At the end of the call, I stated that I’d like to blog the call… it was quickly agreed that was fine. Alicia mentioned her only hesitation was that she might be overwhelmed by requests from others who also want text mining access. Reasonable.

Except that it isn’t. It’s perfectly reasonable for every single person who wants to text mine to want a conversation about access. Elsevier, because they demand control, have set themselves up as the bottleneck. This is really the key point, because the subscription business model implies an imperative to extract income from all possible uses of the content it sets up a need for control of access for differential uses. This means in turn that each different use, and especially each new use, has to be individually negotiated, usually by humans, apparently about six of them. This will fail because it cannot scale in the same way that the demand will.

The technology exists today to make this kind of mass distributed text mining trivial. Publishers could push content to bit torrent servers and then publish regular deltas to notify users of new content. The infrastructure for this already exists. There is no infrastructure investment required. The problems that publishers raise of their servers not coping is one that they have created for themselves. The catch is that distributed systems can’t be controlled from the centre and giving up control requires a different business model. But this is also an opportunity. The publishers also save money  if they give up control – no more need for six people to sit in on each of hundreds of thousands of meetings. I often wonder how much lower subscriptions would be if they didn’t need to cover the cost of access control, sales, and legal teams.

We are increasingly going to see these kinds of failures. Legal and technical incompatibility of resources, contractual requirements at odds with local legal systems, and above all the claim “you can just ask for permission” without the backing of the hundreds or thousands of people that would be required to provide a timely answer. And that’s before we deal with the fact that the most common answer will be “mumble”. A centralised access control system is simply not fit for purpose in a networked world. As demand scales, people making legitimate requests for access will have the effect of a distributed denial of service attack. The clue is in the name; the demand is distributed. If the access control mechanisms are manual, human and centralised, they will fail. But if that’s what it takes to get subscription publishers to wake up to the fact that the networked world is different then so be it.

Enhanced by Zemanta