The Marginal Costs of Article Publishing – Critiquing the Standard Analytics Study

Standard Analytics have released a very useful paper looking at platform costs for scholarly publishing. In the paper (which coincidentally demonstrates part of the platform system they are developing) they make some strong claims about the base cost of publishing scholarly articles. In this post I will critique those claims and attempt to derive a cost that fully represents the base marginal cost of article publishing, while pointing out that such average estimates are probably not very useful. The central point is that the paper shows not marginal costs but (a proportion of) the per particle technical platform costs. It is however the case that their central point, that modular, low cost and flexible platforms that create efficiencies of scale, offer the opportunity for radically cheaper scholarly publishing systems.

Missing platform costs

The first criticism to be made is that there are basic services included in scholarly publishing that have not been included in the calculation. Ironically one of those most important of these is the very system that Standard Analytics are demonstrating with their article, that of hosting and discovery. There is an expectation for modern web based publishers that they will provide some form of web portal for content with some level of functionality. While new options, including overlay journals based on archives, are developing, these do not remove this cost but place it elsewhere in the system. Preservation systems such as CLOCKSS and PORTICO are not designed to be the primary discovery and usage site for articles.

Running a high availability site with medium to high (by scholarly standards) traffic is not free. Calculating a per-article marginal cost for hosting is not straightforward, particularly given a lack of standards for the timeframe over which server costs should be amortised (how long do articles need to be available?) and the limited predictability of future bandwidth and storage costs. Many per-article costs quoted online make the common assumption that storage and server costs are dropping faster than the rate of increase in the size of content stores i.e. that the ongoing marginal cost of hosting an article can be neglected. The article would in any case be substantially improved by including per-article costs of hosting from providers such as Atypon, Highwire etc. I would guess these would add $50-150 to the headline costs depending on volume.

However the calculation of this is difficult as an enormous amount of publisher costs in using third party providers are in updates and enhancements. The article explicitly neglects set-up and development costs for third party peer-review management systems, focussing instead on marginal costs. However updates, enhancements, data migrations and similar processes are a core part of keeping publisher offerings up to date. Neglecting this element, which is similarly neglected in the infrastructure costings by dropping out the cost of developing the putative system, will lead to substantial under estimates not just of amortised startup costs but also of ongoing maintenance.

Missing marginal costs

The article describes its focus as being on marginal costs, but it is actually on technical platform and service costs. The paper neglects the tradeoff between academic labour and publisher systems. Martin Eve in his (now a little dated) series on setting up a journal notes that the annual cash costs for running a journal can be very low ($350 a year in his estimate of cash costs in 2012) but he also notes the time commitments involved. Typesetting can be automated but running it on a manuscript and checking the output might take an hour per article if working at a small scale, adding up to as much as $100 ($200,000 full economic cost of an academic divided by a 2,000 hour work year). And as I’ve noted elsewhere, thinking about “average times” is very misleading here, the real time sink is the unusual articles that require substantially more work. Paying at the high end of the ranges of the typesetting services described in the paper reduces the requirements for publisher (or academic) investment in running these systems and doing quality assurance but it doesn’t remove them. At the other end, taking a word document and printing to PDF will work, but doesn’t generate an article format suitable for a modern web based world.

Of course the article also misses the largest marginal cost, that of peer review itself, both the management and the reviewing. We have very little information in practice on how these costs are spread between professional publisher provided services and the academic community. It is difficult to estimate costs on the academic side (again the issue of whether “averages” are helpful raises its head) and publishers provide little or no information on their costs. More importantly we have no real idea of how investment on one side of the fence reduces (or perhaps increases!) the costs on the other. It might be argued that professional editors within a publisher should be more efficient than academic editors, being professionally focussed on a single task rather than spread between many. And of course the inverse is often argued, that professional editors do not create the same value (or are “doing the wrong thing”) because they are not academic community members. Of course both of these can simultaneously be true, for various interpretations of “efficient” and “value”.

The ongoing cost of development is part of the cost of doing business, open source or third party provided

The biggest issue in the paper is the way that infrastructure costs are presented. Firstly it removes the costs of development of the system from the equation. It might well be the case that grant funded, or donated technologies will evolve where the costs don’t need to be recouped from users. That doesn’t mean those costs don’t exist, just that someone else is footing the bill. But its important to note that Standard Analytics intend to make an offering in this space and they will presumably be looking to recoup at least some of their investment. What they are doing looks interesting and I don’t doubt they could substantially undercut incumbent providers, but it will be worth more than $1 per article, precisely because they’ll be offering additional services and probably customisations.

It is generally the case that in making the choice between a third party provider with a proprietary platform and running an open source platform in house that the balance sheet cash costs are approximately the same. What you pay in license or service fees to a provider tends to cost about the same in cash terms on a day to day basis as having the expertise and capacity in house. Service providers reap efficiencies of scale which they convert into their margin. There can be substantial non-cash benefits to having the capacity in house but its also a liability in terms of overheads and staff obligations. Rarely is Open Source cheap. In the particular case of scholarly publishing platforms there is a possibility for new platforms to be cheaper because they are generally more modern, more flexible and easier to work with than the proprietary systems of third party providers. That doesn’t make them free, and certainly doesn’t mean the cost of running them in house is the same as the cloud provisioning costs. Indeed most publishers are likely to continue to outsource these services, whether the platforms are proprietary or open source. Open Source will mainly serve as a brand advantage, offering confidence that there will be none of the vendor lock-in that plagues the space.

The real distinction is not between Open Source and proprietary but between a new generation of platforms that are built for and of the web, and old systems based on pre-web document management systems. Here is where I agree with the paper. Modular and flexible systems, particularly those that exploit the possibilities of web-native authoring and information management, have a massive potential to reduce the overall costs of scholarly publishing. Lots of the pieces are falling into place, including work being done by Standard Analytics, the Collaborative Knowledge Foundation, PLOS (Aperta), eLife (Lens), Overleaf, Substance, FidusWriter, Scholastica, OpenJournal and others I’ve no doubt forgotten. What is interesting is that everyone misses the fact that these pieces don’t yet join up, that the costs of developing standards to make these pieces fit together will also be part of the puzzle.

Ongoing development, customisation and improvement is a cost that doesn’t go away. Unless we are prepared to draw a line and say that scholarly communications is finished there will always be ongoing costs. These overheads can be hidden in grants or subsidies, or appear on the balance sheet as “innovation” or as part of general staff costs and overheads, or they can be bundled up with per article costs but they don’t go away. How they are accounted is largely an arbitrary choice of the organisation, not a distinction in actual activities, efficiency or margins.

Inefficient vs legacy technology vs unnecessary

It’s easy to add up the obvious costs of bits of a scholarly communication pipeline and observe that they come to less than what people generally charge. Academics often have a habit of simply deciding any bit that isn’t obvious is unnecessary. This is often built on arrogance and ignorance. Equally publishers often have a habit of defending the status quo as the way it has to be without actually addressing whether a given component is necessary. We need a much better conversation about what all the pieces are, and how much value all of them add, from discovery layers and marketing, through typesetting, copy editing, and pagination, to the costs created by the continued insistence on using Word and LaTeX.

A lot of the costs are tied up with inefficiencies and misunderstandings that lead to an attitude of “if I want to do it properly I’ll have to do it myself” on all sides of the academic/publisher/system provider divide. An enormous amount of work arounds, patch jobs and replication go on, and the work arounds generate further work arounds as they fail to engage with the needs of another group of stakeholders. As noted above we have no real understanding of how investments by academic editors in the form of time lead to savings on the publisher side, and vice versa. And a lot of those inefficiencies relate to technology, which as I noted previously is difficult to replace piece by piece.

There are savings to be made, but to make them requires unpicking these three separate issues. What services and outcomes are needed (and by what communities)? Which of those services and outcomes are currently being delivered inefficiently? And of those, where can new technology make a difference? And for all of these what are the social and political challenges in actually achieving change?

Conclusion

This is a very useful new paper that gives up to date prices for vendors for some (but not all) important parts of the scholarly communications pipeline. Adding hosting costs as well as costs of providing indexing and discovery tools would be a valuable addition. The headline prices quoted cannot be relied on, although they provide some useful lower bounds for a traditional technical platform service. It would be very valuable if the differentials in pricing, particularly for typesetting were broken out in terms of differences in service.

What this kind of analysis doesn’t, and can’t, provide is an understanding of how costs are coupled through the pipeline. Many of the costs of typesetting are due to ameliorating the problems of authoring platforms. How do poor manuscript tracking systems add to the costs of managing and conducting review? How could professional editors and support staff work more effectively with academic editors and reviewers to maximise value and reduce costs all around? And how might different service configurations actually align “publisher” and “academic” interests so as to jointly reap the benefits of potential savings rather than obfuscating the costs from each other?

For what its worth I’d guess the likely base level cost of a modern platform for a journal publishing 50-500 articles a year providing submission through publication in JATS XML where the selection (and risk of mistakes) are loaded heavily on the academic community at around $450-550. This is more or less the Ubiquity Press model. If you want to protect the academic community from the more clever forms of fraud, less obvious ethical breaches and the brand risk of making mistakes that price will rise several hundred dollars. On a large scale this is what PLOS ONE and (on a smaller scale and with less of the cash costs on the balance sheet) what PeerJ do. Scale matters, smaller than 50 you are running an operation on spare academic time, probably highly inefficiently. More than 500 requires scale and professionalisation, which in turn moves both risks and costs from the community side to the balance sheet as I have discussed previously.

A good chunk of that $500 is tied up in dealing with legacy technical and workflow issues. New technology will help here but implementing it has serious challenges for both development and adoption. Funding systems as shared infrastructure rather than on a recurrent basis per article also offers savings. The challenge lies in finding technological solutions and economic models that allow research communities to insource what they need, taking advantage of the subsidy we can provide through our time in highly specific niche spaces, while gaining the economies of scale that shared systems can provide. And we need much better numbers to do that. Perhaps more importantly a truly modular system, with the competition between modules that will drive down prices, and show the real potential of Open Source approaches requires much greater community coordination on interchange standards.

The real trick with this kind of analysis is being able to figure out whether there are true like-for-like efficiency savings, or whether the costs are just being pushed elsewhere. This paper tells us about some of the cost elements which will be helpful in tackling both these questions, but it doesn’t provide the answers.

Not what, not who, or how, but Why is Open?

This is an approximation of my talk at OpenCon Cambridge on Thursday 26 November 2015. Any inaccuracies entirely mine. There is also a video recording of the actual talk.

The idea of OpenCon was borne out of the idea that the future belonged to young scholars and that helping them to be effective advocates for the world they want to work in was the best way to make progress. With that in mind I wanted to reflect not on the practical steps, of which there will be much more this afternoon, but on some questions that might help when the going gets tough. When you might question why you would want to continue to try to change the world?

We often talk about the “what” of Open. Open Source, Open Access, Open Data, Open Educational Resources. We parcel the world up into those things which are open by some definition, and those which are not. Definitions are important, particularly for political purposes. We can, and no doubt will, continue to argue about which licenses should be allowed. These dichotomies are almost certainly too harsh. But that’s not what I want to talk about today.

Given the success, some might say too-fast progress of the Open movement we often find ourselves asking who is open. In a world where those we might have previously regarded as the enemy suddenly appear as new allies who should we allow to call themselves open? Open-washing is a real problem, some players use these ideas cynically as marketing tools or smoke screens. But there are people in unusual places genuinely engaged with changing the system. Some more successfully and thoughtfully than others it must be said but should we discourage them by refusing them the label? How we can manage both inclusion and rigour is an important question we need to consider. But this is also not what I want to talk about today.

Nor do I want to talk about the how. This is important, and others will cover it better than I can today. No, today I want to talk about the why of Open.

First there are the big whys, the motivations for greater access to research and scholarship and the greater agency this can bring to people’s lives, for an improved democracy, a better environment, healthier lives and an informed society. All these are good motivations, but with the success of the Open movements we perhaps need to hold ourselves to a higher standard of scrutiny as to whether we are delivering.

Does Open Access truly put research in people’s hands or do we have more work to do on discoverability, comprehensibility? Is Open Government really putting more power in the hands of the disadvantaged or is it just accentuating the gaps between those who already had access to the levers of our democracy and exacerbating a growing wealth divide with an increasing digital one? Could we do more to link the needs of real patients with the direction of medical research? How can we measure progress against these much more challenging goals when we don’t even know how much of the literature is publicly available?

I think we need to do better, not in just articulating the reasons for open approaches but in testing and measuring whether those approaches are delivering. We need to do more than just throw the existing content over the wall of the ivory tower if we want to change the asymmetries of power and privilege that are tied to the existing asymmetries of information. For many of us that’s why we got into this in the first place.

Why brings me to the second set of whys. The smaller ones, but the ones that matter personally, and the ones that sustain us in the face of setbacks. I know many people came into advocacy for Open because of a personal frustration, the inability to do something. In many cases a personal loss played a role. Sometimes that motivation comes out of a deep sense of who we are, and sometimes its a result of happenstance and serendipity. I got into this space by accident, and I kept doing it because it seemed to work. And lets be honest, some of us got into this initially to impress someone.

We don’t often talk about these underlying motivations, but perhaps we need to, because without understanding them we avoid talking about a tension that lies at the heart of the movement. The different pathways that brought us to this space tend to divide into two. On one side we have those who concerned with freedoms, about the freedom to act. Stallmans focus of Free Software around the three freedoms is perhaps the most obvious example of this strand. This is an individualistic, sometimes even libertarian view. On the other side are those who want to build communities of practice, sharing communities, that create value for each other.

Often these strands work together but there is a tension at their heart and it surfaces in our work. When we talk about licensing we talk about the freedoms of users and skip across the freedoms that authors are giving up. Freedom to act (or the lack of it) often brings people together but by definition becoming a community means giving up freedoms. There are benefits to be sure. Simply being a part of this community, attending these meetings is one of them. But if we are to work effectively together, to strengthen what each of us do, then we need to understand our own motivations better. By respecting the desire for freedom as well as recognising which ones we need to collectively give up we will build a stronger movement. And to do that we need to question each other.

And that questioning brings me to the third why. Because in the end what drives all this is the questioning of how the world does (and could) work, how we can make it better, what we can build to help that. It is the open to of new ideas, the open with of inclusion that supports a diversity that can bring those new ideas. At the root of all the frustrations, all the building, all the desire for community, is this question of why the world is as it is. It’s that questioning that makes us human, and as is often the case it is children who are the most human, asking the question repeatedly in that infinite regress of “why….and why is…but why…”.

In the end it isn’t actually a question. It’s the ability to question that matters. It’s a statement.

Why is open.

 

The end of the journal? What has changed, what stayed the same?

goodsThis is an approximate rendering of my comments as part of the closing panel of “The End of Scientific Journal? Transformations in Publishing” held at the Royal Society, London on 27 November 2015. It should be read as a reconstruction of what I might have said rather than an accurate record. The day had focussed on historical accounts of “journals” as mediators of both professional and popular research communications. A note of the meeting will be published. Our panel was set the question of “will the journal still exist in 2035”.

Over the course of 2015 I’ve greatly enjoyed being part of the series of meetings looking at the history of research communications and scientific journals in the past. In many cases we’ve discovered that our modern concerns, today the engagement of the wider public, the challenge of expertise, are not at all new, that many of the same issues were discussed at length in the 17th, 18th and 19th centuries. And then there are moments of whiplash as something incomprehensible streaks past: Pietro Corsi telling us that dictionaries were published as periodicals; Aileen Fyfe explaining that while papers given to at Royal Society meetings were then refereed, the authors could make only make “verbal” not intellectual changes to the text in response; Jon Topham telling us that chemistry and physics were characterised under literature in the journals of the early 19th century.

So if we are to answer the exam question we need to address the charge that Vanessa Heggie gave us in the first panel discussion. What has remained the same? And what has changed? If we are to learn from history then we need to hold ourselves to a high standard in trying to understand what it is (not) telling us. Prediction is always difficult, especially about the future…but it wasn’t Niels Bohr who first said that. A Dane would likely tell us that “det er svært at spÃ¥, især om fremtiden” is a quote from Storm Peterson or perhaps Piet Hein, but it probably has deeper roots. It’s easy to tell ourselves compelling stories, whether they say that “everything has changed” or that “it’s always been that way”, but actually checking and understanding the history matters.

So what has stayed the same? We’ve heard throughout today the importance of groups, communities. Of authors, of those who were (or were trying to be) amongst the small group of paid professors at UK universities. Of the distinctions between communities of amateurs and of professionals. We’ve heard about language communities, of the importance of who you know in being read at the Royal Society and of the development of journals as a means of creating research disciplines. I think this centrality of communities, of groups, of clubs is a strand that links us to the 19th century. And I think that’s true because of the nature of knowledge itself.

Knowledge is a slippery concept, and I’ve made the argument elsewhere, so for now I’ll just assert that it belongs in the bottom right quadrant of Ostrom‘s categorisation of goods. Knowledge is non-rivalrous – if I give it to you I still have it – but also excludable – I can easily prevent you from having it, by not telling you, or by locking it up behind a paywall, or simply behind impenetrable jargon. This is interesting because Buchannan‘s work on the economics of clubs show us that it is precisely the goods in this quadrant which are used to sustain clubs and make them viable.

The survival of journals, or of scholarly societies, disciplines or communities, therefore depends on how they deploy knowledge as a club good. To achieve this deployment it is necessary to make that knowledge, the club good, less exclusive, and more public. What is nice about this view is that it allows us, to borrow Aileen Fyfe’s language, to talk about “public-making” (Jan Velterop has used the old term “publicate” in a similar way) as an activity which includes public engagement, translation, and – to Rebekah Higgitt‘s point – education as well as scholarly publishing as we traditionally understand it as overlapping subsets of this broader activity.

But what has changed? I would argue that the largest change in the 20th century was one of scale. The massive increase in the scale and globalisation of the research enterprise, as well as the rise of literacy meant that traditional modes of coordination, within scholarly societies and communities, and beyond to interested publics were breaking down. To address this coordination problem we took knowledge as a club good and privatised it, introducing copyright and intellectual property as a means of engaging corporate interests to manage the coordination problem for us. It is not an accident that the scale up, the introduction of copyright and IP to scholarly publishing, and scholarly publishing becoming (for the first time) profitable all co-incide. The irony of this, is that by creating larger, and clearly defined markets, we solved the problem of market scale that troubled  early journals that needed to find both popular and expert audiences, by locking wider publics out.

The internet and the web also changed everything, but its not the cost of reproduction that most matters. The critical change for our purpose here is the change in the economics of discovery. As part of our privatisation of knowledge we parcelled it up into journals, an industrial broadcast mechanism in which one person aims with as much precision as possible to reach the right, expert, audience. The web shifts the economics of discovering expertise in a way that makes it viable to discover, not the expert who knows everything about a subject, but the person who just happens to have the right piece of knowledge to solve a specific problem.

These two trends are pulling us in opposite directions. The industrial model means creating specialisation and labelling. The creation of communities and niches that are, for publishers, markets that can be addressed individually. These communities are defined by credentialling and validation of deep expertise in a given subject. The ideas of micro-expertise, of a person with no credentials having the key information or insight radically undermines the traditional dynamics of scholarly group formation. I don’t think it is an accident that those scholarly communities that Michèle Lamont identifies as having the most stable self conception have a tendency to being the most traditional in terms of their communication and public engagement. Lamont identifies history (but not as Berris Charnley reminded me the radicals from the history of science here today!) and (North American analytical) philosophy in this group. I might add synthetic chemistry from my own experience as examples.

It is perhaps indicative of the degree of siloing that I’m a trained biochemist at a history conference, telling you about economics – two things I can’t claim any deep expertise in – and last week I gave a talk from a cultural theory perspective. I am merrily skipping across the surface of these disciplines, dipping in a little to pull out interesting connections and no-one has called me on it*. You are being forced, both by the format of this panel, and the information environment we inhabit, to assess my claims not based on my PhD thesis topic or my status or position, but on how productively my claims and ideas are clashing with yours. We discover each other, not through the silos of our disciplinary clubs and journals, but through the networked affordances that connect me to you, that in this case we could trace explicitly via Berris Charnley and Sally Shuttleworth. That sounds to me rather more like the 19th century world we’ve been hearing about today than the 20th century one that our present disciplinary cultures evolved in.

This restructuring of the economics of discovery has profound implications for our understanding of expertise. And it is our cultures of expertise that form the boundaries of our groups – our knowledge clubs – whether they be research groups, disciplines, journals, discussion meetings or scholarly societies. The web shifts our understanding of public-making. It shifts from the need to define and target the expert audience through broadcast – a one-to-audience interaction – to a many-to-many environent in which we aim to connect with the right person to discover the right contribution. The importance of the groups remains. The means by which they can, and should want to communicate has changed radically.

The challenge lies, not in giving up on our ideas of expertise, but in identifying how we can create groups that both develop shared understanding that enables effective and efficient communication internally but are also open to external contributions. It is not that defining group boundaries doesn’t matter, it is crucial, but that the shape and porosity of those boundaries needs to change. Journals have played a role throughout their history in creating groups, defining boundaries, and validating membership. That role remains important, it is just that the groups, and their cultures, will need to change to compete and survive.

We started 2015 with the idea that the journal was invented in 1665. This morning we heard from Jon Topham that the name was first used in the early 19th century, but for something that doesn’t look much like what we would call a journal today. I believe in 20 years we will still have things called journals, and they will be the means of mediating communications between groups, including professional scholars and interested publics. They’ll look very different from what we have today but their central function, of mediating and expressing identity for groups will remain.

* This is not quite true. Martin Eve has called me on skipping too lightly across the language of a set of theoretical frameworks from the humanities without doing sufficient work to completely understand them. I don’t think it is co-incidental that Martin is a cultural and literary scholar who also happens to be a technologist, computer programmer and deeply interested in policy design and implementation, as well as the intersection of symbolic and financial economies.

PolEcon of OA Publishing: What are the assets of a journal?

Victory Press of Type used by SFPP (Photo credit: Wikipedia)
Victory Press of Type used by SFPP (Photo credit: Wikipedia)

This post wasn’t on the original slate for the Political Economics of Publishing series but it seems apposite as the arguments and consequences of the Editorial Board of Lingua resigning en masse to form a new journal published by Ubiquity Press continue to rumble on.

The resignation of the editorial board of Lingua from the (Elsevier owned) journal to form a new journal, that is intended to really be “the same journal” raises interesting issues of ownership and messaging. Perhaps even more deeply it raises questions of what the real assets of a journal are. The mutual incomprehension on both sides really arises from very different views of what a journal is and therefore of what the assets are, who controls them, and who owns them. The views of the publisher and the editorial board are so incommensurate as to be almost comical. The views, and more importantly the actions, of the group that really matters, the community that underlies the journal, remains to be seen. I will argue that it is that community that is the most important asset of the strange composite object that is “the journal” and that it is control of that asset that determines how these kinds of process (for as many have noted this is hardly the first time this has happened) play out.

The publisher view is a fairly simple one and clearly expressed by Tom Reller in a piece on Elsevier Connect. Elsevier no doubt holds paperwork stating that they own the trademark of the journal name and masthead and other subsidiary rights to represent the journal as continuing the work of the journal first founded in 1949. That journal was published by North Holland, an entity purchased by Elsevier in the 90s. The work of North Holland in building up the name was part of the package purchased by Elsevier, and you can see this clearly in Reller’s language in his update to the post where he says Elsevier sees the work of North Holland as the work of Elsevier. The commercial investment and continuity is precisely what Elsevier purchased and this investment is represented in the holding of the trademarks and trading rights. The investment, first of North Holland, and then Elsevier in building up the value of these holdings was for the purpose of gaining future returns. Whether the returns are from subscription payments or APCs no longer matters very much, what matters is realising them and retaining control of the assets.

As a side note the ownership of these journals founded in the first half of the twentieth century is often much less clear than these claims would suggest. Often the work of an original publisher would have been seen as a collaboration, contracts may not exist and registering of trademarks and copyright may have come much later. I know nothing about the specifics of Lingua but it is not uncommon for later instantiations of an editorial board to have signed over trademarks to the publisher in a way that is legally dubious. The reality of course is that legal action to demonstrate this would be expensive, impractical and pretty risky. A theoretical claim of legal fragility is not much use against the practical fact that big publishers can hire expensive lawyers. The publisher view is that they own those core assets and have invested in them to gain future returns.  They will therefore act to protect those assets.

The view of the editorial board is almost diametrically opposed. They see themselves as representing a community of governance and creating and providing the intellectual prestige of the journal. For the editorial board that community prestige is the core asset. With the political shifts of Open Access and digital scholarship questions of governance have started to play into those issues of prestige. Communities and journals that want to position themselves as forward looking, or supporting their community, are becoming increasingly concerned with access and costs. This is painted as a question of principle, but the core underlying concern is the way that political damage caused by lack of access, or by high (perceived) prices will erode the asset value of the prestige that the editorial board has built up through their labour.

This comes to a head where the editorial board asks Elsevier to hand over the rights to the journal name. For Elsevier this is a demand to simply hand over an asset, the parcel of intellectual property rights. David Mainwaring’s description of the demands of the board as “a gun to the head” gives a sense of how many people in publishing would view that letter. From that perspective it is clearly unreasonable, even deliberately so, intended to create conflict with no expectation of resolution. This is a group demanding, with menaces, the handover of an asset which the publisher has invested in over the years. The paper holdings of trademarks and trading rights represent the core asset and the opportunities for future returns. A unilateral demand to hand them over is only one step from high way robbery. Look closely at the language Reller uses to see how the link between investment and journal name, and therefore those paper holdings is made.

For the editorial board the situation is entirely different. I would guess many researchers would look at that letter and see nothing unreasonable in it at all. The core asset is the prestige and they see Elsevier as degrading that asset, and therefore the value of their contribution over time. For them, this is the end of a long road in which they’ve tried to ensure that their investment is realised through the development of prestige and stature, for them and for the journal. The message they receive from Elsevier is that it doesn’t value the core asset of the journal and that it doesn’t value their investment. To address this they attempt to gain more control, to assert community governance over issues that they hadn’t previously engaged with. These attempts to engage over new issues – price and access – are often seen as naive, or at best fail to connect with the publisher perspective. The approaches are then rebuffed and the editorial group feel they have only a single card left to play, and the tension therefore rises in the language that they use. What the publisher sees as a gun to the head, the editorial board see as their last opportunity to engage on the terms that they see as appropriate.

Of course, this kind of comprehension gap is common to collective action problems that reach across stakeholder groups, and as in most cases that lack of comprehension leads to recrimination and dissmissal of the other parties perspective as on one hand motivated by self-interest on one hand and on the other by naivety and arrogance. There is some justice in these characterisations but regardless of which side of the political fence you may sit it is useful to understand that these incompatible views are driven by differing narratives of value, by an entirely different view as to what the core assets are. On both sides the view is that the other party is dangerously and wilfully degrading the value of the journal.

Both views; that the value of the journal is the prestige and influence realised out of expert editorial work, and that the value is the brand of the masthead and the future income it represents, are limited. They fail to engage with the root of value creation in the authorship of the content. The real core asset is the community of authors. Of course both groups realise this. The editorial board believes that they can carry the community of authors with them. Elsevier believes that the masthead will keep that community loyal. The success or failure of the move depends on which of them is right. That answer is that probably both are to some extent which means the community gets split, the asset degraded in a real-life example of the double-defection strategy in a Prisoner’s Dillemma game.

Such editorial board resignations are not new. There have been a number in the past, some more successful than others. It is important to note that the editorial board is not the community, or representative of them. It is precisely those cases where the editorial board is most directly representing the community of authors where defections are most successful. On the past evidence Elsevier are probably correct to gamble that the journal will at least survive. The factors in favour of the editorial board are that Linguistics is a relatively small, tight knit community, that they have a credible (and APC free) offer on the table that will look and feel a lot like the service offering that they had. I would guess that Lingua authors are focussed on the journal title and only think of the publisher as a distant second issue, if they are even aware of who the publisher is. In that sense the emergence of new lean publishers like Ubiquity Press and consortial sustainability schemes like Open Library of Humanities are a game changer, offering a high quality experience that otherwise looks and feels like a traditional journal process (again, it is crucial to emphasise that lack of APCs to trouble the humanities scholar) while also satisfying the social, funder and institutional pressure for Open Access.

Obviously my sympathies lie with the editorial board. I think they have probably the best chance to make this work we have yet seen. The key is to bring the author community with them. The size and interconnections of this specific community make this possible.

But perhaps more interesting is to look at it from the Elsevier perspective. The internal assessment will be that there were no options here. They’ve weathered similar defections in the past, usually with success and there would be no value in acceding to the demands of the editorial board. The choice was to hold a (possibly somewhat degraded) asset or to give it away. The internal perception will be that the new journal can’t possibly survive, probably that Ubiquity Press will be naively under funded and can’t possibly offer the level of service that the community will expect. Best case scenario is steady as she goes, with a side order of schadenfreude as the new journal fails; worst case, the loss of value to a single masthead. And on the overall profit and loss sheet a single journal doesn’t really matter as its the aggregate value that sells subscription bundles. Logical analysis points to defection as the best move in the prisoner’s dilemma.

Except I think that’s wrong, for two reasons. One is that this is not a single trial prisoners dillemma, its a repeated trial with changing conditions. Second, the asset analysis plays out differently for Elsevier than it does for the editorial board making the repeated trials more important. The asset for the editorial board is the community of the journal. The asset for Elsevier is the author community of all their journals. Thus for the editorial board they are betting everything on one play – they are all in, hence the strength of the rhetoric being deployed. Elsevier need to consider how their choices may play into future conditions.

Again, the standard analyis would be “protect the assets”. Send a strong message to the wider community (including shareholders), that the company will hold its assets. The problem for me is that this is both timid, and in the longer term potentially toxic. It’s timid, compared to what should surely be a robust internal view of the value that Elsevier offers. That the quality services they provided simply cannot be offered sustainably at a lower price. The confident response would be to call the board’s bluff, put the full costing and offer transparently on the table in front of the community and force them to do a compare and contrast. More than just saying “it can’t be done at the price you demand” put out the real costs and the real services.

The more daring move would be to let the editorial board take the name on a “borrow and return” basis, giving Elsevier first right of refusal if (and in the Elsevier view, when) they find that they’re not getting what they need in their new low cost (and therefore, in the Elsevier view, low service) environment. After all, the editorial board already have money to support APCs according to their letter. It’s risky of course, but again it would signal strong confidence in the value of the services offered. Publishers rarely have to do this, but I find it depressing that they almost always shy away from opportunities to really place their value offering in a true market in front of their author communities. To my mind it shows a lack of robust internal confidence in the value they offer.

But beyond the choice of actions there’s a reason why this standard approach is potentially toxic, and potentially more toxic long term even if, perhaps especially if, Elsevier can continue to run the journal with a new board. If Elsevier are to protext the existing asset as they see it, they need to make the case that the journal can continue to run as normal with a new board. The problem is that this case can only be made if the labour of editors is interchangeable, devaluing the contribution of the existing board and by extension the contribution of all other Elsevier editorial boards. If Elsevier can always replace the board of a journal then why would an individual editor, one who believes that it is their special and specific contribution that is building journal prestige, stay engaged? And if its merely to get the line on their CV and they really don’t care, how can Elsevier rely on the quality of their work? Note it is not that Elsevier don’t see the value of that contributed labour, it is clear that editors are part of the value creation chain that adds to Elsevier income, but that the situation forces them to claim that this labour is interchangeable. Elsevier see the masthead as the asset that attracts that labout. The editorial board see their labour and prestige as the asset that attracts the publisher investment in the masthead.

You can see this challenge in Elsevier statements. David Clark, interviewed as part of a Chronicle piece is quoted as follows:

He sees the staff departures as a routine part of the publishing world. “Journals change and editors change,” Mr. Clark said. “That happens normally.”

And Tom Reller in the statement on the Elsevier website:

The editors of Lingua wanted for Elsevier to transfer ownership of the journal to the collective of editors at no cost. Elsevier cannot agree to this as we have invested considerable amount of time, money and other resources into making it a respected journal in its field. We founded Lingua 66 years ago.

You can see here the attempt to discount the specific value of the current editorial board, but in terms that are intended to come across as conciliatory. Elsevier’s comms team are clearly aware of the risk here. Too weak a stance would look weak (and might play badly with institutional share holders) and too strong a stance sends a message to the community that their contribution is not really valued.

This is the toxic heart of the issue. In the end if Elsevier win, then what they’ve shown is that the contribution of the current editorial board doesn’t matter, that the community only cares about the brand. That’s a fine short term win and may even strengthen their hand in subscription negotiations. But it’s utterly toxic to the core message that publishers want to send to the research communities that they serve, that they are merely the platform. It completely undermines the value creation by editorial boards that Elsevier relies on to sell journals (or APCs) and generate their return on investment.

Playing both sides worked in the world before the web, when researchers were increasingly divorced from any connection with the libraries negotiating access to content. Today, context collapse is confronting both groups. Editorial boards suddenly becoming aware that they had acquiesced in giving up control, and frequently legal ownership of “their” journal, at the same time as issues of pricing and cost are finally coming to their attention. Publishers in general, and Elsevier in particular, can’t win a trick in the public arena because their messaging to researchers, lobbying of government, and actions in negotiation are now visible to all the players. But more than that, all those players are starting to pay attention.

The core issue for Elsevier is that if they win this battle, they will show that it is their conception of the core assets of the journal that is dominant. But if that’s true then it means that editorial boards contribute little or no value. That doesn’t mean that a “brand only” strategy couldn’t be pursued, and we will return to the persistence of prestige and brand value in the face of increasing evidence that they don’t reflect underlying value later in the series. But that’s a medium term strategy. In the longer term, if Elsevier and other publishers continue to seek to focus on and hold the masthead and trademarks as the core asset of the journal that they are forced into a messaging and communications stance that will be ultimately disasterous.

There’s no question that Elsevier understands the value that editorial board contributions bring. But continuing down the ownership path through continued rebellions will end up forcing them to keep signalling to senior members of research communities that their personal contribution has no value, that they can easily be replaced with someone else. In the long term that is not going to play out well.

The Limits on “Open”: Why knowledge is not a public good and what to do about it

This is the approximate text of my talk at City University London on 21 October for Open Access Week 2015. If you prefer the “as live” video version then that its available at YouTube. Warning: Nearly 6000 words and I haven’t yet referenced it properly, put the pictures in or done a good edit…

Eight years of Open Access Week, massive progress towards greater access to research, not to mention data and educational resources. The technology landscape has shifted, the assessment landscape has shifted. The policy landscape has certainly shifted. And yet…amongst friends we can express some misgivings can’t we?

Certainly this has been harder and longer than expected, certainly harder and longer than those who have been pushing on wider access for twenty years had hoped, or expected. Of course we have learned, internalised through bitter experience, that collective action problems are hard, that while technical problems are easy, the social ones are hard. That cultural change is slow and painful.

But this isn’t my biggest concern. My real concern is that, just as it seems we have tipped the balance, made progress towards Open Scholarship inevitable, that it is too easy for that momentum to be subverted. For well resourced players to swoop in and take advantage of these shifts. For incumbent players to protect their existing positions, and profit margins, and to prevent new innovative players from eating into their market. Yet at the same time, while we (often) cheer on those small scrappy new players, we become less enamoured as they mature and grow, become more embedded into the traditional landscape and, perhaps, start to look and sound like those old incumbents, or be bought by them.

I believe we can understand what is happening, and understand what we could choose to do about it, through a shift in the way we think about the processes involved. For me Open Access was never the goal, it is a means to an end. That end might be loosely characterised as “Open Knowledge” or a “Knowledge Commons”, a concept that is half socialist utopian and half neo-liberal; that technology can enable the effective sharing of knowledge as a resource, and that that shared resource is a more effective base for value creation. That the sharing economy of Uber and AirBnB would co-exist, indeed would mutually support a sharing economy built on donated time, shared co-creation for its own sake. Under-exploited capital working hand in hand with Clay Shirky’s cognitive surplus to make better markets but also a better world.

Knowledge as a Public Good

Central to this is the idea that Knowledge is a Public Good. It’s important to pin down the technical meaning of this because it is critical for the argument I’m going to make. Elinor Ostrom in her book, Governing the Commons, divides economic goods, units of exchange, on two axes. You can tell this is Social Sciences because there’s a quandrant diagram. Her first axis is whether goods are rivalrous or non-rivalrous. A rivalrous good is like an apple, or a chocolate if you prefer. If you take it from me, I no longer have it, either to exchange or to use. A non rivalrous good is like the flame on a candle. I can use mine to light yours and both of our candles remain lit. The second axis is excludability. How easy is it to prevent someone from using a good? I can lock a book in a room and stop you accessing it. I can’t easily prevent you from using the air or public roads.

Private goods (like money, or food, or your home) are both rivalrous and excludable. Public Goods, like roads, the air, or libraries are neither rivalrous nor excludable. “Natural” incentives exist to create private goods, accumulation seems to be a natural aspect of human character, something that Hirschmann notes in his wonderful book on the arguments of early capitalism. But Public Goods require different forms of provision. Because they are non-rivalrous they can create great benefits, and enable the creation of great value. Because they are non-excludable they come with a natural free-rider problem. Why would anyone pay to create them if anyone can use them without contributing? Public Goods need to be resourced at a global level, most often by government provision out of general taxation. They need to be mandated through legislation. The wearing of seatbelts is a Public Good, as is the provision of courts and institutional systems of the state that support its functioning. Both rely on systems of legislation and compulusion.

It is natural to think about knowledge as a Public Good. It is infinitely shareable and, once given, can’t be taken away. Therefore it follows that in the provision of Knowledge as a Public Good we need to apply those tools that can successfully support the creation of Public Goods. Mandates. Taxpayer funded infrastructures. If we are to remove the barriers to build the knowledge commons we need to nationalise knowledge. Taxpayers fund research, therefore taxpayers should benefit from, and have access to that research. It feels obvious, and natural, and it drives much of the rhetoric of Open Access advocacy. And that rhetoric, focussed as it is on Public Good provision has been enormously successful in providing resources and requirements.

The roots of our culture

So we are done right? The argument is won and it is merely a case of waiting for turnover to occur. So why do we feel so queasy when Springer moves successfully into Open Access? When Academia.edu hoovers up the space that Institutional Repositories were supposed to fill? When Mendeley is purchased by Elsevier? If we follow our own argument, this is what we want. Government has signalled a shift in its provision and the market has responded; commercial players shifting from content provision towards service provision. We can quibble about costs but the fundamentals are right surely?

And yet…(again)…there is something not quite right. There was more to this than just liberating knowledge from product to commons. There is an issue of control. Of governance and direction. A sense that service provision should align better with the values of the scholarly community. But what would that look like? Do we even have shared values?

As advocates of a technically enabled future we often talk about cultural change as though we need new values, new culture. But more often than not what looks new has deep roots. More and more it seems to me that what we need is a return to past values. As a person with a background in the natural sciences my cultural roots are strongly tied to the writings of Robert Boyle and creation of the UK’s Royal Society in the 17th Century. Boyle’s writing frequently focusses on the detailed reporting, the disclosure of failed experiments, sharing of data and reproducibility. The antique language jars but these startlingly modern concerns might have leapt from the page of a training handbook for new graduate students.

Of my being somewhat prolix in many of my Experiments, I have these Reasons to render[…] That in divers cases I thought it necessary to deliver things circumstantially, that the Person I addressed them to might, without mistake, and with as little trouble as is possible, be able to repeat such unusual Experiments

Robert Boyle, New Experiments Physico-Mechanical Touching on the Spring of the Air

In this 350th Anniversary year of the first publication of Philosophical Transactions the focus has been on the invention of the journal as the creation myth of scholarly communications. But to me the journal was merely one of several means to an end; communicating this developing process of enquiry effectively to the right community. Our origin myth, our expression of values, should be roooted in that goal of communication, not its vessel. But this is scientific enquiry. I first found this quote in Shapin and Schaeffer’s The Leviathan and the Air Pump, a text that arguably was a major contributor to the “Science Wars” and which famously ends with the contention that Boyle was wrong and “Hobbes was right”. Are the Humanities different?

I’ve been criticised for not understanding the humanities, and yet I’ve just been appointed to a humanities faculty. I’m not sure whether that means my opinion carries more weight or less but I have come to the conclusion that there is much more in common across scholarship than there is dividing us. What unites us is a rhetorical approach to argument, that builds on evidence using methods accepted within a community. What counts as evidence, and what methods are allowable, differs amongst our various disciplinary communities but the rhetorical structure of our arguments is very similar. Anita de Waard points out that bioscience articles have the same structure as fairy tales, and in fact this seems to hold, although I can’t claim to have done as close an analysis, across all research domains. We tell “stories, that persuade” in Anita’s words, with evidence.

I don’t know enough of the history of early modern scholarship to pin this down but my sense is that the standard story-telling structure of communication in what we now call the sciences was actually borrowed, from what we would now call the humanties, sometime between the mid-17th and early 19th centuries. Regardless Boyle, and modern humanists as well as natural scientists share some ideals in the requirement for review and criticism, and for attacking the ideas and not the person. While Peer Review as practiced today has only recently become regarded as a core criteria, the means by which scholarly work is identified and badged, review and criticism by peers of evidence and argument goes back a lot further.

Knowledge Clubs

We’ve gone from Open Knowledge and Public Goods to review by peers, almost by definition an exclusive club with rules of admittance and engagement. But while an overall commitment to structured arguments built on evidence might be general, the question of what is evidence, what is an allowable approach differs between communities. There is an element of circularity that no amount of Open Access can solve: those that can judge the qualities of a piece of scholarship are those that can understand it are those that are peers. General understanding is hard enough to achieve between cognate disciplines within the natural sciences let alone across wider domains. There is a disciplinary clubbishness, one that tends to exclude outsiders, whether they are other professional scholars, students or representatives of that amorphous blob we call “the public”.

Of course that clubbishness goes back to Boyle and the Royal Society as well. Originally membership of the Society was limited to 50 people. This was an intentional effort to limit it to a group that could actually meet. That could act as real peers in observing and criticising experiments. That many of them were peers in the aristocratic sense was a product of the time. But our modern knowledge of community building and collective action would identify that one of the reasons it was successful as a political body was precisely that it was a club. A club almost exclusively made up of wealthy, white, men. And I say “almost” merely because a small proportion of them did actually have to work for a living.

Ostrom’s work shows us that group size and homogeneity are two of the most important determinants of success in collective action and governance problems. It shouldn’t really be a surprise then that small homogenous groups have been historically successful at both creating knowledge and defining what it is allowed to be. That universities grew out of clerical establishments and some disciplines still seem to act like a priesthood is not surprising, but it doesn’t sit well with either our modern ideas about the value of diversity and inclusion, or with the idea of knowledge as a Public, and therefore universal, Good.

Membership of these clubs need not necessarily be exclusive, but even in the most open of possible worlds, access to understanding and access to the respect and criticism of peers requires work, perhaps years of dedication. It is the tension between the idea that access to understanding must be worked for, and the desire for universal access to the fruits of knowledge, that underlies the mutual incomprehension between Robin Osborne, Professor of History at Cambridge University, and his argument that “Open Access makes no sense”, and Open Access advocates. Osborne’s view is that only those that have been formally trained can possibly understand. The naive Open Access idealist claims that anyone can. Both are wrong. The gap must necessarily be filled by dissecting who can understand what, what forms of access they have, and how the benefits can be maximised.

Osborne is correct in diagnosing the problem; that expanding access requires investment and that that investment needs to be made where it can have the most benefit. I think he is incorrect in how he assesses who might benefit and what those benefits might be. Framed as an investment the resources required in communicating historical research to interested historians has clear benefits. Further investment in more general communication, particularly in new modes with uncertain benefits is more risky, and a case could be made that its not the best use of resources. I think that case is weak, for reasons that many of us have discussed before, but here I want to focus on the risks to our research communities, our knowledge clubs, that a failure to invest could cause.

Cultural Science and the Excellence Trap

I need to take a step sideways at this point, because I want to situate the discussion of culture in a particlar way. Usually when we think about culture, we think of a group of individuals coming together to pursue a common interest. The actions of these individuals combine to create what we see as culture. John Hartley and Jason Potts in their book Cultural Science argue that what is actually happening is almost the opposite: that it is culture that creates groups and in turn that groups create knowledge. Now on the surface this may seem like typical humanistic slight of hand, shift the prism to recast what we know in a different light and proceed to rewrite the text books. But I think this formulation is particularly useful in our context.

Hartley and Potts describe Cultural Science as an evolutionary theory, one in which it is elements of culture that are under selection. The term they use for these elements, the “genes” of their theory is “shared units of meaningfulness”. I struggled with this, as I found it an entirely slippery concept, until a recent workshop working with Martin Eve, Damian Pattinson, Sam Moore and Daniel O’Donnell, where we were trying to tease apart the language we use, the stories we tell, about “excellence” in research and how it differs across disciplinary boundaries.

What struck me was that our shared (or not) conceptions of excellence are a strong determinant of disciplinary boundaries. Whether or not you agree that Hernstein Smith’s, or McClintock’s, or Ostrom’s, or Moser’s work is “excellent”, whether indeed you recognise their names, is a strong signal of what research community you belong to. And the language we use to talk about excellent work is also a strong signal. These are not the “genes” themselves but an obversable manifestation of them, in the same way that bands on a gel, or fluorescence signals in a modern DNA sequencer are an observable traces of underlying genes. And in the same way we can use them as proxies of those elements of culture we are seeking to study.

Culture creates groups in the model of Cultural Science through the gathering and reinforcing of community ties, through shared story telling, which in turn re-creates the culture. It is a cyclic process of reinforcement. Our obsession with winning, with being the top of the ranking, with “excellence” both defines how research communities self identify and is part of the story we tell ourselves about what we are doing. And we transmit that language beyond our communities into the wider world. “While the UK represents just 0.9% of global population, 3.2% of R&D expenditure, and 4.1% of researchers, it accounts for 9.5% of downloads, 11.6% of citations and 15.9% of the world’s most highly-cited articles.”

It is we who create this world in which everyone has to be above average. In which we pursue a goal of focussing resources only on the very best, which if successful would actually destroy our research capacity. Our focus on excellence is both dangerous, leading at best towards building skyscrapers on sand and at worst to systematic fraud and self-deception, but it is also who we are in a very deep sense.

The scaling problem

Cultural Science is an evolutionary model. It is therefore concerned with which cultures (and therefore which groups) survive over time. This does not mean that culture is static. It is not in that sense like an atomistic (and therefore just as implausible) genetic theory of fixed genes under selection, but sees them as dynamic, responsive to the group that is co-creating them, and is created by them. As in any evolutionary theory, there will be stable strategies, stable combinations of cultural elements under given environmental conditions. And as in any evolutionary theory, these stable combinations are susceptible to environmental change.

Our cultures of excellence grew up in an era when research communities were closed and were small. Until perhaps the late 1950s the key players in any given field probably knew each other. Environmental changes then started to shift that. First the massive growth of the research enterprise after the Second World War meant that direct personal contact between researchers within a (broad) field was no longer possible. Both geographic expansion and increase in numbers prevented that, leading to involution, splitting and specialisation. My guess would be the size of those communities stayed more or less constant but their conceptions of excellence, of what good work looks like, became more and more specialised.

As that happened, the shared understanding of excellence was lost, that shared cultural element was lost, but the need to describe our work or claim excellence across those boundaries remained, or even intensified. The rhetoric of excellence remained critical even as the meaning of excellence was lost. Arjun Appadurai in his introductory chapter to The Social Life of Things discusses how parallel markets in a commodity, a local markets where value is tied to use, and larger scale markets where the connection with knowledge of use-value is lost, can operate in parallel. These markets are often driven by prestige. At some level the “market” in excellence across disciplinary boundaries is a knowledge free market in prestige.

Proxies, most of them meaningless, arise to address the problem of unpredictable exchange value. The information value of a journal impact factor, university ranking, or citation count is comparable to that of the magical thinking of graph theories in commodity futures or currency markets. But they fill a deep human need to tell stories that can predict the future. Indeed they function as a kind of cargo-cult level story telling that satisfies our scholarly need to present “evidence” as part of our narrative, while simultaneously failing to qualify for the definition of evidence that would be acceptable in any discipline.

But whether or not there is real meaning in these proxies is secondary. As Appadurai describes markets can be stable and can function without price being connected to reality. It is entirely possible that this expansion could have been managed and a stable and shared means of agreeing on excellence might have emerged. Whether it would have been a good thing or not is a separate issue, but an evolutionarily stable strategy, based on strong in groups with their own internal definitions and some form of cross communication between disciplines and with stakeholders emerging on a shared basis is at least plausible.

The internet changes everything

But then we have the internet. Disciplinary specialisation, the creation of new journals, new departments and sub departments, combined with successfully externalised measures of excellence, would drive, arguably did drive a consolidation of communities, a closing down and hardening of barriers. The fight for resources may not have been pretty between these newly fortified communities, but it might have worked. The internet and in particular the web blew all of that up.

The web, for good or ill, through the way it transforms the economics and geography of both publishing and interactions more generally creates an assumption of access. Perhaps more importantly it restructures the economics of which communities, or clubs, are viable. First, it makes much more niche clubs possible by reducing radically reducing the costs of discovering and interacting with like minded people (those who share cultural elements). But achieving these cost reductions requires some degree of openness. Otherwise there is no discoverability, and no interaction.

Michael Nielsen in Reinventing Discovery and Clay Shirky in Cognitive Surplus, both examine the question of how this change in economics makes micro-contributions and micro-expertise valuable. No longer is it necessary to be a deep expert, to be a participant in Osborne’s community of practice, to contribute to the creation of new knowledge. In the right circumstances the person with just the right insight, just the right piece of knowledge, can contribute that effectively. Massive economic and cultural value is being created daily online by systems that work with these possibilities. This was what gave rise to the neoliberal techno-utopian socialist vision of a possible world, what someone (and I can’t remember who, John Wilbanks?) referred to as “commons-ism”.

But our communities have turned inwards, seeking internal consistencies, in a natural search for stability in a rapidly changing world. As we deepen, and harden, the shared sense of what is excellent work within a discipline, we necessarily fortify precisely those boundaries where the web could bring us into contact with differing conceptions, precisely those that might bring the most benefits. I think it is not accidental that the disciplines that Michèle Lamont in her book How Professors Think identifies history as the discipline with one of the most stable self-conception. She describes a community that is proud of “having prevented the threat posed by post-modernism from com[ing] to fruition in history” because it “demands adherence to a logic of scholarly inquiry shared by [history] scholars generally by which the results of historical inquiry can be tested for their validity very much as they are in other disciplines”. It should not be surprising that the same discipline views with scepticism the idea that value could be added by contributions from the outside, by people not sharing a training in that shared logic.

So we are caught in the cross hairs of two opposing strategies, two different cultural and community approaches, that grow in response to two different environmental changes. It is not self evident which would be more successful and more stable. Or over what time frames. To dissect that we need to return to the economic analysis with a new focus on these communities and clubs.

The economics of clubs

The economics of clubs, what makes them sustainable and which characteristics matter, is a well established field building on the work of Buchannan. I’m not going to pretend to be an expert in this area but I do want to focus on the class of goods that are used in sustaining clubs, and the ways in which clubs manage and invest in their creation. Buchannan’s original work looked at organisations like sports clubs. Clubs that have facilities that members can use. These are (largely) non-rivalrous – we can both use the pool at the same time – but are excludable – the door to the pool can be locked and only members given a key.

This puts us firmly in the bottom right hand corner of the quadrant diagram. Usually this quadrant is referred to as “toll goods” and we would put something like an electronic subscription journal or paid for database in this corner. But they are also called “club goods” and it is goods in this quadrant that typically sustain clubs. They are the membership benefits. A subscription journal is, in this sense at least, a membership club where the member benefit is access to the articles it contains.

But what about our other “knowledge clubs”? What sustains them?

In many cases it is obviously money. Research groups depend on salaries and grants. Open Access journals rely on money from consortia, or from APCs, or receive money indirectly through donated time and platforms. Databases receive grants or charge subscriptions or are sustained by foundation members. This money is exchanged for something, or invested in the creation of something.

My claim is that the central good that is the core of sustaining these clubs is knowledge itself. That knowledge is always created in, and communicated through, clubs; research groups, departments, journals, databases. But if that is true how is it possible that knowledge can be exchanged for money (or prestige or other club goods) if it is a Public Good? By definition the provisioning problem of public goods is precisely that they can’t be exchanged for money because of the freeloader problem.

Knowledge is a Club Good

The anwer, of course, lies in the question. Knowledge is not a Public Good. It is a Club Good. Knowledge as it is created, within a club is excludable, by the simple expedient of not telling anyone else. Through communication and dissemination we make it less exclusive – sharing ideas with colleagues, speaking at a conference, publishing an article, publishing it in an Open Access venue, writing a lay summary, undertaking engagement activities – but we can never entirely eliminate exclusion. We can only ever invest in reducing it.

This may seem depressing. It is an acceptance that we can never make knowledge as free as we might like. But I think it is also liberating. It identifies openness as a process or practice, not a binary state of an object, something to work towards rather than something that we fail at achieving. And it recentres the question of how best to invest limited resources in making things more open, more public, and less exclusive. Which audiences should we target? How best to enable unexpected contributions? How to maximise network benefits?

I am not incidentally in focussing on a more fuzzy process or practice of openness shying away from my positions on licensing or best practice more generally. In fact I think it strengthens my position. Open Licensing, once you’ve decided to invest in making a research article free to read, enhances discoverability and re-usability at no additional cost and, in most cases, no financial downside. But it does open a route for how to make a case for where more restrictive licensing may be sensible. A difficult one to prove but at least one that can be analysed.

In this model clubs create (or refine, curate, present) knowledge in the context of their own culture. Knowledge can circulate within the club as a Club Good, a benefit of membership. Prestige accrues to those whose contribution is viewed internally as excellent and there drives an internal symbolic economy of prestige, as Martin Eve puts it in his book, Open Access in the Humanities. But while this internal economy may be enough to sustain the club in some circumstances (for example a small scholarly society where the journal is a membership benefit) it doesn’t bring in the external resources that are usually required.

To interact beyond their borders clubs also trade knowledge for prestige, and thence for money in the form of grants and salary. Here, that knowledge trade is accompanied by claims of excellence, often associated with the name, or Impact Factor, of the journal, or the name of the publisher for a book. But the claim is empty of meaning. Again, our traditional thinking would be to harden the boundaries, enhance the sense of exclusivity of being in the club, in direct opposition to the possibility of new interactions bringing new goods to the club.

We return then to the question of which strategy is more stable, defining and closing borders, or – in some yet to be well defined way – creating opening and opportunities for interactions. What we have added is an understanding that those openings can be created by investing in making knowledge less “clubbish” and more public. We can layer on top of this the Cultural Science perspective that these openings aim to create interactions with new clubs, and that those clubs that are most open to interaction will be those with shared cultural elements.

The prescription for action

Where does this leave us. I’ve asserted that knowledge is a Club Good. That scholarly communication is a series of efforts at different levels to make that knowledge less exclusive. The new insight is that this process is engineered to return new Club Goods back to the group, primarily prestige but also club facilities and resources. If we are to shift the balance so that clubs make knowledge more public and less exclusive, we need to identify what Club Goods they are getting in return. And we need to understand how the incentives are created to provide the systems that support this “public-making”, literally “publishing” of knowledge.

The scaling problem of the mid to late 20th century was addressed through taking Club Goods and privatising them, corporatising their management in the process. Copyight in scholarly world and the whole apparatus that surrounds scholarly content as a private asset, was essentially invented in the 1950s. This now sits as private capital, largely in the hands of, but very unevenly spread between scholarly publishers. This is not all bad. That capital is what made it possible for scholarly communications to move online in what is in retrospect a staggeringly short time.

Today we have new options for managing that scaling. The change in the economics of club formation, discovery and costs of communication, provide new models. Co-ops and consortia are emerging like Open Library of Humanities and Knowledge Unlatched, membership models that look more club-like such as PeerJ, and cooperative support mechanisms like SCOAP3 and the ArXiv support group. But each of these are relatively small. Governance and management challenges arise when groups sizes grow beyond those that allow personal interaction, often breaking as the group becomes too large to sit around a table.

The web is what has changed the systemic economics in a way that makes these groups viable. To allow them to grow we need to learn that lesson and build new shared infrastructures that continue to drive costs down while connecting clubs together in new ways. Infrastructures that drive discovery, interaction, and at some level make those cultures more interoperable. These infrastructures will have politics, and that politics will be one of openness and porous borders.

We need to architect our disciplinary cultures so as to find a balance between identity, and internal conceptions of excellence that drive internal competition, and openness to contributions from the outside. We can define two extremes, one in which a disciplines self conception is incoherent and therefore in a sense too open to contribution. Lamont diagnoses anthropology as suffering from this, I might point to the constant turmoil in Library Science and Information Studies as perhaps similar. Other disciplines are too stable, too well defined, and resistant to new ideas. Lamont points to history and philosophy – I might point to synthetic chemistry – as disciplines where coherence runs the risk of becoming stasis or isolation.

Somewhere between these extremes there is a space where shared culture and identity are strong enough, and shared understanding of excellence coherent enough, but where new contributions and perspectives (of certain types) can be welcomed and integrated. We can imagine a world where knowledge clubs come together through partially shared conceptions of excellence and culture, but where those overlaps (and lack of overlaps) are understood and productive channels for exchange and communication therefore clear. Of course this conception isn’t restricted to interdiscplinary interactions. It might also form a productive model for deciding how and where to invest in public engagement, and how to talk productively to other stakeholders in our space, how to get beyond the mutual incomprehension that so often characterises these conversations.

To achieve this we may need to reconsider “excellence”. My argument describes our current prestige economy as built on a sharp separate of the degree of knowledge between the internal community conception of excellence and an essentially knowledge-free exchange value market across the communities. To take clubs and wire them into a network, to borrow Nic Suzor’s words, we need to add meaning back to shared conceptions of excellence or value in each specific context. We would need to focus on the elements of overlap, differing qualities, plural, not quality that are unique to those differing contexts. To embed the desire to find and use these overlaps we would need to move away from our obsession with linear rankings, we would have to abandon the rhetoric of excellence. In short we need to re-engineer our shared culture.

So I started with the observation that cultural change is hard. I finish with a call for cultural change. But within a frame that, I think, makes it easier than it would otherwise be. First, through adopting the Cultural Science frame and seeing culture as an evolutionary and competitive process, we can look to those communities most effectively adapting and ask how their cultures can be sustained, enhanced and transmitted. Using that economic framing of goods how can we enhance the club goods that those communities are receiving?

Second, by looking at clubs as created by and co-creating their cultures, and using the language and rhetorics of those clubs as symbols or signals of those cultural elements, we can directly intervene through our use of language. We can consciously create and re-shape the culture we want to see. Qualities, not Quality. “Rhetorics of excellence”. Publics. Even the tone of voice and visible held nose when talking about Impact Factors and University Rankings. The stories we tell ourselves make our culture and make our response to it. Choose to act consciously through the language you use.

Finally, if we adopt this evolutionary frame we recognise that we can shape the environment. Building platforms and infrastructures that tip the balance towards discoverability, interactions and openness will naturally re-shape the investment choices that knowledge clubs make. Shared information systems aid governance and allow clubs to scale up. Platforms that are both general and flexible could completely reshape the economics of scholarly publishing, just as a starting point. We can build it, and they will come, if we focus on building tools that support existing, viable communities, not the imaginary ones we would like to see.

I’ve spent a lot of the last decade worrying about the incentives for individuals. Carrots and sticks, mandates and citation advantages. But even with these in place, the culture of disciplines seemed to be a blocker. I believe that by focussing our attention on communites, groups, clubs and the incentives that they work within we will make more progress, because at some level the incentives for the group are the culture.

There are limits to openness. We can never completely remove exclusion. But we can invest time, effort and resources thoughtfully in dropping those barriers. And in doing so maximise the benefits that network scales can bring. Our challenge is to contruct the world in which we approach that limit as effectively as possible.

Touch points

I have a Moto 360 Android Watch. People often ask how it is, whether it’s useful, and I give some answer along the lines that its great to have a watch that shows me the time in multiple places and is always updated to local, always right. It’s a true answer, but it it’s a partial answer. I can easily enough figure out what the time is somewhere else. What really matters is that those dials are a point of connection with people in those places. If they were to look at their watch now, that is what they would see. It’s a point of contact.

We live in a world of status indicators and updates. I catch myself at certain times of day watching for lights turning from yellow to green, green to yellow, as people start their day, or end it. The morning email missive, the going home tweet, the automated notification of arrival. We live in a world where we could sit down with family or friends at a distance of half the world and watch a movie, or a show, or a game together. The geography has shrunken in ways that we still have not worked out. My sense of place has little to do with where I am, or which countries I’m a citizen of, and much more to do with who is around me, and who I wish was also there. It’s about the interactions of a given moment.

People talk as though online interactions are sterile or shallow, that without the raised eyebrow or subtle inflections of voice there is something missing. This is also true, but also partial. The registers of online media also add, and can be just as rich: the shift from email to chat, from text to voice to video (or back), from synchronous to asynchronous, public to private, all signifying intent, urgency, interest (or perhaps annoyance). And finally silence, redolent of a million different meanings in all its different forms and intensities. We create meaning on the fly. Real meaning, whether that intended or that received, built on a history of actions and re-actions, “speakers” and “listeners”, creating layers of reciprocal affect, all built also on an increasingly dynamic landscape. A loop where more or less the same action might receive more or less the same reaction were it not for the shifting sands it is drawn in.

We are, perhaps, more enriched with sources of nuance, certainly of reach than ever before. More ways to see where someone is. More ways to hear how (they want to say) they are. And if that starts to sound uncomfortable it is hardly a surprise, looking as we are into an uncanny valley of human interactions. A world in which it is possible to reach across the globe and literally touch someone through a timepiece. Is the quiet touch on the wrist, the idea “I was thinking of you” needful, or helpful, or welcome? It could so easily run from wished for to feared; to frightening. And how is that managed, or negotiated, in a world so full of signal? In a world of proprietary locked down platforms? How do we make contact with each other, find each of the right boundaries, when boundaries are tied to those shifting sands?

It turns out to be complicated and contingent. The medium is only the message to first order; and the user is not the content. Each is tied to the other in reciprocal cycles of making meaning.

But it was probably a bit much to expect a watch to sort that out.

 

Abundance Thinking

Last week I was lucky enough to spend five days in North Carolina at the Triangle Scholarly Communications Institute, an Andrew W. Mellon Foundation funded initiative that brings teams together on a retreat style meeting to work on specific projects. More on that, and the work of our team, at a later date but one thing that came out of our work really struck me. When we talk about the web and the internet, particularly in the context of scholarly publishing we talk about how the shift from an environment of scarcity limited by the physical restrictions of the print world to a world of abundance. Often we focus on how thinking shaped in that old world is still limiting us today, often invoking that deity of disruption, Clayton Christenson in the process. So far so obvious.

What struck me as we prepared for our final presentations was that these narratives of scarcity don’t just limit us in the world of publication. I am lucky enough to have been to quite a few meetings where great people are sequestered together to think and discuss. These meetings always generate new ideas, exciting projects and life changing insights that somehow dissolve away as we return to our regular lives. The abundance of these focussed meetings, abundance of time, abundance of expertise, abundance of the attention of smart people gives way to the scarcity of our day to day existence. The development of these new ideas falters as it has to compete with scarce time of individuals. When time can be found it is asynchronous, and patchy. We try to make time but we never seem to be able to find the right kind of time.

Many of us reflected that it was a shame we couldn’t always work like this, focussed periods bringing groups together to do the work. But it struck me that, just as the web provides a platform, an infrastructure, that makes publication cheap, that the Mellon Foundation through the SCI Program has also provided an infrastructure that creates an abundance of time and attention. The marginal cost of each project is minimal compared to investment in the program. It is the program that makes it possible. Could the same be true of that archetypal form of scarcity in research, the grant? Could we imagine infrastructures that make the actual doing of research relatively cheap? Is that possible in a world of expensive reagents and equipment? Are the limitations that we see as so self evident real, or are they imposed by our lack of imagination?

And yet there’s also a dark side to this. It is a privilege to attend these meetings and work with these people. And I mean “privilege” with all the loaded and ambivalent connotations it has today. The language of abundance is the language of the “disruptors” of Silicon Valley, a language of techno-utopianism where, with the money you made from your dating app for Bay Area dogs you can now turn your attention to “solving someone else’s problem” with a new app or a new gadget. It can be well-meaning but it is limited. The true challenge is create the opportunities for abundance where it wasn’t before, supporting the creation of new infrastructure and platforms that truly create abundance for those who will most appreciate it. Tim O’Reilly‘s “X as a Platform” agenda pointed in this direction but in a limited context. The web is not (yet) that platform as the rumbling debate over the Wikipedia Zero program shows.

Each time I look at the question of infrastructures I feel the need to go a layer deeper, that the real solution lies underneath the problems of the layer I’m looking at. At some level this is true, but its also an illusion. The answers to questions of biology do not lie in chemistry, nor do the answers of chemistry do not lie in physics. The answers lie in finding the right level of abstraction and model building (which might be in biology, chemisty, physics or literature depending on the problem). Principles and governance systems are one form of abstraction that might help but its not the whole answer. It seems like if we could re-frame the way we think about these problems, and find new abstractions, new places to stand and see the issues we might be able to break through at least some of those that seem intractable today. How might we recognise the unexpected places where it is possible to create abundance?

If only I could find the time to think that through…

Speaking at City University London for OA Week

Ernesto Priego has invited me to speak at City University in London on Thursday the 22nd October as part of Open Access Week. I wanted to pull together a bunch of the thinking I’ve been doing recently around Open Knowledge in general and how we can get there from here. This is deliberately a bit on the provocative side so do come along to argue! There is no charge but please register for the talk.


The Limits of “Open”: Why knowledge is not a public good and what to do about it

A strong argument could be made that efforts to adopt and require Open Access and Open Data in the 21st Century research enterprise is really only a return to the 17th Century values that underpinned the development of modern scholarship. But if that’s true why does it seem so hard? Is it that those values have been lost, sacrificed to the need to make a limited case for why scholarship matters? Or is something more fundamentally wrong with our community?

Drawing on strands of work from economics, cultural studies, politics and management I will argue that to achieve the goals of Open Knowledge we need to recognise that they are unattainable. That knowledge is not, and never can be, a true public good. If instead we accept that knowledge is by its nature exclusive, and therefore better seen as a club good, we can ask a more productive question.

How is it, or can it be, in the interests of communities to invest in making their knowledge less exclusive and more public? What do they get in return? By placing (or re-placing) the interests of communities at the centre we can understand, and cut through, the apparent dilemma that “information wants to be free” but that it also “wants to be expensive”. By understanding the limits on open open knowledge we can push them, so that, in the limit, they are as close to open as they can be.

A league table by any means will smell just as rank

ladder
Ladder (Wikipedia)

The University Rankings season is upon us with the QS league table released a few weeks back to much hand wringing here in the UK as many science focussed institutions tumbled downwards. The fact that this was due to a changed emphasis in counting humanities and social sciences rather than any change at the universities themselves was at least noted, although how much this was to excuse the drop rather than engage with the issue is unclear.

At around the same time particle physicists and other “big science” communities were up in arms as the Times Higher ranking, being released this week, announced that it would not count articles with huge numbers of authors. Similar to the change in the QS rankings this would tend to disadvantage institutions heavily invested in big science projects, although here the effect would probably be more the signals being sent to communities than a substantial effect on scores or rankings. In the context of these shifts the decision of Japanese government to apparently shut a large proportion of Humanities and Social Sciences departments so as to focus on “areas for which society has strong needs” is…interesting.

Also interesting was the response of Phil Baty, the editor of the THES Rankings to John Butterworth’s objections on twitter.

The response is interesting because it suggests there is a “right way” to manage the “problem”. The issue of course is rather the other way around. There can be no right way to solve the problem independent of an assessment of what it is you are trying to assess. Is it the contribution of the university to the work? Is it some sense of the influence that accrues to the institution for being associated with the article? Is it the degree to which being involved will assist in gaining additional funding?

This, alongside the shifts up and down the QS rankings, illustrates the fundamental problem of rankings. They assume that what is being ranked is obvious, when it is anything but. No linear ranking can ever capture the multifaceted qualities of thousands of institutions but worse than that the very idea of a ranking is built on the assumption that we know what we’re measuring.

Now you might ask why this matters. Surely these are just indicators, mere inputs into decision making, even just a bit of amusing fun that allows Cambridge to tweak the folks at Imperial this year? But there is a problem. And that problem is that these ranking really show a vacuum at the centre of our planning and decision making processes.

What is clear from the discussion above and the hand wringing over how the rankings shift is that the question of what matters is not being addressed. Rather it is swept under the carpet by assuming there is some conception of “quality” or “excellence” that is universally shared. I’ve often said that for me when I hear the word “quality” it is a red flag that means someone wants to avoid talking about values.

What matters in the production of High Energy Physics papers? What do we care about? Is HEP something that all institutions should do or something that should be focussed on a small number of places? But not just HEP, but genomics, history, sociology…or perhaps chemistry. To ask the question “how do we count physics the same as history” is to make a basic category error. Just as it is to assume that one authorship is the same as another.

If the question is was which articles in a year have the most influence, and which institutions contributed, the answer would be very different to the question of which institutions made the most contribution in aggregate to global research outputs. Rankings ignore these questions and try to muddle through with forced compromises like the ones we’re seeing in the THES case.

All that these rankings show is that the way you choose to value things depends how you would (arbitrarily) rank them. Far more interesting is the question of what the rankings tell us about what we really value, and how hard that is in fact to measure.