£160 for a copy of a dissertation
November 29, 2021
As I was clearing out some old documents, I stumbled on this form from 2006:

This was back when Paul Upchurch’s dissertation, then only 13 years old, contained much that still unpublished in more formal venues, notably the description of what was then “Pelorosaurus” becklesii. As a fresh young sauropod researcher I was keen to read this and other parts of what was then almost certainly the most important and comprehensive single publication about sauropods.
I remember contacting Paul directly to ask if he could send a copy, but he didn’t have it in electronic form. So I wrote (on his advice, I think) to Cambridge University Library to request a copy from them. The form is what I got back, saying sure I could have a copy — for £160.08 if I wanted a photocopy, or £337.80 for a microfilm. Since inflation in the UK has run at about 2.83% per year since 2006, that price of £160.08 back then is equivalent to about £243 in today’s money.
Needless to say, I didn’t pursue this any further (and to my shame I’m not even sure I bothered replying to say no thanks). To this day, I have never read Paul’s dissertation — though 28 years after it was completed, it’s obviously less relevant now than it was back in the day.
What is the point of this story? What information pertains?
My point isn’t really “Look how exploitative Cambridge University Library’s pricing is” — I have no experience in running a library, and no realistic sense of what it costs in staff time and materials to make a photocopy of a substantial dissertation. Perhaps the price was only covering costs.
The point instead is: look how things have changed. Newly minted Ph.Ds now routinely deposit copies of their dissertations in repositories where they can be freely read by anyone, anywhere. Here is one recent example from my own university: Logan King’s 2021 dissertation, Macroevolutionary and Ontogenetic Trends in the Anatomy and Morphology of the Non-Avian Dinosaur Endocranium.
This is important. I often think about the Library Loon’s 2012 blog-post Framing Incremental Gains. It’s easy, if you’re in open-access advocacy, to feel the burden of how ssslllooowwwlllyyy things seem to change. Sometimes I’ve heard people claim that nothing has changed in the last 30 years. I completely understand the frustration that leads people to say such things. But it’s not true. Things have changed a lot, and are still changing fast. We seem to be past the tipping point where more than half of all newly published papers are open access. There is a lot to celebrate.
Of course that’s not to deny that there’s plenty more work to be done, and plenty of other ways our present scholarly publishing infrastructure desperately needs changing. But in pushing for that, let’s not neglect how much things have improved.
Elsevier’s increasing control over scholarly infrastructure, and how funders should fix this
May 22, 2016
In a recent blog-post, Kevin Smith tells it like it is: legacy publishers are tightening their grip in an attempt to control scholarly communications. “The same five or six major publishers who dominate the market for scholarly journals are engaged in a race to capture the terms of and platforms for scholarly sharing”, says Smith. “This is a serious threat to academic freedom.”
People can legitimately have different ideas about precisely what it is that Elsevier intends to do with SSRN, now that it’s acquired it. But as we discuss the possible outcomes, we need to keep one principle in mind: it’s simply unrealistic to imagine that Elsevier, in controlling Mendeley and SSRN, will do anything other than what is best for Elsevier.
That’s not a criticism, or even a complaint. It’s a statement of what a for-profit corporation does. It’s in its nature. There’s no need for us to blame Elsevier for this, any more than we blame a fox when it eats a chicken. That’s what it does.
The appropriate response is simply to prevent any more of this kind of thing happening, by taking control of our own scholarly infrastructure.
The big problem with SSRN is the same as the big problem of Mendeley: being privately owned and for-profit, their owners were always going to be susceptible to a good enough offer. People starting private companies are looking to make money from them, and a corporation that comes along with a big offer is a difficult exit strategy to resist. When we entrusted preprints to SSRN, they were always vulnerable to being taken hostage, in a way that arXiv preprints are not.
Again: I am not blaming private companies’ owners for this. It’s in the nature of what a private company is. I recognise that and accept it. The thing is, I interpret it as damage and want to route around it.
So what is the solution?
It’s simple. We, the community, need to own our own infrastructure.
One one level, this is easy. We, the community, know how to do it. We have experience of good and bad infrastructure, we know the difference. We have excellent, clearly articulated principles for open scholarly infrastructure. We have top quality software engineers, interaction designers, UI experts and more.
What we don’t have is funding. And that is crippling.
We can’t build and maintain community-owned infrastructure without funding; and (to a first approximation anyway) no-one is funding it. It’s truly disgraceful that even such a crucial piece of infrastructure of arXiv is constantly struggling for funding. arXiv serves about a million articles per week, and is the primary source of publications in many scientific subfields, yet every year it struggles to bring in the less then a million dollars it costs to run. It’s ridiculous the the Gates Foundation or someone hasn’t come along with a a few tens of millions of dollars and set up a long-term endowment to make arXiv secure.
And when even something as proven as arXiv struggles for funding, what chance does anything else have?
The problem seems to be this: funders have a blind spot when it comes to funding infrastructure. That’s why we have no UK national repository; it’s why there is no longer an independent subject repository for social sciences; it’s why the two main preprint archives for bio-medicine (PeerJ Preprints and BioRxiv) are privately owned, and potentially vulnerable to the offer-you-can’t-refuse from Elsevier or one of the other legacy publishers in the oligopoly(*).
When you think about funders — RCUK, Wellcome, NIH, Gates, all of them — they are great at funding research; and terrible at funding the infrastructure that allows it to have actual benefit. Most funders even seem to have specific policies that they won’t fund infrastructure; those that don’t, simply lack a way to apply for infrastructure funding. It’s a horribly short-sighted approach, and we’re seeing its inevitable fruit in Elsevier’s accumulation of infrastructure.
We’ll look back at funding bodies in 10 or 20 years and say their single biggest mistake was failing to see the need to fund infrastructure.
Please, funders. Fix this. Make whatever changes you need to make, to ensure the the scholarly community owns and controls its own preprint archives, subject repositories, aggregators, text-mining tools, citation graphs, metrics tools and what have you. We’ve already seen what happens when we cede control of the scholarly record to corporations: spiralling prices, poor quality product, arbitrary barriers, and the retardation of all progress. Let’s not make the same mistake again with infrastructure.
(*) Actually, I don’t believe PeerJ’s owners would sell their preprint server to Elsevier for any amount of money — and the same may be true of the BioRxiv for all I know, I’ve never spoken with the owners. But who can tell what might happen?
The University of Bristol’s new #OpenAccess policy
September 17, 2015
A quick note to say that I got an email today — the University of Bristol Staff Bulletin — announcing some extremely welcome news:
(Admittedly it was only the third item on the bulletin, coming in just after “Staff Parking – application deadline Friday 18 September”, but you can’t have everything.)
This is excellent, and the nitty-gritty details are encouraging, too. Although HEFCE recently wound back its own policy, as a transition-period concession, to requiring deposit only at the time of publication, Bristol has quite properly gone with the more rigorous requirement that accepted manuscripts be deposited at the time of acceptance. This is wise for the university — it’s future-proofed against HEFCE’s eventual move back towards the deposit-on-acceptance policy that it wanted — and it’s good for the wider world, too.
Hurrah!
Introducing The One Repo
June 30, 2015
You know what’s wrong with scholarly publishing?
Wait, scrub that question. We’ll be here all day. Let me jump straight to the chase and tell you the specific problem with scholarly publishing that I’m thinking of.
There’s nowhere to go to find all open-access papers, to download their metadata, to access it via an open API, to find out what’s new, to act as a platform for the development of new tools. Yes, there’s PubMed Central, but that’s only for work funded by the NIH. Yes, there’s Google Scholar, but that has no API, and at any moment could go the way of Google Wave and Google Reader when Google loses interest.
Instead, we have something like 4000 repositories out there, balkanised by institution, by geographical region, and by subject area. They have different UIs, different underlying data models, different APIs (if any). They’re built on different software platforms. It’s a jungle out there!
As researchers, we don’t need 4000 repos. You know what we need? One Repo.
Hey! That would be a good name for a project!
I’ve mentioned before how awesome and pro-open my employers, Index Data, are. (For those who are not regular readers, I’m a palaeontologist only in my spare time. By day, I’m a software engineer.) Now we’re working on an index of green/gold OA publishing. Metadata of every article across every repository and publisher. We want it to be complete, in the sense that we will be going aggressively for the long tail as opposed to focusing on some region or speciality, or things that are easily harvestable by OAI-PMH or other standards. We want it to be of a high, consistent quality in terms of metadata. We want it to be up to date. And most importantly, we want it to be fully open for all and any kind of re-use, by any other actor. This will include downloadable data files, OAI-PMH access, search-retrieve web services, embeddable widgets and more. We also envisage a Linked Data representation with a CRUD interface that allows third parties to contribute supplemental information, entity reconciliation, tagging, etc.
Instead of 4000 fragments, one big, meaty chunk of data.
Because we at Index Data have spent the last ten years helping aggregators and publishers and others getting access to difficult-to-access information through all kinds of crazy mechanisms, we have a unique combination of the skills, the tools, and the desire to pursue this venture.
So The One Repo is born. At the noment, we have:
- Harvesting set up for an initial set of 20 repositories.
- A demonstrator of one possible UI.
- A whitepaper describing the motivation and some of the technical aspects.
- A blog about the project’s progress.
- An advisory board of some of the brightest, most experienced and wisest people in the world of open access.
We’ve been flying under the radar for the last month and a bit. Now we’re ready for the world to know what we’re up to.
The One Repo is go!
CC-By documents cannot be re-enclosed if their publisher is acquired
October 29, 2014
Just a quick post today, to refute an incorrect idea about open access that has unfortunately been propagated from time to time. That is the idea that if (say) PLOS were acquired by a barrier-based publisher such as Taylor and Francis, then its papers could be hidden behind paywalls and effectively lost to the world. For example, in Glyn Moody’s article The Open Access Schism, Heather Morrison is quoted as follows:
A major concern about the current move towards CC-BY is that it might allow re-enclosure by companies […] This is a scenario suggested by assistant professor in the School of Information Studies at the University of Ottawa Heather Morrison. As she explains, “There is nothing in the CC BY license that would stop a business from taking all of the works, with attribution, and selling them under a more restrictive license—not only a more restrictive CC-type license (STM’s license is a good indication of what could happen here), but even behind a paywall, then buying out the OA publisher and taking down the OA content.”
This is flatly incorrect.
Reputable open-access publishers not only publish papers on their own sites but also place them in third-party archives, precisely to guard against doomsday scenarios. If (say) PeerJ were made an offer they couldn’t refuse by Elsevier, then the new owners could certainly shut down the PeerJ site; but there’s nothing the could do about the copies of PeerJ articles on PubMed Central, in CLOCKSS and elsewhere. And of course everyone who already has copies of the articles would always be free to distribute them in any way, including posting complete archives on their own websites.
Let’s not accept this kind of scaremongering.
Now it’s NPG’s turn to send takedown notices
May 7, 2014
[NOTE: see the updates at the bottom. In summary, there’s nothing to see here and I was mistaken in posting this in the first place.]
Elsevier’s War On Access was stepped up last year when they started contacting individual universities to prevent them from letting the world read their research. Today I got this message from a librarian at my university:
The irony that this was sent from the Library’s “Open Access Team” is not lost on me. Added bonus irony: this takedown notification pertains to an article about how openness combats mistrust and secrecy. Well. You’d almost think NPG wants mistrust and secrecy, wouldn’t you?
It’s sometimes been noted that by talking so much about Elsevier on this blog, we can appear to be giving other barrier-based publishers a free ride. If we give that impression, it’s not deliberate. By initiating this takedown, Nature Publishing Group has self-identified itself as yet another so-called academic publisher that is in fact an enemy of science.
So what next? Anyone who wants a PDF of this (completely trivial) letter can still get one very easily from my own web-site, so in that sense no damage has been done. But it does leave me wondering what the point of the Institutional Repository is. In practice it seems to be a single point of weakness allowing “publishers” to do the maximum amount of damage with a single attack.
But part of me thinks the thing to do is take the accepted manuscript and format it myself in the exact same way as Nature did, and post that. Just because I can. Because the bottom line is that typesetting is the only actual service they offered Andy, Matt and me in exchange for our right to show our work to the world, and that is a trivial service.
The other outcome is that this hardens my determination never to send anything to Nature again. Now it’s not like my research program is likely to turn up tabloid-friendly results anyway, so this is a bit of a null resolution. But you never know: if I happen to stumble across sauropod feather impressions in an overlooked Wealden fossil, then that discovery is going straight to PeerJ, PLOS, BMC, F1000 Research, Frontiers or another open-access publisher, just like all my other work.
And that’s sheer self-interest at work there, just as much as it’s a statement. I will not let my best work be hidden from the world. Why would anyone?
Let’s finish with another outing for this meme-ready image.
Update (four hours later)
David Mainwaring (on Twitter) and James Bisset (in the comment below) both pointed out that I’ve not seen an actual takedown request from NPG — just the takedown notification from my own library. I assumed that the library were doing this in response to hassle from NPG, but of course it’s possible that my own library’s Open Access Team is unilaterally trying to prevent access to the work of its university’s researchers.
I’ve emailed Lyn Duffy to ask for clarification. In the mean time, NPG’s Grace Baynes has tweeted:
So it looks like this may be even more bizarre than I’d realised.
Further bulletins as events warrant.
Update 2 (two more hours later)
OK, consensus is that I read this completely wrong. Matt’s comment below says it best:
I have always understood institutional repositories to be repositories for author’s accepted manuscripts, not for publisher’s formatted versions of record. By that understanding, if you upload the latter, you’re breaking the rules, and basically pitting the repository against the publisher.
Which is, at least, not a nice thing to do to the respository.
So the conclusion is: I was wrong, and there’s nothing to see here apart from me being embarrassed. That’s why I’ve struck through much of the text above. (We try not to actually delete things from this blog, to avoid giving a false history.)
My apologies to Lyn Duffy, who was just doing her job.
Update 3 (another hour later)
This just in from Lyn Duffy, confirming that, as David and James guessed, NPG did not send a takedown notice:
Dear Mike,
This PDF was removed as part of the standard validation work of the Open Access team and was not prompted by communication from Nature Publishing. We validate every full-text document that is uploaded to Pure to make sure that the publisher permits posting of that version in an institutional repository. Only after validation are full-text documents made publicly available.
In this case we were following the regulations as stated in the Nature Publishing policy about confidentiality and pre-publicity. The policy says, ‘The published version — copyedited and in Nature journal format — may not be posted on any website or preprint server’ (http://www.nature.com/authors/policies/confidentiality.html). In the information for authors about ‘Other material published in Nature’ it says, ‘All articles for all sections of Nature are considered according to our usual conditions of publication’ (http://www.nature.com/nature/authors/gta/others.html#correspondence). We took this to mean that material such as correspondence have the same posting restrictions as other material published by Nature Publishing.
If we have made the wrong decision in this case and you do have permission from Nature Publishing to make the PDF of your correspondence publicly available via an institutional repository, we can upload the PDF to the record.
Kind regards,
Open Access Team
Appendix
Here’s the text of the original notification email so search-engines can pick it up. (If you read the screen-grab above, you can ignore this.)
University of Bristol — Pure
Lyn Duffy has added a comment
Sharing: public databases combat mistrust and secrecy
Farke, A. A., Taylor, M. P. & Wedel, M. J. 22 Oct 2009 In : Nature. 461, 7267, p. 1053Research output: Contribution to journal › Article
Lyn Duffy has added a comment 7/05/14 10:23
Dear Michael, Apologies for the delay in checking your record. It appears that the document you have uploaded alongside this record is the publishers own version/PDF and making this version openly accessible in Pure is prohibited by the publisher, as a result the document has been removed from the record. In this particular instance the publisher would allow you to make accessible the postprint version of the paper, i.e., the article in the form accepted for publication in the journal following the process of peer review. Please upload an acceptable version of the paper if you have one. If you have any questions about this please get back to us, or send an email directly to open-access@bristol.ac.uk Kind regards, Lyn Duffy Library Open Access Team.
HEFCE’s new open-access policy for post-2014 outputs
March 31, 2014
This morning sees the publication of the new Policy for open access in the post-2014 Research Excellence Framework from HEFCE, the Higher Education Funding Council for England. It sets out in details HEFCE’s requirement that papers must be open-access to be eligible for the next (post-2014) Research Excellence Framework (REF).
Here is the core of it, quoted direct from the Executive Summary:
The policy states that, to be eligible for submission to the post-2014 REF, authors’ final peer-reviewed manuscripts must have been deposited in an institutional or subject repository on acceptance for publication. Deposited material should be discoverable, and free to read and download, for anyone with an internet connection […] The policy applies to research outputs accepted for publication after 1 April 2016, but we would strongly urge institutions to implement it now.
There are lots of ifs, buts and maybes, but overall this is excellent news, and solid confirmation that the UK really is committed to an open-access transition. Before we go into those caveats, let’s take a moment to applaud the real, significant progress that this policy represents. For the first time ever, universities’ funding levels, and so individual academics’ careers, will be directly tied to the openness of their output. Congratulations to HEFCE!
Also commendable: the actual policy document is very carefully written, and includes details such as “Outputs whose text is encoded only as a scanned image do not meet the requirement that the text be searchable electronically.” It’s evident that a lot of careful thought has gone into this.
Now for those caveats:
The policy will not apply to monographs, book chapters, other long-form publications, working papers, creative or practice-based research outputs, or data.
This is a shame, but understandable, especially in the case of books. I would have hoped that chapters within edited volumes would have been included. But the main document notes that “Where a higher education institution (HEI) can demonstrate that it has taken steps towards enabling open access for outputs outside the scope of this definition, credit will be given in the research environment component of the post-2014 REF.”
Next disappointment:
The policy allows repositories to respect embargo periods set by publications. Where a publication specifies an embargo period, authors can comply with the policy by making a ‘closed’ deposit on acceptance. Closed deposits must be discoverable to anyone with an Internet connection before the full text becomes available for read and download (which will occur after the embargo period has elapsed). Closed deposits will be admissible to the REF.
I would of course have wanted all embargo periods to be eliminated, or at the very least capped at six months as in the old, pre-watering-down, RCUK policy. But that was too much to hope for in the political environment that publishers have somehow managed to create.
More positively, it’s a good sop that deposit must be made on acceptance — not when the embargo expires, or even on publication, but on acceptance. These “closed deposits” are like a formal promise of openness, with an automated implementation. We don’t have good experimental data on this, but it seems likely that this approach will result in much better compliance rates than just telling authors “you have to come back six to 24 months after publication and make a deposit”.
Third disappointment:
There are a number of exceptions to the various requirements that will be automatically allowed by the policy. These exceptions cover circumstances where deposit was not possible, or where open access to deposited material could not be achieved within the policy requirements. These exceptions will allow institutions to achieve near-total compliance, but the post-2014 REF will also include a mechanism for considering any other exceptional cases where an output could not otherwise meet the requirements.
The exceptions encourage weasel-wordage, of course, and some of the specific exceptions listed in Appendix C are particularly weak: “Author was unable to secure the use of a repository”, “Publication is print-only (no electronic version)”, and the lamentable “Publication does not offer a compliant green or gold option”, which really means “HEFCE authors should not be using this publication”.
But when you read into the details, this approach with specific exceptions is actually rather better than the alternative that had been on the table: a percentage-based target, where some specific proportion of REF submissions would need to be open access. Instead of saying “80% of submissions must be open access” (or some other percentage), HEFCE is saying that it wants them all to be open access except where a specific excuse is given. I’d like them to be much less accommodating with what excuses they’ll accept, but the important thing here is that they have set the default to open.
Now for the most regrettable part of the policy:
While we do not request that outputs are made available under any particular licence, we advise that outputs licensed under a Creative Commons Attribution Non-Commercial Non-Derivative (CC BY-NC-ND) licence would meet this requirement.
I won’t rehearse again all the reasons that Non-Commercial and No-Derivatives clauses are poison, I’ll just note that works published under this licence are not open access according to the original definition of that term, which allows us to “use [OA works] for any other lawful purpose, without financial, legal, or technical barriers”.
Yet even here, the general tenor of the policy is positive. While it accepts NC-ND, the policy adds that “where an HEI can demonstrate that outputs are presented in a form that allows re-use of the work, including via text-mining, credit will be given in the research environment component of the post-2014 REF”.
One last observation: HEFCE should be commended on having provided an excellent, detailed explanation of feedback they received to their consultations. As always, reading such documents can be frustrating because they necessarily contain some views very different from mine; but it’s useful to see the range of opinions laid out so explicitly.
No open-access policy document I’ve ever seen has been perfect, and this one is no exception. But overall, the HEFCE open-access policy is a significant and welcome step forward, and carries the promise of further positive moves in the future.
Generations unite! Open Access can’t fail
November 19, 2013
Yesterday I was at the Berlin 11 satellite conference for students and early-career researchers. It was a privilege to be part of a stellar line-up of speakers, including the likes of SPARC’s Heather Joseph, PLOS’s Cameron Neylon, and eLIFE’s Mark Patterson. But even more than these, there were two people who impressed me so much that I had to give in to my fannish tendencies and have photos taken with them. Here they are.
This is Jack Andraka, who at the age of fifteen invented a new test for pancreatic cancer that is 168 times faster, 1/26000 as expensive and 400 times more sensitive than the current diagnostic tests, and only takes five minutes to run. Of course he’s grown up a bit since then — he’s sixteen now.
Right at the moment Jack’s not getting much science done because he’s sprinting from meeting to meeting. He came to us in Berlin literally straight from an audience with the Pope. He’s met Barack Obama in the oval office. And one of the main burdens of his talk is that he’s not such an outlier as he appears: there are lots of other brilliant kids out there who are capable of doing similarly groundbreaking work — if only they could get access to the published papers they need. (Jack was lucky: his parents are indulgent, and spent thousands of dollars on paywalled papers for him.)
Someone on Twitter noted that every single photo of Jack seems to show him, and the people he’s with, in thumbs-up pose. It’s true: and that is his infectious positivity at work. It’s energising as well as inspiring to be around him.
(Read Jack’s guest post at PLOS on Why Science Journal Paywalls Have to Go)
Here’s the other photo:
This is Bernard Rentier, who is rector of the University of Liège. To put it bluntly, he is the boss of the whole darned university — an academic of the very senior variety that I never meet; and of the vintage that, to put it kindly, can have a tendency to be rather conservative in approach, and cautious about open access.
With Bernard, not a bit of it. He has instituted a superb open-access policy at Liège — one that is now being taken up as the model for the whole of Belgium. Whenever members of the Liège faculty apply for anything — office space, promotions, grants, tenure — their case is evaluated by taking into account only publications that have been deposited in the university’s open-access repository, ORBi.
Needless to say, the compliance rate is superb — essentially 100% since the policy came in. As a result, Liège’s work is more widely used, cited, reused, replicated, rebutted and generally put to work. The world benefits, and the university benefits.
—
Bernard is a spectacular example of someone in a position of great power using that power for good. Meanwhile, at the other end of scale, Jack is someone who — one would have thought — had no power at all. But in part because of work made available through the influence of people like Bernard, it turned out he had the power to make a medical breakthrough.
I came away from the satellite meeting very excited — in fact, by nearly all the presentations and discussions, but most especially by the range represented by Jack and Bernard. People at both ends of their careers; both of them not only promoting open access, but also doing wonderful things with it.
There’s no case against open access, and there never has been. But shifting the inertia of long-established traditions and protocols requires enormous activation energy. With advocates like Jack and Bernard, we’re generating that energy.
Onward and upward!
The SV-POW! open-access decision tree
May 11, 2013
As part of the progressive erosion of RCUK’s initially excellent open-access policy, barrier-based publishers somehow got them to accept their “open-access decision tree“, which you can now find on page 7 of the toothless current version of the policy. The purpose of this manoeuvre by the Publishers Association is to lend an air of legitimacy to continuing to deny citizens access to the research they funded for up to 24 months after publication. It’s to the House of Lords’ enduring shame that they swallowed this, when they must know that there is no justification for embargoes of any length.
More recently, as commentary on the Australian Research Council’s open access policy, the Australian Open Access Support Group (AOASG) published its own rather better decision tree.
But it still doesn’t go nearly far enough. So here is the SV-POW! decision tree, which we encourage you to print out and hang on your office door.
… and don’t forget, when depositing your peer-reviewed accepted manuscripts in a repository, to specify that they are made available under the CC BY licence, which most benefits the field as a whole.
Can repositories solve the access problem?
March 2, 2013
The progressive RCUK policy on open access has recently come under fire, particularly from humanities scholars, for favouring Gold OA over Green. For various reasons — and I won’t, for now, go into the question of which of these reasons are and aren’t sound — they favour an approach to open access where publishers keep final versions of their papers behind paywalls, but drafts are deposited in institutional repositories (IRs) and people who want to read the paper can have access to the drafts.
It’s appealing to think that this relatively lightweight way of solving the access problem can work. Unfortunately, I’m not convinced it can, for several reasons. I’ll discuss these below, not so much with the intention of persuading people that Gold is a better approach, but with the hope that those of you who are Green advocates have seen things that I’ve missed and you’ll be able to explain why it can work after all.
1. Two-class system
Most fundamentally, I worry that Green OA creates a two-class system which I can’t approve of. It does this in two ways.
First, it necessarily creates two classes of papers: author’s draft and publishers’ final versions. These will differ in some respects, and it’s hard in general to know what those respects are. Of course pagination will differ — which means you can’t cite page-numbers reliably. But other changes are possible as well. For example, Matt and I have a paper in press now for which a whole additional figure — an important one — was added at the proofing stage. In our case, the paper will be OA anyway, but if it were not then the authors’ manuscript would be a poor substitute.
And by implication, Green OA creates two classes of researchers — a potentially harmful division between those privileged few who have the “proper” papers and an underclass who have only manuscripts. (It doesn’t help that for stupid historical reasons, our manuscripts are often butt-ugly: double-spaced, line-numbered, all the figures at the end instead of where they’re needed, etc.)
Admittedly, the two-classes-of-researcher problem is not created by Green OA: it already exists, in a worse form where the underclass doesn’t have access to any version of the paper. But whereas Gold OA solves this problem (everyone has exactly the same access to PLOS and PeerJ papers), Green doesn’t.
(To me, it’s obvious that democratising access is a good thing. But now that I’ve made the notion explicit, I can’t help the uncharitable thought that there may be those out there who want to maintain a two-class system — to retain a notion that they are “in” while others are “out”. I hope I’m wrong, and I’m certainly not accusing Green OA advocates of having this motivation. It just seems like it might be implicit in some of the broader struggles over access. Anyway, let’s not confuse this separate potential problem with access with the actual problems with Green OA I’m addressing in this post.)
2. Expense of continuing subscriptions
I find it baffling that people keep talking as though Green OA is cheaper than Gold. It isn’t, at all. As I’ve shown previously, the cost to the world of a paywalled paper (aggregated across all subscriptions) is about $5333. There is no reason to think that will change under the Green model, in which we continue to give the final and best version of our work to publishers.
By contrast, even the publisher-influenced Finch estimates typical Gold APCs as £1500-£2000 (about $2270-$3030), which amounts to 43%-57% as much. (Conveniently, the midpoint of the Finch range, £1750, is about $2650, which is almost exactly half of what we pay by the subscription model.
But the true cost of Gold OA is much, much less. Follow the link for the detail, but one credible banner figure starts with the observation that half of all Gold OA articles are published at no cost to the author and that the average APC of the other half is $906, to arrive at a true average APC of $453 — about one twelfth of the cost for a paywalled article.
So for purely pragmatic financial reasons, Green seems like a silly path. There’s a very short-term saving, sure, as we avoid paying APCs. But we have to look further ahead than the next five years.
3. Embargoes
Now there is nothing intrinsic to Green OA that means embargoes must be in place. It’s perfectly possible, and manifestly desirable, that no-embargo Green-OA mandate should be enacted, requiring that authors’ final manuscripts become available immediately on publication. But for whatever historical reasons (and I admit I find this baffling) there are few or no Green-OA mandates that do this. Even the best of them seem to allow a six-month delay; twelve months is not uncommon (and Michael Eisen worries that the new White House policy will further establish twelve months as the norm.
I will have more to say about embargoes in a subsequent post. (SPOILER: it’s not going to be pretty.) But for now it suffices to say that any system that makes research freely available only a year after it’s published is wholly inadequate. Not to mention stupid. Stupid and inadequate.
So if Green OA is going to be the solution we need, it has to break free from embargoes.
4. Non-open licences
Similarly, there is no intrinsic reason why Green OA should mean non-open licences and Gold OA should mean truly open (BOAI-compliant) open access. And yet history has brought us to a point where that is often how things are. For example, the RCUK policy (even before its progressive erosion got properly under way) says of its Gold arm that “The CC-BY license should be used in this case”, but contains weasel words in its Green arm:
the journal must allow deposit of Accepted Manuscripts that include all changes resulting from peer review (but not necessarily incorporating the publisher’s formatting) in other repositories, without restrictions on non-commercial re-use.
This just won’t do. It’s not open access. To quote Heather Piwowar’s pithy statement once more, “We do basic research not only to know more, but to do more”. Non-commercial licences impede the use of research, and that’s not to the benefit of wider society. (I won’t labour this point now, because I’ll have more to say on non-commercial clauses in a subsequent post.)
So as with embargoes, if Green OA is going to be the solution we need, it has to break free to its habitual acceptance of non-commercial clauses.
5. Practical failings
On top of the fundamental problems already discussed (two-class system, expense of continuing subscriptions, embargoes and non-open licences), the repository system as it exists today suffers from a suite of practical problems that render it pretty inadequate.
- Many institutions don’t even have an IR; or if they do it doesn’t work.
- Many scholars aren’t associated with an institution and so don’t know where they should reposit their manuscripts. (That this is overlooked is a symptom of an unfortunate elitist tendency among academics.) [UPDATE 4th March: thanks to Neil Stewart, whose comment below points out Open Depot as a solution to this.]
- The use of IRs involves an institution-by-institution fragmentation, with different user interfaces, policies, etc.
- For whatever reasons, many scholars do not bother to reposit their manuscripts in institution repositories.
- Even when mandates are in place, compliance is often miserable, to the point where Peter Suber considers the 80% NIH compliance rate as “respectable”. It really isn’t. 100% is acceptable; 99% is respectable.
- Many IRs have abject search facilities, often for example lacking the ability to restrict searches to papers that are actually available.
- Many IRs impose unnecessary restrictions on the use of the materials they contain: for example, Bath’s repo prohibits further redistribution.
- There is no central point for searching all IRs (at least not one that is half-decent; I know about OAIster).
- The quality of metadata within most IRs variable at best
- Use of metadata across IRs is inconsistent — hence many of the problems that render OAIster near-useless.
… and, I am sure, many more that I’ve not thought of right now.
Could these issues be addressed? Yes, probably; but ten years have unfortunately not done much to resolve them, so I don’t feel all that confident that the next ten will.
Do the IR advocates have a plan for solving these problems? Because they are much more political/sociological than technical, and those always seem to be the hardest ones to solve.