Tag Archives: policy

Could the HEFCE policy be a Trojan Horse for gold OA?

The HEFCE Policy for open access in the post-2014 Research Excellence Framework kicks in 9 weeks from now.

The policy states that, to be eligible for submission to the post-2014 REF, authors’ final peer-reviewed manuscripts of journal articles and conference proceedings with an ISSN must have been deposited in an institutional or subject repository on acceptance for publication. Deposited material should be discoverable, and free to read and download, for anyone with an internet connection.

The goal of the policy is to ensure that publicly funded (by HEFCE) research is publicly available. The means HEFCE have chosen to favour is the green route – by putting the AAM into a repository. This does not involve any payment to the publishers. The timing of the policy – at acceptance – is to give us the best chance of obtaining the author’s accepted manuscript (AAM) before it is deleted, forgotten or lost by the author.

Universities across the UK have been preparing. Cambridge has had the ‘Accepted for publication? Send us your manuscript‘ campaign running since May 2014 with a very simple and well liked interface allowing researchers to submit their work. The Open Access team then deposits the item, checks for funding and the publisher policies and then organises payment for open access publication if required.

To give an idea of the numbers we are dealing with at Cambridge, during 2015 the Open Access team deposited 2553 articles into our repository Apollo.

Compliance levels

We have been reporting to Wellcome Trust and the RCUK over the past few years to indicate compliance levels with their policies. However the ‘compliance level’ for the HEFCE policy is a slippery concept. For a start, the policy has not yet come into force. Another complicating factor is the long term nature of the ‘reporting’. We will not truly know how compliant we have been until the time comes to submit to REF – whenever that will be (currently it seems 2021).

At Cambridge have been working on the assumption that because we do not know which outputs will be the ones that we will claim we should collect all eligible articles. However, the number of deposited articles Open Access team received over the past year represents approximately 30% of the full eligible output of the University. This might seem concerning in some ways, but it must be remembered that each researcher in the University will only be reporting four research outputs for the REF.

There are some articles that are obvious contenders for REF. By concentrating on researchers who are publishing in very high impact journals we have been trying to catch those articles we are extremely likely to claim.

During the course of 2015 we discovered 93 papers published in Nature, Science, Cell, The Lancet and PNAS. 33% of these papers were already HEFCE compliant. Of the remaining non-compliant papers we contacted 47 authors, made them aware of the HEFCE open access policy, and invited them to submit their accepted manuscript to the Open Access Service. Less than 40% of those authors who were contacted responded with their accepted manuscript. Therefore, even after direct intervention only 49% papers were HEFCE compliant, which means that still more than half of all eligible papers published in Nature, Science, Cell, The Lancet and PNAS during this period would not have been HEFCE compliant had the policy been in place.

The lack of engagement by members of the academic community with this process is a serious concern – and potentially due to four reasons:

  • Lack of awareness of the policy
  • Putting it off until the policy is in place
  • Deliberately choosing not to submit a work because it is not considered important enough or they do not consider their contribution to be significant enough
  • Some form of conscientious objection to the policy

We should note that the third reason is a matter of some concern to the University as it is not the researcher who decides which articles are put forward for REF. In addition, the University is interested in having a high overall level of compliance for REF as it considers making the research output of the institution available to be important.

Temporary reprieve

Cambridge is no island when it comes to facing significant challenges in capturing all outputs in preparation for HEFCE’s policy. While the highly devolved nature of the institution and the sheer volume of publications may be a problem unique to Cambridge and Oxford, other institutions are still developing the technology they intend to use or are facing staffing issues.

In a concession to serious concern across the sector about the ability to meet the deadline, on 24 July 2015 HEFCE announced that there was a temporary modification to the policy. They now allow research outputs to be made open access up to three months after publication until at least April 2017 (and until such time that the systems to support deposit at acceptance are in place).

This means for the first year of the policy we have a small window after publication to locate articles, determine if they are in our repositories, and if not chase the authors for the Author’s Accepted Manuscript.

The trick is knowing that an article has been published. At Cambridge our ‘best bet’ is to use Symplectic which scrapes various aggregating sources such as Scopus. However Symplectic is hindered by the efficiency of its sources. There is no guarantee that a given article will appear in Symplectic within three months of publication. And even if it is, we have already discussed the low engagement by the research community to approaches from the Open Access team for AAMs.

Subject based repositories

So far this blog has been talking about using institutional repositories for compliance. But the policy specifically states: “The output must have been deposited in an institutional repository, a repository service shared between multiple institutions, or a subject repository“.

The oldest, most established subject repository is arXiv.org and it makes sense for us to consider using arXiv as part of Cambridge’s compliance strategy. After all, some areas of high energy physics, most of computer science and much of mathematics use arXiv as a means to share their research papers. In 2014, the number of articles that were deposited into arXiv.org and subsequently picked up in Symplectic and approved by researchers were 582 – approximately 6.5% of Cambridge’s total eligible articles.

If we are able to claim these articles for HEFCE compliance without any behaviour change requirement from our academic staff then this is an ideal situation. But how do we actually do this? There is a footnote to the HEFCE statement above which says that: “Individuals depositing their outputs in a subject repository are advised to ensure that their chosen repository meets the requirements set out in this policy.” And this is the crunch point. arXiv does not currently identify which version of the work has been deposited, nor does it record the acceptance date of the work. Because of this we are currently not able to simply use the work being uploaded to arXiv.

There is work underway to look at this possibility and what would be required to allow us to use the subject based repositories as a means for compliance. HEFCE themselves have identified under ‘Further areas of work‘ that  “measures to support compliance in subject repositories” is an area of uncertainty and they will work with the community to address this.

Alternative approach?

It is possibly a good moment to take a step back from the minutiae of the means and the timing of the HEFCE policy and focus on the goal that publicly funded research is publicly available. We are in a complex policy environment. HEFCE affects all researchers but many researchers are also funded through COAF or the RCUK with their respective (gold leaning) Open Access policies.

Of the HEFCE eligible articles submitted to to Open Access team in 2015, after working through all the different funder requirements, there was a split of 44% gold Open Access and 56% green Open Access. Of the gold payments the split is approximately 74% for hybrid journals and 26% for fully open access journals.  That said, the three journals with which we have published the most – PLOS ONE, Nature Communications and Scientific Reports – are fully Open Access journals with APCs of $1495, $5200 and $1495 respectively.

A highly relevant question is – outside of the efforts by our Open Access compliance teams, how much Cambridge research is being made open access anyway?

Open access articles

The Web of Science (WoS) allows a filter on ‘Open Access’. It does not appear to list articles that are made open access on a hybrid basis, only picking up fully open access journals. While these are not definitive numbers, it does give us some idea of the scale we are looking at. In 2014 WoS gives us a figure of 981 articles published as open access by a University of Cambridge author in a fully open access journal.

The Springer Compact to which many institutions (including Cambridge) have signed up means that now all articles published by that research community will be made open access. In 2014, the Open Access Service had paid for 21 articles to be made open access. In the same period across the institution we had published 695 articles with Springer. (Note that in 2015 we paid 51 Springer  APCs). This means that for the cost of the Springer subscription and our APC payments for the previous year we will have a good proportion of Cambridge articles published as open access articles.

These two sets of numbers only allow for articles published either in fully open access journals or with Springer. It does not account for the articles where the University (or a Department or individual) pays an APC to make an article available in a hybrid (non Springer) journal. The upshot is – a significant proportion of Cambridge research is published open access.

Skip the AAM on acceptance part?

So what does this published open access research mean for compliance with the HEFCE policy? The updated HEFCE policy has addressed this:

“… we have decided to introduce an exception to the deposit requirements for outputs published via the gold route. This may be used in cases where depositing the output on acceptance is not felt to deliver significant additional benefit. We would strongly encourage these outputs to be deposited as soon as possible after publication, ideally via automated arrangements, but this will not be a requirement of the policy.”

This makes sense from an administrative perspective if the article appears in a journal where there is an embargo period on making the AAM available, forcing the University to pay an APC to make the work Open Access to meet RCUK requirements. It would avoid the palaver of:

  • obtaining the AAM from the author
  • depositing it into the repository
  • having to check to see when the article has been published
  • updating the details and
  • either set the embargo on the AAM or change the attachment in the record to the Open Access final published version

However journals where there is an embargo period on making the AAM available forcing an APC payment is in fact almost a definition of hybrid journals. We know there are issues with hybrid – of the extra expense, of double dipping, of the higher APC charges for hybrid over fully Open Access journals. Putting these aside, what this HEFCE policy change means is that publishers have effectively shifted the HEFCE policy away from a green open access policy to a gold one for a significant proportion of UK research. This is a deliberate tactic, along with the unsubstantiated campaign that green Open Access poses a major threat to scholarly publishing and therefore embargoes should be even longer.

We are already facing the problem that hybrid journals are forcing the move towards green open access being ‘code’ for a 12 month delay. This is the beginning of a very slippery slope. We have been outplayed. It really is time for the RCUK and Wellcome Trust to stop paying for hybrid Open Access.

But I digress.

The cons

The message is confusing enough – three sets of policies and three different requirements in terms of the timing and the means to make work compliant and available. We are trying to make it as simple as possible for researchers – with limited success.

The move to widespread Open Access in the UK is a huge shift for the research community and those that support them. It would be very difficult to debate the ‘against’ argument for the statement that publicly funded research should be publicly available but the devil is very much in the detail.

It would be an incredible shame if the HEFCE policy is hijacked into a partial gold OA policy, but as administrators we are drowning in compliance. There needs to be a broad discussion across the funders to try and address the conflicting compliance requirements and the potentially negative effect these policies are having on the future of open scholarly publishing. 

We welcome the opportunity to discuss these issues with HEFCE, Wellcome Trust and the RCUK. There’s plenty to talk about.

Published 25 January 2016
Written by Dr Danny Kingsley
Creative Commons License

Disruptive innovation: notes from SCONUL winter conference

On Friday 27 November Danny Kingsley attended the SCONUL Winter Conference 2015 which addressed the theme of disruptive innovation and looked at the changes in policy and practice which will shape the scholarly communications environment for years to come. This blog is a summary of her notes from the event. The hastag was #sconul15  and there is a link in Twitter.

Disruptions in scholarly publishing

Dr Koen Becking, President of the Executive Board, Tilburg University, spoke first. He is the lead negotiator with the publishers in the Netherlands. Things are getting tight as we count down to the end of the year given the Dutch negotiations with Elsevier (read more in ‘Dutch boycott of Elsevier – a game changer?‘)

Koen asked: what is the role of a university – is it knowledge to an end, knowledge in relation to learning or knowledge in relation to professional skills? He said that 21st century universities should face society. While Tilburg University is still tied to traditional roots, it is now focused in the idea of ‘third generation university’. The idea of impact on society – the work needs to impact on society.

The Dutch are leading on the open access fight and Koen said they may look at legislation to force the government goal of open access to research articles of 40% by 2016 & 100% by 2024. [Note that the largest Dutch funder NOW has just changed their policy to say funds can no longer be used to pay for hybrid OA and that green OA must be available ‘the moment of’ publication].

Kurt noted that the way the Vice Chancellors got involved in the publisher discussions in the Netherlands was the library director came to him ask about increasing the subscription budget and the he asked why it was going up so much given the publisher’s profit levels. Money talks.

Managing the move away from historic print spend

Liam Earney from Jisc said there were several drivers for the move from historic print spend and we need models that are transparent, equitable, sustainable and acceptable to institutions. They have been running a survey on acceptable metrics on cost allocation (note that Cambridge has participated in this process). Jisc will shortly launch a consultation document on institutions on new proposals.

Liam noted that part of their research found that it was apparent that across Jisc bands and within Jisc bands there are profound differences in what institutions paying for the same material – sometimes by a factor of 100’s of 1000’s pounds different to access the same content in similar institutions.

They also worked out that if they took a mix of historical print spend and a new metric it would take over 50 years to migrate to a new model. This is not realistic.

Jisc is supported by an expert group of senior librarians (including members at Cambridge) who are working on an alternative. Liam noted that historical print spend is harmful to the ability of a consortium to negotiate coherently. Any new solution needs to meet the needs of academics and institutions.

Building a library monograph collection: time for the next paradigm shift?

Diane Brusvoort from the University of Aberdeen comes from the US originally and talked about collaborative collection development – we can move together. Her main argument was that for years we have built libraries on a ‘just in case model’ and we can no longer afford to do that. We need to refine our ‘just in time’ purchasing to take care of faculty requests, also have another strand working across sector to develop the ‘for the ages’ library.

She mentioned the FLorida Academic REpository (FLARE) which is the statewide shared collection of low use print materials from academic libraries in Florida. Libraries look at what is in FLARE and move the FLARE holding into their cataloguing. It is a one copy repository for low use monographs.  The Digital Public Library of America is open to anything that had digitised content can be put in the DPA portal and deals with the problem of items that they are all siloed.

Libraries are also taking books off the shelf when there is an electronic version. This is a pragmatic decision not made because lots of people are reading the electronic one preferentially, it is simply to save shelf space.

Diane noted a benefit of UK compared to UK is the size – it is possible to do collaborative work here in ways you can’t in the US. We need collaborative storage and to create more opportunities for collaborative collections development.

The Metric Tide

Professor James Wilsdon – University of Sussex spoke about the HEFCE report he contributed to The Metric Tide: Report of the Independent Review of the Role of Metrics in Research Assessment and Management. 

This report looked at responsible uses of quantitative data in research management and assessment. He said we should not turn our backs on big data and its possibilities but we know of our experience in the research systems that these can be used as blunt tools in the process. He felt that across the community at large the discussion about metrics was unhelpfully polarised. The debate is open to misunderstanding and need a more sophisticated understanding on ways they can be used more responsibly.

The agreement is that peer review is the ‘least worst’ form of academic governance that we have. Metrics should support not supplant academic management. This underpins the design of assessment exercises like the REF.

James noted that the metrics review team was getting back together that afternoon to discuss ‘section d’ in the report. He referred to this as being ‘like a band reunion’.

A new era for library publishing? The potential for collaboration amongst new university presses

This workshop was presented by Sue White, Director of Computing and Library Services and Graham Stone, Information Resources Manager, University of Huddersfield.

Sue talked about the Northern Collaboration of libraries looking at joining forces to create a publishing group. They started with a meeting in October 2104. There is a lot of uncertainty in the landscape, with a big range of activity from well established university presses to those doing no publishing at all. She said the key challenges to the idea of a joint press was the competition between institutions. But they decided the idea merited further exploration.

Discussions were around the national monograph strategy roadmap  that advocated university publishing models. The Northern Collaboration took a discussion paper to Jisc – and they suggested three areas of activity. They were:

  • Benchmarking and data gathering to see what was happening in the UK.
  • Second to identify best practice and possible workflow efficiencies- common ground.
  • Third was exploring potential for the library publishing coalition.

The project is about sharing and providing networks for publishing ventures. In the past couple of days Jisc has agreed to take the first two forward and welcome input. They want some feedback about taking it forward.

Graham then spoke about the Huddersfield University Press which has been around since 2007 – but was re-launched with an open access flavour. They have been publishing open access materials stuff for three to four years. They publish three formats – monographs, journal publications and sound recordings.

The principles governing the Press is that everything is peer reviewed, as a general rule everything should be open access and they publish by the (ePrints) open access repository which gets lots of downloads. The Press is managed by the library but led by the academics. Business model is a not for profit as it is scholarly communication. If there were any surplus it would be reinvested in the Press. In last four years they have published 12 monographs, of which six are open access.

Potential authors have to come with their own funding. Tends to be an institutional funder sponsored arrangement. The proposal form has a question ‘how is this going to be funded’? This point is ignored for the peer review process. Having money does not guarantee publishing. It means it will be looked at but doesn’t guarantee publishing. The money pays for a small print run, copy editing not staff costs. About a 70,000 word monograph costs in the region of £3000-£4000.

Seven journals are published in the repository – there is an overlay on the repository, preserved in Portico. Discoverable through Google (via OAI-PMH) compliance with repository, Library webscale discovery includes membership of the DOAJ. Their ‘Teaching and lifelong learning’ journal has every tickbox on DOAJ.

The enablers for this Press have been senior support in the university at DVC level and the capacity and enthusiasm of an Information Resource Manager to absorb the Press into existing role. Also having an editorial board with people across the institution. The Press is operating on a shoestring hard. It is difficult to establish reputation and convincing the potential stakeholders and impact. A lack of long term funding means it is difficult to forward plan.

They also noted that there are not very many professional support networks out there and it would be good to have one. They need specialist help with author contracts and licences.

Who will disrupt the disruptors?

The last talk was by Martin Eve, Senior Lecturer in Literature, Technology and Publishing, Birkbeck, University of London.  This was an extremely entertaining and thought provoking talk. The slides are available.

Martin started with critical remarks about the terminology of ‘disruptive’, arguing that often the word is used so the public monopoly can be broken up into private hands. That said, there are parts of the higher education sector which are broken and need disruption.

Disruption is an old word – from Latin used first in 15th century. Now it actually means the point at which an entire industry is shifted out. What we see now is just a series of small increments. The changes happening in the higher education sector are not technological they are social and it is difficult to break that cycle of authority and how it works.

Martin argued that libraries need to be strategic and pragmatic. We have had a century long breakdown of the artificial scarcities in trading of knowledge coming to a head in the digital age. There are new computational practices with no physical or historical analogy – the practices don’t fit well with current understandings. He gave a couple of historical examples where in the 1930s people made similar claims.

The book as a product of scholarly communication is so fetishized that when we want the information we need the real object – we cannot conceive of it in another form.

Universities in the digital age just don’t look like they did before. We have an increasingly educated populace – more people can actually read this stuff so the argument that ‘the public’ can’t understand it is elitist and untrue. Institutional missions need to be to benefit society.

Martin discussed the issues with the academic publication market. A reader always needs a particular article – the traditional discourses around the market play out badly. You don’t know if you need a particular article until you read it and if you do need it you can’t replace it with anything else.

Certain publications can have a rigorous review practice because they are receiving high quality submissions. But they only get high quality submissions if you have lots of them and they get that reputation because of a rigorous review practice. So early players have the advantage.

He noted that different actors care about the academic market in different ways. Researchers produce academic products for themselves – to buy an income and promotion. Publishers frame their services as doing things for authors – but they don’t do enough for readers and libraries. Who pays? Researchers have zero price sensitivity. Libraries are stuck between rock & hard place. They have the cash but are told how to spend it. The whole thing is completely dysfunctional as a market. In the academy, reading is underprivileged. Authorship is rewarded.

Martin then talked about open access and how it affects the Humanities. He noted that monographs are acknowledged as different – eg: HEFCE mandate. There are higher barriers for entry to new publishers – people don’t have a good second book to give away to an OA publisher. There are different employment practices, for example creative writers are often employed on a 0.5 contract – they are writing novels and selling them and commercial publishers get antsy about requirements for open access because there is a cross over with trade books.

The subscription model exists on the principle that if enough people contribute then you have enough centrally to pay for what the costs are. It assumes a rivalrous mode – the assumption is there will always be greedy people who won’t pay in if they don’t get an exclusive benefit.

The Open Library of the Humanities is funded by a library consortium. It is based on arXiv funding model and Knowledge Unlatched. Libraries pay into a central fund in the same way of a subscription. Researcher who publish with us do not have to be at an institution who is funding or even at an institution. There are 128 libraries financially supporting the model (as of Monday should be 150). The rates are very low – each one only has to pay about $5 per article. They are publishing approximately 150 articles per year.

Published 28 November 2015
Written by Dr Danny Kingsley
Creative Commons License

Open Data – moving science forward or a waste of money & time?

On the 4 November the Research Data Facility at Cambridge University invited some inspirational leaders in the area of research data management and asked them to address the question: “is open data moving science forward or a waste of money & time?”. Below are Dr Marta Teperek’s impressions from the event.

Great discussion

Want to initiate a thought-provoking discussion on a controversial subject? The recipe is simple: invite inspirational leaders, bright people with curious minds and have an excellent chair. The outcome is guaranteed.

We asked some truly inspirational leaders in data management and sharing to come to Cambridge to talk to the community about the pros and cons of data sharing. We were honoured to have with us:

  • PRE_IntroSlide_V3_20151123Rafael Carazo-Salas, Group Leader, Department of Genetics, University of Cambridge
    @RafaCarazoSalas
  • Sarah Jones, Senior Institutional Support Officer from the Digital Curation Centre; @sjDCC
  • Frances Rawle, Head of Corporate Governance and Policy, Medical Research Council; @The_MRC
  • Tim Smith, Group Leader, Collaboration and Information Services, CERN/Zenodo; @TimSmithCH
  • Peter Murray-Rust, Molecular Informatics, Dept. of Chemistry, University of Cambridge, ContentMine; @petermurrayrust

The discussion was chaired by Dr Danny Kingsley, the Head of Scholarly Communication at the University of Cambridge (@dannykay68).

What is the definition of Open Data?

IMG_PMRWithText_V1_20151126The discussion started off with a request for a definition of what “open” meant. Both Peter and Sarah explained that ‘open’ in science was not simply a piece of paper saying ‘this is open’. Peter said that ‘open’ meant free to use, free to re-use, and free to re-distribute without permission. Open data needs to be usable, it needs to be described, and to be interpretable. Finally, if data is not discoverable, it is of no use to anyone. Sarah added that sharing is about making data useful. Making it useful also involves the use of open formats, and implies describing the data. Context is necessary for the data to be of any value to others.

What are the benefits of Open Data?

IMG_RCSWithText_V1_20151126Next came a quick question from Danny: “What are the benefits of Open Data”? followed by an immediate riposte from Rafael: “What aren’t the benefits of Open Data?”. Rafael explained that open data led to transparency in research, re-usability of data, benchmarking, integration, new discoveries and, most importantly, sharing data kept it alive. If data was not shared and instead simply kept on the computer’s hard drive, no one would remember it months after the initial publication. Sharing is the only way in which data can be used, cited, and built upon years after the publication. Frances added that research data originating from publicly funded research was funded by tax payers. Therefore, the value of research data should be maximised. Data sharing is important for research integrity and reproducibility and for ensuring better quality of science. Sarah said that the biggest benefit of sharing data was the wealth of re-uses of research data, which often could not be imagined at the time of creation.

Finally, Tim concluded that sharing of research is what made the wheels of science turn. He inspired further discussions by strong statements: “Sharing is not an if, it is a must – science is about sharing, science is about collectively coming to truths that you can then build on. If you don’t share enough information so that people can validate and build up on your findings, then it basically isn’t science – it’s just beliefs and opinions.”

IMG_TSWithText_V1_20151126Tim also stressed that if open science became institutionalised, and mandated through policies and rules, it would take a very long time before individual researchers would fully embrace it and start sharing their research as the default position.

I personally strongly agree with Tim’s statement. Mandating sharing without providing the support for it will lead to a perception that sharing is yet another administrative burden, and researchers will adopt the ‘minimal compliance’ approach towards sharing. We often observe this attitude amongst EPSRC-funded researchers (EPSRC is one of the UK funders with the strictest policy for sharing of research data). Instead, institutions should provide infrastructure, services, support and encouragement for sharing.

Big data

Data sharing is not without problems. One of the biggest issues nowadays it the problem of sharing of big data. Rafael stressed that with big data, it was extremely expensive not only to share, but even to store the data long-term. He stated that the biggest bottleneck in progress was to bridge the gap between the capacity to generate the data, and the capacity to make it useful. Tim admitted that sharing of big data was indeed difficult at the moment, but that the need would certainly drive innovation. He recalled that in the past people did not think that one day it would be possible just to stream videos instead of buying DVDs. Nowadays technologies exist which allow millions of people to watch the webcast of a live match at the same time – the need developed the tools. More and more people are looking at new ways of chunking and parallelisation of data downloads. Additionally, there is a change in the way in which the analysis is done – more and more of it is done remotely on central servers, and this eliminates the technical barriers of access to data.

Personal/sensitive data

IMG_FRWithText_V1_20151126Frances mentioned that in the case of personal and sensitive data, sharing was not as simple as in basic sciences disciplines. Especially in medical research, it often required provision of controlled access to data. It was not only important who would get the data, but also what they would do with it. Frances agreed with Tim that perhaps what was needed is a paradigm shift – that questions should be sent to the data, and not the data sent to the questions.

Shades of grey: in-between “open” and “closed”

Both the audience and the panellists agreed that almost no data was completely “open” and almost no data was completely “shut”. Tim explained that anything that gets research data off the laptop to a shared environment, even if it was shared only with a certain group, was already a massive step forward. Tim said: “Open Data does not mean immediately open to the entire world – anything that makes it off from where it is now is an important step forward and people should not be discouraged from doing so, just because it does not tick all the other checkboxes.” And this is yet another point where I personally agreed with Tim that institutionalising data sharing and policing the process is not the way forward. To the contrary, researchers should be encouraged to make small steps at a time, with the hope that the collective move forward will help achieving a cultural change embraced by the community.

Open Data and the future of publishing

Another interesting topic of the discussion was the future of publishing. Rafael started explaining that the way traditional publishing works had to change, as data was not two-dimensional anymore and in the digital era it could no longer be shared on a piece of paper. Ideally, researchers should be allowed to continue re-analysing data underpinning figures in publications. Research data underpinning figures should be clickable, re-formattable and interoperable – alive.

IMG_DKWithText_V1_20151126Danny mentioned that the traditional way of rewarding researchers was based on publishing and on journal impact factors. She asked whether publishing data could help to start rewarding the process of generating data and making it available. Sarah suggested that rather than having the formal peer review of data, it would be better to have an evaluation structure based on the re-use of data – for example, valuing data which was downloadable, well-labelled, re-usable.

Incentives for sharing research data

IMG_SJWithText_V1_20151126The final discussion was around incentives for data sharing. Sarah was the first one to suggest that the most persuasive incentive for data sharing is seeing the data being re-used and getting credit for it. She also stated that there was also an important role for funders and institutions to incentivise data sharing. If funders/institutions wished to mandate sharing, they also needed to reward it. Funders could do so when assessing grant proposals; institutions could do it when looking at academic promotions.

Conclusions and outlooks on the future

This was an extremely thought-provoking and well-coordinated discussion. And maybe due to the fact that many of the questions asked remained unanswered, both the panellists and the attendees enjoyed a long networking session with wine and nibbles after the discussion.

From my personal perspective, as an ex-researcher in life sciences, the greatest benefit of open data is the potential to drive a cultural change in academia. The current academic career progression is almost solely based on the impact factor of publications. The ‘prestige’ of your publications determines whether you will get funding, whether you will get a position, whether you will be able to continue your career as a researcher. This, connected with a frequently broken peer-review process, leads to a lot of frustration among researchers. What if you are not from the world’s top university or from a famous research group? Will you be able to still publish your work in a high impact factor journal? What if somebody scooped you when you were about to publish results of your five years’ long study? Will you be able to find a new position? As Danny suggested during the discussion, if researchers start publishing their data in the ‘open”’ there is a chance that the whole process of doing valuable research, making it useful and available to others will be rewarded and recognised. This fits well with Sarah’s ideas about evaluation structure based on the re-use of research data. In fact, more and more researchers go to the ‘open’ and use blog posts and social media to talk about their research and to discuss the work of their peers. With the use of persistent links research data can be now easily cited, and impact can be built directly on data citation and re-use, but one could also imagine some sort of badges for sharing good research data, awarded directly by the users. Perhaps in 10 or 20 years’ time the whole evaluation process will be done online, directly by peers, and researchers will be valued for their true contributions to science.

And perhaps the most important message for me, this time as a person who supports research data management services at the University of Cambridge, is to help researchers to really embrace the open data agenda. At the moment, open data is too frequently perceived as a burden, which, as Tim suggested, is most likely due to imposed policies and institutionalisation of the agenda. Instead of a stick, which results in the minimal compliance attitude, researchers need to see the opportunities and benefits of open data to sign up for the agenda. Therefore, the Institution needs to provide support services to make data sharing easy, but it is the community itself that needs to drive the change to “open”. And the community needs to be willing and convinced to do so.

Further resources

  • Click here to see the full recording of the Open Data Panel Discussion.
  • And here you can find a storified version of the event prepared by Kennedy Ikpe from the Open Data Team.

Thank you

We also wanted to express a special ‘thank you’ note to Dan Crane from the Library at the Department of Engineering, who helped us with all the logistics for the event and who made it happen.

Published 27 November 2015
Written by Dr Marta Teperek
Creative Commons License