Funders, academic institutions, journals and data service providers adopt open access policies including the publication of data underlying research results. While these mandates are an important step towards open science, they often neglect that there is no ‘one-size-fits-all’ approach to open research data across academic disciplines. Different disciplines produce different types of data and have various procedures for analysing, archiving and publishing it. Some have established data management procedures, norms or policies, making their research data open by default, while others do not. Consequently the Research Information Network (RIN) states in a report that
‘if the policies and strategies of research funders, universities and service providers are to be effective in optimising the use and exchange of scholarly information, they must be sensitive to the practices and cultures of different research communities.’
This briefing paper presents the current state of open research data across academic disciplines. It describes disciplinary characteristics inhibiting a larger take-up of open research data mandates. These characteristics include data management practices, disciplinary norms of data sharing, career-related factors, infrastructural factors, and legal and ethical questions of public access, Additionally, the paper presents the current strategies and policies established by funders, institutions, journals and data service providers alongside general data policies. It can be found on the website of PASTEUR4OA.
]]>Back in 2001, while still a professor of philosophy at Earlham College, Suber undertook a sabbatical from his teaching duties mainly with the intention of focusing on his academic research. During this time, he became increasingly interested in the web’s power for sharing scholarly writing, starting his own weekly newsletter on the subject. As the popularity of the newsletter rapidly grew, so too did Suber’s interest in open access, leading him to spend ‘every hour of my work day, plus many other hours’ working on the topic. The newsletter soon became a blog entitled ‘Open Access News’, from which most of the book’s contents are taken. Suber left Earlham in 2003 and has worked full-time on open access ever since.
The book covers Suber’s writings from the early days of the newsletter through to 2011 – a time of huge change for open access to knowledge across the world. During this time, open access went from being an extremely niche activity to something that is near impossible for the average researcher to ignore. The book features sections on the case for OA, understandings of OA, disciplinary differences and what the future might hold, all written in an approachable and conversational style.
For policymakers, there is a whole section on funder and university policies for open access that contextualises Suber’s excellent guide (co-authored with his colleague Stuart Shieber) on Good Practices for University Open-Access Policies. Harvard’s 2008 open-access policy was the first OA policy in an American university and the first faculty-led (rather than administrator-led) policy.
Coupled with his concise introductory 2012 book Open Access (also MIT Press) the two works should offer an excellent introduction and a compelling case for open access publishing.
]]>Open Access to research data is fast becoming recognised as complementary to Open Access to research publications, both key components of Open Science. While the PASTEUR4OA project targets the development and reinforcement of Open Access strategies and policies for research publications, the project also encourages the development of such policies for research data
This briefing paper provides an overview of the current situation with regards to Open Access to research data. It considers the benefits and challenges of opening up research data with a particular focus on current funder and institutional policy developments in Europe and further afield and shares resources and initiatives for further study. The paper is available from http://www.pasteur4oa.eu/resources.
]]>The PASTEUR4OA project aims to increase national and institutional policymakers’ as well as research funders understanding and awareness of Open Access (OA). PASTEUR4OA also aims to help develop and/or reinforce OA strategies and policies at a national, university and research funder levels that align with the European Commission’s 2012 Recommendation on Access to and Preservation of Scientific Information and the Open Access Mandate for Horizon 2020.
The project also aims to facilitate coordination among all EU Member States and Aligned Countries by establishing a network of expert organisations across Europe – Knowledge Net – and by collaboratively developing a programme of activities that support policymaking at the national, university and funder levels. Open Knowledge participates as a partner in the project both to strengthen our existing Open Access community and to help increase engagement between our community and policy makers across the EU.
To promote the development and reinforcement of OA policies, PASTEUR4OA has produced a series of advocacy resources that can be used by stakeholders developing new policies or revising existing ones.
This new flyer describes the objectives of the PASTEUR4OA project and outlines the advocacy resources that are available for policymakers and OA stakeholders. More information on these resources is also available on the PASTEUR4OA website.
]]>
While it was agreed that the effectiveness exercise was useful it was recognised that long, comprehensive reports often fail to have the required effect on policy makers. One idea was to carry out some data visualisation work on the ROARMAP data and create both an online data visualisation hub and a series of infographics to feature as part of the advocacy material being developed.
I was chosen to lead on the data visualisation work for PASTEUR4OA, but I hadn’t created a series of visualisations like this before. The prospect was a little daunting! However I was lucky enough to have a more experienced colleague whom I could ask for help and bounce ideas around with.
My main brief was to exploit the ROARMAP database and create visuals to be produced for advocates to use in presentations, literature etc. These visuals would show the statistics in attractive and interesting ways, so for example in the form of maps etc. The visualisations would need to be useful for policy makers, institutions, researchers and individuals interested in Open Access. It was also suggested that we use live data when possible.
Some of the questions I asked myself and others prior to starting the work are listed below:
There are quite a few guides on the process of data visualisation creation but one that I found particularly useful was this overview from Jan Willem Tulp published on the Visual.ly blog. I also appreciated the clarification of roles in the 8 hats of data visualization design by Andy Kirk.
Early on in the process to make sure that I was thinking about the story we wanted to tell I set up a scratch pad in which to record the questions we wanted to ask of the data. So for example: How many Open Access policies are there worldwide? Can I see this on a map? Which countries have the most policies? How many policies are mandatory? How many comply with the Horizon 2020 OA policy? Does mandating deposit result in more items in repositories? How many policies mention Article Processing Charges? Etc.
We also agreed on the data we would be using:
Most of the data would be coming from ROARMAP and we worked closely with the ROARMAP developers, and had had significant input into the data on the site, so we were confident that it was reliable. Usually when selecting sources it is useful to keep in mind a couple of questions: is it a reputable source? Is it openly available? Is it easy to get out and work on? Has it been manipulated? Are there omissions of data? Will you need to combine data sets? The ROARMAP site doesn’t have an API but you can get a JSON feed out of the site, or search for data and create excel dumps.
To begin with I started working on excel dumps from the site. One of the first hurdles I had to jump was getting the country names added to the data. ROARMAP data was categorised using the United Nations geoscheme and the country names were missing. Most of the manipulation could be done in Excel, it is a pretty powerful tool but it requires sensible handling! Some of the useful functions I learnt about include:
Although you don’t need to be an expert in Excel or Google Spreadsheets it does help if you can use the tool fairly confidently. For me much of the confidence came from being able to manipulate how much data was shown on a sheet or page: so being able to hide rows, lock rows, filter data etc. Less is more – and if there is only the data you need on the page then life becomes a lot easier. Another lesson I learnt early on is the need for regular sanity checks to ensure you are being consistent with data and using the right version of the data set. I kept copious amounts of notes on what I’d done to the data – this proved to be very useful if I wanted to go back and repeat a process. Also I’d suggest that you learn early on how to replace a data set within a tool – you don’t want to get pretty far down the line and not be able to update your data set.
Once I had an idea of which questions needed to be answered…I began to experiment with data visualisation tools. There is a great list of tools available on the datavisualisation.ch site. The main ones I tested out were:
I also experimented with the following infographic tools:
Whilst trialing each of these I had a few questions at the back of my mind:
I looked primarily at free services, which obviously have some limitations. Some tools wouldn’t allow me to take the visualisations and embed them elsewhere while others required that I had significant programming skills (in SQL, PHP, Python, R or Matlab) – something I seriously didn’t have time to learn at that point.
Tableau Public came out on top as an all-round tool and I made the decision to stick with one tool for the online visualisations (Tableau public) and one tool for the infographics (here I chose Infogram). Unfortunately both tools didn’t link to live data, in fact none of the free tools seemed to do this in any user-friendly type way.
Whilst I’ve been working on the data visualisations for PASTEUR4OA the number of Open Access policies that have been submitted to ROARMAP has been increasing. While this is great news for the project it has meant that my data is out of date as quickly as I download it. However I’ve discovered that linking to live data isn’t that easy. Few of the free tools allow it and the best way to create visualisations that do this seems to require programming skills. A colleague of mine helped me pull the JSON feed into a Google spreadsheet and then build a map on top of it but the result is slow to load and not particularly attractive. Linking to live data was going to require better skills than those I possessed – so I asked PASTEUR4OA’s project partner POLITO to help us. Their main work so far has been creating Linked Data SPARQL end points for some Open Access data sets but they have also been experimenting with live data visualisations. You can see an example of their efforts so far in this dynamic ball map.
Once I started creating the data visualisations it made sense to have somewhere to store them all. I set up a Github site and worked on a series of pages. Our in-house designer added some PASTEUR4OA styling and the result is available at http://pasteur4oa-dataviz.okfn.org/. The site has information on the questions we have been asking and the data used as well as a FAQ page to explain what the visualisations are for. The visualisations site is linked to from the main menu on the PASTEUR4OA website.
At this point I spent some time thinking about the look and feel of the visualisations. The PASTEUR4OA team suggested we use the ROARMAP colours as a ‘palette’ for the visualisations.
I also added headings, legends and explanations for the online visualisations to explain what questions they were asking. As part of this work a series of infographics (.png files) have been created from the Infogram visualisations with the intention of using them in blog posts, presentations etc. The images are embedded in the main data visualisation website.
Some things I thought about in more detail at this stage:
PASTEUR4OA are also keen to make the data we’d be using openly available so have uploaded versions to Zenodo, a service which allows EU Projects to share and showcase multidisciplinary research results. The data set urls are listed on the data set page on the main visualisation website. Over time we intend to add links from the main data visualisation website to other Open Access open data that believe could be of interest. As mentioned earlier, POLITO will be making some of this data available as linked data. The idea is that developers can use the work we’ve done as ‘proof of concept’ or inspiration and build more visualisations using the data available.
Through carrying out this piece of work for PASTEUR4OA I have learnt many significant lessons about the data visualisation process. Hopefully this blog post has provided a taster of the challenges and benefits such a process brings. As a newbie it wasn’t always easy, but it was certainly interesting and rewarding. If you are thinking about making your own visualisations you might find this complimentary slideset I have created useful.
I believe that the results collected on the PASTEUR4OA data visualisation website are an example of the kind of things those wishing to advocate for Open Access could do without any programming skills. They are there to inspire people, developers, researchers and those new to visualisation and interested in Open Access. It would be great to see some of the visual aids we’ve created in presentations, posters and articles – maybe they can make the (at times!) dry data we were given interesting and accessible.
The participants will exchange ideas and policy practices and discuss concrete actions to improve existing or develop new Open Access policies aligned with the European Commission’s Recommendation to Member States of July 2012 and the Horizon 2020 Open Access requirements. The agendas of these regional workshops will be tailored to the specific needs of the respective target groups (funders or universities/research centers) and the level of Open Access policy development in the region.
The regional workshops will take place on:
For further details see the PASTEUR4OA website.
]]>Image from Pixabay, CC0
This timeline has used What is Open Peer Review as its foundation. Once again any suggestions would be much appreciated.
1665
20th century – Peer review became common for science funding allocations.
1948
1955
1961
1969
1976
1986
1989
1996
1998
1999
2000
2001
2002
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
I am no expert in data visualisation and am very much learning as I go along. I hope to share my experiences in creating the visualisations later down the line. Today I have been working with Tableau Public and have produced a number of different visualisations. With APCs in mind (you may have seen the APC timeline we have been collating) here is an attempt at a visualisation of Open Access Policies (by country) that offer some form of APC grant.
Any feedback welcome – I’m still learning so please be gentle!
]]>Image CC0 Public Domain via Pixabay
Some of you may have seen our Open Access to Research Data Timeline. We now have an APCs timeline we’d appreciate feedback on. Again this timeline builds on Peter Suber’s Open Access timeline – but other important events and papers have been tricky to find.
1930s
1965
1970s
1983
1994
2002
2003
2004
2006
2013
2014
2015
For those interested in subscribing to the Newsletter you can register your details online. You can view past issues from the Mailchimp site.
]]>