Category: Open Science

  • A Better Future for Open Access Publishing in the Netherlands

    A Better Future for Open Access Publishing in the Netherlands

    Today, I learned about an important initiative: Call to Commitment: A Future-Proof Approach to Open Access Publishing in the Netherlands. It raises big questions about how we share research and who has control over it.

    The Problem with Current Deals

    Since 2015, Dutch universities have made deals with big publishers. These deals, called transformative agreements, let researchers publish Open Access (OA) without extra costs. They also give universities access to subscription journals.

    At first, these deals seemed like progress. But they have big downsides. They keep publishers in control and focus on making money. Researchers lose control of their work. The costs are also very high. This system does not match the values of Open Science, which aims for fairness, openness, and equality.

    A New Direction

    The Call to Commitment asks universities to take a different path. Many of these agreements will end soon. This gives us a chance to create a better system. A system that focuses on community needs, not company profits.

    The group behind this call wants universities to follow the Open Science principles from UNESCO. These principles focus on fairness, openness, and public benefit. They can help us build a publishing system that works for everyone.

    What Can Change?

    The Call to Commitment suggests new ways to share research:

    • Institutional Repositories: Free platforms where researchers can share their work without paywalls.
    • Diamond Open Access: Journals run by communities, free for authors and readers, supported by public funding.
    • Preprint Servers: Sites where researchers can share their work before formal publication, encouraging faster sharing and feedback.
    • Thematic Curation: Tools to help different audiences easily find and use research.

    Instead of spending large amounts on current deals, universities could fund these alternatives. This would make research sharing fairer and give control back to the academic community.

    Why It Matters

    How we share research affects everyone. Current systems benefit big companies but limit who can take part. A new system could give all researchers and institutions equal opportunities. It would also make knowledge more available to the public.

    This is our chance to make real change. Dutch universities have always been leaders in Open Access. By acting now, they can set an example for others around the world.

    What Can You Do?

    If you care about Open Science, you can support this initiative. Talk about it with your colleagues, or share it with your institution. Change takes action, and every voice helps.

    Let’s build a future where research is open, fair, and for everyone.

  • Data Scientist Training for Librarians

    Data Scientist Training for Librarians

    DTU in Copenhagen hosts a fantastic data scientist training for librarians.

    http://www.altbibl.io/dtu/

    All you need to know about and understand how a researcher works with their data; cool words are involved like data-scrapin’, data wranglin’ and more. Register here and get data-savvy yourself.

    Data Scientist Training for Librarians (DST4L) is an experimental course, started at the Harvard-Smithsonian Center for Astrophysics John G. Wolbach Library and the Harvard Library to train librarians to respond to the growing data needs of their communities. Data science techniques are becoming increasingly important to all fields of scholarship. In this hands-on course, librarians learn the latest tools for extracting, wrangling, storing, analyzing, and visualizing data. By experiencing the research data lifecycle themselves, librarians develop the data savvy skills that can help transform the services they offer. Read more about the program here.

    Thanks to Foster Open Science, the DTU Library and our contributors, DST4L is coming to Copenhagen. The course is free and open to beginners. Registration opens on June 3rd and closes on June 10th. Space is limited and acceptance will be based on the strength of your application. The course runs from September 9th to the 11th

  • THOR – Technical and Human infrastructure for Open Research

    THOR – Technical and Human infrastructure for Open Research

    THe Onion Router vs Technical and Human infrastructure for Open Research. Despite of the terrible confusing acronyms used in EU projects, I do sympathise with the important initiative linking the world’s body of scientific artifacts.

    http://project-thor.eu

    My current THOR-like activities involve Research Data Management training for PhD’s at TU Delft and University of Twente, where I tell them about using ORCID and connecting their research artifacts; data and publications using DOI’s.
    Also I do have high hopes the data and publication repositories in the Netherlands will incorporate ORCID, if the Dutch universities get to sync their DAI’s to ISNI’s and the ISNI’s to ORCID’s.

    THOR builds on the lessons and recommendations from the ODIN (ORCID and DataCite Interoperability Network) project, and will:

    Leverage two community-driven global PID initiatives for contributors (ORCID) and scientific artefacts (DataCite) to build tools to serve the evolving needs of the research community.
    Deliver PID-based services to submit, identify, attribute, and cite artefacts, starting with four disciplinary communities: Biological and Medical sciences, Environmental and Earth Sciences, Physical Sciences, and Social Sciences and the Humanities
    Create PID integration and interoperability solutions for research institutions, libraries, data centres, publishers, and research funders
    Enhance the expertise of the European research community by running an intensive training programme, and creating a knowledge base for practitioners integrating PIDs into research information systems.

  • Data carpentry

    Data Carpentry aims to teach the skills that will enable researchers to be more effective and productive. The workshop is designed for learners with little to no prior knowledge of programming, shell scripting, or command line tools.

    http://datacarpentry.org

    Upcoming events

    Workshop at Utrecht University
    24-25 June 2015
    Instructors: Karthik Ram, Aleksandra Pawlik
    Sponsored by ELIXIR

  • School of data

    School of data

    School of Data works to empower civil society organizations, journalists and citizens with the skills they need to use data effectively – evidence is power!

    Home 2

    For me it is a great way to learn and understand what a researcher must do with data.

  • Open Access Button

    Open Access Button

    This is an interesting tool every PhD student and researcher should know about, especially when they are frustrated by pay walls for scientific articles. And the anoyances rise when there probably already is an open access version around. The solution: Open Access button. It is a simple bookmarklet you use when browsing the web and stumble upon a pay wall, if you click it it searches the web for an open access alternative.

    Screenshot of the Open Access button in action as a bookmarklet in a browser
    Screenshot of the Open Access button in action as a bookmarklet in a browser

    When there is no OA version available, you can make some noise about your frustration on social media, get in touch with the author and put the article in a wishlist. This wishlist gives you an alert when the article is found als OA alternative weeks or months later.

    It is such a great idea that even our own State Secretary of the Dutch Ministry of Education Culture and Science Sander Dekker is enthusiastic about it.

    I am even more impressed by the Open Access Button, a genuinely grass-roots initiative developed by two students.

    Check it out at www.openaccessbutton.org

  • App and Data Market for Researchers

    App and Data Market for Researchers

    data-one-stop-shop

    As a researcher you have to make a cost estimate of your project. Wouldn’t it be really nice to have a one-stop shop where you can select your research analytics tools like R (as a service) and your datasets, all in the cloud by the way, and pay at a checkout counter.

    This is just an idea where I describe the potential of the SURFmarket’s pilot project “Cloud distribution channel” when adding “Quality Data” next to the so called “Apps”.

    App overview desktop - SURFmarket Cloud distributie kanaal - pilot project

    In the Netherlands there is an increasingly need for an overall overview of the apps, tools, services and data you can utilise as a research in your research project. At the universities “Research Service Centers” are popping up, trying to help and support the researcher in their different phases of their research project, by combining the knowledge and services of the different university departments together in one one-stop-shop at the institutional level. The University library of Maastricht is growing their service center together with all the researchers and staff departments, trying to evolve and improve their offerings. Also the Technical University of Delft is creating a specialised department for Research Support. All these people in the Netherlands who are committed to deliver the support structure for the needs of the researchers, are united in a special interest group (SIG) from SURF, called the SIG Research Support.

    Research Support Center - University Library Maastricht

    Looking at ‘research data’ services in particular is that the University service departments are, for now, concentrating  on services where you can store your research data at an optimum balance between costs and availability. When you have your data sitting there, stored, well described, stamped with a quality brand and ready to be reused and citable, it would be great if you could put it on a market place to cover for the digital curation costs, right?  Well lets assume we live in a paradigm where this is a great idea! What are then the potential possibilities?

    I will put in another ingredient in your mind to this conceptual idea. As an analogy of the iTunes App Store (Trade Mark, and so forth), this market place that I will describe can offer little corners of scientific disciplines where small editorial boards are making collections of tools and high quality datasets for researchers to delve into.

    Lets just focus on some “neutral” databases with factual data, or factoids. I will try to sum-up some of those databases that are out there; public geo spacial data (NL:kadaster), commercial publisher data Web of Science, demographic statistical data (NL:CBS), international bibliographic data (VIAF), etc.

    In a potential use case where I am a researcher, I would like to visit this collection of my discipline and search, filter and look at the datasets available. I would like to select the dataset to use and it shows me the costs. It is maybe even possible to make a query for a sub-section of the data, and reduce the total costs. Then I get two buttons: Try and Buy. The Try button gives me a sample of 10 records of data to look if it is useful enough, or to see if I can find values to connect the data to other datasets. The Buy button sends me to a licence agreement page where it tells me how to cite the dataset or subset, what I can and can’t do with it, if it has an expiration date, can it be reused, combined, made publicly available, etc.

    And just as you this might be great for research! … it is already done by the commercial industry. Bringing data for apps in iTunes style.  has written for the Oreilly blog an article called “Data markets compared” which gives a pretty good overview of the data markets available.

    datamarket

    For the sake of LOCKSS curation principle, I will just cite a big part of that blog content here.

    Data markets compared

    Azure Datamarket Factual Infochimps
    Data sources Broad range Range, with a focus on country and industry stats Geo-specialized, some other datasets Range, with a focus on geo, social and web sources
    Free data Yes Yes Yes
    Free trials of paid data Yes Yes, limited free use of APIs
    Delivery OData API API, downloads API, downloads for heavy users API, downloads
    Application hosting Windows Azure Infochimps Platform
    Previewing Service Explorer Interactive visualization Interactive search
    Tool integration Excel, PowerPivot, Tableau and other OData consumers Developer tool integrations
    Data publishing Via database connection or web service Upload or web/database connection. Via upload or web service. Upload
    Data reselling Yes, 20% commission on non-free datasets Yes. Fees and commissions vary. Ability to create branded data market Yes. 30% commission on non-free datasets.
    Launched 2010 2010 2007 2009

    Other data suppliers

    While this article has focused on the more general purpose marketplaces, several other data suppliers are worthy of note.

    Social dataGnip and Datasift specialize in offering social media data streams, in particular Twitter.

    Linked dataKasabi, currently in beta, is a marketplace that is distinctive for hosting all its data as Linked Data, accessible via web standards such as SPARQL and RDF.

    Wolfram Alpha — Perhaps the most prolific integrator of diverse databases, Wolfram Alpha recently added a Pro subscription level that permits the end user to download the data resulting from a computation.

    cited from: http://strata.oreilly.com/2012/03/data-markets-survey.html

    Perhaps the research community can learn from the commercial industry and engage in collaboration. What is so bad about storing your research data at a commercial storage facility, as long as they use DOI’s for citing, the bucks acquired can be used to curate your data for the generations to come.

    What are your thoughts?

  • Inspire me please

    Inspire me please

    Who will you inspire?Inspire is a fictional webservice for researchers to find blind-spots in the current global scientific body of knowledge. These blind-spots are uncharted question area’s that need answered by theoretic or evidence based research. Basically Inspire can show researchers what work needs to be done, and what work is already done. This makes research more efficient, more fun to find missing pieces of the puzzle together, cooperative and less competitive.

    In January of the year 2008 I have written down this idea, that was related to the SURFshare program at my work at SURF. This ambitious idea could be used to create a pathway of the direction the program could head towards. It should made use of the existing information infrastructure, with Open Access publications as a free body of knowledge, but needed some extra additional components where the state-of-art back then seized to exist. The idea was to stimulate the creation of a free and open academic information infrastructure, and a grand vision could fuel this. This idea never made it, because it was meant too ambitious, to far beyond the grasp, and also it was just one use case, yet research activities contained so much more.

    I have written this down and published this information so it might inspire you, otherwise it would be lost. The original document in Dutch can be found here. Below follows a revised translation.

    Inspire is a webservice that serves the academic community. Its key function is to support researcher in the creative process, by finding blind-spots in the current body of knowledge. Inspire can support the scientist in different search tasks in a research work-flow, by reducing the information overload.

    Ant BridgeWhat can Inspire do for you?:

    • Question spotter: Inspire helps the researcher at the start of research by finding out if the research question hasn’t been asked before. This makes research more efficient, by reducing work that already has been done. Or it shows what research can be done again to double check and evaluate the outcomes of previous research.
    • Blind spotter: Inspire can help point out what questions haven’t been stated before, yet obvious that they need to be stated and researched to find the answer to.
    • Solution pointer: Inspire provides support to look for solutions and methods that are available in other area’s of research. e.g. the idea of path finding technology of peer-to-peer networks in computer science were found in biology, inspired by trail-scent search strategies used by ants.

    Functions

    What are the necessary functionalities of inspire in order to work for you?

    1. Publication list builder: Inspire can generate a publication list of the researcher. (The more information you provide the better the results. Providing name variances, an ISNI and ORCID helps.)
    2. Knowledge profile: Inspire is able to create a knowledge profile. This is a ‘fingerprint’ or concept map of the knowledge and expertise of the researcher. This fingerprint can be used in other Inspire services, for example to measure the proximity of another researchers. This profile is created from extracting information from the publications in the CV.
    3. Related literature (domain related): Inspire is able to offer the researcher new and related literature available within his own knowledge domain. This is done by matching the fingerprint with the body of knowledge within his designated knowledge domain. The researcher is able to receive new literature specific to his interest at this moment in his academic career. This reduces the information overload of unrelated material, and reduces effort for searching himself to get a quick up-date of the state-of-the-art.
    4. Related literature (cross domain): Inspire is able to offer the researcher related literature outside his own knowledge domain. The reason to do this, is to get inspired by the things happening in other research domains, broadening the horizon and creating new research opportunities. This is done by matching the fingerprint with the body of knowledge outside his designated knowledge domain.
    5. People you know: Inspire is able to create a professional network based on people you already know. This is based on your peers (people you cited) and co-authors.
    6. Related people: Inspire is able to build your professional network. It shows people related to you according the proximity of the fingerprints / knowledge profiles. You can meet people you didn’t know they existed before, yet can be very useful to be inspired by.

     

    Use cases:

    According to Joost. G. Kircz (in Modularity: the next form of scientific information presentation? Journal of Documentation. vol.54. No. 2. March 1998. pp. 210-235) we can identify four user types for discovery of scientific information resources. The informed reader, the partially informed reader, the uninformed reader, and the non-reader.

    The informed reader This is the reader who knows what he or she is looking for and is able to find his or her way in the literature quickly. … This type of reader is not interested in the general parts but wants quick, direct access to the specific description of the experiments or theory and the specific results. …

    The partially informed reader This kind of reader is not conversant with the specific research as such, but is interested in the general aspects that might be of use for their own investigations. … This type of reader wants to know how a particular paper fits within the broader spectrum of their own research, what the relations are with other methods in the same field, and what the connections are with related fields. …

    The uninformed reader This is the group of readers who want to learn something new. They are curiosity driven open-minded browsers, who hope to get a fresh idea from fields that are either unknown or of which they have only a rudimentary knowledge. Authors’ names do not ring a bell, prestige and fame are unknown, as is the jargon and the intricate details of the paper. This type of reader needs a clear statement of the goal and embedding of the work, the uniqueness or particularity of this work compared to others and so forth. This method of reading is typical in the exploratory phase of a research project, and demands contextualised and clearly written language.

    The non-reader This category consists mainly of science administrators who want only to know if a researcher is active. Bibliographic information (and nowadays also often impact factors etc.) is normally sufficient. These readers look for a place where the essential record keeping data are given. In addition, this type of reader wants a general statement of the goal of the research and possible claims by the author, why the research was performed and how it fits in the larger body of research within the field; a clear distinction between the author’s own previous work and the new work warranting this new publication, and between their own and other peoples’ work.

     

    Possible applications / scenario’s

    Phase 1

    In this Inspire project we mainly focus on the exploratory phase in the research process where the “uninformed reader” is active.

    To be able to distinguish relevant information sources for the uninformed reader, a concept map fingerprint of his own knowledge domain has to be determined. This fingerprint can be made by linguistic methods like weighted keyword density mapping by scanning the full text papers in his CV. keywords in older research weigh less than newer research; because interests might shift over time.

    Inspire calculates for all available information sources (papers, persons, projects) a concept map fingerprint. Then it calculates the ‘distances’ between concept map fingerprints.

    Visualisation: the uninformed reader is able to vary the distances from his own concept map fingerprint. This way he is able to shift different related information resources in and out the viewpoint. In the first circle will contain his own publications, the second will contain directly related work, cited material and the work of coauthors, further away will lay information resources that are indirectly related, or from other knowledge domains.

    Filters can be applied on the viewpoints, so that the uninformed reader sees only person, only papers, only projects, only research grants, only patents, and so forth. Within the different viewpoints also information resources can be narrowed down by using key-word search.

    Annotations: the information resources can be annotated using social tagging. This tagging helps to create a folksonomy of a particular resource. This helps to improve the concept map fingerprint of that particular information resource.

    Ratings: With a like or dislike the uninformed reader can give a higher or lower weight to the distance of an information resource. This influences the information visible in the differrent viewpoints.

    Phase 2

    Suggestion function: an API that automatically comes with suggestions of related work. for example a programmer can make a addon in popular text editors or blog systems, and use the API to come up with related information resources that are not far distanced from the lines of words that are being written.

    Phase 3

    Relations: This system is usful for the partially informed reader when the relations between information resources are presented in the different viewpoints.

    Phase 4

    Blind spotter: When the above applications are finished, it is possible to create a blind spotting viewpoint. This viewpoint, explained earlier, makes is easier to find research questions that have been questioned before, and that questions have not been asked before.

     

    Comments please

    I hope you enjoyed reading. If you feel like leaving a comment about what triggered you to get inspired by reading this, please don’t hesitate.

  • Research Intelligence

    Research Intelligence

    This presentation reflects upon the report “Users, narcissism and control” from Paul Wouters & Rodrigo Costas (CWTS) [download Users narcissism and control]

    In this presentation we make the bold statement that research policy and reward system is build upon a very small and thin layer of information. Yet the full spectrum contains much richer information. This presentation explains how this rich layer looks like and what information it contains in different layers of abstraction, and how it connects with policy and decision making in the end. This is the business case for the Dutch goverment to introduce performance indicators.

    This presentation suggests that one can do more with these indicators, such as discovering trends, and making analytics. For example making correlations between the downloads and mentions about publications or scientific instruments increases the Shanhai-index by x% in two years time. This can make a shift policy for universities to invest more in open access and in marketing their projects better.

    To support this abundance of information and making cross-domain analysis, one needs a cluster of computing power, storage, and has to arrange licences. This can be done with an array of partners in the Netherlands, creating a middelware infrastructure for Research Information (RI). Universities and research institutes who want to use this RI-middelware infrastructure (RIMI) can subscribe. The RIMI provides raw output and calculated output for services to draw from, services used by the subscribers. These services can be for example, a research information dashboard for individuals, for institutions, for funders and for ministries. It can be used in services that want to visualise semantic-web like relations of the academic information domain. It can be use for resolution mechanisms, research portals, catalogue search systems, etc. The main point is that all these services easily can be build, because the information openly available, in machine reabable formats based on internationally accepted standards so that developers can work with it right away. The information comes from trusted sources, and is as clean as possible, to reduce the noise amplification in the processes later on when the data is compared, combined, calculated and correlated.

    2012 12-12 research analytics – idea from maurice.vanderfeesten

    Disclaimer: This presentation is just an idea, and it contains organisations in fictional situations. It is a proposal, and does not reflect the current situation.

  • Article Level Metrics

    Article Level Metrics

    In response to Martin Fenner’s interview with himself.

    Hi Martin, I’ve got the link of you Blog Post from Najko Jahn – Bielefeld University. I wish you all the best at Plos-One ALM.

    Just for you to know as technical lead: DFG (DE), JISC (UK), DEFF (DK) and SURF (NL) worked together on guidelines for exchanging article level metrics in a transparent manner. The reason to do this is to be able to compare the statistical usage data that comes from various distributed locations, eg. repositories, but also publishers in the future. You might want to look at these KE Guidelines for the aggregation and exchange of Usage Data .
    Dutch repositories already aggregate the Article Level Metics from each repository, and can create a National overview of this data. The EU project OpenAIRE also use these guidelines for Europe-wide data exchange on these metrics. A neat thing is that the Article data can be aggregated on Author level once you know the Author ID. This is being tested in NARCIS right now, results look great.

    Plos-One ALM has been a great source of inspiration and motivation for these kind of projects, so keep up the good work!

    Enhanced by Zemanta