Registration Desk opens at 8:15Main Hall (Uni Mail)
T1 - Metrics (Room R160)Room R160
Altmetrics is a hot buzzword. What does it mean? What's behind the buzz? What are the benefits and risks of including alternative metrics of research impact in our discovery and evaluation systems? What altmetrics tools exist today, what are their strengths and weaknesses, and where is the field going?
Join Heather Piwowar, cofounder of ImpactStory, for this tour of the altmetrics landscape. The session will be relevant to anyone who produces, publishes, or evaluates research: funders, university administrators, journal and repository leaders, and individual scholars.
If you can, bring a laptop to play along during the session. Do you have research products you'd like to experiment with? Prep a digital list of IDs (DOIs, PMIDs, URLs, researcher ORCIDs, etc) and we'll see what we can discover!
T2 - Metadata for the Research Lifecycle (Room M6289)Room M6289
The management and publication of research data is increasingly important, and relies heavily upon appropriate descriptive metadata. The tutorial will provide a walkthrough of the research data management life cycle, from planning through data acquisition to data publication and preservation. Each of these stages requires different tools and methods, for example DMPonline, DataStage, DataBank and DataCite. However, some things stay the same: for example, person, subject, method, instrument, project and funder are likely to be constant throughout the research lifecycle, and metadata describing these can be re-used in different contexts. If tools are connected appropriately, researchers could work in a dynamic web-based data management environment to which they could regularly return, confident that metadata that had already been entered would be available for reuse, and that datasets they describe would be securely backed up.
This tutorial is intended for data service developers and managers. Participants will be expected to adopt a researcher's perspective, and to identify aspects of research data management that can simplify the researcher's life. Overviews of available tools and methods for each stage in the research life cycle – planning, acquisition, publication and preservation – will be alternated with hands-on sessions employing existing tools for handling exemplary data and metadata, including the participants' own data. The use of ontologies and RDF metadata representations will be explored. Participants will be invited to suggest additional tools and use cases. Details will be provided in due course.
(University of Oxford), DrWolfram HORSTMANN
(University of Oxford)
T3 - Metadata: from Records to Graphs (Room R150)Room R150
The tutorial will introduce traditional as well as state of the art approaches to modeling information objects and the processing environments they are part of, both in traditional memory institutions and in the Linked Data web. The central issue is to understand what is changing with RDF-graph based approaches to metadata generation and management. We will do some hands on work on semantic annotation as part of the workshop using the Pundit tool (http://thepund.it)
This tutorial aims to address the following matters:
I. A view upon the software package,
II. Setting and organising the team (software and teamwork bootstrapping), and exploring posible workflows,
III. Data concerning issues.
"A view upon the software package" will answer to the common questions like: what is it, who is behind it, is it suitable for what we do, how long it will take me to set it up, is it simple to use, are there many working with the package?
All these questions will find a set of answers backed with practical work on the software itself starting from installation easy steps up to the moment when we will have a fully workable setup.
"Setting and organizing the team and exploring posible workflows" will be one of the central points of the tutorial because it is also the power behind the long run efficiency of the journal. Certain aspects modeled by the different interaction levels with the software will be explored and some possible workflows will be discussed among the participants.
"Data concerning issues" will be the signal that will mark the ending of the tutorial making all aware of the present actions with regard to data management and where is the value of some possible actions concerning preserving it.
The attendants do not need to have programming skills, but to understand the fundamentals of operating an Operating System - GNU/Linux, Mac OSX or Windows.
Need to have on your belt
Participation to this tutorial requires a laptop and depending on you OS of choice the following software packages already installed:
GNU/Linux (LAMP installed + PhpMyAdmin)
- MAMP (http://www.mamp.info/en/index.html),
- XAMMP (http://www.apachefriends.org/en/xampp-macosx.html).
Windows (you may choose between):
- EasyPHP (http://www.easyphp.org/save-easyphp-latest.php);
- XAMPP (http://www.apachefriends.org/en/xampp-windows.html).
Download the lattest OJS package (Current Development Release) fromhttp://pkp.sfu.ca/ojs_download.
Unarchive the package in htdocs or www directory of your preferrate package.
Having all these set up before the tutorial is mandatory.
T5 - The NISO/OAI ResourceSync Synchronization Framework (Room 1130)Room 1130
This tutorial will provide an overview and a practical introduction to ResourceSync, a framework to synchronize web resources that consists of multiple modular capabilities that a server can selectively implement to enable third party systems to remain synchronized with the its evolving resources. All capabilities leverage document formats introduced by the widely adopted Sitemap protocol.
The editors of the ResourceSync specification are affiliated with the Los Alamos National Laboratory, Cornell University, Old Dominion University, and the University of Michigan. They have been involved in other interoperability specification efforts, including the OAI Protocol for Metadata Harvesting, OAI Object Reuse and Exchange, Memento, and Open Annotation. An international Technical Committee has supported the editors in compiling the draft specification.
The tutorial will:
* Motivate the ResourceSync approach by outlining several synchronization use cases including scholarly article repositories, linked data knowledge bases, and resource aggregators.
* Detail the nature of the various ResourceSync capabilities (Resource List, Resource Dump, Change List, Change Dump) that a server can implement.
* Show how support for these capabilities can be expressed (Capability List) and discovered.
* Describe the extensibility mechanism built into the framework that allows addressing specific needs, such as synchronizing from mirror sites, synchronizing resources by exposing patch information, and synchronizing both metadata and content described by that metadata.
* Provide details about the serialization format used to express ResourceSync capabilities, which is based on document formats introduced by the Sitemap protocol.
* Describe experiences developing general ResourceSync software libraries and particular support for an institutional repository platform.
The intended audience are people involved in both technical and management aspects of digital repositories or in the creation of value-add services across such repositories. The tutorial will assume a basic level of familiarity with fundamental web concepts (URI, resource, representation) and XML, but will be presented in a way that is accessible to people with a non-technical job description that have basic technical knowledge.
MrHerbert Van de SOMPEL
(Los Alamos National Laboratory), MrRichard JONES
(Cottage Labs), DrRobert SANDERSON
(Los Alamos National Laboratory)
T6 - Open Access Café 2013 (Room 1140)Room 1140
Following the success of this session in the OAI7 Workshop, we are pleased to offer again the chance to learn more about Open Access topics that you would like to understand further. We will present a session where you can mingle with various experts, ask them questions in a café-style setting and discuss issues in a very informal atmosphere. There will be time during the session for you to explore 4 or 5 different topics in which you are particularly interested. We will engage experts from across the spectrum in order to cover technical-, cultural- and policy-related issues.
Come along to the Open Access Café, meet people, talk about Open Access -- and drink coffee, of course!
Lunch & RegistrationRestaurant (Uni Mail)
Plenary 1: Technical SessionR380 (Uni Mail)
MrHerbert Van de Sompel
Tech Session: How semantic representations can support scholarly communication
One part of scientific communication is to effectively structure information so that it can be easily consumed. We spend a great deal of effort doing this for humans, however, given the amount of scientific information we are producing we should also spend some effort doing this for machines. In this talk, I review several semantic representations that are making it easier for machines to consume scholarly content. Moreover, I discuss how these representations can facilitate the decoupling of the journal and the ability to remix scholarly content.
Tech Session: W3C Open Annotation effort: Status and Use Cases
The Open Annotation Data Model specifies an interoperable framework for creating associations between related resources - Annotations - using a methodology that conforms to the Architecture of the World Wide Web. Open Annotations can easily be shared between platforms, with sufficient richness of expression to satisfy complex requirements while remaining simple enough to also allow for the most common use cases, such as attaching a piece of text to a single web resource.
In this presentation we will discuss the data model, motivated by scholarly communication use cases for annotation. These include open peer review, nano-publications, personal note-taking, teaching and learning systems, resource organization and many others. We will also discuss the current status of the work, and the next steps in the standardization process.
Tech Session: Naming on the Web: What scholars should want, and what they can have
It's alleged that when Zhou Enlai was asked what he thought of the French Revolution he replied: "It's too early to tell". The scholarly community could be forgiven for wishing they could say the same about the Web, but we don't have that luxury. The incentives for moving scholarship, _all_ scholarship, onto the Web are enormous, and the penalties for failing to do so are rapidly increasing---for the current generation of students, it is increasingly true that "if it's not on the Web, it doesn't exist". Yet how can responsible scholarship depend on such a manifestly uncertain technology? My colleague Michael Sperberg-McQueen once said, in his role as technical advisor to the Model Editions Partnership: "[W]hen I advise people on building systems that will last for the time spans needed for cultural heritage data, I will advise them to build on some system whose design story holds up for more than a minute and a half before an inconsistency is introduced." In this talk I'll try to separate the _necessary_ properties of _any_ web-scale naming system from the _contigent_ socio-technical realities of the Web as it is today. I'll close by attempting to make clear exactly what trade-offs confront us as we try to move scholarly discourse onto the Web in a responsible manner.
CoffeeMain Hall (Uni Mail)
Plenary 2: MetricsR380 (Uni Mail)
Metrics Session: An overview of scholarly impact metrics
The interest in developing scholarly impact metrics is frequently justified by the need to objectively prioritize scarce resources and to better manage scholarly productivity. However, the study of scholarly communication in general, including scholarly impact metrics, has significant relevance to a number of other scientific domains such as computational social science, social network analysis, web science, and complex systems. In this presentation I will provide an overview of established scholarly impact metrics, grounding each in their respective scientific traditions and backgrounds. Changes in scholarly communication patterns, including the move to online environments and the increasing use of social media, have recently prompted a Cambrian explosion of new impact metrics derived from new data sources. These metrics may reflect previously unexplored facets of scholarly communication and impact, and may thus yield a more complete picture of scholarly communication. In my presentation I will provide an overview of these new metrics, and identify the opportunities as well as challenges that they present.
Metrics Session: Discussions of scholarly articles online: who, why and where
The field of altmetrics is based on the online activity around articles. How much of this exists, where is it happening and who is involved? Altmetric.com has been tracking this data for publishers, funders and institutions since July 2011 and in this talk we'll make some general observations and highlight interesting trends.
Metrics Session: Assessing the transparency of peer review in (Open Access) journals
The peer review system is a standard control mechanism in scholarly publishing. Despite its centrality in science, peer-review typically takes place behind closed curtains. For both established and Open Access journals this obscurity may give room to substandard peer-reviews. I contend that transparency concerning the peer-review process at academic journals can be viewed as an indicator of the quality of the peer review. I present results of three studies of the validity and reliability of a straightforward online assessment of transparency of the peer review process at (OA) journals. The assessment can be readily used by different stakeholders (publishers, researchers, librarians, and funders) and entails a list of criteria for transparency concerning the peer review process (e.g., clarity on scope of the journal, rejection rates, decision makers, criteria used by reviewers, publication ethics). Results show good validity and sufficient reliability to use the tool to determine which journals meet common standards of transparent peer-review.
Registration Desk opens at 8:15Main Hall (Uni Mail)
Plenary 3: Data and Document SemanticsR380 (Uni Mail)
Semantic Session: Semantic indexing in PubMed
About 5,000 biomedical journals are indexed and included in the National Library of Medicine’s MEDLINE bibliographic database, available through PubMed. MEDLINE is used internationally to provide access to the world's biomedical journal literature. PubMed supports both text searches and searches based on the indexing of the articles in reference to the Medical Subject Headings (MeSH) thesaurus.
We will review several aspects of semantic indexing in MEDLINE, including traditional MeSH indexing performed manually by human indexers, indexing to UMLS concepts, and indexing of relations or facts (“nano-publications”). We will discuss automatic approaches to MeSH indexing, as well as indexing of specific entities, such as genes. This presentation also explores the relations between semantic indexing and the Semantic Web, as well as applications of semantic indexing, e.g., to literature-based discovery.
Semantic Session: Transformation of keyword indexed collections into semantic repositories
In the information retrieval context, resource collections are frequently classified using simple knowledge models such as thesauri. However, the limited semantics provided restricts their search and browsing capabilities. This work shows a process that improves these capabilities through the conversion of the selected knowledge model into a domain ontology. The process has been tested with the European Urban Knowledge Network and the Urbamet thesauri. Additionally, Urbamet model has been used to create an atlas of urban related resources with advanced search capabilities.
(Univ. of Zaragoza)
Semantic Session: Detecting knowledge-level claims in research articles
The research article genre has developed a specific structure for the effective communication of scientific results. The title, the abstract, more or less standardized section types, etc. are all structural elements in the service of facilitating comprehension and searchability. These elements are apparent for the reader through formatting, and and their markup makes it possible for search algorithms to take advantage of them in relevance ranking. In this presentation we propose another type of content element that can facilitate comprehension and search: knowledge-level claims. We define knowledge-level claims as discourse elements in articles that indicate the status of scientific propositions within the state of the art. Knowledge-level claims significantly contribute to comprehension, they are usually rhetorically salient, but traditionally they are not made prominent through formatting or markup. We show some applications where the automatic detection of knowledge-level claims has been used for enhancing both comprehension and search, and we indicate some further potential applications.
(Xerox Research Centre Europe)
Small Data, or: Bridging the Gap Between Smart and Dumb Research Repositories
Scientific research mostly consists of many tiny niches, with many thousands of small data sets: a ‘long tail’ effect. So we have a ‘Small Data’ problem: how do we connect vastly different experimental results, so that they can be used by other scientists? Currently, there are many large, topically agnostic repositories, requiring little metadata or informatics support, which serve an archival need but provide little opportunity for allowing overarching analytics. On the other end of the scale, highly usable topical repositories require painstaking manual curation, which does not scale. This talk will present a proposal on bridging the chasm between these two approaches, to enable systems for interoperable results reporting. After presenting a general overview of some pertinent developments I’ll focus on two use cases, in electrophysiology and geochemistry, where we will attempt to build a system that allows bridging the gap between such ‘big and dumb’ and ‘small and smart’ solutions.
MrsAnita de Waard
(Vice President Research Data Collaborations, Elsevier)
Coffee and PostersMain Hall (Uni Mail)
Posters sessionMain Hall (Uni Mail)
LunchRestaurant (Uni Mail)
Plenary 4: Research DataR380 (Uni Mail)
Res. Data Session: Research Data Policies: Seachange or Zeitgeist?
What will be the practical implications of research data policies for the everyday life of researchers or institutional and commercial service portfolios? The question is not new: the German Research Foundation DFG addressed data management as part of the recommendations for safeguarding good scientific practice in 1998 , the National Institute of Health NIH in the US presented a draft statement on sharing research data in 2002 and concepts like “Data Deluge”  celebrate a 10th anniversary. Recently, the National Science Foundation NSF in the US introduced a requirement for data management plans , UK’s Engineering and Physical Sciences Research Council EPSRC asks universities to deliver a data roadmap for implementation by 2015  and institutions such as the University of Edinburgh start publishing local policies .
At the same time, disciplinary practice varies significantly. Some life science journals require mandatory data deposits with an article publication. Data journals as well as data publishing platforms are appearing. In other disciplines even mentioning the term ‘data’ causes suspicion. The presentation will analyze selected research data policies in the light of examples of current research practice.
Res. Data Session: Interoperability of Research Data
The advent of the new paradigm of science largely based on data analysis and mining is having a relevant impact on Scholarly Communication. In order to document their work scholars are starting to publish not only research papers but also to make available the experimental datasets and the tools used for achieving such results. The availability of these products facilitates the reproducibility of the results and open the way towards a more wider re-usability of the data for other scientific purposes. However, In order to fully exploit these new possibilities appropriate level data interoperability between the publisher and the consumed of the data have to be assured. This presentation introduces major issues of data interoperability between a data publisher and a data consumer and it discusses how emerging data infrastructures can help in minimizing them.
Res. Data Session: Quality and curation of Research Data
Everyone who wants data wants high-quality data, and curation processes are designed to improve data quality. Since we are all agreed on these things there should be little to discuss. But in practice we use 'quality' to mean different things, and our curation processes emphasise some quality dimensions at the expense of others. This is not always beneficial to the research process.
I will discuss what research tells us about data quality and how this should informat curation practice. Finally I will speculate on how metadata on quality could ease reliable automated data integration and hence promote data reuse.
Res. Data Session: Working with large data sets
Whether large or small, data sets need to be managed. Scale however makes standard data operations more challenging, especially when the data sets expand beyond the capacities of a single data centre. Consequently replication, migration and archival require optimised, sometimes domain specific solutions. Tim will illustrate the HEP approach by describing how CERN collected 100 petabytes of research data and how it organises storage and access by researchers across the globe. He will run through the data reduction and analysis chains up to paper production, and finish with the challenges of archiving these large data sets.
CoffeeMain Hall (Uni Mail)
Breakout GroupsUni Mail
BG1 - Gold OA Infrastructure (Room 1140)Room 1140
Aim: to hear the perspectives of institutions, funders, platforms and publishers, in order to provide information on what infrastructure is likely to be needed for Gold OA.
Likely topics will include micropayments, certification, consortia payment and management models, discovery and tracking of OA material.
- Pierre Mounier (OpenEdition) - France
- Johannes Fournier (DFG) - Germany
- Martin Rasmussen (Copernicus) - Germany
- Paul Ayris (UCL) - UK
This breakout would focus on the use of annotation to help transform scholarly communications. Ideas include:
- Annotation as a form of peer review overlay mechanism
- Annotation of data to support eg discovery
- Annotation activity as an altmetric
MrHerbert Van de SOMPEL
(Los Alamos National Laboratory), MrPaolo CICCARESE
(Harvard), MrRobert SANDERSON
(Los Alamos National Laboratory)
BG3 - Altmetrics (Room R160)Room R160
Aim: To explore and discuss current developments in the context of alternative metrics. The session will focus on issues as methodology, data quality, stability and reliability.
Possible outcomes: Recommendations to improve data quality, acceptance and reliability of altmetrics.
BG4 - Open Access Policy Developments (Room R170)Room R170
Aim: to share and discuss recent policy developments, including the Executive Directive on public access in the U.S. and Horizon 2020 in Europe. The session will examine how policy developments align (and how they don’t), and what we can do to work together globally to achieve the adoption of OA policies.
- European speaker : Iryna Kuchma & Victoria Tsoukala
BG5 - How to make your university into a monograph publisher? (Room 1130)Room 1130
The traditional monograph market is becoming increasingly dysfunctional; failing to meet the needs of researchers in the arts and humanities. Price increases and library budget restrictions are limiting the distribution of newly published monographs while making it harder for authors (especially young, previously unpublished researchers) to get published.
New technology and business models allow us to consider alternatives to the traditional modes of scholarly book publishing. Increasingly academic institutions are finding both the need and the means by which they can take on book publishing activities directly.
To identify and share experiences of scholarly publication activities within academic institutions and libraries.
To consider how these models might be adopted, adapted and developed by others to provide innovative and viable OA book publishing solutions across HSS and STM scholarly, textbook and open course-ware environments.
This breakout session will look at some of the models that are being used to revolutionise monograph publishing. Participants will be encouraged to share their own experiences of book publishing activity at their institutions.
BG6 - Reusing Open Acces materials - a Wikimedia perspective (Room R150)Room R150
- Introduction to Wikimedia, its mission and how it is aligned with Open Access
- Using Wikipedia's popularity to share research
- Publishing to and from Wikipedia
- Reuse of Open Access materials on Wikimedia projects
- Discussion of current barriers to reuse, beyond licensing issues
To share and discuss examples of reuse of Open Access materials, to identify barriers to reuse and to highlight some approaches to overcoming them.
- Daniel Mietchen
- Lane Rasberry
- Samuel Klein
Possible outcomes :
A set of recommendations on how to publish in a way that facilitates reuse
The session will be planned at: http://en.wikipedia.org/wiki/Wikipedia:WikiProject_Open_Access/Wikimedia_at_OAI8 - and you can edit it.
Registration Desk opens at 8:15Main Hall (Uni Mail)
Plenary 5: Arts, Humanities, and Social SciencesR380 (Uni Mail)
Humanities Session: OA Research Monographs in HSS: Opportunities & Challenges
HSS scholars have been generally slower and, arguably, more resistant to OA publications than those in other disciplines. This presentation will look at some of the differences in publishing requirements between scientific and HSS scholars, alternative publishing models being developed to address these, and some of the opportunities OA publishing present for HSS disciplines. Specific attention will be given to monograph publication throughout the presentation.
(Open Book Publishers)
Humanities Session: The Humanities in and for the Digital Age
The spread of digital technologies has presented scholars in the humanities with some extraordinary opportunities, as well as a few challenges, not least for their modes of communicating with one another. This talk will explore some of the changes taking place in the humanities today and their implications for scholars and their institutions. How are humanists' ways of thinking about scholarly communication changing as we do more and more of our work on digital platforms?
(Modern Language Association)
Empowering Development: Why Open is Right for Development
The world we live in today is more connected than ever before. Technology is the great enabler, with citizens around the world accessing information on their mobile phones, smart phones, tablets, and PCs. The World Bank has responded to this change by adopting an "Open Agenda," a sea change in how it makes available its data, information, and knowledge. In the last three years, the World Bank has implemented an Access to Information policy, an Open Access Policy, Creative Commons licensing, and has "opened" statistical databases and other datasets that were previously not available to the public or available via subscription only. At the World Bank we believe that openness drives transparency, which leads to greater accountability, which—in turn—leads to better development results. This presentation reviews the World Bank’s Open Agenda and how it is driving transparency, accountability, and—most important—helping to deliver better development results.
(World Bank Publications)