29–31 Jan 2018
AGH Computer Science Building D-17
Europe/Zurich timezone

Contribution List

57 out of 57 displayed
Export to PDF
  1. Jakub Moscicki (CERN)
    29/01/2018, 10:00
    Welcome
  2. Prof. Piotr Faliszewski (AGH Kraków)
    29/01/2018, 10:30
  3. Mr Grzegorz Bogdał (AGH University of Science and Technology), Mr Piotr Gawryś (AGH University of Science and Technology), Mr Paweł Nowak (AGH University of Science and Technology), Mr Łukasz Plewnia (AGH University of Science and Technology), Leszek Grzanka (AGH University of Science and Technology (PL)), Maciej Malawski (AGH University of Science and Technology)
    29/01/2018, 11:30

    Data analysis in High Energy Physics experiments requires processing of large amounts of data. As the main objective is to find interesting events from among those recorded by detectors, the typical operations involve data filtering by applying cuts and producing of histograms. The typical offline data analysis scenario for TOTEM experiment at LHC, CERN involves processing of 100s of ROOT...

    Go to contribution page
  4. Peter Szegedi (GÉANT)
    29/01/2018, 12:00

    The ”Up to University” (Up2U) Project – coordinated by GÉANT – is to bridge the gaps between secondary school and university by providing European schools with a Next Generation Digital Learning Environment that helps students developing the knowledge, skills and attitudes they need to succeed at university. This student-centered digital learning environment forms the Up2U ecosystem, which...

    Go to contribution page
  5. Michael D'Silva (AARNet)
    29/01/2018, 14:00

    Rocket is the first attempt at handling one of the particular problems that other tools have failed to solve. This presentation will demonstrate AARNet’s experiences and tools used high-speed data transfers of different kinds of research data.

    The research community in Australia is spread far and wide geographically, resulting in some cases to be physically far from one of our three...

    Go to contribution page
  6. Hugo Gonzalez Labrador (CERN)
    29/01/2018, 14:20

    Microservices are an approach to distributed systems that promote the use of finely grained services with their own lifecycles, which collaborate. The use of microservices facilitates embracing new technologies and architectural patterns. Sync and share providers could increase the modularity and facilitating the exchange of components and best practices adopting the use of microservices.
    In...

    Go to contribution page
  7. Enrico Bocchi (CERN)
    29/01/2018, 14:40

    Container technologies are rapidly becoming the preferred way to distribute, deploy, and run services by developers and system administrators. They provide the means to create a light-weight virtualization environment, i.e., a container, which is cheap to create, manage, and destroy, requires a negligible amount of time to set-up, and provides performance equatable with the one of the...

    Go to contribution page
  8. Mr Piotr Mrówczyński (ownCloud GmbH, CERN, KTH Stockholm, TU Berlin)
    29/01/2018, 15:00
    Synchronization/Sharing Technology & Research
    Presentation

    There is growing interest for self-hosted, scalable, fully controlled and secure file sync and share solutions among enterprises. The ownCloud has found its share as free-to-use, open-source solution, which can scale on-premise from a single commodity class server to a cluster of enterprise class machines, and serve from one to thousands of users and PB of data. Over the years, it has grown a...

    Go to contribution page
  9. Dr Hurng-Chun Lee (Donders Institute, Radboud University)
    29/01/2018, 15:40

    Research Data Management (RDM) serves to improve the efficiency and transparency in the scientific process and to fullfil internal and external requirements. Three important goals of RDM are:

    • long-term data preservation,
    • scientific-process documentation,
    • data publication.

    One of the tasks in RDM is to define a workflow for data as part of the research process and data lifecycle. RDM...

    Go to contribution page
  10. Stefano Cozzini
    29/01/2018, 16:05

    We present our recent work [1] where we applied state of the art deep learning techniques for image recognition, automatic categorization, and labeling of nanoscience images obtained by scanning electron microscope (SEM). Roughly 20,000 SEM images were manually classified into 10 categories to form a labeled training set, which can be used as a reference set for future applications of deep...

    Go to contribution page
  11. Mr Armin Burger (European Commission - Joint Research Centre), Mr Paul Hasenohr (European Commission - Joint Research Centre)
    29/01/2018, 16:30

    The Joint Research Centre (JRC) of the European Commission has set up the JRC Earth Observation Data and Processing Platform (JEODPP) as a pilot infrastructure to enable the knowledge production Units to process and analyze big geospatial data in support to EU policy needs. The very heterogeneous data domains and analysis workflows of the various JRC projects require a flexible set-up of the...

    Go to contribution page
  12. Dr. Tilo Steiger (ETH Zuerich)
    29/01/2018, 17:30
  13. Peter van der Reest (DESY)
    29/01/2018, 17:50
    Presentation

    Over two years Data-Cloud team at DESY provides a reliable ownCloud instance for a selected set of users. While service is still officially in a pilot phase, it’s has the same support and priority level as any other production services provided by the IT group. However, before removing it’s “beta” status some extra actions have to be taken: the instance must be fault tolerant and allow...

    Go to contribution page
  14. Gregory Vernon (SWITCH edu-ID)
    29/01/2018, 18:05
    Presentation

    In the summer of 2017, I inheirited SWITCHdrive, SWITCH's ownCloud-based filesharing system. SWITCHdrive is a fairly complex service including a set of docker based microservices. I will describe the continuing story of our experiences with running such an environment. We had some interesting developments in tuning our MariadDB/Galera database infrastructure, and we have also greatly...

    Go to contribution page
  15. Ron Trompert
    29/01/2018, 18:20
    Presentation

    The past year we were able add a number of extra features to the SURFdrive service in order to make it more attractive to users and institutes and there is more to come. Another thing that we have observed that several institutes and research groups have a need for a SURFdrive in a version more tailored to their needs. SURFdrive is fine as it is but it is a one size fits all solution....

    Go to contribution page
  16. Hugo Gonzalez Labrador (CERN)
    29/01/2018, 18:35
    Presentation

    CERNBox is a cloud synchronisation service for end-users: it allows synchronising and sharing files on all major desktop and mobile platforms (Linux, Windows, MacOSX, Android, iOS) aiming to provide universal access and offline availability to any data stored in the CERN EOS infrastructure.
    With 12000 users registered in the system, CERNBox has responded to the high demand in our diverse...

    Go to contribution page
  17. Andrey Erokhin (St Petersburg State University (RU))
    29/01/2018, 18:50
    Site reports
    Presentation

    The report focuses on the deployment of the CERN SWAN-like environment on top of existing EOS storage. Our setup consists of a local cluster with Kubernetes to run JupyterHub and single-user Jupyter notebooks plus a dedicated server with CERNBox. The current setup is tested by our colleagues in the Laboratory of ultra-high energy physics of the St. Petersburg State University, but there are...

    Go to contribution page
  18. Guido Aben (AARNet)
    30/01/2018, 09:00

    What is the DLCF?​

    The Data LifeCycle Framework (DLCF) is an Australian nationwide strategy to connect research resources and activities; predominantly those funded by national eInfrastructure funding.​​

    The goal of the DLCF is to smooth over the complexity faced by ordinary researchers, when they have to piece together their own digital workflow from all the bits and pieces made available...

    Go to contribution page
  19. Tibor Simko (CERN), Diego Rodriguez Rodriguez (Universidad de Oviedo (ES))
    30/01/2018, 09:20

    The revalidation, reuse and reinterpretation of data analyses requires having access to the original virtual environments, datasets, software, instructions and workflow steps which were used by the researcher to produce the original scientific results in the first place. The CERN Analysis Preservation pilot project is developing a set of tools that assist the particle...

    Go to contribution page
  20. Mr Vladislav Makarenko (Max-Planck Digital Library)
    30/01/2018, 09:40

    Keeper is a central service for scientists of the Max Planck Society and their project partners for storing and archiving all relevant data of scientific projects. Keeper facilitates the storage and distribution of project data among the project members during or after a particular project phase and seamlessly integrates into the everyday work of scientists. The main goal of the Keeper service...

    Go to contribution page
  21. Diogo Castro (CERN)
    30/01/2018, 10:00

    SWAN (Service for Web-based ANalysis) is a CERN service that allows users to perform interactive data analysis in the cloud, in a "software as a service" model. It is built upon the widely-used Jupyter notebooks, allowing users to write - and run - their data analysis using only a web browser. By connecting to SWAN, users have immediate access to storage, software and computing resources that...

    Go to contribution page
  22. Mr Jason Coposky (iRODS Consortium)
    30/01/2018, 10:50

    iRODS is Open Source Data Management that can be deployed seamlessly onto your existing infrastructure, creating a unified namespace, and a metadata catalog of all the data objects, storage, and users on your system. iRODS allows access to distributed storage assets under the unified namespace and frees organizations from getting locked into single-vendor storage solutions. iRODS can...

    Go to contribution page
  23. Mr Stefano Baldi (Cynny Space)
    30/01/2018, 11:10

    The purpose of the presentation we’ll propose during the CS3 conference in Krakow is to highlight the technological features of Cynny Space’s cloud object storage solution and the results of performance and usability for a sync & share use case.

    1) Software specifically designed for ARM® architecture

    The object storage solution is specifically designed and developed on storage nodes composed...

    Go to contribution page
  24. Peter van der Reest (DESY)
    30/01/2018, 11:30

    Over two years Data-Cloud team at DESY uses dCache as a backend storage for the ownCloud instance used in a production. As being a highly scalable storage system, dCache is widely used by many sites to store hundreds of petabytes of scientific data. However, the cloud-backend usage scenarios have added new requirements, like high availability and downtime less updates any software or hardware...

    Go to contribution page
  25. Luca Mascetti (CERN)
    30/01/2018, 11:50

    EOS, the high-performance CERN IT distributed storage for High-Energy Physics provides now more than 250PB of raw disks and supports several work-flows from LHC data-taking and reconstruction to physics analysis.
    The software is developed at CERN since 2010, is available under GPL license and it is also used in several external institutes and organisations.

    EOS is the key component behind the...

    Go to contribution page
  26. Lukasz Dutka
    30/01/2018, 12:10

    Onedata is a complete high-performance storage solution that unifies data access across globally distributed environments and multiple types of underlying storages, such as NFS, Lustre, GPFS, Amazon S3, CEPH, as well as other POSIX-compliant file systems. It allows users to share, collaborate and perform computations on their data.

    Globally Onedata comprises of: Onezones, distributed...

    Go to contribution page
  27. Frank Karlitschek (Nextcloud)
    30/01/2018, 14:00

    We are heading into a world were the files of most users are hosted by 4 big companies. This is the case for most home users, companies but also education and research institutions. If we want to keep our sovereignty over our data, protect our privacy and prevent vendor lock-in then we need open source self hosted and federated alternatives.
    A new challenge is the increasing blending of...

    Go to contribution page
  28. David Jericho (AARNet)
    30/01/2018, 14:10

    “Sync and Share is Dead. Long Live Sync and Share." discusses the increasing disinterest users have in simple file storage, Simple storage is a commodity service, with Google, DropBox, and other big players who can legitimately resolve concerns about data centre security, legal control, administration and audit, and standards compliance. The competitive advantage for any given data storage...

    Go to contribution page
  29. Jakub Moscicki (CERN)
    30/01/2018, 14:20

    Over the last years we have witnessed a global transformation of the IT industry with the advent of commercial (“public”) cloud services on a massive scale. Global Internet industry firms such as Amazon, Google, Microsoft massively invest in networking infrastructure and data-centers around the globe to provide ubiquitous cloud service platforms for any kind of service imaginable: storage,...

    Go to contribution page
  30. 30/01/2018, 14:30
  31. Mr Charles du Jeu (Pydio)
    30/01/2018, 15:30

    On-premise EFSS is now an established market, and open source solutions have been key-players in the last couple of years. For many enterprises or labs, the need for privacy and handling large volumes of data are show-stoppers for using saas-based solutions. Still, for these users, the experience speaks by itself: even with good software, it is hard to deploy a scalable and reliable system...

    Go to contribution page
  32. Mr Jonathan Xu (Seafile Ltd.)
    30/01/2018, 15:35

    Seafile is an open source file sync and share solution. Thanks to its high performance, scalability and reliability, it has been successfully used by many organizations in Europe, North America and China.

    In this presentation, we'll provide a review of Seafile's development in 2017, and what we plan to accomplish in the future. We'll also present a site report from China with heavy usage,...

    Go to contribution page
  33. Frank Karlitschek (Nextcloud)
    30/01/2018, 15:55

    This talks covers the current state and functionality of Nextcloud. Especially the new and innovativ features of Nextcloud 12 and 13 are discussed and presented in detail. Examples are End 2 End encryption, collaboration and communication features and security and performance improvements. The second part of the talk presents the roadmap and strategic direction of Nextcloud for the coming...

    Go to contribution page
  34. Christian Schmitz (ownCloud Inc), Felix Boehm (Owncloud Inc)
    30/01/2018, 16:15

    ownCloud has been an excitingly successful service in the EFSS space since its breakthrough in 2013. Since customers deploy the solution in vastly different environments as public, private or hybrid cloud and utilizing different infrastructure components and identity providers, operational experience showed challenges with the previous design decisions.

    This talk will reflect on the past...

    Go to contribution page
  35. Christian Sprajc (PowerFolder)
    30/01/2018, 16:35

    Blockchain is currently one of the hot topics. Developed as part of the cryptocurrency Bitcoin as a web-based, decentralized, public and most important all secure accounting system, this database principle could not only revolutionize the worldwide financial economy in the future; Blockchain is already an topic in electromobility, health care or supply-chain-management - just to name a...

    Go to contribution page
  36. Mr Lorenzo Posani (Cubbit), Mr Marco Moschettini (Cubbit)
    30/01/2018, 16:55
    File Sync&Share Products for Home, Lab and Enterprise
    Presentation

    Cubbit is a hybrid cloud infrastructure comprised of a network of p2p-interacting IoT devices (swarm) coordinated by a central optimization server. The storage architecture is designed to reverse the traditional paradigm of cloud storage from "one data center to rule them all" to "a small device in everyone’s house".

    Any IoT device that supports an Unix-based OS can join the swarm and...

    Go to contribution page
  37. Jörn Dreyer (ownCloud GmbH), Thomas Müller
    30/01/2018, 17:40

    Over the past year we dropped the requirement that ownCloud should run on every PHP platform. This allows us to research architectural changes, like push notifications, microservices, dockerized deployments, HSM integration and storing metadata in POSIX or object storages. On the client side we are exploring E2EE, virtual filesystems and delta sync. Together with feedback from our community...

    Go to contribution page
  38. Mr Enno Gröper (Humboldt-Universität zu Berlin)
    30/01/2018, 18:00

    The talk will introduce the main concepts of Shibboleth, advantages and disadvantages and show the integration of Shibboleth with a Sync and Share service (webapp with own session handling, not designed for using the Shibboleth session as webapp session) with Seafile as an example.
    Furthermore it will discuss the problems of Shibboleth federations and possible mitigations.
    A special focus will...

    Go to contribution page
  39. Björn Schießle
    30/01/2018, 18:20

    The typical Nextcloud setup for large installations includes a storage and a database cluster attached to multiple application servers behind a load balancer. This allows organisations to scale Nextcloud for thousands of users. But at some point the shared components like the storage, database and load balancer become a expensive bottleneck. Therefore Nextcloud introduced "Global Scale", a new...

    Go to contribution page
  40. Gregory Vernon (SWITCH edu-ID)
    30/01/2018, 18:40

    Managing the database where you store your application data is always an
    interesting challenge. As the scale of your service grows, so does the
    challenge of keeping a healthy database service. However with just a few tools
    and techniques it is possible to implement some serious performance
    improvements with just a little bit of effort. Using the performance tools
    included with MariaDB, at...

    Go to contribution page
  41. Peter Szegedi (GÉANT)
    31/01/2018, 09:00

    As part of the collaboration effort between GÉANT, CERN and ownCloud, in January 2015, an idea (aka. OpenCloudMesh, OCM) has been initiated to interconnect the individual on-premises private cloud domains at the server side in order to provide federated sharing and syncing functionality between the different administrative domains. The federated sharing protocol, in the first place, can be...

    Go to contribution page
  42. Jan Wiegmann
    31/01/2018, 09:20

    Open Cloud Mesh (OCM) is a joint international initiative under the umbrella of the GÉANT Association that is built on the open Federated Cloud Sharing application programming interface (API). Taking Universal File Access beyond the borders of individual clouds and into a globally interconnected mesh of research clouds without sacrificing any of the advantages in privacy, control and security...

    Go to contribution page
  43. Diogo Castro (CERN)
    31/01/2018, 09:40

    This presentation gives details and demonstrates the new SWAN sharing interface. See also: "SWAN: Service for Web-based Analysis" in "Cloud Infrastructure&Software Stacks for Data Science" session.

    Go to contribution page
  44. Mr Patrick Maier, Mr Tom Needham
    31/01/2018, 09:50

    Current sharing in ownCloud does not allow seamless access of shared data. Media disruptions and inefficient communication methods reduce productivity for teams through a lack of information. Sharing 3 introduces a new bi-directional request-accept flow for streamlining collaboration within the ownCloud platform. This gives users further control over their data, allows them to request access...

    Go to contribution page
  45. Peter Szegedi (GÉANT)
    31/01/2018, 10:05

    This panel discussion session will be focusing on the actual use cases that can drive the adoption and further development of the OCM protocol. Panellists will be requested to provide their views and vision for the future with regard to interoperability between private cloud domains.

    Go to contribution page
  46. Giuseppe Lo Presti (CERN)
    31/01/2018, 10:40

    In this contribution, the evolution of CERNBox as a collaborative
    platform is presented.

    Powered by EOS and ownCloud, CERNBox is now the reference storage
    solution for the CERN user community, with an ever-growing user base
    that is now beyond 12K users.

    While offline sync is instrumental for such a widespread usage, online
    applications are becoming more and more important for the...

    Go to contribution page
  47. Jan Holesovsky (Collabora Productivity)
    31/01/2018, 11:00

    Come and hear about what is Collabora Online and how it integrates into many File Sync&Share Products to create a powerful, secure, real-time document editing experience. Hear about the improvements over the last year, catch a glimpse of where we are going next, and hear how you can get it integrated into your product - if you haven't integrated it yet.

    Go to contribution page
  48. Olga Golubeva (Ascensio System SIA)
    31/01/2018, 11:20
    Sharing and Collaborative Platforms
    Presentation

    Global academic community tends to show more interest in using cloud technologies for scientific data processing, which is determined by the need for quick joint access to the data.
    This presentation will deal with the question of the convenient and effective cloud editing of documents as the main form of storing and exchanging the information.
    ONLYOFFICE, a project by Latvian software...

    Go to contribution page
  49. Dr Jean-Thomas Acquaviva (DDN Storage)
    31/01/2018, 12:00

    IME I/O Acceleration layer is one of the latest efforts of DDN in order to satisfy the never ending needs for performance of the HPC community. We propose to discuss some of the latest advancements of IME product in respect to the larger evolution of Software Defined Storage has it is observed outside of the HPC market.

    The arrival of the Flash has pushed existing HPC file systems to their...

    Go to contribution page
  50. Maciej Brzezniak (PSNC Poznan Poland)
    31/01/2018, 12:20

    Handling 100s of Terabytes of data at the speed of 10s of GB/s is nothing new in HPC. However, high performance and large capacity of the storage systems rarely go together with their ease of use. HPC storage systems are specifically difficult to access from outside the HPC cluster. While researchers and engineers tolerate the fact that they need to use rigid tools and applications such as...

    Go to contribution page
  51. Thomas Müller
    31/01/2018, 12:40

    To allow better scalability of ownCloud in large installations we spent some time to leverage the ownCloud integration with s3 based objectstores like Ceph and Scality.
    At the ownCloud Conference 2017 we have been presenting the vision where to go.
    At CS3 we will present the results!

    Go to contribution page
  52. Ricardo Makino (RNP), Rodrigo Azevedo (RNP)
    31/01/2018, 14:00

    The National Education and Research Network (RNP) is an organization that plans, designs, implements and operates the national network infrastructure under contract with the Ministry of Science, Technology, Innovation and Communications (MCTIC). A current government program includes five ministries - MCTI, Education (MEC), Culture (MinC), Health (MS) and Defense (MD), and annually define the...

    Go to contribution page
  53. Gregory Vernon (SWITCH edu-ID)
    31/01/2018, 14:20

    SWITCH has been running cloud-based filesharing services since 2012, starting with an experiment where we hosted FileSender in the Amazon Cloud. After this experience, we decided to build a cloud service for ourselves, SWITCHengines, which runs upon an OpenStack Infrastructure. The challenge with our SWITCHengines infrastructure and filesharing is the Ceph storage that we use for our user...

    Go to contribution page
  54. Mr Matthias Wobben (Nextcloud GmbH)
    31/01/2018, 14:40

    Nextcloud can be scaled from very small to very big installations. This talk gives an insiders look on how to deploy, run and scale Nextcloud in different scenarios. Discussed will be a very big installation in the research space, an installation in a global enterprise and the implementation of Nextcloud at one of the largest service providers in the world. The different infrastructural...

    Go to contribution page
  55. Crystal Chua (AARNet)
    31/01/2018, 15:00

    This talk covers a journey through fuzz-testing CERN's EOS file system with AFL, from compiling EOS with afl-gcc/afl-g++, to learning to use AFL, and finally, making sense of the results obtained.
    Fuzzing is a software testing process that aims to find bugs, and subsequently potential security vulnerabilities, by attempting to trigger unexpected behaviour with random inputs. It is particularly...

    Go to contribution page
  56. 31/01/2018, 17:00
  57. D Pennings (360 ICT), Martijn Kint

    We started looking at the site reports of CS3 with the goal of designing a large OwnCloud/Nextcloud solution. We learned that the main product used for sync and share is Nextcloud/Owncloud and looking at that large user base implementations, saw that most are implementations are large monolithic installs. The site reports also showed that these large installs have some weaknesses like scaling...

    Go to contribution page