Conveners
Software Engineering, Data Stores and Databases
- David Lange (Lawrence Livermore Nat. Laboratory (US))
Software Engineering, Data Stores and Databases
- Simone Campana (CERN)
Software Engineering, Data Stores and Databases
- Benedikt Hegner (CERN)
Software Engineering, Data Stores and Databases
- Simone Campana (CERN)
Software Engineering, Data Stores and Databases
- David Lange (Lawrence Livermore Nat. Laboratory (US))
Software Engineering, Data Stores and Databases
- Benedikt Hegner (CERN)
Vasil Georgiev Vasilev
(CERN)
5/21/12, 1:30โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
Cling (http://cern.ch/cling) is a C++ interpreter, built on top of clang (http://clang.llvm.org) and LLVM (http://llvm.org). Like its predecessor CINT, cling offers an interactive, terminal-like prompt. It enables exploratory programming with rapid edit / run cycles.
The ROOT team has more than 15 years of experience with C++ interpreters, and this has been fully exploited in the design of...
Roberto Agostino Vitillo
(LBNL)
5/21/12, 1:55โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
Modern superscalar, out-of-order microprocessors dominate large scale server computing. Monitoring their activity, during program execution, has become complicated due to the complexity of the microarchitectures and their IO interactions. Recent processors have thousands of performance monitoring events. These are required to actually provide coverage for all of the complex interactions and...
Frederik Beaujean
(Max Planck Institute for Physics)
5/21/12, 2:20โฏPM
Software Engineering, Data Stores and Databases (track 5)
Poster
The Bayesian Analysis Toolkit (BAT) is a C++ library designed to analyze data through
the application of Bayes' theorem.
For parameter inference, it is necessary to draw samples from the posterior distribution
within the given statistical model. At its core, BAT uses an adaptive Markov Chain Monte Carlo (MCMC) algorithm.
As an example of a challenging task, we consider the analysis of...
Maria Alandes Pradillo
(CERN)
5/21/12, 2:45โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The EMI Quality Model has been created to define, and later review, the EMI (European Middleware Initiative) software product and process quality. A quality model is based on a set of software quality metrics and helps to set clear and measurable quality goals for software products and processes. The EMI Quality Model follows the ISO/IEC 9126 Software Engineering โ Product Quality to identify...
Mrs
Jianlin Zhu
(Huazhong Normal University (CN))
5/21/12, 3:10โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The Data-Acquisition System designed by ALICE , which is the experiment dedicated to the study of strongly interacting matter and the quark-gluon plasma at the CERN LHC(Large Hadron Collider), handles the data flow from the sub-detector electronics to the archiving on tape. The software framework of the ALICE data-acquisition system is called DATE (ALICE Data Acquisition and Test Environment)...
Dr
Andreas Pfeiffer
(CERN)
5/21/12, 4:35โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The CMS experiment is made of many detectors which in total sum up to more than 75 million channels. The online database stores the configuration data used to configure the various parts of the detector and bring it in all possible running states. The database also stores the conditions data, detector monitoring parameters of all channels (temperatures, voltages), detector quality information,...
Dr
David Malon
(Argonne National Laboratory (US))
5/21/12, 5:00โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The volume and diversity of metadata in an experiment of the size and scope of ATLAS is considerable. Even the definition of metadata may seem context-dependent: data that are primary for one purpose may be metadata for another. Trigger information and data from the Large Hadron Collider itself provide cases in point, but examples abound.
Metadata about logical or physics constructs, such...
Dave Dykstra
(Fermi National Accelerator Lab. (US))
5/21/12, 5:25โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
Non-relational "NoSQL" databases such as Cassandra and CouchDB are best known for their ability to scale to large numbers of clients spread over a wide area. The Frontier distributed database caching system, used in production by the Large Hadron Collider CMS and ATLAS detector projects, is based on traditional SQL databases but also has the same high scalability and wide-area distributability...
Mario Lassnig
(CERN)
5/21/12, 5:50โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The Distributed Data Management System DQ2 is responsible for the global management of petabytes of ATLAS physics data. DQ2 has a critical dependency on Relational Database Management Systems (RDBMS), like Oracle, as RDBMS are well suited to enforce data integrity in online transaction processing application. Despite these advantages, concerns have been raised recently on the scalability of...
Marco Clemencic
(CERN)
5/22/12, 1:30โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The LHCb experiment has been using the CMT build and configuration tool for its software since the first versions, mainly because of its multi-platform build support and its powerful configuration management functionality. Still, CMT has some limitations in terms of build performance and the increased complexity added to the tool to cope with new use cases added latterly. Therefore, we have...
Andrew Hanushevsky
(STANFORD LINEAR ACCELERATOR CENTER),
Wei Yang
(SLAC National Accelerator Laboratory (US))
5/22/12, 1:55โฏPM
Software Engineering, Data Stores and Databases (track 5)
Poster
For more than a year, the ATLAS Western Tier 2 (WT2) at SLAC National Accelerator has been successfully operating a two tiered storage system based on Xrootd's flexible cross-cluster data placement framework, the File Residency Manager. The architecture allows WT2 to provide both, high performance storage at the higher tier to ATLAS analysis jobs, as well as large, low cost disk capacity at...
Jakob Blomer
(Ludwig-Maximilians-Univ. Muenchen (DE))
5/22/12, 2:20โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The CernVM File System (CernVM-FS) is a read-only file system used to access HEP experiment software and conditions data. Files and directories are hosted on standard web servers and mounted in a universal namespace. File data and meta-data are downloaded on demand and locally cached. CernVM-FS has been originally developed to decouple the experiment software from virtual machine hard disk...
Alexander Mazurov
(Universita di Ferrara (IT))
5/22/12, 2:45โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The LHCb software is based on the Gaudi framework, on top of which are built several large and complex software applications. The LHCb experiment is now in the active phase of collecting and analyzing data and significant performance problems arise in the Gaudi based software beginning from High Level Trigger (HLT) programs and ending with data analysis frameworks (DaVinci). Itโs not easy to...
Christophe Haen
(Univ. Blaise Pascal Clermont-Fe. II (FR))
5/22/12, 3:10โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The LHCb online system relies on a large and heterogeneous IT infrastructure made from thousands of servers on which many different applications are running. They run a great variety of ย tasks : critical ones such as data taking and secondary ones like web servers. The administration of such a system and making sure it is working properly represents a very important workload for the ย small...
Mr
Tigran Mkrtchyan
(DESY/dCache.ORG)
5/22/12, 4:35โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
dCache is a high performance scalable storage system widely used by HEP community. In addition to set of home grown protocols we also provide industry standard access mechanisms like WebDAV and NFSv4.1. This support places dCache as a direct competitor to commercial solutions. Nevertheless conforming to a protocol is not enough; our implementations must perform comparably or even better than...
Raffaello Trentadue
(Universita e INFN (IT))
5/22/12, 5:00โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The LCG Persistency Framework consists of three software packages (POOL, CORAL and COOL) that address the data access requirements of the LHC experiments in several different areas. The project is the result of the collaboration between the CERN IT Department and the three experiments (ATLAS, CMS and LHCb) that are using some or all of the Persistency Framework components to access their data....
Alastair Dewhurst
(STFC - Science & Technology Facilities Council (GB))
5/22/12, 5:25โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The ATLAS experiment deployed Frontier technology world-wide during the the initial year of LHC collision data taking to enable user analysis jobs running on the World-wide LHC Computing Grid to access database resident data. Since that time, the deployment model has evolved to optimize resources, improve performance, and streamline maintenance of Frontier and related infrastructure.
In this...
Giacomo Govi
(Fermi National Accelerator Lab. (US))
5/22/12, 5:50โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
Data management for a wide category of non-event data plays a critical role in the operation of the CMS experiment. The processing chain (data taking-reconstruction-analysis) relies in the prompt availability of specific, time dependent data describing the state of the various detectors and their calibration parameters, which are treated separately from event data. The Condition Database...
Sven Kreiss
(New York University (US))
5/24/12, 1:30โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
RooStats is a project providing advanced statistical tools required for the analysis of
LHC data, with emphasis on discoveries, confidence intervals, and combined measurements in the both the Bayesian and Frequentist approaches. The tools are built on top of the RooFit data modeling language and core ROOT mathematics libraries and persistence technology.
These tools have been developed in...
Marek Gayer
(CERN)
5/24/12, 1:55โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
We present our effort for the creation of a new software library of geometrical primitives, which are used for solid modelling in Monte Carlo detector simulations. We plan to replace and unify current geometrical primitive classes in the CERN software projects Geant4 and ROOT with this library. Each solid is represented by a C++ class with methods suited for measuring distances of particles...
Bertrand Bellenot
(CERN)
5/24/12, 2:20โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
A JavaScript version of the ROOT I/O subsystem is being developed, in order to be able to browse (inspect) ROOT files in a platform independent way. This allows the content of ROOT files to be displayed in most web browsers, without having to install ROOT or any other software on the server or on the client. This gives a direct access to ROOT files from new (e.g. portable) devices in a light...
David Tuckett
(CERN)
5/24/12, 2:45โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
Improvements in web browser performance and web standards compliance, as well as the availability of comprehensive JavaScript libraries, provides an opportunity to develop functionally rich yet intuitive web applications that allow users to access, render and analyse data in novel ways. However, the development of such large-scale JavaScript web applications presents new challenges, in...
Durga Rajaram
(IIT, Chicago)
5/24/12, 3:10โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The Muon Ionization Cooling Experiment (MICE) has developed the MICE Analysis User Software (MAUS) to simulate and analyse experimental data. It serves as the primary codebase for the experiment, providing for online data quality checks and offline batch simulation and reconstruction. The code is structured in a Map-Reduce framework to allow parallelization whether on a personal machine or in...
Andrzej Nowak
(CERN openlab)
5/24/12, 4:35โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
As the mainstream computing world has shifted from multi-core to many-core platforms, the situation for software developers has changed as well. With the numerous hardware and software options available, choices balancing programmability and performance are becoming a significant challenge. The expanding multiplicative dimensions of performance offer a growing number of possibilities that need...
Mr
Thomas Hauth
(KIT - Karlsruhe Institute of Technology (DE))
5/24/12, 5:00โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The processing of data acquired by the CMS detector at LHC is carried out with an object-oriented C++ software framework: CMSSW. With the increasing luminosity delivered by the LHC, the treatment of recorded data requires extraordinary large computing resources, also in terms of CPU usage. A possible solution to cope with this task is the exploitation of the features offered by the latest...
Stefan Lohn
(Universitaet Bonn (DE))
5/24/12, 5:25โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
Chip multiprocessors are going to support massive parallelism to provide
further processing capacities by adding more and more physical and logical
cores. Unfortunately the growing number of cores come along with slower
advances in speed and size of the main memory, the cache hierarchy, the front-
side bus or processor interconnections. Parallelism can only result in
performance...
Wim Lavrijsen
(Lawrence Berkeley National Lab. (US))
5/24/12, 5:50โฏPM
Software Engineering, Data Stores and Databases (track 5)
Parallel
The Python programming language allows objects and classes to respond dynamically to the execution environment. Most of this, however, is made possible through language hooks which by definition can not be optimized and thus tend to be slow. The PyPy implementation of Python includes a tracing just in time compiler (JIT), which allows similar dynamic responses but at the interpreter-, rather...