Next generation of CMS data replication system

Description

The CMS experiment at the LHC proton-proton collider developed PhEDEx as a reliable and scalable dataset replication system. Designed in 2004, and written mainly in Perl, it is still used today for managing multi-PB transfer loads per week, across a complex topology of dozen of Grid-compliant computing centres.

Its architecture, instead of having a central brain making global decisions on all CMS replica allocation, has a data management layer composed of a set of loosely coupled and stateless software agents - each managing highly specific parts of replication operations at each site in the distribution network - which communicate asynchronously through a central blackboard architecture. The system is resilient and robust against a large variety of possible failure reasons, and it has been designed by assuming a transfer will fail (thus implementing fail-over tactics) and being completely agnostic on the lower-level file transfer mechanism (thus focusing on full dataset management and detailed transfer bookkeeping). Collision data and derived data collected at LHC that allowed to achieve the Higgs boson discovery by ATLAS and CMS experiments at LHC were transferred in the CMS worldwide domain using this toolkit.

Tasks

The proposed project aims at reviewing the basic PhEDEX functionality in a context of up-coming exa-byte HL-HLC era and exploring its implementation in Go programming language.

The motivation for the effort is many fold:

We propose the following steps:

Expected results

Working implementation of the system and fully tested in distributed environment.

Requirements

Solid knowledge of Go-langauge, some knowledge of perl (be able to read legacy code and translate it into Go), some knowledge of databases (Sqlite, MySQL, ORACLE), git

Mentors

Corresponding Project

Participating Organizations