|Title||Project description||Project duration||Contact for further details||Status|
|Using data analytics for WLCG data transfer optimization||
The overall success of LHC data processing depends heavily on the stable, reliable and fast data distribution performed by the WLCG File Transfer Service (FTS). FTS transfers around 15 PB of data each month representing millions of files per day. The efficient functioning of this service is crucial for successful ...
|The potential of HTTP proxy caches for LHC computing||
Managing storage is one of the major contributors to operational costs on the LHC’s grid infrastructure (WLCG). The task is to design and prototype an HTTP proxy cache system, built using standard components, intended to allow pure unmanaged cache storage at a grid site or to accelerate data access in cloud ...
|Distributed storage systems for big data||
The group maintains a framework called dmlite which is used to integrate various types of storage with different protocol frontends. It is the basis of a number of the group’s products such as the Disk Pool Manager (DPM), a grid storage system which holds over 50PB of storage in the ...
|3 to 9 months depending on the selected firstname.lastname@example.org||Submitted|
|Adding Webhooks, or similar, support to DPM||
Cloud storages - as Dropbox and Google Drive - implement an API that allows an authorized party to be notified via callbacks - Webhooks in Dropbox and Push Notifications in Google Drive - when some event occurs in the storage. For instance, when a file is uploaded, modified or destroyed.
|QA in distributed cloud architecture: evolution of smashbox framework||
Cloud synchronization and sharing is an area in evolution with innovative services being built on top of different platforms. CERNBOX is a service ran at CERN to provide at the same time synchronisation services (based on the OwnCloud software) and high-performance data access and sharing (based on EOS, the CERN ...
|3-12 months depending on the agreed email@example.com||Submitted|
|QA in distributed cloud architecture: injection-fault testing||
Clients of the sync&share system (CERNBOX) are particularly exposed to "operational failures" due to heterogeneity of hardware, OS and network environments.
Sync&share system operates in very heterogenous network environment: from fast, reliable network inside the computing center to unreliable, high-latency ad-hoc connections such as from air-ports etc.
|The CERN volunteer computing platform||
CERN-IT is developing a volunteer computing solution intended to be a common platform for the LHC experiments’ activities in this area and which should help to maximise the number of cycles they can acquire. The task is to accompany this project through its initial prototyping, work on all problems discovered ...
|Dynamic storage federations||
The group runs a project whose goal is the dynamic federation of
|From 3 to 9 months depending on a selected firstname.lastname@example.org||Submitted|
|File Transfer Service (FTS) extensions||
The File Transfer Service (FTS) manages the global distribution
|From 3 months, depending on task email@example.com||Submitted|
|Performance optimization in a High Throughput Computing environment||
Profiling of computing resources respect to WLCG experiment workloads is a crucial factor to select the most effective resources and to be able to optimise their usage.
|6 to 12 firstname.lastname@example.org||Submitted|
|Cloud data analysis||
Cloud synchronization and sharing is a promising area for the preparation of powerful transformative data services.
The goal of this project is to prepare CERNBOX to be used in connection with heavy-duty activities (large-scale batch processing) on the current LXBATCH infrastructure (LSF) and on its future evolution (HT-Condor): physicists can ...
|Advanced Notifications for Network Incidents||
One of the main challenges in LHCOPN/LHCONE networking is the network diagnostics and advanced notifications on the issues seen in the network. Currently, most of the issues are only visible by the applications and need to be debugged after the incident and performance degradation has already occurred. This is primarily ...
|e-learning - video production and Academic Training video archive promotion||
The Academic Training (AT) video archive in CDS contains a wealth of knowledge that we could promote in youtube as part of CERN's mission around Education. To prepare:
|6 months||Maria Dimou||Submitted|
|Optimisation of experiment workflows in the Worldwide LHC Computing Grid||
The LHC experiments perform the vast majority of the data processing and analysis on the Worldwide LHC Computing Grid (WLCG), which provides a globally distributed infrastructure with more than 500k cores to analyse the tens of PB of data collected each year. Profiling of the computing infrastructure with respect to the ...
|3 to 6 months||Andrea.Sciaba@cern.ch||Submitted|
|e-learning - IT Collaboration, Devices & Applications - document with the user in mind||
The CERN IT Collaboration, Devices & Applications (CDA) group in general and the Integrated Collaboration (IC) section host services used widely at CERN and beyond. Examples from IC:
|6 months Jan-June 2017 - 1 day/week, often from home||Maria Dimou||Submitted|
|Analysis of the I/O performance of LHC computing jobs at the CERN computing centre||
The LHC experiments execute a significant fraction of their data reconstruction, simulation and analysis on the CERN computing batch resources. One of the most important features of these data processing jobs is their I/O pattern in accessing the local storage system, EOS, which is based on the xrootd protocol. In ...
|3 to 6 months||Andrea.Sciaba@cern.ch||Submitted|