In order to stimulate an open dialogue with companies interested in the ARCHIVER project during the OMC process, the Technical Summaries of each Deployment Scenario are made publicly accessible:
Problem Definition:
We want to substitute the current in-house tape library storage for both initial and derived datasets. Each instance of the service to be purchased is the 5-year safe-keeping of a yearly dataset and its derived datasets according to the specifications below.
Problem Definition:
We want to substitute the current in-house tape library storage by an off-premise commercial service. Each instance of the service to be purchased consists in the 5-year safe-keeping of a yearly dataset from a single source according to the specifications below.
Problem Definition:
We want to substitute the current in-house tape library storage, Hierarchical Storage Manager, disk storage and data distribution. Each instance of the service to be purchased is the 5-year safe-keeping and data distribution of a yearly dataset and its derived datasets from at most two sources according to the specifications below.
Problem Definition:
PETRA III is the worldwide most brilliant storage ring based X-ray sources for high energy photons. 22 beamlines distributed over three experimental halls are concurrently available for users. The European XFEL is a world's largest X-ray laser generating 27 000 ultrashort X-ray per second and with a brilliance that is a billion times higher than that of the best conventional X-ray radiation sources. 6 beamlines are available for user experiments.
Problem Definition:
In 2020 the BaBar Experiment infrastructure at SLAC will be decommissioned. As a result, the 2 PBs of BaBar data can no longer be stored at the host laboratory and alternative solutions need to be found. We want to ensure that a complete copy of Babar data will be retained in such a way that will allow researchers to re-use and re-analyse their data.
Problem Definition:
FIRE is the ‘FIle REplication’ service, our current archiving system. It consists of two complete copies of the data, one is a distributed object store (across three of our own dedicated data centres), the other is a tape archive hosted at Hinxton.
Problem Definition:
As we expand our computational activities into the cloud, we expect that it won’t be cost-effective to put all our data into the cloud immediately. Our data volume is doubling every two years, so half our data is less than two years old, and that data is likely to be actively used much of the time. Even the older data has a long tail of access, so very little of our data is ‘never’ used anymore.
The CERN Open Data portal disseminates more than one petabytes of primary and derived datasets from particle physics as they were released by LHC collaborations. The portal offers datasets together with accompanying software examples, virtual machine images, condition databases, configuration files and necessary associated documentation to enable non-specialists to use the data.
Problem definition:
We need to archive the CERN Digital Memory which consists of the digital production of the institution for the 21st century (including new types like web sites, social media, emails, etc) as well as the analogue documents produced by the institution in the 20th century, composed of digitised papers (physical archive) and various multimedia: audio (e.g. recordings of meetings), still images and moving images.