Загрузка страницы

Processing LHC Data

The LHC produces 600 million collisions every second in each detector, which generates approximately one petabyte of data per second. None of today's computing systems are capable of recording such rates. Hence sophisticated selection systems are used for a first fast electronic pre-selection, only passing one out of 10 000 events. Tens of thousands of processor cores then select 1% of the remaining events. Even after such a drastic data reduction, the four big experiments, ALICE, ATLAS, CMS and LHCb, together need to store over 25 petabytes per year. The LHC data are aggregated in the CERN Data Centre, where initial data reconstruction is performed, and a copy is archived to long-term tape storage. Another copy is sent to several large scale data centres around the world. Subsequently hundreds of thousands of computers from around the world come into action: harnessed in a distributed computing service, they form the Worldwide LHC Computing Grid (WLCG), which provides the resources to store, distribute, and process the LHC data. WLCG combines the power of more than 170 collaborating centres in 36 countries around the world, which are linked to CERN. Every day WLCG processes more than 1.5 million 'jobs', corresponding to a single computer running for more than 600 years.

Produced by: CERN IT department
Director: The Little Big Studio
You can follow us in:

http://www.cern.ch
http://www.youtube.com/cerntv
http://google.com/+CERN
http://www.facebook.com/cern
http://twitter.com/cern/

Copyright © 2013 CERN. Terms of use: http://copyright.web.cern.ch

Видео Processing LHC Data канала CERN
Показать
Комментарии отсутствуют
Введите заголовок:

Введите адрес ссылки:

Введите адрес видео с YouTube:

Зарегистрируйтесь или войдите с
Информация о видео
22 апреля 2013 г. 18:27:19
00:05:10
Яндекс.Метрика