File-based data flow in the CMS Filter Farm
Author(s)
Darlea, G.-L.; Gomez-Ceballos, Guillelmo; Paus, Christoph M. E.; Veverka, Jan
DownloadFile-based data.pdf (1.145Mb)
PUBLISHER_CC
Publisher with Creative Commons License
Creative Commons Attribution
Terms of use
Metadata
Show full item recordAbstract
During the LHC Long Shutdown 1, the CMS Data Acquisition system underwent a partial redesign to replace obsolete network equipment, use more homogeneous switching technologies, and prepare the ground for future upgrades of the detector front-ends. The software and hardware infrastructure to provide input, execute the High Level Trigger (HLT) algorithms and deal with output data transport and storage has also been redesigned to be completely file- based. This approach provides additional decoupling between the HLT algorithms and the input and output data flow. All the metadata needed for bookkeeping of the data flow and the HLT process lifetimes are also generated in the form of small "documents" using the JSON encoding, by either services in the flow of the HLT execution (for rates etc.) or watchdog processes. These "files" can remain memory-resident or be written to disk if they are to be used in another part of the system (e.g. for aggregation of output data). We discuss how this redesign improves the robustness and flexibility of the CMS DAQ and the performance of the system currently being commissioned for the LHC Run 2.
Date issued
2015-12Department
Massachusetts Institute of Technology. Department of Physics; Massachusetts Institute of Technology. Laboratory for Nuclear ScienceJournal
Journal of Physics: Conference Series
Publisher
IOP Publishing
Citation
Andre, J-M et al. “File-Based Data Flow in the CMS Filter Farm.” Journal of Physics: Conference Series 664.8 (2015): 082033. © 2017 IOP Publishing
Version: Final published version
ISSN
1742-6588
1742-6596