scholarly journals High Energy Physics Research with the CMS Experiment at CERN

2013 ◽  
Author(s):  
Gail G. Hanson
2020 ◽  
Vol 245 ◽  
pp. 01005
Author(s):  
Wassef Karimeh ◽  
Maroun Chammoun ◽  
Ivan Shvetsov ◽  
Andromachi Tsirou ◽  
Piero Giorgio Verdini

Detector Control Systems (DCS) for modern High-Energy Physics (HEP) experiments are based on complex distributed (and often redundant) hardware and software implementing real-time operational procedures meant to ensure that the detector is always in a safe state, thus maximizing the lifetime of the detector. Display, archival and often analysis of the environmental data are also part of the tasks assigned to DCS systems. The CMS Tracker Control System (TCS) is a resilient system that has been designed to safely operate the silicon tracking detector in the CMS experiment. It has been built on top of an industrial Supervisory Control and Data Acquisition (SCADA) software product WinCC OA extended with a framework developed at CERN, JCOP, along with CMS and Tracker specific components. The TCS is at present undergoing major architecture redesign which is critical to ensure efficient control of the detector and its future upgrades for the next fifteen years period. In this paper, we will present an overview of the Tracker DCS and the architecture of the software components as well as the associated deliverables.


2011 ◽  
Vol 26 (05) ◽  
pp. 309-317
Author(s):  
◽  
DAN GREEN

The Large Hadron Collider (LHC) began 7 TeV C.M. energy operation in April, 2010. The CMS experiment immediately analyzed the earliest data taken in order to "rediscover" the Standard Model (SM) of high energy physics. By the late summer, all SM particles were observed and CMS began to search for physics beyond the SM and beyond the present limits set at the Fermilab Tevatron. The first LHC run ended in Dec., 2010 with a total integrated luminosity of about 45 pb-1 delivered to the experiments.


2019 ◽  
Vol 214 ◽  
pp. 01003
Author(s):  
Sioni Summers ◽  
Andrew Rose

Track reconstruction at the CMS experiment uses the Combinatorial Kalman Filter. The algorithm computation time scales exponentially with pileup, which will pose a problem for the High Level Trigger at the High Luminosity LHC. FPGAs, which are already used extensively in hardware triggers, are becoming more widely used for compute acceleration. With a combination of high performance, energy efficiency, and predictable and low latency, FPGA accelerators are an interesting technology for high energy physics. Here, progress towards porting of the CMS track reconstruction to Maxeler Technologies’ Dataflow Engines is shown, programmed with their high level language MaxJ. The performance is compared to CPUs, and further steps to optimise for the architecture are presented.


2019 ◽  
Vol 214 ◽  
pp. 03004
Author(s):  
Brian Paul Bockelman ◽  
Edgar Fajardo Hernandez ◽  
Diego Davila Foyo ◽  
Kenyi Hurtado Anampa ◽  
Farrukh Aftab Khan ◽  
...  

The CMS experiment has an HTCondor Global Pool, composed of more than 200K CPU cores available for Monte Carlo production and the analysis of da.The submission of user jobs to this pool is handled by either CRAB, the standard workflow management tool used by CMS users to submit analysis jobs requiring event processing of large amounts of data, or by CMS Connect, a service focused on final stage condor-like analysis jobs and applications that already have a workflow job manager in place. The latest scenario canbring cases in which workflows need further adjustments in order to efficiently work in a globally distributed pool of resources. For instance, the generation of matrix elements for high energy physics processes via Madgraph5_aMC@NLO and the usage of tools not (yet) fully supported by the CMS software, such as Ten-sorFlow with GPUsupport, are tasks with particular requirements. A special adaption, either at the pool factory level (advertising GPU resources) or at the execute level (e.g: to handle special parameters that describe certain needs for the remote execute nodes during submission) is needed in order to adequately work in the CMS global pool. This contribution describes the challenges and efforts performed towards adaptingsuch workflows so they can properly profit from the Global Pool via CMS Connect.


Author(s):  
Preeti Kumari ◽  
◽  
Kavita Lalwani ◽  
Ranjit Dalal ◽  
Ashutosh Bhardwaj ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document