The use of statistical methods for the search for new physics at the LHC (in Russian)
We review statistical methods used for the search for new physics at LHC.
đĄ Research Summary
**
The manuscript âThe use of statistical methods for the search for new physics at the LHCâ is a comprehensive lectureâstyle review aimed at graduate students, postâdoctoral researchers, and analysts working on the CMS and ATLAS experiments. It is written in Russian but the content is presented here in English. The authors, S.I. Bityukov and N.V. Krasnikov, organize the material into seven chapters that progress from the fundamentals of probability theory to the practical implementation of statistical tools in highâenergy physics (HEP) data analysis.
ChapterâŻ1 â Foundations of Probability
The first chapter introduces the Kolmogorov axioms, conditional probability, Bayesâ theorem, random variables, probability density functions (PDFs), cumulative distribution functions (CDFs), and characteristic functions. The exposition uses intuitive examples such as coin tossing and drawing balls from an urn to illustrate discrete and continuous probability concepts. This groundwork is essential for later sections where more sophisticated statistical models are built.
ChapterâŻ2 â Core Statistical Paradigms
ChapterâŻ2 contrasts the frequentist and Bayesian approaches. In the frequentist subsection the authors discuss confidence intervals, the construction of upper limits for Poissonâdistributed counts, and the concept of expected limits (often denoted CLs in LHC analyses). The Bayesian subsection explains prior distributions, posterior inference, and credible intervals, emphasizing how the choice of prior influences the final result. Numerical illustrations show how both paradigms are applied to typical LHC counting experiments.
ChapterâŻ3 â Parameter Estimation
Maximumâlikelihood estimation (MLE) is presented as the central frequentist tool. The authors derive the likelihood function for signalâplusâbackground models, introduce the profileâlikelihood ratio as a test statistic, and discuss its asymptotic chiâsquare behavior (Wilksâ theorem). The chapter also covers the Fisher information matrix, the construction of confidence regions for multiâparameter problems, and the treatment of nuisance parameters.
ChapterâŻ4 â Systematic Uncertainties
Systematics are treated in depth. The authors describe three main strategies: (i) treating background normalisation as an auxiliary measurement, (ii) incorporating nuisance parameters directly into the likelihood with constraint terms (Gaussian, logânormal, etc.), and (iii) the KuzinsâHayland averaging method for correlated systematics. Detailed examples show how to propagate detectorârelated uncertainties (energy scale, resolution) into the final signal strength estimate.
ChapterâŻ5 â Hypothesis Testing
The fifth chapter focuses on testing the null hypothesis âno new physicsâ against the alternative âsignal presentâ. It reviews pâvalue computation, the NeymanâPearson lemma, the construction of test statistics such as q0 (backgroundâonly) and qÎź (signalâplusâbackground), and the use of the CLs method to avoid exclusion of signal models with low sensitivity. The âlookâelsewhere effectâ is explained, together with techniques (trials factor, global pâvalue) to correct for scanning over multiple mass hypotheses. Bayesian model comparison via Bayes factors is also presented.
ChapterâŻ6 â Combination of Results
Combining measurements from different channels, experiments, or even different colliders is essential for increasing sensitivity. The authors discuss analytic combination of Gaussian results, the general BLUE (Best Linear Unbiased Estimator) method, combined likelihood fits (both frequentist and Bayesian), and the treatment of correlated systematic uncertainties across experiments. Realâworld examples from CMSâATLAS joint Higgs searches illustrate the impact of each combination technique on the final exclusion limits.
ChapterâŻ7 â Statistical Software for HEP
The final chapter surveys the software ecosystem built around ROOT. RooStats is described as the primary library for constructing likelihood models, performing profileâlikelihood scans, and computing confidence intervals. The Bayesian Analysis Toolkit (BAT) is introduced for MarkovâChain MonteâCarlo (MCMC) sampling of posterior distributions. Sample code snippets, workflow diagrams, and bestâpractice recommendations enable readers to translate the theoretical material directly into analysis code.
Critical Assessment
The manuscript excels as an educational resource: it balances rigorous mathematical definitions with concrete HEP examples, making it suitable for newcomers who need a âoneâstopâ reference. By covering both frequentist and Bayesian paradigms, it respects the diversity of statistical cultures within the LHC collaborations. The inclusion of systematicâuncertainty handling and resultâcombination strategies reflects the real challenges faced in modern searches for physics beyond the Standard Model.
However, the text is primarily a review and does not present original research. Proofs of advanced theorems (e.g., the full derivation of Wilksâ theorem for nonâregular models) are omitted, which may limit its usefulness for theorists interested in methodological development. Moreover, emerging machineâlearningâbased inference techniques (e.g., likelihoodâfree inference, neural density estimation) are not covered, reflecting the manuscriptâs focus on traditional statistical tools.
Conclusion
Overall, the paper provides a thorough, wellâstructured overview of the statistical machinery employed in LHC newâphysics searches. Its pedagogical style, extensive numerical examples, and direct connection to the ROOTâbased software stack make it a valuable reference for graduate students, postâdoctoral researchers, and analysts engaged in highâenergy physics data interpretation.
Comments & Academic Discussion
Loading comments...
Leave a Comment