SciPy 2025

Breaking New Ground: Scalable Simulation-Based Inference at the LHC with Scientific Python
07-09, 10:45–11:15 (US/Pacific), Ballroom

Simulation-Based Inference (SBI) is a powerful class of machine learning based methods for statistical inference, with applications in many scientific domains including particle physics, cosmology and astrophysics. While demonstrating significant promise in small-scale studies over the last 10 years, a full-scale deployment for a particle physics experiment remained elusive due to the computational challenges involved. Using novel ideas, modern distributed computing resources, and tools like JAX and Tensorflow, we built the first end-to-end SBI workflow exclusively using the Scientific Python ecosystem that is scalable for measurements at the Large Hadron Collider (LHC). The new techniques were used to measure the lifetime of the Higgs boson with the ATLAS experiment at the LHC with unprecedented precision.

In this talk, I will present the many challenges encountered while scaling the new analysis method to a full-scale measurement and demonstrate how the power, versatility and the rich set of tools in the Scientific Python ecosystem played a critical role in overcoming them.


High-energy physics experiments, such as ATLAS at the Large Hadron Collider (LHC), rely on advanced simulation software to accurately model the dense and complex physical interactions occurring within particle detectors. These simulators generate particle collision events using an implicit likelihood model that is analytically intractable. Traditionally, statistical inference in the presence of intractable models relies on dimensionality reduction techniques, which result in a loss of measurement sensitivity. Simulation-Based Inference (SBI) is an emerging new set of techniques that use deep-learning to construct surrogate models for the analytically intractable likelihood functions directly using high-dimensional data. This leads to significantly improved precision in the measurements of the parameters of the model being tested, making the new set of techniques promising for searches of new physics at the energy frontier.

While powerful, these techniques are also computationally demanding and have thus remained inaccessible for full measurements at the LHC since their first proposal in 2015. However, in 2024, using SBI techniques, we published the most precise measurement of the Higgs boson lifetime using ATLAS data. For this application, several novel ideas were developed that not only scale better to a full LHC measurement, but also offer efficient computational workflows using the Scientific Python ecosystem. The technique and workflow developed are described in four publications, and apart from offering a deeper insight on the Higgs boson properties, have opened up the applicability of SBI techniques for many other precision measurements at the LHC.

In this talk, I will describe the significant role that the open-source Scientific Python ecosystem played in the development of the full framework at the ATLAS experiment. Using a live demo, I will demonstrate how we went from ideas to a working at-scale implementation, using tools like TensorFlow and JAX. I will highlight the various challenges, both foreseen and un-foreseen, that we encountered along the way and how the versatility and power of the Python tools helped in overcoming each of them and making the final measurement of the Higgs boson lifetime using the full ATLAS model possible.

This talk will be of interest to people doing precision measurements in a complex likelihood-free setting, which could include any domain from particle physics to biology. Attendees will primarily learn about the versatility of libraries like JAX for efficient statistical inference, including the use of just-in-time compilation, auto-vectorization, and auto-differentiation. For people who don’t fall into this target audience, they will learn about how the various available tools in the Scientific Python ecosystem are being used in some of the cutting-edge research and what future developments can help tackle emerging challenges.

See also:

I am a postdoctoral researcher at the University of Wisconsin-Madison, affiliated with the Data Science Institute and the Department of Physics. My research focuses on the intersection of modern statistical tools and high-energy physics.