Graph Neural Networks for Charged Particle Tracking
The reconstruction of charged particle trajectories ("tracking") in particle physics detectors is one of the computationally most challenging tasks of the field, limiting the amount of high-quality data that can even be recorded. Applied to particle collider experiments such as the CMS experiment, this task is different from many other problems that involve trajectories: There are millions of particle collisions per second, each with thousands of individual particles that need to be tracked, there is no time information (the particles travel too fast), and we do not observe a continuous trajectory but instead only 5-15 points ("hits") along the way in different detector layers. The task can be described as a combinatorically very challenging "connect-the-dots" problem, essentially turning a cloud of points (hits) in 3D space into a set of O(1000) trajectories.
Unlike traditional tracking algorithms built around Kalman filters, this project uses graph neural networks for significant speed increases. A conceptually simple way to turn tracking into a machine learning task is to create a fully connected graph of all points and then train an edge classifier to reject any edge that doesn't connect points that belong to the same particle. In this way, only the individual trajectories remain as components of the initial fully connected graph. In this project, we instead explore the idea of object condensation or learned clustering, where a network maps all hits to a latent space, learning to place hits from the same track close to each other, such that simple operations can recover the hits belonging to the same tracks.
Coordinating Software Training and Education efforts in High Energy Physics
Experimental high energy physics at large experiments is tasked with analyzing petabytes of data, necessitating an ever-evolving, ever more complex software stack. Delivering the best possible science depends crucially on the software skills of a large workforce of researchers. Keeping up with the latest big data tools and technology requires extensive training, covering everything from programming best practices to the latest industry tools and experiment-specific software frameworks.
From 2020 to 2022, I led the Belle II Software Training and Documentation group (Belle II Structure) that organizes training events and provides training material, primarily focusing on getting researchers up to speed with the Belle II software framework. In 2020 and since 2022, I have also been coordinating software training across experiments as one of the conveners of the HSF Training Group. I have also taught the basics of programming paradigms and software design patterns to more than 500 participants.
Past projects
\(\bar B\longrightarrow D^*\ell^-\bar\nu_\ell\) Decays with Hadronic Tagging at Belle
The decay \(\bar B\longrightarrow D^*\ell^-\bar\nu_\ell\) is used to precisely determine the CKM matrix element \(|V_{cb}|\), an important ingredient for tests of the flavor sector of the Standard Model. It is also the normalization channel for measurements of \(R(D^*)\), one of the key quantities of the flavor anomalies that recently sparked a flurry of interest in the field. Improving our understanding of \(\bar B\longrightarrow D^*\ell^-\bar\nu_\ell\) might help to understand and improve analyses of \(R(D^*)\) as well.
Reconstruction of a tag side \(B\) meson in addition to the semileptonically decaying \(B\) allows for a very clean data sample. Using the large Belle dataset but applying Belle II software for analysis, we can improve upon previous studies: The Belle II Full Event Interpretation, a machine learning algorithm to reconstruct the tag side \(B\) meson is almost two times more efficient than previously used algorithms. However, careful calibration studies are needed to address inconsistencies in its efficiency between data and Monte Carlo simulation.
Clustering of kinematic graphs
ClusterKinG
: clustering of a a three dimensional parameter space results in three distinct clusters. Benchmark points are highlighted with enlarged markers.Belle II Software Integration and Performance Testing
Construction of Angular Observables Sensitive to New Physics in \(\bar B\longrightarrow D^* \tau^-\bar\nu_\tau\) Decays and Measurements of Differential Cross Sections of \(\bar B\longrightarrow D^*\ell^-\bar\nu_\ell\) Decays with Hadronic Tagging at Belle
Recent measurements of \(\bar B\longrightarrow D^{(*)}\ell^-\bar\nu_\ell\) at Belle, BaBar and LHCb challenge lepton universality and thus the Standard Model at a combined confidence level close to four standard deviations. New measurements of differential decay rates could contribute to the understanding of these anomalies.
The differential cross section of the decay \(\bar B\longrightarrow D^*(\rightarrow D\pi)\ell^-\bar\nu_\ell\) is parametrized according to different dependencies on the three decay angles and the coupling constants of potential new physics contributions. Observables using binned measurements of the differential cross section are characterized and explicitly constructed. Based on an estimate for the obtainable sensitivity, optimal binnings for such measurements are discussed. The discriminatory power of the thus constructed observables is discussed based on a basis of dimension six operators with renormalizable couplings contributing to \(\bar B\longrightarrow D^*\ell^-\bar\nu_\ell\).
Furthermore, continuing work on an analysis of the \(\bar B\longrightarrow D^*(\rightarrow D\pi)\ell^-\bar\nu_\ell\) decay channel for \(\ell = e, \mu\) using data from the Belle detector at KEKB is presented. The events are selected from 772 million \(e^+e^- \longrightarrow \Upsilon(4S) \longrightarrow B\bar B\) events, where one \(B\) meson is fully reconstructed in hadronic modes. Unfolded differential decay rates in four kinematic variables are presented separately for \(\ell= e, \mu\) and a combined fit, allowing for precise calculations of \(|V_{cb}|\) and \(B\longrightarrow D^*\) form factors. The new lepton flavor specific results are also expected to impact the discussion about potential light lepton flavor universality violations prompted by measurements of \(B\longrightarrow K^{(*)}\ell\ell\) decays.
- Full text [PDF, 9MB]
Complex Organic Molecules in Protoplanetary Disks
Complex Organic Molecules (COMs) in protoplanetary disks have been the subject of extensive studies using chemical reaction networks (CRNs) (e.g. Walsh et al., 2014). The accuracy of these models depends on our knowledge of the relevant chemical processes. Some classes of reactions have been comprehensively studied, resulting in large databases like the UMIST database of astrochemistry, which lists more than 6000 gas-phase reactions. However, other classes of reactions, such as grain-surface reactions, still pose challenges.
By expanding the previously studied CRNs with additional grain-surface reactions that are currently studied in new laboratory experiments (and have so far mostly been considered in the context of meteorites), we tried to improve the description of COM formation in protoplanetary disks. More specifically, I have been using the existing simulation code to investigate the influence of physical and chemical parameters, such as temperature, density and activation energies, on the time evolution of the chemistry found on grains. Trying to automize some time-consuming manual tasks necessary for such studies, I wrote a framework to repeatedly run the simulation with different settings and to visualize the resulting datasets. This framework will enable future students to conduct similar studies much more efficiently, thereby opening new research possibilities.
- Experience Report Prof. Nomura
- Final presentation (aimed at a general audience) [PowerPoint pptx, 5MB]
Performance monitoring for LHCb DAQ
In 2020 the Data Acquisition (DAQ) of the LHCb experiment will be updated to feature a trigger-free readout. This requires an event builder network consisting of about 500 nodes with a total network capacity of 4 TBytes/s. DAQPIPE (Data Acquisition Protocol Independent Performance Evaluator) is a tool to simulate and evaluate the performance of such a DAQ system. The current implementation of DAQPIPE only gives rough feedback about the event building rate.
The aim of this 10-week summer student project was to implement network monitoring for a more detailed performance evaluation of different transport protocols and to spot potential bottlenecks. First, several existing performance monitors were tested. To that end DAQPIPE was run together with Tau and the obtained performance data was plotted with ParaProf, JumpShot and Vampir. In the second stage of the project, a light-weight performance analysis tool was written from scratch by wrapping around the C++
MPI communication library to collect data.
Truth-level based estimation of the sensitivity to pMSSM models in events with one hard lepton
Elliptic Functions
Central subject of the this are so called elliptic functions, meromorphic functions that are periodic in two directions, i.e. invariant under a translation of their argument by two linearly independent complex numbers.
Among others, elliptic functions are of great use in number theory, in particular there are interesting connections to sums of divisors of natural numbers. Furthermore they are used in the theory of elliptic curves and elliptic integrals.