Enhancing low energy reconstruction and classification in KM3NeT/ORCA with transformers

Reading time: 5 minute
...

📝 Original Info

  • Title: Enhancing low energy reconstruction and classification in KM3NeT/ORCA with transformers
  • ArXiv ID: 2511.18999
  • Date: 2025-11-24
  • Authors: ** 논문에 명시된 저자 정보가 제공되지 않았습니다. (저자명 및 소속을 확인 후 추가해 주세요.) **

📝 Abstract

The current KM3NeT/ORCA neutrino telescope, still under construction, has not yet reached its full potential in neutrino reconstruction capability. When training any deep learning model, no explicit information about the physics or the detector is provided, thus they remain unknown to the model. This study leverages the strengths of transformers by incorporating attention masks inspired by the physics and detector design, making the model understand both the telescope design and the neutrino physics measured on it. The study also shows the efficacy of transformers on retaining valuable information between detectors when doing fine-tuning from one configurations to another.

💡 Deep Analysis

Figure 1

📄 Full Content

KM3NeT/ORCA is a neutrino telescope [1] placed in the Mediterranean sea, at a depth of 2450 km about 40 km from Toulon (France), with the objective of measuring the neutrino mass hierarchy using atmospheric neutrinos [2]. The detector design is a tridimensional array of PhotoMultiplier Tubes (PMTs) hosted within Digital Optical Modules (DOMs) [3] arranged along vertical detection units (DUs). The telescope collects light from Cherenkov photons emitted along the path of charged particles in neutrino interactions, creating a time-ordered sequence of light pulses with position and timing information that can be used to reconstruct neutrino event kinematics.

In this study, a novel deep learning architecture named transformer [4], that handles sequential data as those observed in a neutrino telescope is presented. In section 2, the model is introduced and the use of attention mask inspired by physics and detector design is motivated. In section 3, the challenges of reconstructing neutrino physics, both due to the physics itself and the limited telescope size, and how transformers overcome them are described.

Thanks to the light pattern originated from a neutrino interaction (fig. 1), transformers are very well suited for reconstructing physics in neutrino telescopes. In KM3NeT/ORCA telescope, the raw observations can be arranged as a time-ordered sequence of light pulses with position and time information from the PMT that recorded it. A transformer model handles sequential data and processes their components in parallel, being able to capture complex patterns and establish relationships between the pulses through the use of the self-attention mechanism [4],

where Q, K and V are matrices built from the input data, and d is the latent space size of the model. If no prior information is given to the model, eq. ( 1) provides information about how the components of the input sequence are correlated among themselves. Nevertheless, no information about the physics is directly injected in the model.

To overcome this burden, physics constraints and detector information are artificially included in eq. ( 1) through the use of attention masks U [5,6],

These attention masks are N ×N matrices with N being the input sequence length, that encode different correlations between the light pulses of an event, for instance, the space-time relativistic distance to identify pulses from the same source [6], the euclidean distance to quantify spatial proximity, or a local-coincidence mask to determine which pulses come from the same PMT, DOM or DU. This approach also allows into discrimination optical background hits from those coming from a physics source, called triggered hits, while increasing context length.

When talking about reconstruction in a neutrino telescope, the main difficulty found is based on its detection principle. Since the PMTs detects light, only charged particles can easily be reconstructed, whereas non-charged particles are invisible to it, leading to an intrinsic bias when trying to reconstruct the whole interaction.

The KM3NeT/ORCA telescope grows in size by adding DUs to already deployed ones, increasing its capability and sensitivity to measure neutrino properties. When a deep learning model is initialized, it does not contain physics information, it learns as it trains. However, this is not optimal because a model can already have learnt that information from a previous configuration. Besides, if a model is first trained on a larger telescope it retains valuable information from DUs not yet deployed as shown in fig. 2 by the area under the Receiver Operating Characteristic (ROC) curve, a metric that indicates the model’s ability to distinguish the two neutrino events: ν C C µ /ν C C e . Figure 2 shows an improvement of over 20% achieved with a very limited training sample (100 events per class), by using interpolated information from a larger configuration, contrary to a model trained from scratch, whose performances are only comparable when using a large training sample (1M events per class), where limitations from the detector itself play a higher role rather than the model capabilities. Another benefit of the use of deep learning for classification is that the training does not rely on any reconstructed variable, it computes the score it from what the raw data. On the contrary, classical classification methods rely on decision-tree-based classifiers trained using reconstruction variables. These techniques are prone to be strongly biased by reconstruction algorithms and must be updated whenever a new version of the reconstruction software is released.

The major difficulty of neutrino telescopes is that they are built to measure neutrino properties, but neutrinos are invisible to them. Therefore, reconstruction algorithms based on a maximum-likelihood fit (MLF) are developed to reconstruct only what is observed, which in most cases is based on either a track or a shower hypothesis, whereas in realit

📸 Image Gallery

AUROC.png EuCAIF_logo.png Event.png angular_resolution.png energy_fit.png

Reference

This content is AI-processed based on open access ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut