Imaging atmospheric Cherenkov telescopes (IACT) data require an important analysis in order to reconstruct events and obtain a photon list. The state-of-the-art reconstruction is made of several steps including image analysis, features extraction and machine learning.
Since the 2012 ImageNet breakthrough, deep learning advances have shown dramatic improvements in data analysis across a variety of fields. Convolutional neural networks look particularly suited to the task of analysing IACT camera images for event reconstruction as
they provide a way to reconstruct photon list directly from raw images, skipping the pre-processing steps. Moreover, despite demanding important computing resources to be trained and optimised, neural networks show very good performances during execution, making them viable for real-time analysis for the future generation of IACT. Here we present GammaLearn, a python framework providing the tools and environment to easily train neural networks on IACT data. Relying on PyTorch, it allows the use of indexed convolution on images with non-cartesian pixel lattices predominant in IACT for the low-level operations and offers a simple configuration file-based workflow, producing the trained model, training estimators as well as higher level results. The proposed framework is modular and straightforward to customize by end users. It has been tested and validated on
the analysis of the Cherenkov Telescope Array simulated data.