Multi-scale cross-attention transformer encoder for event classification
Multi-scale cross-attention transformer encoder for event classification
We deploy an advanced Machine Learning (ML) environment, leveraging a multi-scale cross-attention encoder for event classification, towards the identification of the gg→H→hh→bb¯bb¯ process at the High Luminosity Large Hadron Collider (HL-LHC), where h is the discovered Standard Model (SM)-like Higgs boson and H a heavier version of it (with mH>2mh). { In the ensuing boosted Higgs regime, the final state consists of two fat jets. Our multi-modal network can extract information from the jet substructure and the kinematics of the final state particles through self-attention transformer layers. The diverse learned information is subsequently integrated to improve classification performance using an additional transformer encoder with cross-attention heads.} We ultimately prove that our approach surpasses in performance current alternative methods used to establish sensitivity to this process, whether solely based on kinematic analysis or else on a combination of this with mainstream ML approaches. {Then, we employ various interpretive methods to evaluate the network results, including attention map analysis and visual representation of Gradient-weighted Class Activation Mapping (Grad-CAM). Finally, we note that the proposed network is generic and can be applied to analyse any process carrying information at different scales.} Our code is publicly available for generic use.
hep-ph
Hammad, A.
08f4a260-b5e7-404f-aa33-131cb5097394
Moretti, S.
b57cf0f0-4bc3-4e02-96e3-071255366614
Nojiri, M.
6f682a41-7d59-4b29-8320-4ec5491d3f40
Hammad, A.
08f4a260-b5e7-404f-aa33-131cb5097394
Moretti, S.
b57cf0f0-4bc3-4e02-96e3-071255366614
Nojiri, M.
6f682a41-7d59-4b29-8320-4ec5491d3f40
[Unknown type: UNSPECIFIED]
Abstract
We deploy an advanced Machine Learning (ML) environment, leveraging a multi-scale cross-attention encoder for event classification, towards the identification of the gg→H→hh→bb¯bb¯ process at the High Luminosity Large Hadron Collider (HL-LHC), where h is the discovered Standard Model (SM)-like Higgs boson and H a heavier version of it (with mH>2mh). { In the ensuing boosted Higgs regime, the final state consists of two fat jets. Our multi-modal network can extract information from the jet substructure and the kinematics of the final state particles through self-attention transformer layers. The diverse learned information is subsequently integrated to improve classification performance using an additional transformer encoder with cross-attention heads.} We ultimately prove that our approach surpasses in performance current alternative methods used to establish sensitivity to this process, whether solely based on kinematic analysis or else on a combination of this with mainstream ML approaches. {Then, we employ various interpretive methods to evaluate the network results, including attention map analysis and visual representation of Gradient-weighted Class Activation Mapping (Grad-CAM). Finally, we note that the proposed network is generic and can be applied to analyse any process carrying information at different scales.} Our code is publicly available for generic use.
Text
2401.00452v1
- Author's Original
More information
Accepted/In Press date: 31 December 2023
Additional Information:
11 figures and a table
Keywords:
hep-ph
Identifiers
Local EPrints ID: 486082
URI: http://eprints.soton.ac.uk/id/eprint/486082
PURE UUID: 9d9f5a1a-0971-48db-ae0d-477d3246ed0d
Catalogue record
Date deposited: 09 Jan 2024 17:32
Last modified: 18 Mar 2024 02:57
Export record
Altmetrics
Contributors
Author:
A. Hammad
Author:
M. Nojiri
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics