maciek-g commited on
Commit
07b1bb7
1 Parent(s): 634a26e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -3
README.md CHANGED
@@ -1,3 +1,21 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # AnalysisObjectTransformer Model
2
+
3
+ This repository contains the implementation of the AnalysisObjectTransformer model, a deep learning architecture designed for event classification with reconstructed-object inputs. MultiHeadAttention is used to extract the correlation between jets (hadrons) in the final state. Achieves state-of-the-art performance on final states which can be summarized as jets accompanied by missing transverse energy.
4
+
5
+ ## Model Overview
6
+
7
+ The AnalysisObjectTransformer model is structured to process jet-level features (energy, mass, area, btag score) in any order (permutation invariance) and event-level features (angle analysis of missing energy and leading jets) to classify signal from background processes to enhance the sensitivity to rare BSM signatures.
8
+
9
+ ### Components
10
+
11
+ - **Embedding Layers**: Transform input data into a higher-dimensional space for subsequent processing.
12
+ - **Attention Blocks (AttBlock)**: Utilize multi-head attention to capture dependencies between different elements of the input data.
13
+ - **Class Blocks (ClassBlock)**: Extend attention mechanisms to incorporate class tokens, enabling the model to focus on class-relevant features. Implementation based on "Going deeper with transformers": https://arxiv.org/abs/2103.17239
14
+ - **MLP Head**: A sequence of fully connected layers that maps the output of the transformer blocks to the final prediction targets.
15
+
16
+ ## Usage
17
+
18
+ ```python
19
+ from particle_transformer import AnalysisObjectTransformer
20
+
21
+ model = AnalysisObjectTransformer(input_dim_obj=..., input_dim_event=..., embed_dims=..., linear_dims1=..., linear_dims2=..., mlp_hidden_1=..., mlp_hidden_2=..., num_heads=...)