EveNet: Towards a Generalist Event Transformer for Unified Understanding and Generation of Collider Data
Ting-Hsiang Hsu2, Qibin Liu3, Yuan-Tang Chou1, Wei-Po Wang2, Yue Xu1, Haoran Zhao1, Bai-Hong Zhou3, Yi Ren Wu2*, Shu Li3, Benjamin Nachman4, Shih-Chieh Hsu1, Vinicius Massami Mikuni5, Yulei Zhang1
1Department of Physics, University of Washington, Seattle, USA
2Department of Physics, National Taiwan University, Taipei, Taiwan
3Department of Physics, Shanghai Jiao Tong University, Shanghai, China
4Department of Physics, Stanford University, California, USA
5Physics Division, Lawrence Berkeley National Laboratory, California, USA
6National Energy Research Scientific Computing Center, California, USA
* Presenter:Yi Ren Wu, email:b11202011@ntu.edu.tw
With the increasing size of the machine learning (ML) model and vast datasets, the foundation model has transformed how we apply ML to solve real-world problems. Multimodal language models like chatGPT and Llama have expanded their capability to specialized tasks with common pre-train. Similarly, in high-energy physics (HEP), common tasks in the analysis face recurring challenges that demand scalable, data-driven solutions. In this talk, we present a foundation model for high-energy physics. Our model leverages extensive simulated datasets in pre-training to address common tasks across analyses, offering a unified starting point for specialized applications. We demonstrate the benefit of using such a pre-train model in improving search sensitivity, anomaly detection, event reconstruction, feature generation, and beyond. By harnessing the power of pre-trained models, we could push the boundaries of discovery with greater efficiency and insight.


Keywords: Machine Learning , Collider Experiments, Foundation Model