Not enough data to create a plot.
Try a different view from the menu above.
Farrell, Steven
FAIR Universe HiggsML Uncertainty Challenge Competition
Bhimji, Wahid, Calafiura, Paolo, Chakkappai, Ragansu, Chang, Po-Wen, Chou, Yuan-Tang, Diefenbacher, Sascha, Dudley, Jordan, Farrell, Steven, Ghosh, Aishik, Guyon, Isabelle, Harris, Chris, Hsu, Shih-Chieh, Khoda, Elham E, Lyscar, Rémy, Michon, Alexandre, Nachman, Benjamin, Nugent, Peter, Reymond, Mathis, Rousseau, David, Sluijter, Benjamin, Thorne, Benjamin, Ullah, Ihsan, Zhang, Yulei
The FAIR Universe -- HiggsML Uncertainty Challenge focuses on measuring the physics properties of elementary particles with imperfect simulators due to differences in modelling systematic errors. Additionally, the challenge is leveraging a large-compute-scale AI platform for sharing datasets, training models, and hosting machine learning competitions. Our challenge brings together the physics and machine learning communities to advance our understanding and methodologies in handling systematic (epistemic) uncertainties within AI techniques.
Comprehensive Performance Modeling and System Design Insights for Foundation Models
Subramanian, Shashank, Rrapaj, Ermal, Harrington, Peter, Chheda, Smeet, Farrell, Steven, Austin, Brian, Williams, Samuel, Wright, Nicholas, Bhimji, Wahid
Generative AI, in particular large transformer models, are increasingly driving HPC system design in science and industry. We analyze performance characteristics of such transformer models and discuss their sensitivity to the transformer type, parallelization strategy, and HPC system features (accelerators and interconnects). We utilize a performance model that allows us to explore this complex design space and highlight its key components. We find that different transformer types demand different parallelism and system characteristics at different training regimes. Large Language Models are performant with 3D parallelism and amplify network needs only at pre-training scales with reduced dependence on accelerator capacity and bandwidth. On the other hand, long-sequence transformers, representative of scientific foundation models, place a more uniform dependence on network and capacity with necessary 4D parallelism. Our analysis emphasizes the need for closer performance modeling of different transformer types keeping system features in mind and demonstrates a path towards this. Our code is available as open-source.
Hierarchical Graph Neural Networks for Particle Track Reconstruction
Liu, Ryan, Calafiura, Paolo, Farrell, Steven, Ju, Xiangyang, Murnane, Daniel Thomas, Pham, Tuan Minh
We introduce a novel variant of GNN for particle tracking--called Hierarchical Graph Neural Network (HGNN). The architecture creates a set of higher-level representations which correspond to tracks and assigns spacepoints to these tracks, allowing disconnected spacepoints to be assigned to the same track, as well as multiple tracks to share the same spacepoint. We propose a novel learnable pooling algorithm called GMPool to generate these higher-level representations called "super-nodes", as well as a new loss function designed for tracking problems and HGNN specifically. On a standard tracking problem, we show that, compared with previous ML-based tracking algorithms, the HGNN has better tracking efficiency performance, better robustness against inefficient input graphs, and better convergence compared with traditional GNNs.
Track Seeding and Labelling with Embedded-space Graph Neural Networks
Choma, Nicholas, Murnane, Daniel, Ju, Xiangyang, Calafiura, Paolo, Conlon, Sean, Farrell, Steven, Prabhat, null, Cerati, Giuseppe, Gray, Lindsey, Klijnsma, Thomas, Kowalkowski, Jim, Spentzouris, Panagiotis, Vlimant, Jean-Roch, Spiropulu, Maria, Aurisano, Adam, Hewes, V, Tsaris, Aristeidis, Terao, Kazuhiro, Usher, Tracy
To address the unprecedented scale of HL-LHC data, the Exa.TrkX project is investigating a variety of machine learning approaches to particle track reconstruction. The most promising of these solutions, graph neural networks (GNN), process the event as a graph that connects track measurements (detector hits corresponding to nodes) with candidate line segments between the hits (corresponding to edges). Detector information can be associated with nodes and edges, enabling a GNN to propagate the embedded parameters around the graph and predict node-, edge- and graph-level observables. Previously, message-passing GNNs have shown success in predicting doublet likelihood, and we here report updates on the state-of-the-art architectures for this task. In addition, the Exa.TrkX project has investigated innovations in both graph construction, and embedded representations, in an effort to achieve fully learned end-to-end track finding. Hence, we present a suite of extensions to the original model, with encouraging results for hitgraph classification. In addition, we explore increased performance by constructing graphs from learned representations which contain non-linear metric structure, allowing for efficient clustering and neighborhood queries of data points. We demonstrate how this framework fits in with both traditional clustering pipelines, and GNN approaches. The embedded graphs feed into high-accuracy doublet and triplet classifiers, or can be used as an end-to-end track classifier by clustering in an embedded space. A set of post-processing methods improve performance with knowledge of the detector physics. Finally, we present numerical results on the TrackML particle tracking challenge dataset, where our framework shows favorable results in both seeding and track finding.