Goto

Collaborating Authors

 glitch


MEDIC: a network for monitoring data quality in collider experiments

Bassa, Juvenal, Chattopadhyay, Arghya, Malik, Sudhir, Rivera, Mario Escabi

arXiv.org Artificial Intelligence

Data Quality Monitoring (DQM) is a crucial component of particle physics experiments and ensures that the recorded data is of the highest quality, and suitable for subsequent physics analysis. Due to the extreme environmental conditions, unprecedented data volumes, and the sheer scale and complexity of the detectors, DQM orchestration has become a very challenging task. Therefore, the use of Machine Learning (ML) to automate anomaly detection, improve efficiency, and reduce human error in the process of collecting high-quality data is unavoidable. Since DQM relies on real experimental data, it is inherently tied to the specific detector substructure and technology in operation. In this work, a simulation-driven approach to DQM is proposed, enabling the study and development of data-quality methodologies in a controlled environment. Using a modified version of Delphes -- a fast, multi-purpose detector simulation -- the preliminary realization of a framework is demonstrated which leverages ML to identify detector anomalies as well as localize the malfunctioning components responsible. We introduce MEDIC (Monitoring for Event Data Integrity and Consistency), a neural network designed to learn detector behavior and perform DQM tasks to look for potential faults. Although the present implementation adopts a simplified setup for computational ease, where large detector regions are deliberately deactivated to mimic faults, this work represents an initial step toward a comprehensive ML-based DQM framework. The encouraging results underline the potential of simulation-driven studies as a foundation for developing more advanced, data-driven DQM systems for future particle detectors.


Learning and Interpreting Gravitational-Wave Features from CNNs with a Random Forest Approach

Tian, Jun, Wang, He, He, Jibo, Pan, Yu, Cao, Shuo, Jiang, Qingquan

arXiv.org Artificial Intelligence

Convolutional neural networks (CNNs) have become widely adopted in gravitational wave (GW) detection pipelines due to their ability to automatically learn hierarchical features from raw strain data. However, the physical meaning of these learned features remains underexplored, limiting the interpretability of such models. In this work, we propose a hybrid architecture that combines a CNN-based feature extractor with a random forest (RF) classifier to improve both detection performance and interpretability. Unlike prior approaches that directly connect classifiers to CNN outputs, our method introduces four physically interpretable metrics - variance, signal-to-noise ratio (SNR), waveform overlap, and peak amplitude - computed from the final convolutional layer. These are jointly used with the CNN output in the RF classifier to enable more informed decision boundaries. Tested on long-duration strain datasets, our hybrid model outperforms a baseline CNN model, achieving a relative improvement of 21\% in sensitivity at a fixed false alarm rate of 10 events per month. Notably, it also shows improved detection of low-SNR signals (SNR $\le$ 10), which are especially vulnerable to misclassification in noisy environments. Feature attribution via the RF model reveals that both CNN-extracted and handcrafted features contribute significantly to classification decisions, with learned variance and CNN outputs ranked among the most informative. These findings suggest that physically motivated post-processing of CNN feature maps can serve as a valuable tool for interpretable and efficient GW detection, bridging the gap between deep learning and domain knowledge.


Vision Transformer for Transient Noise Classification

Srivastava, Divyansh, Niedzielski, Andrzej

arXiv.org Artificial Intelligence

Transient noise (glitches) in LIGO data hinders the detection of gravitational waves (GW). The Gravity Spy project has categorized these noise events into various classes. With the O3 run, there is the inclusion of two additional noise classes and thus a need to train new models for effective classification. We aim to classify glitches in LIGO data into 22 existing classes from the first run plus 2 additional noise classes from O3a using the Vision Transformer (ViT) model. We train a pre-trained Vision Transformer (ViT-B/32) model on a combined dataset consisting of the Gravity Spy dataset with the additional two classes from the LIGO O3a run. We achieve a classification efficiency of 92.26%, demonstrating the potential of Vision Transformer to improve the accuracy of gravitational wave detection by effectively distinguishing transient noise. Key words: gravitational waves --vision transformer --machine learning


Chinese 'Virtual Human' Salespeople Are Outperforming Their Real Human Counterparts

WIRED

The salesperson hawking Brother printers on Taobao works hard--like, really hard. At any time of the day, even when there's no audience on the Chinese ecommerce platform, the same woman wearing a white shirt and black skirt is always livestreaming, boasting about the various features of different office printers. She has a phone in one hand and often checks it as if to read a sales script or monitor the viewer comments coming in. "My friends, I've gotta plug this game-changing office tool that can double your workplace efficiency, " the salesperson said during one recent broadcast, trying to achieve the delicate balance between friendliness and precision that has come to define the billion-dollar livestream ecommerce industry in China. Occasionally, she greeted the invisible audience.


Glitches in Decision Tree Ensemble Models

Chandra, Satyankar, Gupta, Ashutosh, Mallik, Kaushik, Shankaranarayanan, Krishna, Varshney, Namrita

arXiv.org Machine Learning

Many critical decision-making tasks are now delegated to machine-learned models, and it is imperative that their decisions are trustworthy and reliable, and their outputs are consistent across similar inputs. We identify a new source of unreliable behaviors-called glitches-which may significantly impair the reliability of AI models having steep decision boundaries. Roughly speaking, glitches are small neighborhoods in the input space where the model's output abruptly oscillates with respect to small changes in the input. We provide a formal definition of glitches, and use well-known models and datasets from the literature to demonstrate that they have widespread existence and argue they usually indicate potential model inconsistencies in the neighborhood of where they are found. We proceed to the algorithmic search of glitches for widely used gradient-boosted decision tree (GBDT) models. We prove that the problem of detecting glitches is NP-complete for tree ensembles, already for trees of depth 4. Our glitch-search algorithm for GBDT models uses an MILP encoding of the problem, and its effectiveness and computational feasibility are demonstrated on a set of widely used GBDT benchmarks taken from the literature.


Apple to fix iPhone dictation bug that replaces word 'racist' with 'Trump'

The Guardian

Apple has promised to fix a bug in its iPhone automatic dictation tool after some users reported it had suggested to them "Trump" when they said the word "racist". The glitch was first highlighted in a viral post on TikTok, when the speech-to-text tool sometimes briefly flashed up the word "Trump" when they said "racist", and was later repeated by others on social media. "We are aware of an issue with the speech recognition model that powers dictation and we are rolling out a fix," an Apple spokesperson said. The company blamed the bug on its tool displaying words that have "phonetic overlap" before the "intended word" is identified, which in this case included words with the "r" consonant. However, the glitch caused outrage among some conservative commentators in the US, who have long accused big tech companies of political bias against those on the right.


DeepExtractor: Time-domain reconstruction of signals and glitches in gravitational wave data with deep learning

Dooney, Tom, Narola, Harsh, Bromuri, Stefano, Curier, R. Lyana, Broeck, Chris Van Den, Caudill, Sarah, Tan, Daniel Stanley

arXiv.org Artificial Intelligence

Gravitational wave (GW) interferometers, detect faint signals from distant astrophysical events, such as binary black hole mergers. However, their high sensitivity also makes them susceptible to background noise, which can obscure these signals. This noise often includes transient artifacts called "glitches" that can mimic astrophysical signals or mask their characteristics. Fast and accurate reconstruction of both signals and glitches is crucial for reliable scientific inference. In this study, we present DeepExtractor, a deep learning framework designed to reconstruct signals and glitches with power exceeding interferometer noise, regardless of their source. We design DeepExtractor to model the inherent noise distribution of GW interferometers, following conventional assumptions that the noise is Gaussian and stationary over short time scales. It operates by predicting and subtracting the noise component of the data, retaining only the clean reconstruction. Our approach achieves superior generalization capabilities for arbitrary signals and glitches compared to methods that directly map inputs to the clean training waveforms. We validate DeepExtractor's effectiveness through three experiments: (1) reconstructing simulated glitches injected into simulated detector noise, (2) comparing performance with the state-of-the-art BayesWave algorithm, and (3) analyzing real data from the Gravity Spy dataset to demonstrate effective glitch subtraction from LIGO strain data. DeepExtractor achieves a median mismatch of only 0.9% for simulated glitches, outperforming several deep learning baselines. Additionally, DeepExtractor surpasses BayesWave in glitch recovery, offering a dramatic computational speedup by reconstructing one glitch sample in approx. 0.1 seconds on a CPU, compared to BayesWave's processing time of approx. one hour per glitch.


Enhancing the Reliability in Machine Learning for Gravitational Wave Parameter Estimation with Attention-Based Models

Iwanaga, Hibiki, Matsuyama, Mahoro, Itoh, Yousuke

arXiv.org Artificial Intelligence

We introduce a technique to enhance the reliability of gravitational wave parameter estimation results produced by machine learning. We develop two independent machine learning models based on the Vision Transformer to estimate effective spin and chirp mass from spectrograms of gravitational wave signals from binary black hole mergers. To enhance the reliability of these models, we utilize attention maps to visualize the areas our models focus on when making predictions. This approach enables demonstrating that both models perform parameter estimation based on physically meaningful information. Furthermore, by leveraging these attention maps, we demonstrate a method to quantify the impact of glitches on parameter estimation. We show that as the models focus more on glitches, the parameter estimation results become more strongly biased. This suggests that attention maps could potentially be used to distinguish between cases where the results produced by the machine learning model are reliable and cases where they are not.


A Neural Network-Based Search for Unmodeled Transients in LIGO-Virgo-KAGRA's Third Observing Run

Raikman, Ryan, Moreno, Eric A., Govorkova, Katya, Soni, Siddharth, Marx, Ethan, Benoit, William, Gunny, Alec, Chatterjee, Deep, Reissel, Christina, Desai, Malina M., Omer, Rafia, Saleem, Muhammed, Harris, Philip, Katsavounidis, Erik, Coughlin, Michael W., Rankin, Dylan

arXiv.org Artificial Intelligence

This paper presents the results of a Neural Network (NN)-based search for short-duration gravitational-wave transients in data from the third observing run of LIGO, Virgo, and KAGRA. The search targets unmodeled transients with durations of milliseconds to a few seconds in the 30-1500 Hz frequency band, without assumptions about the incoming signal direction, polarization, or morphology. Using the Gravitational Wave Anomalous Knowledge (GWAK) method, three compact binary coalescences (CBCs) identified by existing pipelines are successfully detected, along with a range of detector glitches. The algorithm constructs a low-dimensional embedded space to capture the physical features of signals, enabling the detection of CBCs, detector glitches, and unmodeled transients. This study demonstrates GWAK's ability to enhance gravitational-wave searches beyond the limits of existing pipelines, laying the groundwork for future detection strategies.


ChatGPT's refusal to acknowledge 'David Mayer' down to glitch, says OpenAI

The Guardian

Last weekend the name was all over the internet – just not on ChatGPT. David Mayer became famous for a moment on social media because the popular chatbot appeared to want nothing to do with him. Legions of chatbot wranglers spent days trying – and failing – to make ChatGPT write the words "David Mayer". But the chatbot refused to comply, with replies alternating between "something seems to have gone wrong" to "I'm unable to produce a response" or just stopping at "David". This produced a blizzard of online speculation about Mayer's identity.