Goto

Collaborating Authors

 rsc



Time-Aware Synthetic Control

Rho, Saeyoung, Illick, Cyrus, Narasipura, Samhitha, Abadie, Alberto, Hsu, Daniel, Misra, Vishal

arXiv.org Machine Learning

The synthetic control (SC) framework is widely used for observational causal inference with time-series panel data. SC has been successful in diverse applications, but existing methods typically treat the ordering of pre-intervention time indices interchangeable. This invariance means they may not fully take advantage of temporal structure when strong trends are present. We propose Time-Aware Synthetic Control (TASC), which employs a state-space model with a constant trend while preserving a low-rank structure of the signal. TASC uses the Kalman filter and Rauch-Tung-Striebel smoother: it first fits a generative time-series model with expectation-maximization and then performs counterfactual inference. We evaluate TASC on both simulated and real-world datasets, including policy evaluation and sports prediction. Our results suggest that TASC offers advantages in settings with strong temporal trends and high levels of observation noise.



StableMotion: Repurposing Diffusion-Based Image Priors for Motion Estimation

Wang, Ziyi, Li, Haipeng, Sui, Lin, Zhou, Tianhao, Jiang, Hai, Nie, Lang, Liu, Shuaicheng

arXiv.org Artificial Intelligence

We present StableMotion, a novel framework leverages knowledge (geometry and content priors) from pretrained large-scale image diffusion models to perform motion estimation, solving single-image-based image rectification tasks such as Stitched Image Rectangling (SIR) and Rolling Shutter Correction (RSC). Specifically, StableMotion framework takes text-to-image Stable Diffusion (SD) models as backbone and repurposes it into an image-to-motion estimator. To mitigate inconsistent output produced by diffusion models, we propose Adaptive Ensemble Strategy (AES) that consolidates multiple outputs into a cohesive, high-fidelity result. Additionally, we present the concept of Sampling Steps Disaster (SSD), the counterintuitive scenario where increasing the number of sampling steps can lead to poorer outcomes, which enables our framework to achieve one-step inference. StableMotion is verified on two image rectification tasks and delivers state-of-the-art performance in both, as well as showing strong generalizability. Supported by SSD, StableMotion offers a speedup of 200 times compared to previous diffusion model-based methods.


BTS: Bridging Text and Sound Modalities for Metadata-Aided Respiratory Sound Classification

Kim, June-Woo, Toikkanen, Miika, Choi, Yera, Moon, Seoung-Eun, Jung, Ho-Young

arXiv.org Artificial Intelligence

Respiratory sound classification (RSC) is challenging due to varied acoustic signatures, primarily influenced by patient demographics and recording environments. To address this issue, we introduce a text-audio multimodal model that utilizes metadata of respiratory sounds, which provides useful complementary information for RSC. Specifically, we fine-tune a pretrained text-audio multimodal model using free-text descriptions derived from the sound samples' metadata which includes the gender and age of patients, type of recording devices, and recording location on the patient's body. Our method achieves state-of-the-art performance on the ICBHI dataset, surpassing the previous best result by a notable margin of 1.17%. This result validates the effectiveness of leveraging metadata and respiratory sound samples in enhancing RSC performance. Additionally, we investigate the model performance in the case where metadata is partially unavailable, which may occur in real-world clinical setting.


Regularized Spectral Clustering under the Degree-Corrected Stochastic Blockmodel

Neural Information Processing Systems

Spectral clustering is a fast and popular algorithm for finding clusters in networks. Recently, Chaudhuri et al. [1] and Amini et al. [2] proposed inspired variations on the algorithm that artificially inflate the node degrees for improved statistical performance. The current paper extends the previous statistical estimation results to the more canonical spectral clustering algorithm in a way that removes any assumption on the minimum degree and provides guidance on the choice of the tuning parameter. Moreover, our results show how the "star shape" in the eigenvectors-a common feature of empirical networks-can be explained by the Degree-Corrected Stochastic Blockmodel and the Extended Planted Partition model, two statistical models that allow for highly heterogeneous degrees. Throughout, the paper characterizes and justifies several of the variations of the spectral clustering algorithm in terms of these models.


RSC: Accelerating Graph Neural Networks Training via Randomized Sparse Computations

Liu, Zirui, Chen, Shengyuan, Zhou, Kaixiong, Zha, Daochen, Huang, Xiao, Hu, Xia

arXiv.org Artificial Intelligence

The training of graph neural networks (GNNs) is extremely time consuming because sparse graph-based operations are hard to be accelerated by hardware. Prior art explores trading off the computational precision to reduce the time complexity via sampling-based approximation. Based on the idea, previous works successfully accelerate the dense matrix based operations (e.g., convolution and linear) with negligible accuracy drop. However, unlike dense matrices, sparse matrices are stored in the irregular data format such that each row/column may have different number of non-zero entries. Thus, compared to the dense counterpart, approximating sparse operations has two unique challenges (1) we cannot directly control the efficiency of approximated sparse operation since the computation is only executed on non-zero entries; (2) sub-sampling sparse matrices is much more inefficient due to the irregular data format. To address the issues, our key idea is to control the accuracy-efficiency trade off by optimizing computation resource allocation layer-wisely and epoch-wisely. Specifically, for the first challenge, we customize the computation resource to different sparse operations, while limit the total used resource below a certain budget. For the second challenge, we cache previous sampled sparse matrices to reduce the epoch-wise sampling overhead. Finally, we propose a switching mechanisms to improve the generalization of GNNs trained with approximated operations. To this end, we propose Randomized Sparse Computation, which for the first time demonstrate the potential of training GNNs with approximated operations. In practice, rsc can achieve up to $11.6\times$ speedup for a single sparse operation and a $1.6\times$ end-to-end wall-clock time speedup with negligible accuracy drop.


Introducing the AI Research SuperCluster -- Meta's cutting-edge AI supercomputer for AI research

#artificialintelligence

Developing the next generation of advanced AI will require powerful new computers capable of quintillions of operations per second. Today, Meta is announcing that we've designed and built the AI Research SuperCluster (RSC) -- which we believe is among the fastest AI supercomputers running today and will be the fastest AI supercomputer in the world when it's fully built out in mid-2022. Our researchers have already started using RSC to train large models in natural language processing (NLP) and computer vision for research, with the aim of one day training models with trillions of parameters. RSC will help Meta's AI researchers build new and better AI models that can learn from trillions of examples; work across hundreds of different languages; seamlessly analyze text, images, and video together; develop new augmented reality tools; and much more. Our researchers will be able to train the largest models needed to develop advanced AI for computer vision, NLP, speech recognition, and more.


Best Artificial Intelligence Stocks To Buy Now? 4 To Check Out

#artificialintelligence

As more and more companies adopt artificial intelligence (AI) in their operations, investors may be considering AI stocks in the stock market. After all, AI remains to be an increasingly relevant area of research now. For the uninitiated, AI deals with machines that attempt to emulate human intelligence. Most of us may think of sentient robots and futuristic gadgets upon hearing the term. But in reality, AI often revolves around complex algorithms and software which enable enterprises to make more effective decisions.


Meta's RSC supercomputer brings revolutionary power -- and privacy and bias concerns

#artificialintelligence

Last month, Meta (formerly Facebook) announced that it had developed a supercomputer known as the AI Research SuperCluster (RSC). The company claims that once completed by the end of the year, it will be one of the world's fastest AI supercomputers. The technology and social media giant says the RSC will build a strong backbone of compute capability and AI to power the metaverse. The company claims that the RSC's advanced compute capabilities will help address concerns the company has been criticized for, such as identifying harmful content and algorithmic bias. To address these concerns, Meta detailed in a blog post how it plans to safeguard the privacy of user data that this AI-powered RSC will be trained on.