Goto

Collaborating Authors

 axion




Physicist Frank Wilczek's unique insights on the nature of reality

New Scientist

In June, at a conference set in the picturesque Italian town of Campagna, south-east of Naples, two physicists in a seemingly endless argument over a long-sought theory of fundamental reality caught my attention. From the sidelines, an unassuming figure politely interrupted them. "I've got a slide that might help. Can I put it up?" asked Frank Wilczek. The slide, concisely describing the realms in which this theory may act, swiftly ended the dispute.


Paxion: Patching Action Knowledge in Video-Language Foundation Models

Wang, Zhenhailong, Blume, Ansel, Li, Sha, Liu, Genglin, Cho, Jaemin, Tang, Zineng, Bansal, Mohit, Ji, Heng

arXiv.org Artificial Intelligence

Action knowledge involves the understanding of textual, visual, and temporal aspects of actions. We introduce the Action Dynamics Benchmark (ActionBench) containing two carefully designed probing tasks: Action Antonym and Video Reversal, which targets multimodal alignment capabilities and temporal understanding skills of the model, respectively. Despite recent video-language models' (VidLM) impressive performance on various benchmark tasks, our diagnostic tasks reveal their surprising deficiency (near-random performance) in action knowledge, suggesting that current models rely on object recognition abilities as a shortcut for action understanding. To remedy this, we propose a novel framework, Paxion, along with a new Discriminative Video Dynamics Modeling (DVDM) objective. The Paxion framework utilizes a Knowledge Patcher network to encode new action knowledge and a Knowledge Fuser component to integrate the Patcher into frozen VidLMs without compromising their existing capabilities. Due to limitations of the widely-used Video-Text Contrastive (VTC) loss for learning action knowledge, we introduce the DVDM objective to train the Knowledge Patcher. DVDM forces the model to encode the correlation between the action text and the correct ordering of video frames. Our extensive analyses show that Paxion and DVDM together effectively fill the gap in action knowledge understanding (~50% to 80%), while maintaining or improving performance on a wide spectrum of both object- and action-centric downstream tasks. The code and data will be made publicly available for research purposes at https://github.com/MikeWangWZHL/Paxion.git.


Evolution of ML Fact Store

#artificialintelligence

At Netflix, we aim to provide recommendations that match our members' interests. To achieve this, we rely on Machine Learning (ML) algorithms. ML algorithms can be only as good as the data that we provide to it. This post will focus on the large volume of high-quality data stored in Axion -- our fact store that is leveraged to compute ML features offline. We built Axion primarily to remove any training-serving skew and make offline experimentation faster. We will share how its design has evolved over the years and the lessons learned while building it.