LazyVLM: Neuro-Symbolic Approach to Video Analytics
Jian, Xiangru, Pang, Wei, Dong, Zhengyuan, Zhang, Chao, Özsu, M. Tamer
–arXiv.org Artificial Intelligence
Current video analytics approaches face a fundamental trade-off between flexibility and efficiency. End-to-end Vision Language Models (VLMs) often struggle with long-context processing and incur high computational costs, while neural-symbolic methods depend heavily on manual labeling and rigid rule design. In this paper, we introduce LazyVLM, a neuro-symbolic video analytics system that provides a user-friendly query interface similar to VLMs, while addressing their scalability limitation. LazyVLM enables users to effortlessly drop in video data and specify complex multi-frame video queries using a semi-structured text interface for video analytics. To address the scalability limitations of VLMs, LazyVLM decomposes multi-frame video queries into fine-grained operations and offloads the bulk of the processing to efficient relational query execution and vector similarity search. We demonstrate that LazyVLM provides a robust, efficient, and user-friendly solution for querying open-domain video data at scale.
arXiv.org Artificial Intelligence
May-28-2025
- Country:
- Europe > Switzerland (0.04)
- North America > United States
- New York > New York County
- New York City (0.04)
- Washington > King County
- Seattle (0.04)
- New York > New York County
- Genre:
- Research Report (0.41)
- Technology: