Goto

Collaborating Authors

 led


TinyML for Speech Recognition

Barovic, Andrew, Moin, Armin

arXiv.org Artificial Intelligence

--We train and deploy a quantized 1D convolutional neural network model to conduct speech recognition on a highly resource-constrained IoT edge device. This can be useful in various Internet of Things (IoT) applications, such as smart homes and ambient assisted living for the elderly and people with disabilities, just to name a few examples. In this paper, we first create a new dataset with over one hour of audio data that enables our research and will be useful to future studies in this field. Second, we utilize the technologies provided by Edge Impulse to enhance our model's performance and achieve a high Accuracy of up to 97% on our dataset. For the validation, we implement our prototype using the Arduino Nano 33 BLE Sense microcontroller board. This microcontroller board is specifically designed for IoT and AI applications, making it an ideal choice for our target use case scenarios. While most existing research focuses on a limited set of keywords, our model can process 23 different keywords, enabling complex commands. Natural Language Processing (NLP) and Speech Recognition are crucial domains in Artificial Intelligence (AI). While NLP deals with enabling computers to analyze, understand, reason on, and generate human language in textual form, speech recognition is concerned with that in spoken form.


The Best Artificial Christmas Trees, as Blind-Judged By Interior Designers

WIRED

WIRED brought 10 of the most popular artificial Christmas trees into a studio and got three interior designers to pick the best through blind judging. For extra trimming, we checked in on how those trees fared once they were taken home and decorated. Shopping for an artificial Christmas tree can be overwhelming, especially when you're doing it online. You'll find yourself staring at product photos, wondering: How realistic does it look? Will it shed all over my living room? Can you see daylight through the branches? Are the branches strong enough to hold that lopsided homemade macaroni ornament you've hung on your tree since 2004? We got tired of guessing, so we did a little experiment. We brought 10 of the most popular artificial trees from three top brands (Balsam Hill, King of Christmas, and National Tree Company) and hauled them to a photo studio in Kansas.


The Great Tree Test: Best Artificial Christmas Trees 2025

WIRED

We brought 10 of the most popular artificial Christmas trees into a studio, had volunteers assemble them, then got three interior designers to pick the best through blind judging. All products featured on WIRED are independently selected by our editors. However, we may receive compensation from retailers and/or from purchases of products through these links. You can spend hours scrolling through lists of the best artificial Christmas trees and still end up wondering what to buy. How real does it look? Are the branches strong enough to hold that lopsided homemade macaroni ornament you've hung on your tree since 2004? We decided to settle the debate once and for all by bringing the best-selling artificial trees from three leading brands into a studio for a blind-judged contest. We got 10 trees from Balsam Hill, King of Christmas, and National Tree Company, then found 10 assemblers to put the trees together and fluff them.


Dual-Mode Visual System for Brain-Computer Interfaces: Integrating SSVEP and P300 Responses

Kasawala, Ekgari, Mouli, Surej

arXiv.org Artificial Intelligence

In brain-computer interface (BCI) systems, steady-state visual evoked potentials (SSVEP) and P300 responses have achieved widespread implementation owing to their superior information transfer rates (ITR) and minimal training requirements. These neurophysiological signals have exhibited robust efficacy and versatility in external device control, demonstrating enhanced precision and scalability. However, conventional implementations predominantly utilise liquid crystal display (LCD)-based visual stimulation paradigms, which present limitations in practical deployment scenarios. This investigation presents the development and evaluation of a novel light-emitting diode (LED)-based dual stimulation apparatus designed to enhance SSVEP classification accuracy through the integration of both SSVEP and P300 paradigms. The system employs four distinct frequencies, 7 Hz, 8 Hz, 9 Hz, and 10 Hz, corresponding to forward, backward, right, and left directional controls, respectively. Oscilloscopic verification confirmed the precision of these stimulation frequencies. Real-time feature extraction was accomplished through the concurrent analysis of maximum Fast Fourier Transform (FFT) amplitude and P300 peak detection to ascertain user intent. Directional control was determined by the frequency exhibiting maximal amplitude characteristics. The visual stimulation hardware demonstrated minimal frequency deviation, with error differentials ranging from 0.15%to 0.20%across all frequencies. The implemented signal processing algorithm successfully discriminated all four stimulus frequencies whilst correlating them with their respective P300 event markers. Classification accuracy was evaluated based on correct task intention recognition. The proposed hybrid system achieved a mean classification accuracy of 86.25%, coupled with an average ITR of 42.08 bits per minute (bpm).


Self-supervised Learning Of Visual Pose Estimation Without Pose Labels By Classifying LED States

Carlotti, Nicholas, Nava, Mirko, Giusti, Alessandro

arXiv.org Artificial Intelligence

We introduce a model for monocular RGB relative pose estimation of a ground robot that trains from scratch without pose labels nor prior knowledge about the robot's shape or appearance. At training time, we assume: (i) a robot fitted with multiple LEDs, whose states are independent and known at each frame; (ii) knowledge of the approximate viewing direction of each LED; and (iii) availability of a calibration image with a known target distance, to address the ambiguity of monocular depth estimation. Training data is collected by a pair of robots moving randomly without needing external infrastructure or human supervision. Our model trains on the task of predicting from an image the state of each LED on the robot. In doing so, it learns to predict the position of the robot in the image, its distance, and its relative bearing. At inference time, the state of the LEDs is unknown, can be arbitrary, and does not affect the pose estimation performance. Quantitative experiments indicate that our approach: is competitive with SoA approaches that require supervision from pose labels or a CAD model of the robot; generalizes to different domains; and handles multi-robot pose estimation.


Vibe2Spike: Batteryless Wireless Tags for Vibration Sensing with Event Cameras and Spiking Networks

Scott, Danny, LaForest, William, Das, Hritom, Polykretis, Ioannis, Schuman, Catherine D., Rizzo, Charles, Plank, James, Swaminathan, Sai

arXiv.org Artificial Intelligence

--The deployment of dense, low-cost sensors is critical for realizing ubiquitous smart environments. However, existing sensing solutions struggle with the energy, scalability, and reliability trade-offs imposed by battery maintenance, wireless transmission overhead, and data processing complexity. In this work, we present Vibe2Spike, a novel battery-free, wireless sensing framework that enables vibration-based activity recognition using visible light communication (VLC) and spiking neural networks (SNNs). Our system uses ultra-low-cost tags composed only of a piezoelectric disc, a Zener diode, and an LED, which harvest vibration energy and emit sparse visible light spikes without requiring batteries or RF radios. These optical spikes are captured by event cameras and classified using optimized SNN models evolved via the EONS framework. We evaluate Vibe2Spike across five device classes, achieving 94.9% average classification fitness while analyzing the latency-accuracy trade-offs of different temporal binning strategies. Vibe2Spike demonstrates a scalable, and energy-efficient approach for enabling intelligent environments in a batteryless manner . The promise of ubiquitous smart environments--industrial workshops predicting machinery faults, kitchens monitoring appliance usage, or buildings diagnosing structural wear--remains hindered by the energy-cost-reliability tradeoffs of conventional sensing systems. While vibrations offer rich insights into machine health, human activity, etc., deploying sensors at scale requires solutions that are batteryless (to eliminate maintenance), wireless (to simplify installation), and ultra-low-cost (to enable dense deployments).


Omnidirectional vision sensors based on catadioptric systems with discrete infrared photoreceptors for swarm robotics

Contreras-Monsalvo, Jose Fernando, Dossetti, Victor, Soto-Cruz, Blanca Susana

arXiv.org Artificial Intelligence

In this work, we fabricated and studied two designs for omnidirectional vision sensors for swarm robotics, based on catadioptric systems consisting of a mirror with rotational symmetry, eight discrete infrared photodiodes and a single LED, in order to provide localization and navigation abilities for mobile robotic agents. We considered two arrangements for the photodiodes: one in which they point upward into the mirror, and one in which they point outward, perpendicular to the mirror. To determine which design offers a better field of view on the plane, as well as detection of distance and orientation between two agents, we developed a test rail with three degrees of freedom to experimentally and systematically measure the signal registered by the photodiodes of a given sensor (in a single readout) from the light emitted by another as functions of the distance and orientation. Afterwards, we processed and analyzed the experimental data to develop mathematical models for the mean response of a photodiode in each design. Finally, by numerically inverting the models, we compared the two designs in terms of their accuracy. Our results show that the design with the photodiodes pointing upward resolves better the distance, while the other resolves better the orientation of the emitting agent, both providing an omnidirectional field of view. Keywords: computer vision, catadioptric sensors, swarm robotics 1. Introduction Localization is a key factor in the implementation of navigation and motion control in mobile autonomous robotic systems. Several methods can be implemented for this purpose.


A Survey on Event-based Optical Marker Systems

Tofighi, Nafiseh Jabbari, Robic, Maxime, Morbidi, Fabio, Vasseur, Pascal

arXiv.org Artificial Intelligence

The advent of event-based cameras, with their low latency, high dynamic range, and reduced power consumption, marked a significant change in robotic vision and machine perception. In particular, the combination of these neuromorphic sensors with widely-available passive or active optical markers (e.g. AprilTags, arrays of blinking LEDs), has recently opened up a wide field of possibilities. This survey paper provides a comprehensive review on Event-Based Optical Marker Systems (EBOMS). We analyze the basic principles and technologies on which these systems are based, with a special focus on their asynchronous operation and robustness against adverse lighting conditions. We also describe the most relevant applications of EBOMS, including object detection and tracking, pose estimation, and optical communication. The article concludes with a discussion of possible future research directions in this rapidly-emerging and multidisciplinary field.


BiasBench: A reproducible benchmark for tuning the biases of event cameras

Ziegler, Andreas, Joseph, David, Gossard, Thomas, Moldovan, Emil, Zell, Andreas

arXiv.org Artificial Intelligence

Event-based cameras are bio-inspired sensors that detect light changes asynchronously for each pixel. They are increasingly used in fields like computer vision and robotics because of several advantages over traditional frame-based cameras, such as high temporal resolution, low latency, and high dynamic range. As with any camera, the output's quality depends on how well the camera's settings, called biases for event-based cameras, are configured. While frame-based cameras have advanced automatic configuration algorithms, there are very few such tools for tuning these biases. A systematic testing framework would require observing the same scene with different biases, which is tricky since event cameras only generate events when there is movement. Event simulators exist, but since biases heavily depend on the electrical circuit and the pixel design, available simulators are not well suited for bias tuning. To allow reproducibility, we present BiasBench, a novel event dataset containing multiple scenes with settings sampled in a grid-like pattern. We present three different scenes, each with a quality metric of the downstream application. Additionally, we present a novel, RL-based method to facilitate online bias adjustments.


A Monocular Event-Camera Motion Capture System

Bauersfeld, Leonard, Scaramuzza, Davide

arXiv.org Artificial Intelligence

Motion capture systems are a widespread tool in research to record ground-truth poses of objects. Commercial systems use reflective markers attached to the object and then triangulate pose of the object from multiple camera views. Consequently, the object must be visible to multiple cameras which makes such multi-view motion capture systems unsuited for deployments in narrow, confined spaces (e.g. ballast tanks of ships). In this technical report we describe a monocular event-camera motion capture system which overcomes this limitation and is ideally suited for narrow spaces. Instead of passive markers it relies on active, blinking LED markers such that each marker can be uniquely identified from the blinking frequency. The markers are placed at known locations on the tracking object. We then solve the PnP (perspective-n-points) problem to obtain the position and orientation of the object. The developed system has millimeter accuracy, millisecond latency and we demonstrate that its state estimate can be used to fly a small, agile quadrotor.