self-supervised
- North America > United States > Wisconsin > Dane County > Madison (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Europe > Switzerland (0.04)
- Asia > India (0.04)
Self-supervised learning through the eyes of a child
Within months of birth, children develop meaningful expectations about the world around them. How much of this early knowledge can be explained through generic learning mechanisms applied to sensory data, and how much of it requires more substantive innate inductive biases? Addressing this fundamental question in its full generality is currently infeasible, but we can hope to make real progress in more narrowly defined domains, such as the development of high-level visual categories, thanks to improvements in data collecting technology and recent progress in deep learning. In this paper, our goal is precisely to achieve such progress by utilizing modern self-supervised deep learning methods and a recent longitudinal, egocentric video dataset recorded from the perspective of three young children (Sullivan et al., 2020). Our results demonstrate the emergence of powerful, high-level visual representations from developmentally realistic natural videos using generic self-supervised learning objectives.
propagation on the DAVIS dataset (Table 1), in comparison to a SOT A3 self-supervised method [49] and the ImageNet pre-trained representation
Model J (Mean) Self-supervised, SOT A [49] 43.0 ImageNet Representation 49.4 Self-supervised, Ours 57.7 The shared affinity matrix bridges these tasks, and facilitates iterative improvements. These contributions are significant in the field of self-supervised learning. The contributions of this work are also demonstrated by our ablation study, i.e., Table 2 in the paper. We note that these components are novel and have not been explored in prior work. In the following, we address the other comments by reviewers.
Self-supervised learning for soccer ball detection and beyond: interview with winners of the RoboCup 2025 best paper award
This is the focus of work by and, which won the best paper award at the recent RoboCup symposium . The symposium takes place alongside the annual RoboCup competition, which this year was held in Salvador, Brazil. We caught up with some of the authors to find out more about the work, how their method can be transferred to applications beyond RoboCup, and their future plans for the competition. Could you start by giving us a brief description of the problem that you were trying to solve in your paper "Self-supervised Feature Extraction for Enhanced Ball Detection on Soccer Robots"? The main challenge we faced was that deep learning generally requires a large amount of labeled data. This is not a major problem for common tasks that have already been studied, because you can usually find labeled datasets online.
- South America > Brazil > Bahia > Salvador (0.24)
- Europe > Italy > Basilicata (0.05)
- Europe > Switzerland > Zürich > Zürich (0.04)
- Asia > China > Beijing > Beijing (0.04)
Self-supervised learning predicts plant growth trajectories from multi-modal industrial greenhouse data
Riesselman, Adam J, Cofer, Evan M, LaRue, Therese, Meeussen, Wim
Quantifying organism-level phenotypes, such as growth dynamics and biomass accumulation, is fundamental to understanding agronomic traits and optimizing crop production. However, quality growing data of plants at scale is difficult to generate. Here we use a mobile robotic platform to capture high-resolution environmental sensing and phenotyping measurements of a large-scale hydroponic leafy greens system. We describe a self-supervised modeling approach to build a map from observed growing data to the entire plant growth trajectory. We demonstrate our approach by forecasting future plant height and harvest mass of crops in this system. This approach represents a significant advance in combining robotic automation and machine learning, as well as providing actionable insights for agronomic research and operational efficiency.
- North America > United States > Montana (0.04)
- North America > Canada (0.04)
- Europe > France (0.04)
- Asia > Middle East > Jordan (0.04)
Self-supervised learning through the eyes of a child
Within months of birth, children develop meaningful expectations about the world around them. How much of this early knowledge can be explained through generic learning mechanisms applied to sensory data, and how much of it requires more substantive innate inductive biases? Addressing this fundamental question in its full generality is currently infeasible, but we can hope to make real progress in more narrowly defined domains, such as the development of high-level visual categories, thanks to improvements in data collecting technology and recent progress in deep learning. In this paper, our goal is precisely to achieve such progress by utilizing modern self-supervised deep learning methods and a recent longitudinal, egocentric video dataset recorded from the perspective of three young children (Sullivan et al., 2020). Our results demonstrate the emergence of powerful, high-level visual representations from developmentally realistic natural videos using generic self-supervised learning objectives.
Not All Data are Good Labels: On the Self-supervised Labeling for Time Series Forecasting
Yang, Yuxuan, Zhang, Dalin, Liang, Yuxuan, Lu, Hua, Chen, Gang, Li, Huan
Time Series Forecasting (TSF) is a crucial task in various domains, yet existing TSF models rely heavily on high-quality data and insufficiently exploit all available data. This paper explores a novel self-supervised approach to re-label time series datasets by inherently constructing candidate datasets. During the optimization of a simple reconstruction network, intermediates are used as pseudo labels in a self-supervised paradigm, improving generalization for any predictor. We introduce the Self-Correction with Adaptive Mask (SCAM), which discards overfitted components and selectively replaces them with pseudo labels generated from reconstructions. Additionally, we incorporate Spectral Norm Regularization (SNR) to further suppress overfitting from a loss landscape perspective. Our experiments on eleven real-world datasets demonstrate that SCAM consistently improves the performance of various backbone models. This work offers a new perspective on constructing datasets and enhancing the generalization of TSF models through self-supervised learning.
- Pacific Ocean > North Pacific Ocean > San Francisco Bay (0.04)
- North America > United States > California > San Francisco County > San Francisco (0.04)
- Europe > Denmark > North Jutland > Aalborg (0.04)
- (3 more...)
- Information Technology > Data Science > Data Mining (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.67)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)
Review for NeurIPS paper: Self-supervised learning through the eyes of a child
Weaknesses: - I expected to see the linear evaluation performance on ImageNet can be impressive. However, it's a pity to see this transfer learning's performance is poor with only TC-S: 20.9% at best. This seriously limits the impact of this work. If the model can only perform well on some easy datasets that are close to the SAYCam, we cannot get too much benefits from learning on such datasets, especially with access to so many big datasets. Maybe the authors can change the SAYCam to other standard videos (Charades) and see if they can have good transfer learning performances.
Review for NeurIPS paper: Self-supervised learning through the eyes of a child
This is an interesting paper combining machine learning and psychology, and brings interesting insights about what can be learned from naturalistic, egocentric, real-world datasets, and how the learned representations can be used on downstream tasks. It's well-written and clearly presented, and likely of interest to the general NeurIPS audience. Reviewers 3 and 4 initially had concerns about the model's performance, and suggestions about using other datasets. After discussion and the rebuttal, they were able to be convinced that these run counter to the main motivation of the study. They subsequently raised their scores and all reviewers - and myself - are in agreement to accept.
Self-supervised learning through the eyes of a child
Within months of birth, children develop meaningful expectations about the world around them. How much of this early knowledge can be explained through generic learning mechanisms applied to sensory data, and how much of it requires more substantive innate inductive biases? Addressing this fundamental question in its full generality is currently infeasible, but we can hope to make real progress in more narrowly defined domains, such as the development of high-level visual categories, thanks to improvements in data collecting technology and recent progress in deep learning. In this paper, our goal is precisely to achieve such progress by utilizing modern self-supervised deep learning methods and a recent longitudinal, egocentric video dataset recorded from the perspective of three young children (Sullivan et al., 2020). Our results demonstrate the emergence of powerful, high-level visual representations from developmentally realistic natural videos using generic self-supervised learning objectives.