Kumar, Himanshu
PNN: A Novel Progressive Neural Network for Fault Classification in Rotating Machinery under Small Dataset Constraint
Chopra, Praveen, Kumar, Himanshu, Yadav, Sandeep
Fault detection in rotating machinery is a complex task, particularly in small and heterogeneous dataset scenarios. Variability in sensor placement, machinery configurations, and structural differences further increase the complexity of the problem. Conventional deep learning approaches often demand large, homogeneous datasets, limiting their applicability in data-scarce industrial environments. While transfer learning and few-shot learning have shown potential, however, they are often constrained by the need for extensive fault datasets. This research introduces a unified framework leveraging a novel progressive neural network (PNN) architecture designed to address these challenges. The PNN sequentially estimates the fixed-size refined features of the higher order with the help of all previously estimated features and appends them to the feature set. This fixed-size feature output at each layer controls the complexity of the PNN and makes it suitable for effective learning from small datasets. The framework's effectiveness is validated on eight datasets, including six open-source datasets, one in-house fault simulator, and one real-world industrial dataset. The PNN achieves state-of-the-art performance in fault detection across varying dataset sizes and machinery types, highlighting superior generalization and classification capabilities.
NanoVLMs: How small can we go and still make coherent Vision Language Models?
Agarwalla, Mukund, Kumar, Himanshu, Dandekar, Raj, Dandekar, Rajat, Panat, Sreedath
Vision-Language Models (VLMs), such as GPT-4V and Llama 3.2 vision, have garnered significant research attention for their ability to leverage Large Language Models (LLMs) in multimodal tasks. However, their potential is constrained by inherent challenges, including proprietary restrictions, substantial computational demands, and limited accessibility. Smaller models, such as GIT and BLIP, exhibit marked limitations, often failing to generate coherent and consistent text beyond a few tokens, even with extensive training. This underscores a pivotal inquiry: how small can a VLM be and still produce fluent and consistent text? Drawing inspiration from the exceptional learning process of 3-4 year old children, who rely heavily on visual cues for understanding and communication, we introduce two novel datasets: ShortDesc (featuring concise image descriptions) and LongDesc (containing more detailed image descriptions). These datasets consist of image-text pairs where the text is restricted to the simple vocabulary and syntax typically used by young children, generated with a scaled- down model, GPT-4o. Using these datasets, we demonstrate that it is possible to train VLMs that are significantly smaller, up to 10 times smaller than state of the art(SOTA) small VLMs while maintaining architectural simplicity. To evaluate the outputs, we leverage GPT-4o to grade the text, as if stories written by students, on creativity, meaningfulness, and consistency, assigning scores out of 10. This method addresses limitations of standard benchmarks by accommodating unstructured outputs and providing a multidimensional evaluation of the model capabilities. Our findings contribute to the development of lightweight, accessible multimodal models for resource constrained environments.
3rd Workshop on Maritime Computer Vision (MaCVi) 2025: Challenge Results
Kiefer, Benjamin, Žust, Lojze, Muhovič, Jon, Kristan, Matej, Perš, Janez, Teršek, Matija, Desai, Uma Mudenagudi Chaitra, Wiliem, Arnold, Kreis, Marten, Akalwadi, Nikhil, Quan, Yitong, Zhong, Zhiqiang, Zhang, Zhe, Liu, Sujie, Chen, Xuran, Yang, Yang, Fabijanić, Matej, Ferreira, Fausto, Lee, Seongju, Lee, Junseok, Lee, Kyoobin, Yao, Shanliang, Guan, Runwei, Huang, Xiaoyu, Ni, Yi, Kumar, Himanshu, Feng, Yuan, Cheng, Yi-Ching, Lin, Tzu-Yu, Lee, Chia-Ming, Hsu, Chih-Chung, Sheikh, Jannik, Michel, Andreas, Gross, Wolfgang, Weinmann, Martin, Šarić, Josip, Lin, Yipeng, Yang, Xiang, Jiang, Nan, Lu, Yutang, Feng, Fei, Awad, Ali, Lucas, Evan, Saleem, Ashraf, Cheng, Ching-Heng, Lin, Yu-Fan, Lin, Tzu-Yu, Hsu, Chih-Chung
The 3rd Workshop on Maritime Computer Vision (MaCVi) 2025 addresses maritime computer vision for Unmanned Surface Vehicles (USV) and underwater. This report offers a comprehensive overview of the findings from the challenges. We provide both statistical and qualitative analyses, evaluating trends from over 700 submissions. All datasets, evaluation code, and the leaderboard are available to the public at https://macvi.org/workshop/macvi25.
Robust Loss Functions under Label Noise for Deep Neural Networks
Ghosh, Aritra, Kumar, Himanshu, Sastry, P. S.
In many applications of classifier learning, training data suffers from label noise. Deep networks are learned using huge training data where the problem of noisy labels is particularly relevant. The current techniques proposed for learning deep networks under label noise focus on modifying the network architecture and on algorithms for estimating true labels from noisy labels. An alternate approach would be to look for loss functions that are inherently noise-tolerant. For binary classification there exist theoretical results on loss functions that are robust to label noise. In this paper, we provide some sufficient conditions on a loss function so that risk minimization under that loss function would be inherently tolerant to label noise for multiclass classification problems. These results generalize the existing results on noise-tolerant loss functions for binary classification. We study some of the widely used loss functions in deep networks and show that the loss function based on mean absolute value of error is inherently robust to label noise. Thus standard back propagation is enough to learn the true classifier even under label noise. Through experiments, we illustrate the robustness of risk minimization with such loss functions for learning neural networks.
Robust Loss Functions under Label Noise for Deep Neural Networks
Ghosh, Aritra (Microsoft, Bangalore) | Kumar, Himanshu (Indian Institute of Science, Bangalore) | Sastry, P. S. (Indian Institute of Science, Bangalore)
In many applications of classifier learning, training data suffers from label noise. Deep networks are learned using huge training data where the problem of noisy labels is particularly relevant. The current techniques proposed for learning deep networks under label noise focus on modifying the network architecture and on algorithms for estimating true labels from noisy labels. An alternate approach would be to look for loss functions that are inherently noise-tolerant. For binary classification there exist theoretical results on loss functions that are robust to label noise. In this paper, we provide some sufficient conditions on a loss function so that risk minimization under that loss function would be inherently tolerant to label noise for multiclass classification problems. These results generalize the existing results on noise-tolerant loss functions for binary classification. We study some of the widely used loss functions in deep networks and show that the loss function based on mean absolute value of error is inherently robust to label noise. Thus standard back propagation is enough to learn the true classifier even under label noise. Through experiments, we illustrate the robustness of risk minimization with such loss functions for learning neural networks.