Goto

Collaborating Authors

 Concord


Accelerating Local AI on Consumer GPUs: A Hardware-Aware Dynamic Strategy for YOLOv10s

Masum, Mahmudul Islam, Islam, Miad

arXiv.org Artificial Intelligence

Abstract--As local AI grows in popularity, there is a critical gap between the benchmark performance of object detectors and their practical viability on consumer-grade hardware. While models like YOLOv10s promise real-time speeds, these metrics are typically achieved on high-power, desktop-class GPUs. This paper reveals that on resource-constrained systems, such as laptops with RTX 4060 GPUs, performance is not compute-bound but is instead dominated by system-level bottlenecks, as illustrated by a simple bottleneck test. T o overcome this hardware-level constraint, we introduce a Two-Pass Adaptive Inference algorithm, a model-independent approach that requires no architectural changes. This study mainly focuses on'adaptive' inference strategies and undertakes a comparative analysis of architectural early-exit and resolution-adaptive routing, highlighting their respective trade-offs within a unified evaluation framework. The system uses a fast, low-resolution pass and only escalates to a high-resolution model pass when detection confidence is low. On a 5000-image COCO dataset, our method achieves a 1.85x speedup over a PyT orch Early-Exit baseline, with a modest mAP loss of 5.51%. This work provides a practical and reproducible blueprint for deploying high-performance, real-time AI on consumer-grade devices by shifting the focus from pure model optimization to hardware-aware inference strategies that maximize throughput.


Open Banking Foundational Model: Learning Language Representations from Few Financial Transactions

Polleti, Gustavo, Santana, Marlesson, Fontes, Eduardo

arXiv.org Artificial Intelligence

We introduced a multimodal foundational model for financial transactions that integrates both structured attributes and unstructured textual descriptions into a unified representation. By adapting masked language modeling to transaction sequences, we demonstrated that our approach not only outperforms classical feature engineering and discrete event sequence methods but is also particularly effective in data-scarce Open Banking scenarios. To our knowledge, this is the first large-scale study across thousands of financial institutions in North America, providing evidence that multimodal representations can generalize across geographies and institutions. These results highlight the potential of self-supervised models to advance financial applications ranging from fraud prevention and credit risk to customer insights


Do LLMs Really Forget? Evaluating Unlearning with Knowledge Correlation and Confidence Awareness

Wei, Rongzhe, Niu, Peizhi, Hsu, Hans Hao-Hsun, Wu, Ruihan, Yin, Haoteng, Ghassemi, Mohsen, Li, Yifan, Potluru, Vamsi K., Chien, Eli, Chaudhuri, Kamalika, Milenkovic, Olgica, Li, Pan

arXiv.org Artificial Intelligence

Machine unlearning techniques aim to mitigate unintended memorization in large language models (LLMs). However, existing approaches predominantly focus on the explicit removal of isolated facts, often overlooking latent inferential dependencies and the non-deterministic nature of knowledge within LLMs. Consequently, facts presumed forgotten may persist implicitly through correlated information. To address these challenges, we propose a knowledge unlearning evaluation framework that more accurately captures the implicit structure of real-world knowledge by representing relevant factual contexts as knowledge graphs with associated confidence scores. We further develop an inference-based evaluation protocol leveraging powerful LLMs as judges; these judges reason over the extracted knowledge subgraph to determine unlearning success. Our LLM judges utilize carefully designed prompts and are calibrated against human evaluations to ensure their trustworthiness and stability. Extensive experiments on our newly constructed benchmark demonstrate that our framework provides a more realistic and rigorous assessment of unlearning performance. Moreover, our findings reveal that current evaluation strategies tend to overestimate unlearning effectiveness. Our code is publicly available at https://github.com/Graph-COM/Knowledge_Unlearning.git.


Towards Autonomous Robotic Electrosurgery via Thermal Imaging

Riaziat, Naveed D., Chen, Joseph, Krieger, Axel, Brown, Jeremy D.

arXiv.org Artificial Intelligence

Electrosurgery is a surgical technique that can improve tissue cutting by reducing cutting force and bleeding. However, electrosurgery adds a risk of thermal injury to surrounding tissue. Expert surgeons estimate desirable cutting velocities based on experience but have no quantifiable reference to indicate if a particular velocity is optimal. Furthermore, prior demonstrations of autonomous electrosurgery have primarily used constant tool velocity, which is not robust to changes in electrosurgical tissue characteristics, power settings, or tool type. Thermal imaging feedback provides information that can be used to reduce thermal injury while balancing cutting force by controlling tool velocity. We introduce Thermography for Electrosurgical Rate Modulation via Optimization (ThERMO) to autonomously reduce thermal injury while balancing cutting force by intelligently controlling tool velocity. We demonstrate ThERMO in tissue phantoms and compare its performance to the constant velocity approach. Overall, ThERMO improves cut success rate by a factor of three and can reduce peak cutting force by a factor of two. ThERMO responds to varying environmental disturbances, reduces damage to tissue, and completes cutting tasks that would otherwise result in catastrophic failure for the constant velocity approach.


Hierarchical Fallback Architecture for High Risk Online Machine Learning Inference

Polleti, Gustavo, Santana, Marlesson, Del Sant, Felipe Sassi, Fontes, Eduardo

arXiv.org Artificial Intelligence

These systems can fail unexpectedly in a variety of different ways. Notably, applications Open Banking powered machine learning applications require novel that rely on online inference are subject to their inability robustness approaches to deal with challenging stress and failure to keep up with the expected operating procedures while, now scenarios. In this paper we propose an hierarchical fallback architecture additionally, having to make tedious computational tasks for these for improving robustness in high risk machine learning AI/ML applications, typically resulting in timeouts, infrastructure applications with a focus in the financial domain. We define generic outages and, often, failures in external dependencies such as third failure scenarios often found in online inference that depend on party data providers (external API calls) [7]. When the underlying external data providers and we describe in detail how to apply the machine learning applications are presented with strong robustness hierarchical fallback architecture to address them. Finally, we offer requirements, fallback or fall-over strategies are needed to keep a real world example of its applicability in the industry for near-real operations running, even in the event of unexpected failures. In time transactional fraud risk evaluation using Open Banking data finance, specifically applications that require real time risk mitigation and under extreme stress scenarios.


Crafting Large Language Models for Enhanced Interpretability

Sun, Chung-En, Oikarinen, Tuomas, Weng, Tsui-Wei

arXiv.org Artificial Intelligence

We introduce the Concept Bottleneck Large Language Model (CB-LLM), a pioneering approach to creating inherently interpretable Large Language Models (LLMs). Unlike traditional black-box LLMs that rely on post-hoc interpretation methods with limited neuron function insights, CB-LLM sets a new standard with its built-in interpretability, scalability, and ability to provide clear, accurate explanations. This innovation not only advances transparency in language models but also enhances their effectiveness. Our unique Automatic Concept Correction (ACC) strategy successfully narrows the performance gap with conventional black-box LLMs, positioning CB-LLM as a model that combines the high accuracy of traditional LLMs with the added benefit of clear interpretability -- a feature markedly absent in existing LLMs.


Advancing Household Robotics: Deep Interactive Reinforcement Learning for Efficient Training and Enhanced Performance

Soni, Arpita, Alla, Sujatha, Dodda, Suresh, Volikatla, Hemanth

arXiv.org Artificial Intelligence

The market for domestic robots made to perform household chores is growing as these robots relieve people of everyday responsibilities. Domestic robots are generally welcomed for their role in easing human labor, in contrast to industrial robots, which are frequently criticized for displacing human workers. But before these robots can carry out domestic chores, they need to become proficient in several minor activities, such as recognizing their surroundings, making decisions, and picking up on human behaviors. Reinforcement learning, or RL, has emerged as a key robotics technology that enables robots to interact with their environment and learn how to optimize their actions to maximize rewards. However, the goal of Deep Reinforcement Learning is to address more complicated, continuous action-state spaces in real-world settings by combining RL with Neural Networks. The efficacy of DeepRL can be further augmented through interactive feedback, in which a trainer offers real-time guidance to expedite the robot's learning process. Nevertheless, the current methods have drawbacks, namely the transient application of guidance that results in repeated learning under identical conditions. Therefore, we present a novel method to preserve and reuse information and advice via Deep Interactive Reinforcement Learning, which utilizes a persistent rule-based system. This method not only expedites the training process but also lessens the number of repetitions that instructors will have to carry out. This study has the potential to advance the development of household robots and improve their effectiveness and efficiency as learners.


GM is developing a drone-killing off-road pickup for the US Army

FOX News

A General Motors pickup has never hauled something like this. GM Defense is collaborating with military contractor Black Sage Technologies to integrate a drone defense system into the Infantry Squad Vehicle (ISV) that GM Defense recently began supplying to the US Army. The ISV is based on the last-generation Chevrolet Colorado ZR2 midsize pickup and manufactured in Concord, N.C., using frames supplied by NASCAR's Hendrick Motorsports. The midsize truck was engineered for high-speed off-road driving and designed to fit inside a CH-47 Chinook helicopter, slung from a UH-60 Blackhawk helicopter, or air-dropped from a cargo plane by parachute for quick deployment into the field. The vehicle can be outfitted to fit nine troops, but there are several configurations that mix passenger, cargo and arms carrying capabilities.