Goto

Collaborating Authors

 requester






Stable Acoustic Relay Assignment with High Throughput via Lase Chaos-based Reinforcement Learning

Chen, Zengjing, Wang, Lu, Xing, Chengzhi

arXiv.org Artificial Intelligence

Underwater Acoustic Networks (UANs) have gained significant attention from both industry and academia due to their indisputable advantages in improving link reliability, increasing system capacity, expanding transmission range and so on. Acoustic communication is most widely used underwater communication as sound wave is not absorbed by water so easily like electromagnetic wave and optical wave [1]. UANs typically consist of acoustic-linked seabed sensors, autonomous underwater vehicles, and ground stations that provide links to onshore control centers. Due to the battery-powered network nodes, shallow water acoustic channel characteristics, such as low available bandwidth and highly varying multi-path, maximizing throughput while minimizing consumption has become a very challenging task [2]. Recent studies have discussed the challenges and opportunities of underwater cognitive communication [3], proposed cooperative automatic repeat request protocols for higher channel quality [4], and analyzed the impact of low transmission rates and long preambles on medium access control protocols [5]. Artificial intelligence (AI) has experienced significant growth in popularity in recent years, and many industries and research fields have explored its potential applications, including information theory, game theory, biological systems, and so on [6-9].


InfantAgent-Next: A Multimodal Generalist Agent for Automated Computer Interaction

Lei, Bin, Kang, Weitai, Zhang, Zijian, Chen, Winson, Xie, Xi, Zuo, Shan, Xie, Mimi, Payani, Ali, Hong, Mingyi, Yan, Yan, Ding, Caiwen

arXiv.org Artificial Intelligence

This paper introduces \textsc{InfantAgent-Next}, a generalist agent capable of interacting with computers in a multimodal manner, encompassing text, images, audio, and video. Unlike existing approaches that either build intricate workflows around a single large model or only provide workflow modularity, our agent integrates tool-based and pure vision agents within a highly modular architecture, enabling different models to collaboratively solve decoupled tasks in a step-by-step manner. Our generality is demonstrated by our ability to evaluate not only pure vision-based real-world benchmarks (i.e., OSWorld), but also more general or tool-intensive benchmarks (e.g., GAIA and SWE-Bench). Specifically, we achieve $\mathbf{7.27\%}$ accuracy on OSWorld, higher than Claude-Computer-Use. Codes and evaluation scripts are open-sourced at https://github.com/bin123apple/InfantAgent.


Seeing, Saying, Solving: An LLM-to-TL Framework for Cooperative Robots

Choe, Dan BW, Sangeetha, Sundhar Vinodh, Emanuel, Steven, Chiu, Chih-Yuan, Coogan, Samuel, Kousik, Shreyas

arXiv.org Artificial Intelligence

Increased robot deployment, such as in warehousing, has revealed a need for seamless collaboration among heterogeneous robot teams to resolve unforeseen conflicts. To address this challenge, we propose a novel, decentralized framework for robots to request and provide help. The framework begins with robots detecting conflicts using a Vision Language Model (VLM), then reasoning over whether help is needed. If so, it crafts and broadcasts a natural language (NL) help request using a Large Language Model (LLM). Potential helper robots reason over the request and offer help (if able), along with information about impact to their current tasks. Helper reasoning is implemented via an LLM grounded in Signal Temporal Logic (STL) using a Backus-Naur Form (BNF) grammar to guarantee syntactically valid NL-to-STL translations, which are then solved as a Mixed Integer Linear Program (MILP). Finally, the requester robot chooses a helper by reasoning over impact on the overall system. We evaluate our system via experiments considering different strategies for choosing a helper, and find that a requester robot can minimize overall time impact on the system by considering multiple help offers versus simple heuristics (e.g., selecting the nearest robot to help).


CARE: Compatibility-Aware Incentive Mechanisms for Federated Learning with Budgeted Requesters

Liu, Xiang, Chan, Hau, Li, Minming, Zeng, Xianlong, Fu, Chenchen, Wu, Weiwei

arXiv.org Artificial Intelligence

Federated learning (FL) is a promising approach that allows requesters (\eg, servers) to obtain local training models from workers (e.g., clients). Since workers are typically unwilling to provide training services/models freely and voluntarily, many incentive mechanisms in FL are designed to incentivize participation by offering monetary rewards from requesters. However, existing studies neglect two crucial aspects of real-world FL scenarios. First, workers can possess inherent incompatibility characteristics (e.g., communication channels and data sources), which can lead to degradation of FL efficiency (e.g., low communication efficiency and poor model generalization). Second, the requesters are budgeted, which limits the amount of workers they can hire for their tasks. In this paper, we investigate the scenario in FL where multiple budgeted requesters seek training services from incompatible workers with private training costs. We consider two settings: the cooperative budget setting where requesters cooperate to pool their budgets to improve their overall utility and the non-cooperative budget setting where each requester optimizes their utility within their own budgets. To address efficiency degradation caused by worker incompatibility, we develop novel compatibility-aware incentive mechanisms, CARE-CO and CARE-NO, for both settings to elicit true private costs and determine workers to hire for requesters and their rewards while satisfying requester budget constraints. Our mechanisms guarantee individual rationality, truthfulness, budget feasibility, and approximation performance. We conduct extensive experiments using real-world datasets to show that the proposed mechanisms significantly outperform existing baselines.


Blockchain-based Crowdsourced Deep Reinforcement Learning as a Service

Alagha, Ahmed, Otrok, Hadi, Singh, Shakti, Mizouni, Rabeb, Bentahar, Jamal

arXiv.org Artificial Intelligence

Deep Reinforcement Learning (DRL) has emerged as a powerful paradigm for solving complex problems. However, its full potential remains inaccessible to a broader audience due to its complexity, which requires expertise in training and designing DRL solutions, high computational capabilities, and sometimes access to pre-trained models. This necessitates the need for hassle-free services that increase the availability of DRL solutions to a variety of users. To enhance the accessibility to DRL services, this paper proposes a novel blockchain-based crowdsourced DRL as a Service (DRLaaS) framework. The framework provides DRL-related services to users, covering two types of tasks: DRL training and model sharing. Through crowdsourcing, users could benefit from the expertise and computational capabilities of workers to train DRL solutions. Model sharing could help users gain access to pre-trained models, shared by workers in return for incentives, which can help train new DRL solutions using methods in knowledge transfer. The DRLaaS framework is built on top of a Consortium Blockchain to enable traceable and autonomous execution. Smart Contracts are designed to manage worker and model allocation, which are stored using the InterPlanetary File System (IPFS) to ensure tamper-proof data distribution. The framework is tested on several DRL applications, proving its efficacy.


Blockchain-assisted Demonstration Cloning for Multi-Agent Deep Reinforcement Learning

Alagha, Ahmed, Bentahar, Jamal, Otrok, Hadi, Singh, Shakti, Mizouni, Rabeb

arXiv.org Artificial Intelligence

Multi-Agent Deep Reinforcement Learning (MDRL) is a promising research area in which agents learn complex behaviors in cooperative or competitive environments. However, MDRL comes with several challenges that hinder its usability, including sample efficiency, curse of dimensionality, and environment exploration. Recent works proposing Federated Reinforcement Learning (FRL) to tackle these issues suffer from problems related to model restrictions and maliciousness. Other proposals using reward shaping require considerable engineering and could lead to local optima. In this paper, we propose a novel Blockchain-assisted Multi-Expert Demonstration Cloning (MEDC) framework for MDRL. The proposed method utilizes expert demonstrations in guiding the learning of new MDRL agents, by suggesting exploration actions in the environment. A model sharing framework on Blockchain is designed to allow users to share their trained models, which can be allocated as expert models to requesting users to aid in training MDRL systems. A Consortium Blockchain is adopted to enable traceable and autonomous execution without the need for a single trusted entity. Smart Contracts are designed to manage users and models allocation, which are shared using IPFS. The proposed framework is tested on several applications, and is benchmarked against existing methods in FRL, Reward Shaping, and Imitation Learning-assisted RL. The results show the outperformance of the proposed framework in terms of learning speed and resiliency to faulty and malicious models.