manager
FlowReasoner: Reinforcing Query-Level Meta-Agents
Gao, Hongcheng, Liu, Yue, He, Yufei, Dou, Longxu, Du, Chao, Deng, Zhijie, Hooi, Bryan, Lin, Min, Pang, Tianyu
This paper proposes a query-level meta-agent named FlowReasoner to automate the design of query-level multi-agent systems, i.e., one system per user query. Our core idea is to incentivize a reasoning-based meta-agent via external execution feedback. Concretely, by distilling DeepSeek R1, we first endow the basic reasoning ability regarding the generation of multi-agent systems to FlowReasoner. Then, we further enhance it via reinforcement learning (RL) with external execution feedback. A multi-purpose reward is designed to guide the RL training from aspects of performance, complexity, and efficiency. In this manner, FlowReasoner is enabled to generate a personalized multi-agent system for each user query via deliberative reasoning. Experiments on both engineering and competition code benchmarks demonstrate the superiority of FlowReasoner. Remarkably, it surpasses o1-mini by 10.52% accuracy across three benchmarks. The code is available at https://github.com/sail-sg/FlowReasoner.
DynTaskMAS: A Dynamic Task Graph-driven Framework for Asynchronous and Parallel LLM-based Multi-Agent Systems
Yu, Junwei, Ding, Yepeng, Sato, Hiroyuki
The emergence of Large Language Models (LLMs) in Multi-Agent Systems (MAS) has opened new possibilities for artificial intelligence, yet current implementations face significant challenges in resource management, task coordination, and system efficiency. While existing frameworks demonstrate the potential of LLM-based agents in collaborative problem-solving, they often lack sophisticated mechanisms for parallel execution and dynamic task management. This paper introduces DynTaskMAS, a novel framework that orchestrates asynchronous and parallel operations in LLM-based MAS through dynamic task graphs. The framework features four key innovations: (1) a Dynamic Task Graph Generator that intelligently decomposes complex tasks while maintaining logical dependencies, (2) an Asynchronous Parallel Execution Engine that optimizes resource utilization through efficient task scheduling, (3) a Semantic-Aware Context Management System that enables efficient information sharing among agents, and (4) an Adaptive Workflow Manager that dynamically optimizes system performance. Experimental evaluations demonstrate that DynTaskMAS achieves significant improvements over traditional approaches: a 21-33% reduction in execution time across task complexities (with higher gains for more complex tasks), a 35.4% improvement in resource utilization (from 65% to 88%), and near-linear throughput scaling up to 16 concurrent agents (3.47X improvement for 4X agents). Our framework establishes a foundation for building scalable, high-performance LLM-based multi-agent systems capable of handling complex, dynamic tasks efficiently.
Towards Agentic Schema Refinement
Rissaki, Agapi, Fountalis, Ilias, Vasiloglou, Nikolaos, Gatterbauer, Wolfgang
Understanding the meaning of data is crucial for performing data analysis, yet for the users to gain insight into the content and structure of their database, a tedious data exploration process is often required [2, 16]. A common industry practice taken on by specialists such as Knowledge Engineers is to explicitly construct an intermediate layer between the database and the user -- a semantic layer -- abstracting away certain details of the database schema in favor of clearer data semantics [3, 10]. In the era of Large Language Models (LLMs), industry practitioners and researchers attempt to circumvent this costly process using LLM-powered Natural Language Interfaces [4, 6, 12, 18, 19, 22]. The promise of such Text-to-SQL solutions is to allow users without technical expertise to seamlessly interact with databases. For example, a new company employee could effectively issue queries in natural language without programming expertise or even explicit knowledge of the database structure, e.g., knowing the names of entities or properties, the exact location of data sources, etc.
- North America > United States (0.47)
- Asia > Myanmar > Tanintharyi Region > Dawei (0.04)
- Asia > Middle East > Jordan (0.04)
Optimizing Collaboration of LLM based Agents for Finite Element Analysis
This paper investigates the interactions between multiple agents within Large Language Models (LLMs) in the context of programming and coding tasks. We utilize the AutoGen framework to facilitate communication among agents, evaluating different configurations based on the success rates from 40 random runs for each setup. The study focuses on developing a flexible automation framework for applying the Finite Element Method (FEM) to solve linear elastic problems. Our findings emphasize the importance of optimizing agent roles and clearly defining their responsibilities, rather than merely increasing the number of agents. Effective collaboration among agents is shown to be crucial for addressing general FEM challenges. This research demonstrates the potential of LLM multi-agent systems to enhance computational automation in simulation methodologies, paving the way for future advancements in engineering and artificial intelligence.
- Oceania > New Zealand (0.04)
- Europe > Switzerland > Basel-City > Basel (0.04)
VerilogCoder: Autonomous Verilog Coding Agents with Graph-based Planning and Abstract Syntax Tree (AST)-based Waveform Tracing Tool
Ho, Chia-Tung, Ren, Haoxing, Khailany, Brucek
Due to the growing complexity of modern Integrated Circuits (ICs), automating hardware design can prevent a significant amount of human error from the engineering process and result in less errors. Verilog is a popular hardware description language for designing and modeling digital systems; thus, Verilog generation is one of the emerging areas of research to facilitate the design process. In this work, we propose VerilogCoder, a system of multiple Artificial Intelligence (AI) agents for Verilog code generation, to autonomously write Verilog code and fix syntax and functional errors using collaborative Verilog tools (i.e., syntax checker, simulator, and waveform tracer). Firstly, we propose a task planner that utilizes a novel Task and Circuit Relation Graph retrieval method to construct a holistic plan based on module descriptions. To debug and fix functional errors, we develop a novel and efficient abstract syntax tree (AST)-based waveform tracing tool, which is integrated within the autonomous Verilog completion flow. The proposed methodology successfully generates 94.2% syntactically and functionally correct Verilog code, surpassing the state-of-the-art methods by 33.9% on the VerilogEval-Human v2 benchmark.
- Information Technology (0.46)
- Semiconductors & Electronics (0.34)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.94)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.94)
Multi-Agent Causal Discovery Using Large Language Models
Le, Hao Duong, Xia, Xin, Chen, Zhang
Large Language Models (LLMs) have demonstrated significant potential in causal discovery tasks by utilizing their vast expert knowledge from extensive text corpora. However, the multi-agent capabilities of LLMs in causal discovery remain underexplored. This paper introduces a general framework to investigate this potential. The first is the Meta Agents Model, which relies exclusively on reasoning and discussions among LLM agents to conduct causal discovery. The second is the Coding Agents Model, which leverages the agents' ability to plan, write, and execute code, utilizing advanced statistical libraries for causal discovery. The third is the Hybrid Model, which integrates both the Meta Agents Model and Coding Agents Model approaches, combining the statistical analysis and reasoning skills of multiple agents. Our proposed framework shows promising results by effectively utilizing LLMs' expert knowledge, reasoning capabilities, multi-agent cooperation, and statistical causal methods. By exploring the multi-agent potential of LLMs, we aim to establish a foundation for further research in utilizing LLMs multi-agent for solving causal-related problems.
- Workflow (1.00)
- Research Report > New Finding (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.67)
Who's hiring who? Hiring activity related to artificial intelligence decreased by 13% in the medical industry in Q3 2022
The global medical industry experienced a 13% rise in new job postings related to artificial intelligence in Q3 2022 compared with the previous quarter, according to GlobalData's Jobs Analytics. This compares to a 13% increase in the previous quarter and a 39% increase versus Q3 2021. Notably, Software and Web Developers, Programmers, and Testers jobs accounted for a 16% share of the global medical industry's artificial intelligence-related total new job postings in Q3 2022, down 11% over the prior quarter. Software and Web Developers, Programmers, and Testers, with a share of 16%, emerged as the top artificial intelligence-related job roles within the medical industry in Q3 2022, with new job postings drop by 11% quarter-on-quarter. General and Operations Managers came in second with a share of 5% in Q3 2022, with new job postings rise by 4% over the previous quarter.
Four Cool Artificial Intelligence Technologies
According to the National Interagency Fire Center, wildfires have burned 2,990,255 acres this year (June 17). Adding to firefighters' challenges, using current resources, it often takes hours to map a growing wildfire's perimeter and heat spots. It sometimes takes days using fuel property data – often 3-5 years old – to help predict fire behavior. Time is not on their side and the situation on the ground is always changing. Lockheed Martin is using AI/ML to help get critical data to firefighters faster.
Petuum and Inception Institute for AI Partner for Advanced AI
Petuum, the creator of the world's first composable platform for MLOps, and the Inception Institute for Artificial Intelligence (IIAI), have agreed to partner on the development of revolutionary AI applications. Petuum has recently announced a limited release of the composable platform, which includes the AI OS, Universal Pipelines, Deployment Manager, and Experiment Manager, for select private beta partners. Through the partnership with Petuum, IIAI's enterprise AI/ML teams will operationalize and scale their applications into production. Founded in 2018, IIAI's mission is to build full-stack AI solutions and operating systems for enterprise businesses and developers. Besides being the research arm for G42, IIAI is also empowering stakeholders with AI applications and incubating new technology at the cutting edge of ML innovation.
- Africa > Middle East (0.26)
- Europe > Middle East (0.20)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.17)
GPT-3 Powers The Next Generation Of Apps - AI Summary
Nine months since the launch of our first commercial product, the OpenAI API, more than 300 applications are now using GPT-3, and tens of thousands of developers around the globe are building on our platform. They also want a way to edit their address in checkout and save multiple payment methods." "GPT-3's ability to identify themes from natural language and generate summaries allows Viable to give product, customer experience, and marketing teams at companies across industries a better understanding of their customers' wants and needs," said Daniel Erickson, CEO of Viable. Algolia uses GPT-3 in their Algolia Answers product to offer relevant, lightning-fast semantic search for their customers. When the OpenAI API launched, Algolia partnered with OpenAI to integrate GPT-3 with their advanced search technology in order to create their new Answers product that better understands customers' questions and connects them to the specific part of the content that answers their questions. Algolia Answers helps publishers and customer support help desks query in natural language and surface nontrivial answers. "We've seen great results from Algolia Answers on questions that are difficult to answer with textual search alone," said Peter Buffington, Product Manager at ABC Australia. "It was able to return very relevant, evergreen content from our news archives for questions such as'Why does a volcano erupt?'" "GPT-3 allows Algolia to answer more complex queries than ever before with our Algolia Answers product, identifying deeper contextual information to improve the quality of results and deliver them in seconds," said Dustin Coates, Product and GTM Manager at Algolia. We require developers to implement safety measures such as rate limits, user verification and testing, or human-in-the-loop requirements before they move into production. Nine months since the launch of our first commercial product, the OpenAI API, more than 300 applications are now using GPT-3, and tens of thousands of developers around the globe are building on our platform. They also want a way to edit their address in checkout and save multiple payment methods."
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.94)