LM Agents for Coordinating Multi-User Information Gathering
Jhamtani, Harsh, Andreas, Jacob, Van Durme, Benjamin
–arXiv.org Artificial Intelligence
This paper introduces PeopleJoin, a benchmark for evaluating LM-mediated collaborative problem solving. Given a user request, PeopleJoin agents must identify teammates who might be able to assist, converse with these teammates to gather information, and finally compile a useful answer or summary for the original user. PeopleJoin comprises two evaluation domains: PeopleJoin-QA, focused on questions about tabular data, and PeopleJoin-DocCreation, focused on document creation tasks. The two domains are adapted from existing NLP benchmarks for database question answering and multi-document summarization; here, however, the information needed to complete these tasks is distributed across synthetic ``organizations'' of 2--20 users, simulating natural multi-user collaboration scenarios. We implemented several popular LM agent architectures, evaluating their accuracy and efficiency at completing tasks, and highlight new research questions that can be studied using PeopleJoin.
arXiv.org Artificial Intelligence
Feb-17-2025
- Country:
- North America > United States (1.00)
- Genre:
- Research Report > New Finding (0.66)
- Industry:
- Banking & Finance > Economy (1.00)
- Education (0.67)
- Government > Regional Government
- Health & Medicine (1.00)
- Law (0.67)
- Transportation (0.67)
- Technology: