Agent 与自动化 3.0 · 值得看 2026-03-26 · 论文

多智能体共识机制研究

当前基于 LLM 的多智能体系统能够可靠地达成共识吗?在存在恶意智能体的情况下,共识机制是否鲁棒? 这篇论文研究了一个基础问题:当多个 LLM 智能体需要达成一致决策时,它们能否可靠地完成这一任务?特别是在存在可能破坏共识的拜占庭智能体的情况下。...

打开原文回到归档

来源: https://arxiv.org/abs/2603.01213

[2603.01213] Can AI Agents Agree?

Skip to main content

Learn about arXiv becoming an independent nonprofit.

We gratefully acknowledge support from the Simons Foundation, member institutions, and all contributors. Donate

> cs > arXiv:2603.01213

Help | Advanced Search

All fields Title Author Abstract Comments Journal reference ACM classification MSC classification Report number arXiv identifier DOI ORCID arXiv author ID Help pages Full text

Search

open search

GO

open navigation menu

quick links

Login Help Pages About

-->

Computer Science > Multiagent Systems

arXiv:2603.01213 (cs)[Submitted on 1 Mar 2026 (v1), last revised 12 Mar 2026 (this version, v2)] Title:Can AI Agents Agree? Authors:Frédéric Berdoz, Leonardo Rugli, Roger Wattenhofer View a PDF of the paper titled Can AI Agents Agree?, by Fr\'ed\'eric Berdoz and 2 other authors View PDF

Abstract:Large language models are increasingly deployed as cooperating agents, yet their behavior in adversarial consensus settings has not been systematically studied. We evaluate LLM-based agents on a Byzantine consensus game over scalar values using a synchronous all-to-all simulation. We test consensus in a no-stake setting where agents have no preferences over the final value, so evaluation focuses on agreement rather than value optimality. Across hundreds of simulations spanning model sizes, group sizes, and Byzantine fractions, we find that valid agreement is not reliable even in benign settings and degrades as group size grows. Introducing a small number of Byzantine agents further reduces success. Failures are dominated by loss of liveness, such as timeouts and stalled convergence, rather than subtle value corruption. Overall, the results suggest that reliable agreement is not yet a dependable emergent capability of current LLM-agent groups even in no-stake settings, raising caution for deployments that rely on robust coordination.

Subjects:

Multiagent Systems (cs.MA); Machine Learning (cs.LG)

Cite as: arXiv:2603.01213[cs.MA]

  (or arXiv:2603.01213v2[cs.MA] for this version)

  https://doi.org/10.48550/arXiv.2603.01213

Focus to learn more

arXiv-issued DOI via DataCite

Submission history From: Frédéric Berdoz[view email][v1] Sun, 1 Mar 2026 18:18:59 UTC (93 KB)[v2] Thu, 12 Mar 2026 08:32:01 UTC (93 KB)

Full-text links: Access Paper:

View a PDF of the paper titled Can AI Agents Agree?, by Fr\'ed\'eric Berdoz and 2 other authorsView PDFTeX Source

view license

Current browse context: cs.MA

< prev

  |   next >

new | recent | 2026-03

Change to browse by:

cs cs.LG

References & Citations

NASA ADSGoogle Scholar Semantic Scholar

export BibTeX citation Loading...

BibTeX formatted citation ×

loading...

Data provided by:

Bookmark

Bibliographic Tools

Bibliographic and Citation Tools

Bibliographic Explorer Toggle

Bibliographic Explorer (What is the Explorer?)

Connected Papers Toggle

Connected Papers (What is Connected Papers?)

Litmaps Toggle

Litmaps (What is Litmaps?)

scite.ai Toggle

scite Smart Citations (What are Smart Citations?)

Code, Data, Media

Code, Data and Media Associated with this Article

alphaXiv Toggle

alphaXiv (What is alphaXiv?)

Links to Code Toggle

CatalyzeX Code Finder for Papers (What is CatalyzeX?)

DagsHub Toggle

DagsHub (What is DagsHub?)

GotitPub Toggle

Gotit.pub (What is GotitPub?)

Huggingface Toggle

Hugging Face (What is Huggingface?)

ScienceCast Toggle

ScienceCast (What is ScienceCast?)

Demos

Demos

Replicate Toggle

Replicate (What is Replicate?)

Spaces Toggle

Hugging Face Spaces (What is Spaces?)

Spaces Toggle

TXYZ.AI (What is TXYZ.AI?)

Related Papers

Recommenders and Search Tools

Link to Influence Flower

Influence Flower (What are Influence Flowers?)

Core recommender toggle

CORE Recommender (What is CORE?)

Author Venue Institution Topic

About arXivLabs

arXivLabs: experimental projects with community collaborators arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website. Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them. Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)

mathjaxToggle();

About Help

contact arXivClick here to contact arXiv Contact

subscribe to arXiv mailingsClick here to subscribe Subscribe

Copyright Privacy Policy

Web Accessibility Assistance

arXiv Operational Status