Papers
arxiv:2402.05120

More Agents Is All You Need

Published on Feb 3
· Featured in Daily Papers on Feb 9
Authors:
,
,
,
,

Abstract

We find that, simply via a sampling-and-voting method, the performance of large language models (LLMs) scales with the number of agents instantiated. Also, this method is orthogonal to existing complicated methods to further enhance LLMs, while the degree of enhancement is correlated to the task difficulty. We conduct comprehensive experiments on a wide range of LLM benchmarks to verify the presence of our finding, and to study the properties that can facilitate its occurrence. Our code is publicly available at: https://anonymous.4open.science/r/more_agent_is_all_you_need.

Community

This ought to be the way we create training data, iterate through existing dataset, create 15+ derivatives and use the voting results for DPO. Ideally each time the model improves, use it to generate the new data, rather than using a stale model. That way we sacrifice training speed but gain 15x inference time. I wonder how many times you can rinse and repeat before getting diminishing results

I'm not sure that more agents is "all you need" when there are diminishing returns with increasing the number agents (i.e., the performance gains are asymptotic). If you effectively hit a ceiling in performance at n=30 agents, and the performance is below the current SOTA, then how is it "all you need"? I feel like this paper should be called "crowd wisdom" or "agentic tree-of-thought," since this is essentially and agent-based implementation of ToT.

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

I'm not sure that more agents is "all you need" when there are diminishing returns with increasing the number agents (i.e., the performance gains are asymptotic). If you effectively hit a ceiling in performance at n=30 agents, and the performance is below the current SOTA, then how is it "all you need"? I feel like this paper should be called "crowd wisdom" or "agentic tree-of-thought," since this is essentially and agent-based implementation of ToT.

... "is all you need" has become a bit of a joke. It's no different to the way that a lot of researchers "fudge" their charts and use unnecessarily complex terminology to hide the simplicity of the underlying method. Showmanship aside, the take away I get from this research, is that we haven't trained the models well enough. When we create 15+ derivatives of the same prompt, the votes should be uniform. Then we know that we have trained it sufficiently, because it is always outputting the best answer.

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2402.05120 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2402.05120 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2402.05120 in a Space README.md to link it from this page.

Collections including this paper 26