GoAgent: Group-of-Agents Communication Topology Generation for LLM-based Multi-Agent Systems

GoAgent: Teaching AI Agents to “Band Together,” 93.8% Accuracy, 17% Lower Token Cost

Even the most capable single AI Agent can struggle with complex tasks. As a result, Multi-Agent Systems (MAS) have emerged, solving problems through the collaboration of multiple Agents. But this also introduces a core challenge: how should these Agents be organized for maximum efficiency?

ArXiv URL:http://arxiv.org/abs/2603.19677v1

At present, most methods are like recruiting people one by one and then deciding, pair by pair, who should talk to whom. This “node-centric” approach often leads to chaotic collaboration workflows and high communication costs.

Now, a paper called GoAgent proposes a disruptive idea: stop connecting individual nodes one by one—just work in “groups” directly! The study uses predefined “collaboration groups” as the basic building blocks, not only significantly improving performance on complex tasks, but also reducing Token consumption by about 17%.

From “Solo Combat” to “Team Collaboration”: GoAgent’s New Paradigm

When traditional methods build communication networks, they consider only one Agent at a time; this is known as the node-centric paradigm. The drawbacks are obvious: it cannot capture the collaboration structure required by the task from a macro perspective, easily leads to unnecessary communication, and can even result in missing key roles.

Refer to caption

Figure 1: (a) Node-centric paradigm vs. (b) GoAgent’s group-centric paradigm

GoAgent pioneers a new group-centric paradigm. It no longer treats individual Agents as atomic units, but instead uses “collaboration groups” as the basic unit for building communication topologies.

A group may include a “problem decomposer,” a “code implementer,” and a “code reviewer,” already forming an efficient workflow internally. The model only needs to decide which groups are needed for the current task and how information should flow between them.

This approach has two major advantages:

  1. Ensures internal cohesion: The collaboration pattern within each group is predefined, ensuring efficient execution of subtasks.

  2. Simplifies macro-level coordination: The model only needs to focus on higher-level dependencies between groups, greatly reducing construction complexity.

Refer to caption

Figure 2: Overview of GoAgent’s overall architecture

GoAgent’s entire workflow can be divided into several key steps: first, it encodes the task and discovers candidate collaboration groups; then, it selects and connects these groups step by step in an autoregressive manner; finally, it uses a clever mechanism to compress noise in communication.

How Does GoAgent Build an “Elite Team”?

The process by which GoAgent generates a communication topology is like a smart project manager assembling a team.

1. Collaboration Group Discovery

First, the researchers use large language models such as GPT-4 to pre-generate a series of candidate “expert groups” based on the needs of specific domains, such as code generation and mathematical reasoning.

Each group has a clear definition, including:

These candidate groups form a “talent pool” waiting to be called upon.

2. Autoregressive Group Generation

Next, GoAgent’s core generation model gets to work. It incrementally builds the final communication graph:

At each step $t$, the model does two things:

This process is autoregressive, meaning each decision depends on the previous state, until the model generates a special \(END\) token indicating that graph construction is complete.

Rejecting “Too Many Voices”: Compressing Communication with an Information Bottleneck

As the collaboration graph continues to grow, the historical dialogue between Agents becomes increasingly verbose, inevitably containing a large amount of “noise” unrelated to the current decision. If this noise is allowed to spread, it not only increases Token consumption but may also interfere with the judgment of subsequent Agents.

To address this problem, GoAgent introduces a key mechanism: the Conditional Information Bottleneck (CIB).

You can think of it as an intelligent filter. Before each group or connection prediction, the historical information first passes through the CIB layer. The goal of CIB is to:

Its optimization objective can be expressed as:

\[\min\mathcal{L}_{\text{CIB}}=-I(\mathbf{c};y \mid \mathbf{z}_{\mathcal{Q}})+\beta I(\mathbf{x};\mathbf{c} \mid \mathbf{z}_{\mathcal{Q}})\]

Here, $y$ is the prediction target (the next group or connection), $\mathbf{x}$ is the original historical information, $\mathbf{c}$ is the compressed essential information, and $\mathbf{z}_{\mathcal{Q}}$ is the representation of the task itself.

The first term (Predictive Term) ensures that the compressed information is sufficient for accurate prediction, while the second term (Compression Term) is responsible for discarding irrelevant information as much as possible. In this way, GoAgent ensures that every decision is based on the most refined and relevant signals, thereby improving the efficiency and robustness of the entire system.

Experimental Results: More Accurate, More Efficient

GoAgent performs exceptionally well on six mainstream benchmarks, including MMLU, GSM8K, and HumanEval, demonstrating its effectiveness.

1. Leading Accuracy Across the Board

As shown in the summary table below, GoAgent achieves SOTA performance on all tasks, with an average accuracy of 93.84%. Its advantage is even more pronounced on more complex tasks such as MMLU (commonsense reasoning) and HumanEval (code generation), where it improves by 1.96% and 2.47%, respectively, over the previous SOTA methods.

This confirms the study’s core hypothesis: building at the group level can form a more coherent and specialized collaboration flow, thereby better solving complex problems.

2. Significantly Lower Token Consumption

Efficiency is another key metric for evaluating the quality of a multi-agent system. As shown in the figure below, while achieving higher accuracy, GoAgent also successfully reduced Token consumption by about 17%.

Refer to caption

Figure 3: Token consumption comparison

This is mainly due to two factors: first, the group-based structure avoids a large amount of redundant point-to-point communication; second, the CIB mechanism effectively filters historical noise, reducing the context length passed to the LLM.

Ablation experiments further confirm that both the “group-centric” design and the “CIB” mechanism are crucial to the final performance improvement, and neither can be omitted.

Conclusion

The introduction of GoAgent opens up a new path for designing efficient and powerful multi-agent systems. It successfully elevates the construction paradigm from the chaotic “node-centric” approach to an orderly “group-centric” one, simulating real-world team collaboration and teaching AI Agents to work in “bands.”

More accurate and more economical, GoAgent proves that good top-level design is more important than simply stacking Agent capabilities. In the future, we may see more AI collaboration systems emerge based on this kind of “organizational architecture” thinking.