MAPEX: A Multi-Agent Pipeline for Keyphrase Extraction


TL;DR

This paper proposes MAPEX, a multi-agent collaborative framework for keyphrase extraction. By designing a dynamic dual-path strategy for documents of different lengths—knowledge-driven and topic-guided—it significantly improves the zero-shot keyphrase extraction performance of large language models (LLMs).

Key Definitions

This paper introduces or adopts the following key concepts:

Current unsupervised keyphrase extraction techniques are mainly divided into traditional methods (statistics, graph, embedding) and language-model-based prompt-based methods. With the rise of large language models (LLMs), prompt-based methods have become mainstream. However, existing methods generally suffer from a key bottleneck: most of them adopt a single, fixed inference flow and prompt strategy, without distinguishing between document length or the underlying LLM model. This one-size-fits-all design cannot fully exploit the reasoning and generation potential of LLMs in diverse scenarios, limiting their generalization ability on keyphrase extraction tasks.

This paper aims to address the above problem: how to design a more flexible and powerful framework that fully leverages the capabilities of LLMs, especially in handling the challenges posed by documents of different lengths, thereby improving the accuracy and robustness of zero-shot keyphrase extraction.

Method

This paper proposes MAPEX (Multi-Agent Pipeline for Keyphrase Extraction), a multi-agent pipeline framework. The framework works through three Agents and a dual-path strategy dynamically selected according to document length.

Figure illustration

Agent Roles and Responsibilities

The MAPEX framework includes three Agents that work collaboratively, with their behavior guided by carefully designed prompts.

Figure illustration

  1. Expert Recruiter: This Agent first analyzes the document content to determine its professional domain. It then assigns an appropriate expert role to the “domain expert” Agent (for example, “software engineer”) and provides the rationale for the assignment. This makes the subsequent extraction process more professionally informed.
  2. Candidate Extractor: This Agent is responsible for generating a broad pool of candidate keyphrases from the original document. It is not assigned any specific role, with the goal of ensuring diversity in the initial candidate set and avoiding the omission of important lexical variants due to the limitations of a specific expert role.

Core Innovation: Dual-Path Strategy

To effectively handle documents of different lengths, MAPEX introduces a length threshold \(ℓ\) and dispatches documents to two different processing paths based on this threshold. The core motivation behind this design is that external knowledge provides significant semantic supplementation for short texts, but its advantage weakens for long texts due to context window limitations and semantic dilution.

  1. Based on this enhanced knowledge, the “domain expert” reranks the candidate terms and outputs an initial keyphrase list.

Post-processing

After the “domain expert” generates the initial results, the framework performs a post-processing step to improve the quality of the final output. This step includes three subtasks:

  1. Remove redundancy: Delete duplicate phrases.
  2. Standardization: Unify abbreviations and their full forms.
  3. Filter hallucinations: Remove phrases that do not appear in the original text, ensuring that all keyphrases are derived from the document.

Advantages

Experimental Results

Overall Performance

Experiments were conducted on six benchmark datasets, including Inspec and SemEval-2010, using three different LLMs such as Mistral-7B and Qwen2-7B as the underlying models.


Method Model Inspec SemEval-10 SemEval-17 DUC-2001 NUS Krapivin Average
Traditional Unsupervised                
SIFRank - 35.15 29.56 38.38 26.68 27.69 22.09 30.03
MDERank - 36.31 30.13 42.17 27.35 29.80 25.12 31.81
PromptRank BART 39.56 32.55 45.34 31.33 30.65 27.48 34.49
LLM Baselines                
Base Mistral-7B 37.98 29.59 42.16 28.52 27.91 25.26 31.90
Hybrid Mistral-7B 38.45 30.11 42.94 29.07 28.98 26.17 32.62
MAPEX (this paper) Mistral-7B 40.31 32.88 45.92 30.64 33.34 28.75 35.31
(Other LLM results)


Ablation Study

To verify the contribution of each module, this paper conducted an ablation study on the Mistral-7B model.

Performance gains relative to baseline methods (a) Performance gains relative to the baseline

Performance differences across different paths (b) Performance differences between the knowledge-driven and topic-guided paths

Final Conclusion

The MAPEX framework proposed in this paper successfully addresses the “one-size-fits-all” problem of traditional LLM methods in keyword extraction tasks by introducing multi-agent collaboration and a dynamic dual-path strategy. Experiments demonstrate that the framework has strong generalization ability and versatility, achieving performance that surpasses existing SOTA methods across multiple LLMs, especially when handling complex and long documents.