2025.emnlp-main.503@ACL

Total: 1

#1 Enhancing LLM Text Detection with Retrieved Contexts and Logits Distribution Consistency [PDF] [Copy] [Kimi] [REL]

Authors: Zhaoheng Huang, Yutao Zhu, Ji-Rong Wen, Zhicheng Dou

Large language models (LLMs) can generate fluent text, raising concerns about misuse in online comments and academic writing, leading to issues like corpus pollution and copyright infringement. Existing LLM text detection methods often rely on features from the logit distribution of the input text. However, the distinction between the LLM-generated and human-written texts may rely on only a few tokens due to the short length or insufficient information in some texts, leading to minimal and hard-to-detect differences in logit distributions. To address this, we propose HALO, an LLM-based detection method that leverages external text corpora to evaluate the difference in the logit distribution of input text under retrieved human-written and LLM-rewritten contexts. HALO also complements basic detection features and can serve as a plug-and-play module to enhance existing detection methods. Extensive experiments on five public datasets with three widely-used source LLMs show that our proposed detection method achieves state-of-the-art performance in AUROC, both in cross-domain and domain-specific scenarios.

Subject: EMNLP.2025 - Main