Continual Learning (CL) is essential for enabling self-evolving large language models (LLMs) to adapt and remain effective amid rapid knowledge growth. Yet, despite its importance, little attention has been given to establishing statistical reliability guarantees for LLMs under CL, particularly in the setting of continual domain pretraining (CDP). Conformal Prediction (CP) has shown promise in offering correctness guarantees for LLMs, but it faces major challenges in CDP: testing data often stems from unknown or shifting domain distributions, under which CP may no longer provide valid guarantees. Moreover, when high coverage is required, CP can yield excessively large prediction sets for unanswerable queries, reducing informativeness. To address these challenges, we introduce an adaptive rejection and non-exchangeable CP framework. Our method first estimates the distribution of questions across domains in the test set using transformer-based clustering, then reweights or resamples the calibration data accordingly. Building on this, adaptive rejection CP allows the LLM to selectively abstain from answering when its confidence or competence shifts significantly. Extensive experiments demonstrate that our framework enhances both the effectiveness and reliability of CP under CDP scenarios. Our code is available at: https://anonymous.4open.science/r/CPCL-8C12/
翻译:持续学习对于使自演进大语言模型能够适应知识快速增长的挑战并保持有效性至关重要。然而,尽管其重要性不言而喻,在持续学习环境下为大语言模型建立统计可靠性保障的研究却鲜有涉及,尤其是在持续领域预训练场景中。保形预测已被证明能为大语言模型提供正确性保证,但在持续领域预训练中面临重大挑战:测试数据往往来源于未知或动态变化的领域分布,在此情况下保形预测可能无法继续提供有效保证。此外,当需要高覆盖率时,保形预测可能对无法回答的查询产生过大的预测集,从而降低信息量。为解决这些挑战,我们提出了一种自适应拒绝与非可交换保形预测框架。我们的方法首先利用基于Transformer的聚类估计测试集中问题在跨领域的分布,然后相应地对校准数据进行重加权或重采样。在此基础上,自适应拒绝保形预测允许大语言模型在其置信度或能力发生显著变化时有选择地拒绝回答。大量实验表明,我们的框架在持续领域预训练场景下显著提升了保形预测的有效性和可靠性。我们的代码公开于:https://anonymous.4open.science/r/CPCL-8C12/