系统仿真学报 ›› 2026, Vol. 38 ›› Issue (2): 399-415.doi: 10.16182/j.issn1004731x.joss.25-0996

• 博弈与推演评估 • 上一篇    

基于演化博弈的生成式人工智能幻觉应对分析

闫强, 张倩语, 魏娜   

  1. 北京邮电大学 经济管理学院,北京 100876
  • 收稿日期:2025-10-16 修回日期:2025-12-03 出版日期:2026-02-18 发布日期:2026-02-11
  • 通讯作者: 魏娜
  • 第一作者简介:闫强(1972-),男,教授,博士,研究方向为人工智能风险治理和智能人机交互与个体决策。
  • 基金资助:
    国家社会科学基金(24BZZ076)

Evolutionary Game-based Analysis of Responses to Hallucinations in Generative Artificial Intelligence

Yan Qiang, Zhang Qianyu, Wei Na   

  1. School of Economics and Management of Beijing University of Posts and Telecommunications, Beijing 100876, China
  • Received:2025-10-16 Revised:2025-12-03 Online:2026-02-18 Published:2026-02-11
  • Contact: Wei Na

摘要:

生成式人工智能尤其是大语言模型的广泛应用引发了“幻觉”这一社会风险。为解决现有研究主要聚焦于技术层面的幻觉缓解机制或政策层面的监管框架设计,缺乏对“大模型-用户-监管者”三方主体在有限理性条件下策略互动演化逻辑的系统性理论阐释,将演化博弈理论引入生成式人工智能治理领域,构建了融合大模型诚实性策略、用户反馈行为、监管者干预的三方动态博弈模型,揭示了多元主体在成本-收益权衡下的策略选择动态演化路径及其稳定性条件。结果表明:系统在合理参数条件下可收敛至“大模型诚实回答-用户积极反馈-监管者积极监管”的理想均衡;用户积极反馈初始意愿通过双重信号效应同步加速大模型诚实化进程与监管响应强度;激励机制呈现非对称敏感性,用户对正向激励最为敏感,监管惩罚对模型合规形成刚性约束,协同收益则在长期发挥稳定作用。

关键词: 生成式人工智能, 大语言模型, AI幻觉, 演化博弈, 协同治理

Abstract:

The accelerated deployment of generative artificial intelligence, particularly large language models, has amplified the social risks of hallucinations, posing systemic threats to the credibility of the information ecosystem, the effectiveness of users’ cognitive decision-making, and the governance security in the public domain. Research primarily focuses on hallucination mitigation mechanisms at the technical level or the design of regulatory frameworks at the policy level, lacking a systematic theoretical analysis of the evolutionary logic of strategic interactions among the “large language models, users, and regulators” under conditions of bounded rationality. By introducing evolutionary game theory into the field of generative artificial intelligence governance, a tripartite dynamic game model integrating the honesty strategies of large language models, user feedback behaviors, and regulatory interventions was constructed. This model revealed the dynamic evolutionary paths of strategy selection and their stability conditions for multiple actors under cost-benefit trade-offs. Research shows that under reasonable parameters, the system can converge to the optimal equilibrium of “honest responding of large language models, active feedback of users, and proactive oversight of regulators”. The initial willingness of users to provide positive feedback accelerates both the honesty process of large language models and the intensity of regulatory responses simultaneously through the dual signal effect. Incentive mechanisms exhibit asymmetric sensitivity: Users are most sensitive to positive incentives; regulatory penalties form rigid constraints on model compliance, and the collaborative benefits play a stable role in the long term. Accordingly, it is necessary to strengthen user feedback incentives, advance regulatory technology empowerment, and optimize institutional collaborative mechanisms. These measures aim to build a governance ecosystem characterized by tripartite collaboration, cost hedging, and risk sharing, thereby providing theoretical support and policy paths for the construction of trustworthy AI and the governance of hallucinations.

Key words: generative artificial intelligence, large language model, AI hallucination, evolutionary game theory, collaborative governance

中图分类号: