结合A2C和手牌估值方法的麻将博弈研究
Research on mahjong game combining A2C with hand value evaluation method
衣御寒 1王亚杰 1吴燕燕 1刘松 1张兴慧 1蒋传禹1
作者信息
- 1. 沈阳航空航天大学 工程训练中心, 沈阳 110136
- 折叠
摘要
针对大众麻将中对手牌信息利用不充分的问题,提出了手牌估值方法,并设计了基础麻将程序(MJE).为进一步提升麻将AI的博弈能力,使用深度强化学习方法设计了麻将AI(MJE-RL).首先,通过MJE自对弈生成深度学习的训练数据.其次,根据训练集、测试集和对比实验的结果,选择效果最好的模型作为强化学习的预训练模型.最后,使用优势演说-评论家模型作为强化学习的主要框架,将训练好的深度学习模型作为演说家部分进行决策,通过MJE-RL与MJE的对弈不断提升麻将AI的博弈能力.实验结果显示,MJE-RL的胜率比MJE高4.08%,点炮率比MJE低3.02%,表明MJE-RL在攻守两端都有提升,达到了提升麻将AI牌力的目的.
Abstract
To address the underutilizing hand information in popular mahjong, this paper designs a hand valuation method and a basic mahjong program ( MJE) .Mahjong AI ( MJE-RL) is designed by using the deep reinforcement learning approach to further improve its gaming ability.First, the training data of deep learning is generated by MJE' s self-play.Second, the best model is selected as the pre-training model of reinforcement learning, according to the results of training set, test set and comparison experiment.Finally, the Advantage Actor-Critic ( A2C) model is employed as the main framework of reinforcement learning.The well-trained deep learning model is used as the Actor to make decisions, and the game ability of mahjong AI is constantly improved by playing between MJE-RL and MJE.Our experimental results indicate the winning rate of MJE-RL is 4 .08% higher than that of MJE and the rate of Win by Discard is 3.02% lower than that of MJE.Meanwhile, it is shown that MJE-RL markedly improves both offensive and defensive fronts, demonstrating improved overall strength of mahjong AI.
关键词
麻将/非完备信息/深度强化学习/A2CKey words
popular mahjong/incomplete information/deep reinforcement learning/A2C引用本文复制引用
基金项目
辽宁省兴辽英才计划项目(XLYC1906003)
出版年
2024