[an error occurred while processing this directive] [an error occurred while processing this directive] [an error occurred while processing this directive]
[an error occurred while processing this directive]
Information Science and Engineering

Optimization of MCTS algorithm for Tibetan Jiu Chess by incorporating prior knowledge

  • Yajie WANG ,
  • Feng GU ,
  • Song LIU ,
  • Jingyi YANG ,
  • Shipeng WANG
Expand
  • Engineering Training Center,Shenyang Aerospace University,Shenyang 110136,China

Received date: 2024-12-24

  Revised date: 2025-01-09

  Accepted date: 2025-01-12

  Online published: 2025-08-19

Abstract

Tibetan Jiu Chess, a traditional folk chess game, is a complete information game that carries the profound Tibetan civilization and splendid culture. In view of the complexity of the rule system and the diversity of the game changes, the traditional game search algorithm is unable to cope with the vast game board and complex strategies. In order to improve the intelligence level of Tibetan Jiu Chess, a Monte Carlo tree search (MCTS) algorithm optimization strategy incorporating prior knowledge was proposed. The strategy was based on deep reinforcement learning in the key phases of layout planning and move strategy,and the strategy selection optimization function and evaluation function were designed by integrating the prior knowledge of domain experts. The search process of MCTS was efficiently guided by functions,and the best model for high-quality tessellation could be trained. Experimental results show that the improved MCTS algorithm achieves significant performance in the game.

Cite this article

Yajie WANG , Feng GU , Song LIU , Jingyi YANG , Shipeng WANG . Optimization of MCTS algorithm for Tibetan Jiu Chess by incorporating prior knowledge[J]. Journal of Shenyang Aerospace University, 2025 , 42(4) : 59 -67 . DOI: 10.3969/j.issn.2095-1248.2025.04.009

[an error occurred while processing this directive]
[1]
李霞丽,吴立成,李永集.基于棋型的藏族“久” 棋计算机博弈研究[J].智能系统学报201813(4):577-583.

[2]
Silver D Huang A Maddison C J,et al.Mastering the game of go with deep neural networks and tree search[J].Nature2016529(7587):484-489.

[3]
McGrath T Kapishnikov A Tomašev N,et al.Acquisition of chess knowledge in AlphaZero[J].Proceedings of the National Academy of Sciences of the United States of America2022119(47):e2206625119.

[4]
Schmid M Moravčík M Burch N,et al.Student of games:a unified learning algorithm for both perfect and imperfect information games[J].Science Advances20239(46):3256.

[5]
Li X L Deng S T.Review of research on computer games for Tibetan Jiu chess[C]//IEEE 14th International Conference on Dependable, Autonomic and Secure Computing. Auckland:IEEE,2016: 97-99.

[6]
Li X L Wang S Lv Z Y,et al.Strategy research based on chess shapes for Tibetan Jiu computer game[J].ICGA Journal201840(3):318-328.

[7]
张小川,杨小漫,涂飞,等.融合经验知识与深度强化学习的久棋Alpha-Beta算法优化研究[J].重庆理工大学学报(自然科学)202438(5):115-120.

[8]
Naderzadeh Y Grosu D Chinnam R B.PPB-MCTS:a novel distributed-memory parallel partial-backpropagation Monte Carlo tree search algorithm[J].Journal of Parallel and Distributed Computing2024193:104944.

[9]
Wang Y J Liang K Qiao J L,et al.The application of improved UCT combined with neural network in Tibetan Jiu chess[J].International Journal of Wireless and Mobile Computing202223(1):22.

[10]
Li X L Chen Y D Zhang Y Y,et al.A phased game algorithm combining deep reinforcement learning and UCT for Tibetan Jiu chess[C]//2023 IEEE 47th Annual Computers,Software,and Applications Conference.Torino:IEEE, 2023:390-395.

[11]
中国大学生计算机博弈大赛组委会. 久棋规则及棋谱简介[EB/OL].(2019-04-29)[2024-05-06].

[12]
Wang Q He Y Q Tang C L.Mastering construction heuristics with self-play deep reinforcement learning[J].Neural Computing and Applications202335(6):4723-4738.

[13]
Shen Q W Ding M Li S Q,et al.Research on jiuqi game strategy based on chess shape[C]//2020 3rd International Conference on Algorithms,Computing and Artificial Intelligence.Sanya:ACM,2020:1-5.

[14]
张小川,刘溜,陈龙,等.一种非遗藏族久棋项目计算机博弈智能体的评估方法[J].重庆理工大学学报(自然科学)202135(12):119-126.

[15]
Khan M Olivier J.Regression to the mean:Estimation and adjustment under the bivariate normal distribution[J].Communications in Statistics-Theory and Methods202352(19):6972-6990.

[16]
Dong S Wang P Abbas K.A survey on deep learning and its applications[J].Computer Science Review202140:100379.

[17]
Silver D Schrittwieser J Simonyan K,et al.Mastering the game of go without human knowledge[J].Nature2017550(7676):354-359.

[18]
Tan M X Le Q V.Efficient Net:rethinking model scaling for convolutional neural networks[EB/OL].(2020-09-11)[2024-06-05].

[19]
Silver D Hubert T Schrittwieser J,et al.A general reinforcement learning algorithm that masters chess,shogi,and go through self-play[EB/OL].(2020-09-11)[2024-06-05].

[20]
Ribeiro E S Araújo L R G Chaves G T L,et al.Distance-based loss function for deep feature space learning of convolutional neural networks[J].Computer Vision and Image Understanding2024249:104184.

Outlines

/

[an error occurred while processing this directive]