Addictive Games: Case Study on Multi-Armed Bandit Game
Abstract
:1. Introduction
2. Theoretical Framework
2.1. Multiarmed Bandit
2.2. Reward Mechanism in Games
2.3. Motions in Mind and Internal Energy Change in Games
3. Methodology
3.1. Energy Difference in Games
3.2. Upper Confidence Bound Method
- Initialize the number of round, random generator, and arm choices (line 12–17). Then, try it for each arm (line 18).
- Calculate the score for each arm randomly (line 13–15) and according to Formula (7) (line 20–21), of which the arm with the largest score is then selected.
- Then, based on the observed selection results, update t (line 16) and (line 22).
Algorithm 1: UCB Algorithm (Modified from original source [18] to the source code given at https://github.com/KANG-XIAOHAN/Multi-Armed accessed on 9 December 2021). |
3.3. Experiment Setup
4. Results and Analysis
4.1. Psychological Gap Expressed by Energy Difference
4.2. Link between Satisfaction and Competitive in Game Playing
5. Discussion
5.1. Application with Player Fairness Domain
5.2. Why Is the Multiarmed Bandit Addictive?
5.3. Limitation
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Skinner, B.F. The Behavior of Organisms: And Experimental Analysis; Appleton-Century-Crofts: New York, NY, USA, 1938; p. 457. [Google Scholar]
- Pfaus, J.G.; Everitt, B.J. The psychopharmacology of sexual behavior. Psychopharmacology 1995, 65, 743–758. [Google Scholar]
- Demos, K.E.; Heatherton, T.F.; Kelley, W.M. Individual differences in nucleus accumbens activity to food and sexual images predict weight gain and sexual behavior. J. Neurosci. 2012, 32, 5549–5552. [Google Scholar] [CrossRef] [PubMed]
- Balleine, B.W.; Daw, N.D.; O’Doherty, J.P. Multiple forms of value learning and the function of dopamine. In Neuroeconomics; Elsevier: Amsterdam, The Netherlands, 2009; pp. 367–387. [Google Scholar]
- Flagel, S.B.; Clark, J.J.; Robinson, T.E.; Mayo, L.; Czuj, A.; Willuhn, I.; Akers, C.A.; Clinton, S.M.; Phillips, P.E.; Akil, H. A selective role for dopamine in stimulus—Reward learning. Nature 2011, 469, 53–57. [Google Scholar] [CrossRef] [PubMed]
- Volkow, N.D.; Wang, G.J.; Fowler, J.S.; Tomasi, D.; Telang, F. Addiction: Beyond dopamine reward circuitry. Proc. Natl. Acad. Sci. USA 2011, 108, 15037–15042. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Green, C.S.; Bavelier, D. Learning, attentional control, and action video games. Curr. Biol. 2012, 22, R197–R206. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Bellis, M. A Brief History of Slot Machines. 2019. Available online: https://www.thoughtco.com/history-of-slot-machines-liberty-bell-1992409 (accessed on 10 December 2021).
- Iida, H.; Takeshita, N.; Yoshimura, J. A metric for entertainment of boardgames: Its implication for evolution of chess variants. In Entertainment Computing; Springer: Berlin/Heidelberg, Germany, 2003; pp. 65–72. [Google Scholar]
- Xiaohan, K.; Khalid, M.N.A.; Iida, H. Player Satisfaction Model and its Implication to Cultural Change. IEEE Access 2020, 8, 184375–184382. [Google Scholar] [CrossRef]
- Malone, T.W. What makes things fun to learn? Heuristics for designing instructional computer games. In Proceedings of the 3rd ACM SIGSMALL Symposium and the First SIGPC Symposium on Small Systems, New York, NY, USA, 18–19 September 1980; pp. 162–169. [Google Scholar]
- Bartle, R. Hearts, clubs, diamonds, spades: Players who suit MUDs. J. Mud Res. 1996, 1, 19. [Google Scholar]
- Yee, N. Motivations of play in MMORPGs. 2005. Available online: https://summit.sfu.ca/item/212 (accessed on 10 December 2021).
- Multi-Armed Bandit Wiki. Available online: https://en.wikipedia.org/wiki/Multi-armed_bandit/ (accessed on 9 December 2021).
- McCall, B.P. Multi-Armed Bandit Allocation Indices (JC Gittins). SIAM Rev. 1991, 33, 154. [Google Scholar] [CrossRef]
- Berry, D.A.; Fristedt, B. Bandit Problems: Sequential Allocation of Experiments (Monographs on Statistics and Applied Probability); Chapman and Hall: London, UK, 1985; Volume 5, p. 7. [Google Scholar]
- Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
- Zhang, S. Reinforcement Learning: An Introduction. Available online: https://github.com/ShangtongZhang/reinforcement-learning-an-introduction/blob/master/chapter02/ten_armed_testbed.py (accessed on 9 December 2021).
- Skinner, B.F. ‘Superstition’ in the pigeon. J. Exp. Psychol. 1948, 38, 168. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Mcgonigal, J. Game Can Make a Better World. 2010. Available online: https://www.ted.com/talks/jane_mcgonigal_gaming_can_make_a_better_world (accessed on 10 December 2021).
- Wang, J. Research on mergers and acquisitions of the international hotel group–Take Marriott M & A Starwood as an example. In Proceedings of the 2018 3rd International Conference on Humanities Science, Management and Education Technology (HSMET 2018), Nanjing, China, 8–10 June 2018; Atlantis Press: Paris, France; pp. 743–746. [Google Scholar]
- Kerr, J.; Slocum, J.W., Jr. Managing corporate culture through reward systems. Acad. Manag. Perspect. 2005, 19, 130–138. [Google Scholar] [CrossRef]
- Iigaya, K.; Hauser, T.U.; Kurth-Nelson, Z.; O’Doherty, J.P.; Dayan, P.; Dolan, R.J. The value of what’s to come: Neural mechanisms coupling prediction error and the utility of anticipation. Sci. Adv. 2020, 6, eaba3828. [Google Scholar] [CrossRef] [PubMed]
- Fiorillo, C.D.; Tobler, P.N.; Schultz, W. Discrete coding of reward probability and uncertainty by dopamine neurons. Science 2003, 299, 1898–1902. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Shizgal, P.; Arvanitogiannis, A. Gambling on dopamine. Science 2003, 299, 1856–1858. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Iida, H.; Khalid, M.N.A. Using games to study law of motions in mind. IEEE Access 2020, 8, 138701–138709. [Google Scholar] [CrossRef]
- Ferster, C.B.; Skinner, B.F. Schedules of Reinforcement. 1957. Available online: https://psycnet.apa.org/doiLanding?doi=10.1037%2F10627-000 (accessed on 10 December 2021).
- Carathéodory, C. Untersuchungen über die Grundlagen der Thermodynamik. Math. Ann. 1909, 67, 355–386. [Google Scholar] [CrossRef] [Green Version]
- Lai, T.L.; Robbins, H. Consistency and asymptotic efficiency of slope estimates in stochastic approximation schemes. Z. Wahrscheinlichkeitstheorie Verwandte Geb. 1981, 56, 329–360. [Google Scholar] [CrossRef]
- Lai, T.L.; Robbins, H. Asymptotically efficient adaptive allocation rules. Adv. Appl. Math. 1985, 6, 4–22. [Google Scholar] [CrossRef] [Green Version]
- Higgins, E.T. Self-discrepancy: A theory relating self and affect. Psychol. Rev. 1987, 94, 319. [Google Scholar] [CrossRef] [PubMed]
- Agarwal, S.; Khalid, M.N.A.; Iida, H. Game refinement theory: Paradigm shift from performance optimization to comfort in mind. Entertain. Comput. 2019, 32, 100314. [Google Scholar] [CrossRef]
- Lewis, Z.H.; Swartz, M.C.; Lyons, E.J. What’s the point?: A review of reward systems implemented in gamification interventions. Games Health J. 2016, 5, 93–99. [Google Scholar] [CrossRef] [PubMed]
Notation | Game Context | Notation | Physics Context |
---|---|---|---|
y | solved uncertainty | x | displacement |
t | progress or length | t | time |
p | win rate | v | velocity |
m | win hardness | M | mass |
a | acceleration | g | gravitational acceleration |
Notation | Game-Side | Player-Side |
---|---|---|
intuitive probability based game velocity | entry difficulty | |
return rate based game velocity | engagement(confident) |
Arm Setting Distribution | m | Arm Numbers |
---|---|---|
(0,1)(0,1)(0,1) | 0 | 3 |
(−1.03,1)(−1.22,1)(−1.75,1) | 0.1 | 3 |
(−0.77,1)(−0.68,1)(−1.12,1) | 0.2 | 3 |
(−0.14,1)(−0.51,1)(−0.99,1) | 0.3 | 3 |
(−1.03,1)(−0.55,1)(0.71,1) | 0.4 | 3 |
(0.30,1)(−0.56,1)(0.22,1) | 0.5 | 3 |
(2.04,1)(0.20,1)(−0.71,1) | 0.6 | 3 |
(0.61,1)(0.73,1)(0.25,1) | 0.7 | 3 |
(4.13,1)(0.77,1)(0.31,1) | 0.8 | 3 |
(4.16,1)(1.32,1)(0.80,1) | 0.9 | 3 |
(4.27,1)(4.27,1)(4.27,1) | 1.0 | 3 |
Arm Setting Distribution | m | Arm Numbers |
---|---|---|
(−4.5,1)(−4.5,1)(−4.5,1)(−4.5,1)(−4.5,1)(−4.5,1)(−4.5,1)(−4.5,1)(−4.5,1)(−4.5,1) | 1.0 | 10 |
(−2.07,1)(−0.94,1) (−1.42,1) (−4.77,1) (−0.90,1) (−1.28,1) (−1.07,1) (−1.04,1) (−1.36,1) (−1.46,1) | 0.9 | 10 |
(−0.43,1)(−0.78,1)(−3.46,1)(−1.01,1)(−0.75,1)(−0.65,1)(−1.21,1)(−1.22,1) (−0.47,1) (−0.59,1) | 0.8 | 10 |
(−0.87,1)(0.20,1)(−0.80,1)(−0.86,1)(−1.07,1)(0.17,1)(−1.40,1)(−0.21,1)(0.19,1)(−1.62,1) | 0.7 | 10 |
(−0.09,1)(0.75,1)(0.52,1)(1.36,1)(−0.83,1)(−1.53,1)(−2.22,1)(−0.58,1)(−1.18,1)(−0.09,1) | 0.6 | 10 |
(−0.92,1)(1.13,1)(−0.80,1)(−0.82,1)(0.50,1)(0.19,1)(0.53,1)(0.78,1)(0.24,1)(−0.94,1) | 0.5 | 10 |
(0.74,1)(0.82,1)(0.11,1)(0.17,1)(0.65,1)(0.06,1)(−0.55,1)(0.31,1)(−0.23,1)(0.62,1) | 0.4 | 10 |
(−1.99,1)(−0.90,1)(−0.29,1)(−1.55,1)(−1.10,1)(−0.75,1)(−0.50,1)(−0.68,1)(−0.42,1) (−1.27,1) | 0.3 | 10 |
(0.39,1)(0.69,1)(0.39,1)(1.77,1)(0.89,1)(1.60,1)(0.92,1)(0.79,1)(1.03,1)(0.73,1) | 0.2 | 10 |
(2.18,1)(1.11,1)(1.91,1)(1.25,1)(1.76,1)(1.22,1)(0.53,1)(1.01,1)(1.33,1) (2.50,1) | 0.1 | 10 |
(0,1)(0,1)(0,1)(0,1)(0,1)(0,1)(0,1)(0,1)(0,1) (0,1) | 0 | 10 |
m | Actual Probability | Intuitive Probability | Energy Difference |
---|---|---|---|
1.0 | 0.00000 | 0.00000 | 0.00000 |
0.9 | 0.00001 | 0.08100 | 0.01206 |
0.8 | 0.00014 | 0.23491 | 0.08444 |
0.7 | 0.00150 | 0.34777 | 0.15776 |
0.6 | 0.33433 | 0.38396 | 0.03282 |
0.5 | 0.66537 | 0.50869 | −0.04202 |
0.4 | 0.66776 | 0.59327 | −0.00998 |
0.3 | 0.99994 | 0.69022 | 0.29504 |
0.2 | 0.99983 | 0.81218 | 0.24744 |
0.1 | 0.99995 | 0.89007 | 0.17408 |
0.0 | 0.99997 | 1.00000 | −0.00001 |
m | Actual Probability | Intuitive Probability | Energy Difference |
---|---|---|---|
1.0 | 0.00000 | 0.00000 | 0.00000 |
0.9 | 0.00002 | 0.17683 | 0.05147 |
0.8 | 0.00013 | 0.30717 | 0.13074 |
0.7 | 0.99744 | 0.58729 | 0.27960 |
0.6 | 0.99973 | 0.88805 | 0.17603 |
0.5 | 0.99978 | 0.91979 | 0.13527 |
0.4 | 0.99962 | 0.81684 | 0.24365 |
0.3 | 0.00003 | 0.46733 | 0.23266 |
0.2 | 0.99989 | 0.96995 | 0.05632 |
0.1 | 0.99988 | 0.99971 | −0.00033 |
0.0 | 0.99990 | 1.00000 | −0.00019 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kang, X.; Ri, H.; Khalid, M.N.A.; Iida, H. Addictive Games: Case Study on Multi-Armed Bandit Game. Information 2021, 12, 521. https://doi.org/10.3390/info12120521
Kang X, Ri H, Khalid MNA, Iida H. Addictive Games: Case Study on Multi-Armed Bandit Game. Information. 2021; 12(12):521. https://doi.org/10.3390/info12120521
Chicago/Turabian StyleKang, Xiaohan, Hong Ri, Mohd Nor Akmal Khalid, and Hiroyuki Iida. 2021. "Addictive Games: Case Study on Multi-Armed Bandit Game" Information 12, no. 12: 521. https://doi.org/10.3390/info12120521