Hanabi (card game) explained

Hanabi
Designer:Antoine Bauza
Publisher:R&R Games, Cocktail Games, Abacus Spiele
Image Caption:The box cover of Hanabi
Players:2 to 5
Ages:8 and up
Setup Time:5 minutes
Playing Time:20–30 minutes
Random Chance:Medium
Skills:Deduction, Memory, Cooperation, Planning

Hanabi (from Japanese 花火, fireworks) is a cooperative card game created by French game designer Antoine Bauza and published in 2010.[1] Players are aware of other players' cards but not their own, and attempt to play a series of cards in a specific order to set off a simulated fireworks show. The types of information that players may give to each other is limited, as is the total amount of information that can be given during the game. In 2013, Hanabi won the Spiel des Jahres, an industry award for best board game of the year.

Gameplay

The Hanabi deck contains cards in five suits (white, yellow, green, blue, and red): three 1s, two each of 2s, 3s, and 4s, and one 5. The game begins with 8 available information tokens and 3 fuse tokens. To start the game, players are dealt a hand containing five cards (four for 4 or 5 players). As in blind man's bluff, players can see each other's cards but they cannot see their own. Play proceeds around the table; each turn, a player must take one of the following actions:

The game ends immediately when either all fuse tokens are used up, resulting in a game loss, or all 5s have been played successfully, leading to a game win. Otherwise, play continues until the deck runs out, and for one full round after that. At the end of the game, the values of the highest cards in each suit are summed, resulting in a total score out of a possible 25 points.

Reception

Hanabi received positive reviews. Board Game Quest awarded the game four and a half stars, praising its uniqueness, accessibility and engagement.[2] Similarly, The Opinionated Gamers also praised the game's engagement and addictiveness.[3] It won several awards, including the 2013 Spiel des Jahres winner and 2013 Fairplay À la carte Award winner.[4] [5] Hanabi also placed sixth place in the 2013 Deutscher Spiele Preis.[6]

Computer Hanabi

Hanabi is a cooperative game of imperfect information.

Computer programs which play Hanabi can either engage in self-play or "ad hoc team play". In self-play, multiple instances of the program play with each other on a team. They thus share a carefully honed strategy for communication and play, though of course they are not allowed to illegally share any information about each game with other instances of the program.

In ad hoc team play, the program plays with other arbitrary programs or human players.

A variety of computer programs have been developed by hand-coding rule-based strategies. The best programs, such as WTFWThat, achieved near-perfect results in self-play with five players, with an average score of 24.9 out of 25.[7] [8]

AI challenge

In 2019, DeepMind proposed Hanabi as an ideal game with which to establish a new benchmark for Artificial intelligence research in cooperative play.[9] [10]

In self-play mode, the challenge is to develop a program which can learn from scratch to play well with other instances of itself. Such programs achieve only about 15 points per game as of 2019, far worse than hand-coded programs. However, this gap has narrowed significantly as of 2020, with the Simplified Action Decoder achieving scores around 24.[11]

Ad hoc team play is a far greater challenge for AI, because "Hanabi elevates reasoning about the beliefs and intentions of other agents to the foreground". Playing at human levels with ad hoc teams requires the algorithms to learn and develop communication conventions and strategies over time with other players via a theory of mind. Computer programs developed for self-play fail badly when playing on ad hoc teams, since they don't know how to learn to adapt to the way other players play. Hu et al. demonstrated that learning symmetry-invariant strategies helps AI agents avoid learning uninterpretable conventions, improving their performance when matched with separately trained AI agents (scoring around 22), and with humans (scoring around 16 vs. a baseline self-play model that scored around 9).

Deepmind released an open source code framework to facilitate research, called the Hanabi Learning Environment.

See also

External links

Notes and References

  1. Web site: Hanabi Board Game BoardGameGeek. boardgamegeek.com. 2016-01-24.
  2. Web site: Mastrangeli. Tony. 2014-02-25. Hanabi Review. 2022-02-24. Board Game Quest. en-US.
  3. Web site: Wray. Chris. 2015-12-29. SdJ Re-Reviews #35: Hanabi. 2022-02-25. The Opinionated Gamers. en.
  4. Web site: Fairplay Online: À la carte prize 2013. (in German)
  5. Web site: Spiel des Jahres official site: 2013 winner.
  6. Web site: Preisträger – SPIEL Messe . 2022-12-22 . . 2020-11-03 . https://web.archive.org/web/20201103224339/https://www.spiel-messe.com/de/deutscher-spielepreis/preistraeger/ . dead .
  7. Cox. Christopher. De Silva. Jessica. Deorsey. Philip. Kenter. Franklin H. J.. Retter. Troy. Tobin. Josh. December 2014. How to Make the Perfect Fireworks Display: Two Strategies for Hanabi. Mathematics Magazine. en. 88. 5. 323–336. 10.4169/math.mag.88.5.323. 124445429. 0025-570X.
  8. Bowling. Michael. Bellemare. Marc G.. Larochelle. Hugo. Mourad. Shibl. Dunning. Iain. Hughes. Edward. Moitra. Subhodeep. Dumoulin. Vincent. Parisotto. Emilio. 2019-02-01. The Hanabi Challenge: A New Frontier for AI Research. cs.LG. en. 1902.00506v1.
  9. Web site: The next big challenge for Google's A.I. is a card game you've never heard of. www.digitaltrends.com. 9 February 2019. 2019-07-04.
  10. Web site: A cooperative benchmark: Announcing the Hanabi Learning Environment. www.marcgbellemare.info. 2019-07-04.
  11. Web site: Hu . Hengyuan . Lerer . Adam . Peysakhovich . Alex . Foerster . Jakob . "Other-Play" for Zero-Shot Coordination . International Conference on Machine Learning, 2020.