blog.mathieuacher.com/GTP2AndChess

Preview meta tags from the blog.mathieuacher.com website.

Linked Hostnames

7

Search Engine Appearance

Google

https://blog.mathieuacher.com/GTP2AndChess

GPT-2 and Chess

Shawn Presser has released an intringuing chess engine based on deep learning-based language model (GPT-2). The model was trained on the Kingbase dataset (3.5 million chess games in PGN notation) in 24 hours using 146 TPUs (ouch!). The engine is purely based on text prediction with no concept of chess. Though GPT-2 has already delivered promising/bluffing results for text generation, one can be skeptical and wonder whether it does work for chess.



Bing

GPT-2 and Chess

https://blog.mathieuacher.com/GTP2AndChess

Shawn Presser has released an intringuing chess engine based on deep learning-based language model (GPT-2). The model was trained on the Kingbase dataset (3.5 million chess games in PGN notation) in 24 hours using 146 TPUs (ouch!). The engine is purely based on text prediction with no concept of chess. Though GPT-2 has already delivered promising/bluffing results for text generation, one can be skeptical and wonder whether it does work for chess.



DuckDuckGo

https://blog.mathieuacher.com/GTP2AndChess

GPT-2 and Chess

Shawn Presser has released an intringuing chess engine based on deep learning-based language model (GPT-2). The model was trained on the Kingbase dataset (3.5 million chess games in PGN notation) in 24 hours using 146 TPUs (ouch!). The engine is purely based on text prediction with no concept of chess. Though GPT-2 has already delivered promising/bluffing results for text generation, one can be skeptical and wonder whether it does work for chess.

  • General Meta Tags

    8
    • title
      GPT-2 and Chess – Mathieu Acher – Professor in Computer Science
    • charset
      utf-8
    • Content-Type
      text/html; charset=utf-8
    • X-UA-Compatible
      IE=edge
    • viewport
      width=device-width, initial-scale=1.0, maximum-scale=1.0
  • Open Graph Meta Tags

    2
    • og:description
      Shawn Presser has released an intringuing chess engine based on deep learning-based language model (GPT-2). The model was trained on the Kingbase dataset (3.5 million chess games in PGN notation) in 24 hours using 146 TPUs (ouch!). The engine is purely based on text prediction with no concept of chess. Though GPT-2 has already delivered promising/bluffing results for text generation, one can be skeptical and wonder whether it does work for chess.
    • og:title
      GPT-2 and Chess
  • Link Tags

    2
    • alternate
      /feed.xml
    • stylesheet
      /style.css

Emails

1

Links

18