Loading…
Mastering the game of Go with deep neural networks and tree search
The game of Go has long been viewed as the most challenging of classic games for artificial intelligence owing to its enormous search space and the difficulty of evaluating board positions and moves. Here we introduce a new approach to computer Go that uses ‘value networks’ to evaluate board positio...
Saved in:
Published in: | Nature (London) 2016-01, Vol.529 (7587), p.484-489 |
---|---|
Main Authors: | , , , , , , , , , , , , , , , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c756t-29169ab0b0ca120ba3231833251d5d051b8e8ba996b8a08665069a80d90d86693 |
---|---|
cites | cdi_FETCH-LOGICAL-c756t-29169ab0b0ca120ba3231833251d5d051b8e8ba996b8a08665069a80d90d86693 |
container_end_page | 489 |
container_issue | 7587 |
container_start_page | 484 |
container_title | Nature (London) |
container_volume | 529 |
creator | Silver, David Huang, Aja Maddison, Chris J. Guez, Arthur Sifre, Laurent van den Driessche, George Schrittwieser, Julian Antonoglou, Ioannis Panneershelvam, Veda Lanctot, Marc Dieleman, Sander Grewe, Dominik Nham, John Kalchbrenner, Nal Sutskever, Ilya Lillicrap, Timothy Leach, Madeleine Kavukcuoglu, Koray Graepel, Thore Hassabis, Demis |
description | The game of Go has long been viewed as the most challenging of classic games for artificial intelligence owing to its enormous search space and the difficulty of evaluating board positions and moves. Here we introduce a new approach to computer Go that uses ‘value networks’ to evaluate board positions and ‘policy networks’ to select moves. These deep neural networks are trained by a novel combination of supervised learning from human expert games, and reinforcement learning from games of self-play. Without any lookahead search, the neural networks play Go at the level of state-of-the-art Monte Carlo tree search programs that simulate thousands of random games of self-play. We also introduce a new search algorithm that combines Monte Carlo simulation with value and policy networks. Using this search algorithm, our program AlphaGo achieved a 99.8% winning rate against other Go programs, and defeated the human European Go champion by 5 games to 0. This is the first time that a computer program has defeated a human professional player in the full-sized game of Go, a feat previously thought to be at least a decade away.
A computer Go program based on deep neural networks defeats a human professional player to achieve one of the grand challenges of artificial intelligence.
AlphaGo computer beats Go champion
The victory in 1997 of the chess-playing computer Deep Blue in a six-game series against the then world champion Gary Kasparov was seen as a significant milestone in the development of artificial intelligence. An even greater challenge remained — the ancient game of Go. Despite decades of refinement, until recently the strongest computers were still playing Go at the level of human amateurs. Enter AlphaGo. Developed by Google DeepMind, this program uses deep neural networks to mimic expert players, and further improves its performance by learning from games played against itself. AlphaGo has achieved a 99% win rate against the strongest other Go programs, and defeated the reigning European champion Fan Hui 5–0 in a tournament match. This is the first time that a computer program has defeated a human professional player in even games, on a full, 19 x 19 board, in even games with no handicap. |
doi_str_mv | 10.1038/nature16961 |
format | article |
fullrecord | <record><control><sourceid>gale_proqu</sourceid><recordid>TN_cdi_proquest_miscellaneous_1761459006</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A661111782</galeid><sourcerecordid>A661111782</sourcerecordid><originalsourceid>FETCH-LOGICAL-c756t-29169ab0b0ca120ba3231833251d5d051b8e8ba996b8a08665069a80d90d86693</originalsourceid><addsrcrecordid>eNqN001v1DAQBmALgei2cOKOLLgUQcrYcRznWCoolYqQ-DhbTjLJpuSrtqPCv8fLFshWKdQ-WIkevx5ZY0KeMDhiEKvXvfGTRSYzye6RFROpjIRU6X2yAuAqAhXLPbLv3AUAJCwVD8kel4plIPiKvPlgnEfb9DX1a6S16ZAOFT0d6FXj17REHGmPkzVtWPzVYL85avqSeotIHRpbrB-RB5VpHT6-Xg_I13dvv5y8j84_np6dHJ9HRZpIH_EslGhyyKEwjENuYh4zFcc8YWVShspyhSo3WSZzZUBJmUDwCsoMyvCVxQfkcJs72uFyQud117gC29b0OExOs1QykWQAMtDnN-jFMNk-VPdLcRGHm_uratOibvpq8NYUm1B9LCULI1X8n0oIphRXTAQVLagaewxXN_RYNeH3Tupd_Dz_2YIvxuZSz0NvRfOkowUUZoldUyyWeqcN8xNe7GwIxuN3X5vJOX32-dNu-P_sPPfl1hZ2cM5ipUfbdMb-0Az05ino2VMI-ul1D0x5h-Uf-7v3A3i1BW7c9D_aWZMs5P0EwikIgw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1761243103</pqid></control><display><type>article</type><title>Mastering the game of Go with deep neural networks and tree search</title><source>Nature</source><creator>Silver, David ; Huang, Aja ; Maddison, Chris J. ; Guez, Arthur ; Sifre, Laurent ; van den Driessche, George ; Schrittwieser, Julian ; Antonoglou, Ioannis ; Panneershelvam, Veda ; Lanctot, Marc ; Dieleman, Sander ; Grewe, Dominik ; Nham, John ; Kalchbrenner, Nal ; Sutskever, Ilya ; Lillicrap, Timothy ; Leach, Madeleine ; Kavukcuoglu, Koray ; Graepel, Thore ; Hassabis, Demis</creator><creatorcontrib>Silver, David ; Huang, Aja ; Maddison, Chris J. ; Guez, Arthur ; Sifre, Laurent ; van den Driessche, George ; Schrittwieser, Julian ; Antonoglou, Ioannis ; Panneershelvam, Veda ; Lanctot, Marc ; Dieleman, Sander ; Grewe, Dominik ; Nham, John ; Kalchbrenner, Nal ; Sutskever, Ilya ; Lillicrap, Timothy ; Leach, Madeleine ; Kavukcuoglu, Koray ; Graepel, Thore ; Hassabis, Demis</creatorcontrib><description>The game of Go has long been viewed as the most challenging of classic games for artificial intelligence owing to its enormous search space and the difficulty of evaluating board positions and moves. Here we introduce a new approach to computer Go that uses ‘value networks’ to evaluate board positions and ‘policy networks’ to select moves. These deep neural networks are trained by a novel combination of supervised learning from human expert games, and reinforcement learning from games of self-play. Without any lookahead search, the neural networks play Go at the level of state-of-the-art Monte Carlo tree search programs that simulate thousands of random games of self-play. We also introduce a new search algorithm that combines Monte Carlo simulation with value and policy networks. Using this search algorithm, our program AlphaGo achieved a 99.8% winning rate against other Go programs, and defeated the human European Go champion by 5 games to 0. This is the first time that a computer program has defeated a human professional player in the full-sized game of Go, a feat previously thought to be at least a decade away.
A computer Go program based on deep neural networks defeats a human professional player to achieve one of the grand challenges of artificial intelligence.
AlphaGo computer beats Go champion
The victory in 1997 of the chess-playing computer Deep Blue in a six-game series against the then world champion Gary Kasparov was seen as a significant milestone in the development of artificial intelligence. An even greater challenge remained — the ancient game of Go. Despite decades of refinement, until recently the strongest computers were still playing Go at the level of human amateurs. Enter AlphaGo. Developed by Google DeepMind, this program uses deep neural networks to mimic expert players, and further improves its performance by learning from games played against itself. AlphaGo has achieved a 99% win rate against the strongest other Go programs, and defeated the reigning European champion Fan Hui 5–0 in a tournament match. This is the first time that a computer program has defeated a human professional player in even games, on a full, 19 x 19 board, in even games with no handicap.</description><identifier>ISSN: 0028-0836</identifier><identifier>EISSN: 1476-4687</identifier><identifier>DOI: 10.1038/nature16961</identifier><identifier>PMID: 26819042</identifier><identifier>CODEN: NATUAS</identifier><language>eng</language><publisher>London: Nature Publishing Group UK</publisher><subject>631/378/1788 ; 639/705/1042 ; 639/705/117 ; Algorithms ; Analysis ; Artificial intelligence ; Computer games ; Computers ; Europe ; Evaluation ; Games ; Games, Recreational ; Go (Game) ; Humanities and Social Sciences ; Humans ; Monte Carlo Method ; Monte Carlo simulation ; multidisciplinary ; Neural networks ; Neural Networks (Computer) ; Product development ; Reinforcement (Psychology) ; Science ; Software ; Supervised Machine Learning ; Technology application</subject><ispartof>Nature (London), 2016-01, Vol.529 (7587), p.484-489</ispartof><rights>Springer Nature Limited 2016</rights><rights>COPYRIGHT 2016 Nature Publishing Group</rights><rights>Copyright Nature Publishing Group Jan 28, 2016</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c756t-29169ab0b0ca120ba3231833251d5d051b8e8ba996b8a08665069a80d90d86693</citedby><cites>FETCH-LOGICAL-c756t-29169ab0b0ca120ba3231833251d5d051b8e8ba996b8a08665069a80d90d86693</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/26819042$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Silver, David</creatorcontrib><creatorcontrib>Huang, Aja</creatorcontrib><creatorcontrib>Maddison, Chris J.</creatorcontrib><creatorcontrib>Guez, Arthur</creatorcontrib><creatorcontrib>Sifre, Laurent</creatorcontrib><creatorcontrib>van den Driessche, George</creatorcontrib><creatorcontrib>Schrittwieser, Julian</creatorcontrib><creatorcontrib>Antonoglou, Ioannis</creatorcontrib><creatorcontrib>Panneershelvam, Veda</creatorcontrib><creatorcontrib>Lanctot, Marc</creatorcontrib><creatorcontrib>Dieleman, Sander</creatorcontrib><creatorcontrib>Grewe, Dominik</creatorcontrib><creatorcontrib>Nham, John</creatorcontrib><creatorcontrib>Kalchbrenner, Nal</creatorcontrib><creatorcontrib>Sutskever, Ilya</creatorcontrib><creatorcontrib>Lillicrap, Timothy</creatorcontrib><creatorcontrib>Leach, Madeleine</creatorcontrib><creatorcontrib>Kavukcuoglu, Koray</creatorcontrib><creatorcontrib>Graepel, Thore</creatorcontrib><creatorcontrib>Hassabis, Demis</creatorcontrib><title>Mastering the game of Go with deep neural networks and tree search</title><title>Nature (London)</title><addtitle>Nature</addtitle><addtitle>Nature</addtitle><description>The game of Go has long been viewed as the most challenging of classic games for artificial intelligence owing to its enormous search space and the difficulty of evaluating board positions and moves. Here we introduce a new approach to computer Go that uses ‘value networks’ to evaluate board positions and ‘policy networks’ to select moves. These deep neural networks are trained by a novel combination of supervised learning from human expert games, and reinforcement learning from games of self-play. Without any lookahead search, the neural networks play Go at the level of state-of-the-art Monte Carlo tree search programs that simulate thousands of random games of self-play. We also introduce a new search algorithm that combines Monte Carlo simulation with value and policy networks. Using this search algorithm, our program AlphaGo achieved a 99.8% winning rate against other Go programs, and defeated the human European Go champion by 5 games to 0. This is the first time that a computer program has defeated a human professional player in the full-sized game of Go, a feat previously thought to be at least a decade away.
A computer Go program based on deep neural networks defeats a human professional player to achieve one of the grand challenges of artificial intelligence.
AlphaGo computer beats Go champion
The victory in 1997 of the chess-playing computer Deep Blue in a six-game series against the then world champion Gary Kasparov was seen as a significant milestone in the development of artificial intelligence. An even greater challenge remained — the ancient game of Go. Despite decades of refinement, until recently the strongest computers were still playing Go at the level of human amateurs. Enter AlphaGo. Developed by Google DeepMind, this program uses deep neural networks to mimic expert players, and further improves its performance by learning from games played against itself. AlphaGo has achieved a 99% win rate against the strongest other Go programs, and defeated the reigning European champion Fan Hui 5–0 in a tournament match. This is the first time that a computer program has defeated a human professional player in even games, on a full, 19 x 19 board, in even games with no handicap.</description><subject>631/378/1788</subject><subject>639/705/1042</subject><subject>639/705/117</subject><subject>Algorithms</subject><subject>Analysis</subject><subject>Artificial intelligence</subject><subject>Computer games</subject><subject>Computers</subject><subject>Europe</subject><subject>Evaluation</subject><subject>Games</subject><subject>Games, Recreational</subject><subject>Go (Game)</subject><subject>Humanities and Social Sciences</subject><subject>Humans</subject><subject>Monte Carlo Method</subject><subject>Monte Carlo simulation</subject><subject>multidisciplinary</subject><subject>Neural networks</subject><subject>Neural Networks (Computer)</subject><subject>Product development</subject><subject>Reinforcement (Psychology)</subject><subject>Science</subject><subject>Software</subject><subject>Supervised Machine Learning</subject><subject>Technology application</subject><issn>0028-0836</issn><issn>1476-4687</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2016</creationdate><recordtype>article</recordtype><recordid>eNqN001v1DAQBmALgei2cOKOLLgUQcrYcRznWCoolYqQ-DhbTjLJpuSrtqPCv8fLFshWKdQ-WIkevx5ZY0KeMDhiEKvXvfGTRSYzye6RFROpjIRU6X2yAuAqAhXLPbLv3AUAJCwVD8kel4plIPiKvPlgnEfb9DX1a6S16ZAOFT0d6FXj17REHGmPkzVtWPzVYL85avqSeotIHRpbrB-RB5VpHT6-Xg_I13dvv5y8j84_np6dHJ9HRZpIH_EslGhyyKEwjENuYh4zFcc8YWVShspyhSo3WSZzZUBJmUDwCsoMyvCVxQfkcJs72uFyQud117gC29b0OExOs1QykWQAMtDnN-jFMNk-VPdLcRGHm_uratOibvpq8NYUm1B9LCULI1X8n0oIphRXTAQVLagaewxXN_RYNeH3Tupd_Dz_2YIvxuZSz0NvRfOkowUUZoldUyyWeqcN8xNe7GwIxuN3X5vJOX32-dNu-P_sPPfl1hZ2cM5ipUfbdMb-0Az05ino2VMI-ul1D0x5h-Uf-7v3A3i1BW7c9D_aWZMs5P0EwikIgw</recordid><startdate>20160128</startdate><enddate>20160128</enddate><creator>Silver, David</creator><creator>Huang, Aja</creator><creator>Maddison, Chris J.</creator><creator>Guez, Arthur</creator><creator>Sifre, Laurent</creator><creator>van den Driessche, George</creator><creator>Schrittwieser, Julian</creator><creator>Antonoglou, Ioannis</creator><creator>Panneershelvam, Veda</creator><creator>Lanctot, Marc</creator><creator>Dieleman, Sander</creator><creator>Grewe, Dominik</creator><creator>Nham, John</creator><creator>Kalchbrenner, Nal</creator><creator>Sutskever, Ilya</creator><creator>Lillicrap, Timothy</creator><creator>Leach, Madeleine</creator><creator>Kavukcuoglu, Koray</creator><creator>Graepel, Thore</creator><creator>Hassabis, Demis</creator><general>Nature Publishing Group UK</general><general>Nature Publishing Group</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7QG</scope><scope>7QL</scope><scope>7QP</scope><scope>7QR</scope><scope>7RV</scope><scope>7SN</scope><scope>7SS</scope><scope>7ST</scope><scope>7T5</scope><scope>7TG</scope><scope>7TK</scope><scope>7TM</scope><scope>7TO</scope><scope>7U9</scope><scope>7X2</scope><scope>7X7</scope><scope>7XB</scope><scope>88A</scope><scope>88E</scope><scope>88G</scope><scope>88I</scope><scope>8AF</scope><scope>8AO</scope><scope>8C1</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FH</scope><scope>8FI</scope><scope>8FJ</scope><scope>8FK</scope><scope>8G5</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>ATCPS</scope><scope>AZQEC</scope><scope>BBNVY</scope><scope>BEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>BHPHI</scope><scope>BKSAR</scope><scope>C1K</scope><scope>CCPQU</scope><scope>D1I</scope><scope>DWQXO</scope><scope>FR3</scope><scope>FYUFA</scope><scope>GHDGH</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>H94</scope><scope>HCIFZ</scope><scope>K9.</scope><scope>KB.</scope><scope>KB0</scope><scope>KL.</scope><scope>L6V</scope><scope>LK8</scope><scope>M0K</scope><scope>M0S</scope><scope>M1P</scope><scope>M2M</scope><scope>M2O</scope><scope>M2P</scope><scope>M7N</scope><scope>M7P</scope><scope>M7S</scope><scope>MBDVC</scope><scope>NAPCQ</scope><scope>P5Z</scope><scope>P62</scope><scope>P64</scope><scope>PATMY</scope><scope>PCBAR</scope><scope>PDBOC</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PSYQQ</scope><scope>PTHSS</scope><scope>PYCSY</scope><scope>Q9U</scope><scope>R05</scope><scope>RC3</scope><scope>S0X</scope><scope>SOI</scope><scope>7X8</scope></search><sort><creationdate>20160128</creationdate><title>Mastering the game of Go with deep neural networks and tree search</title><author>Silver, David ; Huang, Aja ; Maddison, Chris J. ; Guez, Arthur ; Sifre, Laurent ; van den Driessche, George ; Schrittwieser, Julian ; Antonoglou, Ioannis ; Panneershelvam, Veda ; Lanctot, Marc ; Dieleman, Sander ; Grewe, Dominik ; Nham, John ; Kalchbrenner, Nal ; Sutskever, Ilya ; Lillicrap, Timothy ; Leach, Madeleine ; Kavukcuoglu, Koray ; Graepel, Thore ; Hassabis, Demis</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c756t-29169ab0b0ca120ba3231833251d5d051b8e8ba996b8a08665069a80d90d86693</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2016</creationdate><topic>631/378/1788</topic><topic>639/705/1042</topic><topic>639/705/117</topic><topic>Algorithms</topic><topic>Analysis</topic><topic>Artificial intelligence</topic><topic>Computer games</topic><topic>Computers</topic><topic>Europe</topic><topic>Evaluation</topic><topic>Games</topic><topic>Games, Recreational</topic><topic>Go (Game)</topic><topic>Humanities and Social Sciences</topic><topic>Humans</topic><topic>Monte Carlo Method</topic><topic>Monte Carlo simulation</topic><topic>multidisciplinary</topic><topic>Neural networks</topic><topic>Neural Networks (Computer)</topic><topic>Product development</topic><topic>Reinforcement (Psychology)</topic><topic>Science</topic><topic>Software</topic><topic>Supervised Machine Learning</topic><topic>Technology application</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Silver, David</creatorcontrib><creatorcontrib>Huang, Aja</creatorcontrib><creatorcontrib>Maddison, Chris J.</creatorcontrib><creatorcontrib>Guez, Arthur</creatorcontrib><creatorcontrib>Sifre, Laurent</creatorcontrib><creatorcontrib>van den Driessche, George</creatorcontrib><creatorcontrib>Schrittwieser, Julian</creatorcontrib><creatorcontrib>Antonoglou, Ioannis</creatorcontrib><creatorcontrib>Panneershelvam, Veda</creatorcontrib><creatorcontrib>Lanctot, Marc</creatorcontrib><creatorcontrib>Dieleman, Sander</creatorcontrib><creatorcontrib>Grewe, Dominik</creatorcontrib><creatorcontrib>Nham, John</creatorcontrib><creatorcontrib>Kalchbrenner, Nal</creatorcontrib><creatorcontrib>Sutskever, Ilya</creatorcontrib><creatorcontrib>Lillicrap, Timothy</creatorcontrib><creatorcontrib>Leach, Madeleine</creatorcontrib><creatorcontrib>Kavukcuoglu, Koray</creatorcontrib><creatorcontrib>Graepel, Thore</creatorcontrib><creatorcontrib>Hassabis, Demis</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Animal Behavior Abstracts</collection><collection>Bacteriology Abstracts (Microbiology B)</collection><collection>Calcium & Calcified Tissue Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>ProQuest Nursing and Allied Health Source</collection><collection>Ecology Abstracts</collection><collection>Entomology Abstracts (Full archive)</collection><collection>Environment Abstracts</collection><collection>Immunology Abstracts</collection><collection>Meteorological & Geoastrophysical Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Nucleic Acids Abstracts</collection><collection>Oncogenes and Growth Factors Abstracts</collection><collection>Virology and AIDS Abstracts</collection><collection>Agricultural Science Collection</collection><collection>ProQuest Health and Medical</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Biology Database (Alumni Edition)</collection><collection>Medical Database (Alumni Edition)</collection><collection>Psychology Database (Alumni)</collection><collection>Science Database (Alumni Edition)</collection><collection>STEM Database</collection><collection>ProQuest Pharma Collection</collection><collection>Public Health Database</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Natural Science Collection</collection><collection>Hospital Premium Collection</collection><collection>Hospital Premium Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>Research Library (Alumni Edition)</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>Agricultural & Environmental Science Collection</collection><collection>ProQuest Central Essentials</collection><collection>Biological Science Collection</collection><collection>eLibrary</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest Natural Science Collection</collection><collection>Earth, Atmospheric & Aquatic Science Collection</collection><collection>Environmental Sciences and Pollution Management</collection><collection>ProQuest One Community College</collection><collection>ProQuest Materials Science Collection</collection><collection>ProQuest Central</collection><collection>Engineering Research Database</collection><collection>Health Research Premium Collection</collection><collection>Health Research Premium Collection (Alumni)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>AIDS and Cancer Research Abstracts</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Health & Medical Complete (Alumni)</collection><collection>Materials Science Database</collection><collection>Nursing & Allied Health Database (Alumni Edition)</collection><collection>Meteorological & Geoastrophysical Abstracts - Academic</collection><collection>ProQuest Engineering Collection</collection><collection>ProQuest Biological Science Collection</collection><collection>Agriculture Science Database</collection><collection>Health & Medical Collection (Alumni Edition)</collection><collection>PML(ProQuest Medical Library)</collection><collection>ProQuest Psychology Journals</collection><collection>ProQuest research library</collection><collection>ProQuest Science Journals</collection><collection>Algology Mycology and Protozoology Abstracts (Microbiology C)</collection><collection>ProQuest Biological Science Journals</collection><collection>Engineering Database</collection><collection>Research Library (Corporate)</collection><collection>Nursing & Allied Health Premium</collection><collection>ProQuest advanced technologies & aerospace journals</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>Environmental Science Database</collection><collection>Earth, Atmospheric & Aquatic Science Database</collection><collection>Materials science collection</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest One Psychology</collection><collection>Engineering collection</collection><collection>Environmental Science Collection</collection><collection>ProQuest Central Basic</collection><collection>University of Michigan</collection><collection>Genetics Abstracts</collection><collection>SIRS Editorial</collection><collection>Environment Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>Nature (London)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Silver, David</au><au>Huang, Aja</au><au>Maddison, Chris J.</au><au>Guez, Arthur</au><au>Sifre, Laurent</au><au>van den Driessche, George</au><au>Schrittwieser, Julian</au><au>Antonoglou, Ioannis</au><au>Panneershelvam, Veda</au><au>Lanctot, Marc</au><au>Dieleman, Sander</au><au>Grewe, Dominik</au><au>Nham, John</au><au>Kalchbrenner, Nal</au><au>Sutskever, Ilya</au><au>Lillicrap, Timothy</au><au>Leach, Madeleine</au><au>Kavukcuoglu, Koray</au><au>Graepel, Thore</au><au>Hassabis, Demis</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Mastering the game of Go with deep neural networks and tree search</atitle><jtitle>Nature (London)</jtitle><stitle>Nature</stitle><addtitle>Nature</addtitle><date>2016-01-28</date><risdate>2016</risdate><volume>529</volume><issue>7587</issue><spage>484</spage><epage>489</epage><pages>484-489</pages><issn>0028-0836</issn><eissn>1476-4687</eissn><coden>NATUAS</coden><abstract>The game of Go has long been viewed as the most challenging of classic games for artificial intelligence owing to its enormous search space and the difficulty of evaluating board positions and moves. Here we introduce a new approach to computer Go that uses ‘value networks’ to evaluate board positions and ‘policy networks’ to select moves. These deep neural networks are trained by a novel combination of supervised learning from human expert games, and reinforcement learning from games of self-play. Without any lookahead search, the neural networks play Go at the level of state-of-the-art Monte Carlo tree search programs that simulate thousands of random games of self-play. We also introduce a new search algorithm that combines Monte Carlo simulation with value and policy networks. Using this search algorithm, our program AlphaGo achieved a 99.8% winning rate against other Go programs, and defeated the human European Go champion by 5 games to 0. This is the first time that a computer program has defeated a human professional player in the full-sized game of Go, a feat previously thought to be at least a decade away.
A computer Go program based on deep neural networks defeats a human professional player to achieve one of the grand challenges of artificial intelligence.
AlphaGo computer beats Go champion
The victory in 1997 of the chess-playing computer Deep Blue in a six-game series against the then world champion Gary Kasparov was seen as a significant milestone in the development of artificial intelligence. An even greater challenge remained — the ancient game of Go. Despite decades of refinement, until recently the strongest computers were still playing Go at the level of human amateurs. Enter AlphaGo. Developed by Google DeepMind, this program uses deep neural networks to mimic expert players, and further improves its performance by learning from games played against itself. AlphaGo has achieved a 99% win rate against the strongest other Go programs, and defeated the reigning European champion Fan Hui 5–0 in a tournament match. This is the first time that a computer program has defeated a human professional player in even games, on a full, 19 x 19 board, in even games with no handicap.</abstract><cop>London</cop><pub>Nature Publishing Group UK</pub><pmid>26819042</pmid><doi>10.1038/nature16961</doi><tpages>6</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0028-0836 |
ispartof | Nature (London), 2016-01, Vol.529 (7587), p.484-489 |
issn | 0028-0836 1476-4687 |
language | eng |
recordid | cdi_proquest_miscellaneous_1761459006 |
source | Nature |
subjects | 631/378/1788 639/705/1042 639/705/117 Algorithms Analysis Artificial intelligence Computer games Computers Europe Evaluation Games Games, Recreational Go (Game) Humanities and Social Sciences Humans Monte Carlo Method Monte Carlo simulation multidisciplinary Neural networks Neural Networks (Computer) Product development Reinforcement (Psychology) Science Software Supervised Machine Learning Technology application |
title | Mastering the game of Go with deep neural networks and tree search |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T05%3A32%3A41IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Mastering%20the%20game%20of%20Go%20with%20deep%20neural%20networks%20and%20tree%20search&rft.jtitle=Nature%20(London)&rft.au=Silver,%20David&rft.date=2016-01-28&rft.volume=529&rft.issue=7587&rft.spage=484&rft.epage=489&rft.pages=484-489&rft.issn=0028-0836&rft.eissn=1476-4687&rft.coden=NATUAS&rft_id=info:doi/10.1038/nature16961&rft_dat=%3Cgale_proqu%3EA661111782%3C/gale_proqu%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c756t-29169ab0b0ca120ba3231833251d5d051b8e8ba996b8a08665069a80d90d86693%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=1761243103&rft_id=info:pmid/26819042&rft_galeid=A661111782&rfr_iscdi=true |