Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 3 de 3
Filtrar
Más filtros

Banco de datos
País/Región como asunto
Tipo del documento
País de afiliación
Intervalo de año de publicación
1.
Nature ; 529(7587): 484-9, 2016 Jan 28.
Artículo en Inglés | MEDLINE | ID: mdl-26819042

RESUMEN

The game of Go has long been viewed as the most challenging of classic games for artificial intelligence owing to its enormous search space and the difficulty of evaluating board positions and moves. Here we introduce a new approach to computer Go that uses 'value networks' to evaluate board positions and 'policy networks' to select moves. These deep neural networks are trained by a novel combination of supervised learning from human expert games, and reinforcement learning from games of self-play. Without any lookahead search, the neural networks play Go at the level of state-of-the-art Monte Carlo tree search programs that simulate thousands of random games of self-play. We also introduce a new search algorithm that combines Monte Carlo simulation with value and policy networks. Using this search algorithm, our program AlphaGo achieved a 99.8% winning rate against other Go programs, and defeated the human European Go champion by 5 games to 0. This is the first time that a computer program has defeated a human professional player in the full-sized game of Go, a feat previously thought to be at least a decade away.


Asunto(s)
Juegos Recreacionales , Redes Neurales de la Computación , Programas Informáticos , Aprendizaje Automático Supervisado , Computadores , Europa (Continente) , Humanos , Método de Montecarlo , Refuerzo en Psicología
2.
Neural Netw ; 23(2): 239-43, 2010 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-19932002

RESUMEN

A Recurrent Neural Network (RNN) is a powerful connectionist model that can be applied to many challenging sequential problems, including problems that naturally arise in language and speech. However, RNNs are extremely hard to train on problems that have long-term dependencies, where it is necessary to remember events for many timesteps before using them to make a prediction. In this paper we consider the problem of training RNNs to predict sequences that exhibit significant long-term dependencies, focusing on a serial recall task where the RNN needs to remember a sequence of characters for a large number of steps before reconstructing it. We introduce the Temporal-Kernel Recurrent Neural Network (TKRNN), which is a variant of the RNN that can cope with long-term dependencies much more easily than a standard RNN, and show that the TKRNN develops short-term memory that successfully solves the serial recall task by representing the input string with a stable state of its hidden units.


Asunto(s)
Memoria a Corto Plazo , Redes Neurales de la Computación , Algoritmos , Humanos , Pruebas Neuropsicológicas , Factores de Tiempo
3.
Neural Comput ; 20(11): 2629-36, 2008 Nov.
Artículo en Inglés | MEDLINE | ID: mdl-18533819

RESUMEN

In this note, we show that exponentially deep belief networks can approximate any distribution over binary vectors to arbitrary accuracy, even when the width of each layer is limited to the dimensionality of the data. We further show that such networks can be greedily learned in an easy yet impractical way.


Asunto(s)
Aprendizaje , Redes Neurales de la Computación , Algoritmos , Humanos , Dinámicas no Lineales
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA