Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 12 de 12
Filtrar
5.
Mem Cognit ; 18(4): 331-8, 1990 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-2381312

RESUMEN

To interpret utterances in conversations, listeners must often make reference to the common ground they share with speakers. For example, when faced with an utterance such as "That game was a disaster," listeners must decide whether they share common assumptions about what outcome would be disastrous. In our experiments, we examine how common ground, as encoded in community membership, is used to constrain judgments about the interpretations of ambiguous utterances. In Experiment 1, subjects were sensitive to community membership when they were asked to evaluate the interpretations at a leisurely pace. Experiment 2 replicated this result with greater time pressure. Experiment 3 demonstrated that judgments based on assessments of community membership were equivalent to those based on certain knowledge, except when the judgements were challenged. The results suggest that models of memory retrieval during language comprehension should make mention of access to common ground.


Asunto(s)
Formación de Concepto , Relaciones Interpersonales , Disposición en Psicología , Percepción del Habla , Adulto , Humanos
6.
Neural Comput ; 11(8): 2017-59, 1999 Nov 15.
Artículo en Inglés | MEDLINE | ID: mdl-10578043

RESUMEN

Reinforcement learning is the problem of generating optimal behavior in a sequential decision-making environment given the opportunity of interacting with it. Many algorithms for solving reinforcement-learning problems work by computing improved estimates of the optimal value function. We extend prior analyses of reinforcement-learning algorithms and present a powerful new theorem that can provide a unified analysis of such value-function-based reinforcement-learning algorithms. The usefulness of the theorem lies in how it allows the convergence of a complex asynchronous reinforcement-learning algorithm to be proved by verifying that a simpler synchronous algorithm converges. We illustrate the application of the theorem by analyzing the convergence of Q-learning, model-based reinforcement learning, Q-learning with multistate updates, Q-learning for Markov games, and risk-sensitive reinforcement learning.


Asunto(s)
Algoritmos , Inteligencia Artificial , Aprendizaje , Refuerzo en Psicología
8.
Am J Philol ; 94: 243-55, 1973.
Artículo en Inglés | MEDLINE | ID: mdl-11616517
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA