Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 153
Filtrar
1.
Artículo en Inglés | MEDLINE | ID: mdl-38700375

RESUMEN

INTRODUCTION: Little research has been done on how people mentally simulate future suicidal thoughts and urges, a process we term suicidal prospection. METHODS: Participants were 94 adults with recent suicidal thoughts. Participants completed a 42-day real-time monitoring study and then a follow-up survey 28 days later. Each night, participants provided predictions for the severity of their suicidal thoughts the next day and ratings of the severity of suicidal thoughts over the past day. We measured three aspects of suicidal prospection: predicted levels of desire to kill self, urge to kill self, and intent to kill self. We generated prediction errors by subtracting participants' predictions of the severity of their suicidal thoughts from their experienced severity. RESULTS: Participants tended to overestimate (although the average magnitude was small and the modal error was zero) the severity of their future suicidal thoughts. The best fitting models suggested that participants used both their current suicidal thinking and previous predictions of their suicidal thinking to generate predictions of their future suicidal thinking. Finally, the average severity of predicted future suicidal thoughts predicted the number of days participants thought about suicide during the follow-up period. CONCLUSIONS: This study highlights prospection as a psychological process to better understand suicidal thoughts and behaviors.

2.
Behav Neurosci ; 2024 May 16.
Artículo en Inglés | MEDLINE | ID: mdl-38753400

RESUMEN

Psychopathology is vast and diverse. Across distinct disease states, individuals exhibit symptoms that appear counter to the standard view of rationality (expected utility maximization). We argue that some aspects of psychopathology can be described as resource-rational, reflecting a rational trade-off between reward and cognitive resources. We review work on two theories of this kind: rational inattention, where a capacity limit applies to perceptual channels, and policy compression, where the capacity limit applies to action channels. We show how these theories can parsimoniously explain many forms of psychopathology, including affective, primary psychotic, and neurodevelopmental disorders, as well as many effects of psychoactive medications on these disorders. While there are important disorder-specific differences and the theories are by no means universal, we argue that resource rationality offers a useful new perspective on psychopathology. By emphasizing the role of cognitive resource constraints, this approach offers a more inclusive picture of rationality. Some aspects of psychopathology may reflect rational trade-offs rather than the breakdown of rationality. (PsycInfo Database Record (c) 2024 APA, all rights reserved).

3.
PLoS Comput Biol ; 20(4): e1012057, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38669280

RESUMEN

Policy compression is a computational framework that describes how capacity-limited agents trade reward for simpler action policies to reduce cognitive cost. In this study, we present behavioral evidence that humans prefer simpler policies, as predicted by a capacity-limited reinforcement learning model. Across a set of tasks, we find that people exploit structure in the relationships between states, actions, and rewards to "compress" their policies. In particular, compressed policies are systematically biased towards actions with high marginal probability, thereby discarding some state information. This bias is greater when there is redundancy in the reward-maximizing action policy across states, and increases with memory load. These results could not be explained qualitatively or quantitatively by models that did not make use of policy compression under a capacity limit. We also confirmed the prediction that time pressure should further reduce policy complexity and increase action bias, based on the hypothesis that actions are selected via time-dependent decoding of a compressed code. These findings contribute to a deeper understanding of how humans adapt their decision-making strategies under cognitive resource constraints.


Asunto(s)
Toma de Decisiones , Recompensa , Humanos , Toma de Decisiones/fisiología , Biología Computacional , Masculino , Adulto , Femenino , Refuerzo en Psicología , Modelos Psicológicos , Adulto Joven , Cognición/fisiología
4.
Cogn Psychol ; 150: 101653, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38503178

RESUMEN

In order to efficiently divide labor with others, it is important to understand what our collaborators can do (i.e., their competence). However, competence is not static-people get better at particular jobs the more often they perform them. This plasticity of competence creates a challenge for collaboration: For example, is it better to assign tasks to whoever is most competent now, or to the person who can be trained most efficiently "on-the-job"? We conducted four experiments (N=396) that examine how people make decisions about whom to train (Experiments 1 and 3) and whom to recruit (Experiments 2 and 4) to a collaborative task, based on the simulated collaborators' starting expertise, the training opportunities available, and the goal of the task. We found that participants' decisions were best captured by a planning model that attempts to maximize the returns from collaboration while minimizing the costs of hiring and training individual collaborators. This planning model outperformed alternative models that based these decisions on the agents' current competence, or on how much agents stood to improve in a single training step, without considering whether this training would enable agents to succeed at the task in the long run. Our findings suggest that people do not recruit and train collaborators based solely on their current competence, nor solely on the opportunities for their collaborators to improve. Instead, people use an intuitive theory of competence to balance the costs of hiring and training others against the benefits to the collaboration.

5.
bioRxiv ; 2024 Feb 06.
Artículo en Inglés | MEDLINE | ID: mdl-38370735

RESUMEN

Associative learning depends on contingency, the degree to which a stimulus predicts an outcome. Despite its importance, the neural mechanisms linking contingency to behavior remain elusive. Here we examined the dopamine activity in the ventral striatum - a signal implicated in associative learning - in a Pavlovian contingency degradation task in mice. We show that both anticipatory licking and dopamine responses to a conditioned stimulus decreased when additional rewards were delivered uncued, but remained unchanged if additional rewards were cued. These results conflict with contingency-based accounts using a traditional definition of contingency or a novel causal learning model (ANCCR), but can be explained by temporal difference (TD) learning models equipped with an appropriate inter-trial-interval (ITI) state representation. Recurrent neural networks trained within a TD framework develop state representations like our best 'handcrafted' model. Our findings suggest that the TD error can be a measure that describes both contingency and dopaminergic activity.

6.
Biol Cybern ; 118(1-2): 1-5, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38337064

RESUMEN

Neuroscience and artificial intelligence (AI) share a long, intertwined history. It has been argued that discoveries in neuroscience were (and continue to be) instrumental in driving the development of new AI technology. Scrutinizing these historical claims yields a more nuanced story, where AI researchers were loosely inspired by the brain, but ideas flowed mostly in the other direction.


Asunto(s)
Inteligencia Artificial , Encéfalo , Neurociencias , Humanos , Encéfalo/fisiología , Neurociencias/tendencias , Historia del Siglo XX , Animales , Historia del Siglo XXI
7.
Nat Hum Behav ; 8(5): 917-931, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38332340

RESUMEN

Computational phenotyping has emerged as a powerful tool for characterizing individual variability across a variety of cognitive domains. An individual's computational phenotype is defined as a set of mechanistically interpretable parameters obtained from fitting computational models to behavioural data. However, the interpretation of these parameters hinges critically on their psychometric properties, which are rarely studied. To identify the sources governing the temporal variability of the computational phenotype, we carried out a 12-week longitudinal study using a battery of seven tasks that measure aspects of human learning, memory, perception and decision making. To examine the influence of state effects, each week, participants provided reports tracking their mood, habits and daily activities. We developed a dynamic computational phenotyping framework, which allowed us to tease apart the time-varying effects of practice and internal states such as affective valence and arousal. Our results show that many phenotype dimensions covary with practice and affective factors, indicating that what appears to be unreliability may reflect previously unmeasured structure. These results support a fundamentally dynamic understanding of cognitive variability within an individual.


Asunto(s)
Cognición , Fenotipo , Humanos , Cognición/fisiología , Masculino , Femenino , Estudios Longitudinales , Toma de Decisiones/fisiología , Adulto , Adulto Joven , Aprendizaje/fisiología , Afecto/fisiología , Memoria/fisiología , Individualidad
8.
Psychol Rev ; 131(2): 578-597, 2024 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-37166847

RESUMEN

Two of the main impediments to learning complex tasks are that relationships between different stimuli, including rewards, can be uncertain and context-dependent. Reinforcement learning (RL) provides a framework for learning, by predicting total future reward directly (model-free RL), or via predictions of future states (model-based RL). Within this framework, "successor representation" (SR) predicts total future occupancy of all states. A recent theoretical proposal suggests that the hippocampus encodes the SR in order to facilitate prediction of future reward. However, this proposal does not take into account how learning should adapt under uncertainty and switches of context. Here, we introduce a theory of learning SRs using prediction errors which includes optimally balancing uncertainty in new observations versus existing knowledge. We then generalize that approach to a multicontext setting, allowing the model to learn and maintain multiple task-specific SRs and infer which one to use at any moment based on the accuracy of its predictions. Thus, the context used for predictions can be determined by both the contents of the states themselves and the distribution of transitions between them. This probabilistic SR model captures animal behavior in tasks which require contextual memory and generalization, and unifies previous SR theory with hippocampal-dependent contextual decision-making. (PsycInfo Database Record (c) 2024 APA, all rights reserved).


Asunto(s)
Aprendizaje , Refuerzo en Psicología , Animales , Humanos , Recompensa , Incertidumbre , Generalización Psicológica
9.
bioRxiv ; 2023 Nov 13.
Artículo en Inglés | MEDLINE | ID: mdl-38014354

RESUMEN

Dopamine release in the nucleus accumbens has been hypothesized to signal reward prediction error, the difference between observed and predicted reward, suggesting a biological implementation for reinforcement learning. Rigorous tests of this hypothesis require assumptions about how the brain maps sensory signals to reward predictions, yet this mapping is still poorly understood. In particular, the mapping is non-trivial when sensory signals provide ambiguous information about the hidden state of the environment. Previous work using classical conditioning tasks has suggested that reward predictions are generated conditional on probabilistic beliefs about the hidden state, such that dopamine implicitly reflects these beliefs. Here we test this hypothesis in the context of an instrumental task (a two-armed bandit), where the hidden state switches repeatedly. We measured choice behavior and recorded dLight signals reflecting dopamine release in the nucleus accumbens core. Model comparison based on the behavioral data favored models that used Bayesian updating of probabilistic beliefs. These same models also quantitatively matched the dopamine measurements better than non-Bayesian alternatives. We conclude that probabilistic belief computation plays a fundamental role in instrumental performance and associated mesolimbic dopamine signaling.

10.
Cognition ; 241: 105609, 2023 Dec.
Artículo en Inglés | MEDLINE | ID: mdl-37708602

RESUMEN

How do people judge responsibility in collaborative tasks? Past work has proposed a number of metrics that people may use to attribute blame and credit to others, such as effort, competence, and force. Some theories consider only the actual effort or force (individuals are more responsible if they put forth more effort or force), whereas others consider counterfactuals (individuals are more responsible if some alternative behavior on their or their collaborator's part could have altered the outcome). Across four experiments (N=717), we found that participants' judgments are best described by a model that considers both actual and counterfactual effort. This finding generalized to an independent validation data set (N=99). Our results thus support a dual-factor theory of responsibility attribution in collaborative tasks.

11.
PLoS Comput Biol ; 19(9): e1011067, 2023 09.
Artículo en Inglés | MEDLINE | ID: mdl-37695776

RESUMEN

To behave adaptively, animals must learn to predict future reward, or value. To do this, animals are thought to learn reward predictions using reinforcement learning. However, in contrast to classical models, animals must learn to estimate value using only incomplete state information. Previous work suggests that animals estimate value in partially observable tasks by first forming "beliefs"-optimal Bayesian estimates of the hidden states in the task. Although this is one way to solve the problem of partial observability, it is not the only way, nor is it the most computationally scalable solution in complex, real-world environments. Here we show that a recurrent neural network (RNN) can learn to estimate value directly from observations, generating reward prediction errors that resemble those observed experimentally, without any explicit objective of estimating beliefs. We integrate statistical, functional, and dynamical systems perspectives on beliefs to show that the RNN's learned representation encodes belief information, but only when the RNN's capacity is sufficiently large. These results illustrate how animals can estimate value in tasks without explicitly estimating beliefs, yielding a representation useful for systems with limited capacity.


Asunto(s)
Aprendizaje , Refuerzo en Psicología , Animales , Teorema de Bayes , Recompensa , Redes Neurales de la Computación
12.
Elife ; 122023 07 12.
Artículo en Inglés | MEDLINE | ID: mdl-37435811

RESUMEN

Rate-distortion theory provides a powerful framework for understanding the nature of human memory by formalizing the relationship between information rate (the average number of bits per stimulus transmitted across the memory channel) and distortion (the cost of memory errors). Here, we show how this abstract computational-level framework can be realized by a model of neural population coding. The model reproduces key regularities of visual working memory, including some that were not previously explained by population coding models. We verify a novel prediction of the model by reanalyzing recordings of monkey prefrontal neurons during an oculomotor delayed response task.


Asunto(s)
Memoria a Corto Plazo , Neuronas , Humanos , Memoria a Corto Plazo/fisiología , Neuronas/fisiología , Corteza Prefrontal/fisiología
13.
Nat Hum Behav ; 7(9): 1481-1489, 2023 09.
Artículo en Inglés | MEDLINE | ID: mdl-37488401

RESUMEN

Studies of human exploration frequently cast people as serendipitously stumbling upon good options. Yet these studies may not capture the richness of exploration strategies that people exhibit in more complex environments. Here we study behaviour in a large dataset of 29,493 players of the richly structured online game 'Little Alchemy 2'. In this game, players start with four elements, which they can combine to create up to 720 complex objects. We find that players are driven not only by external reward signals, such as an attempt to produce successful outcomes, but also by an intrinsic motivation to create objects that empower them to create even more objects. We find that this drive for empowerment is eliminated when playing a game variant that lacks recognizable semantics, indicating that people use their knowledge about the world and its possibilities to guide their exploration. Our results suggest that the drive for empowerment may be a potent source of intrinsic motivation in richly structured domains, particularly those that lack explicit reward signals.


Asunto(s)
Juegos de Video , Humanos , Conducta Exploratoria , Motivación , Logro , Recompensa
14.
J Exp Psychol Gen ; 152(11): 3074-3086, 2023 Nov.
Artículo en Inglés | MEDLINE | ID: mdl-37307336

RESUMEN

People make fast and reasonable predictions about the physical behavior of everyday objects. To do so, people may use principled mental shortcuts, such as object simplification, similar to models developed by engineers for real-time physical simulations. We hypothesize that people use simplified object approximations for tracking and action (the body representation), as opposed to fine-grained forms for visual recognition (the shape representation). We used three classic psychophysical tasks (causality perception, time-to-collision, and change detection) in novel settings that dissociate body and shape. People's behavior across tasks indicates that they rely on coarse bodies for physical reasoning, which lies between convex hulls and fine-grained shapes. Our empirical and computational findings shed light on basic representations people use to understand everyday dynamics, and how these representations differ from those used for recognition. (PsycInfo Database Record (c) 2023 APA, all rights reserved).

15.
J Cogn Neurosci ; 35(9): 1508-1520, 2023 09 01.
Artículo en Inglés | MEDLINE | ID: mdl-37382476

RESUMEN

Exploration is an important part of decision making and is crucial to maximizing long-term rewards. Past work has shown that people use different forms of uncertainty to guide exploration. In this study, we investigate the role of the pupil-linked arousal system in uncertainty-guided exploration. We measured participants' (n = 48) pupil dilation while they performed a two-armed bandit task. Consistent with previous work, we found that people adopted a hybrid of directed, random, and undirected exploration, which are sensitive to relative uncertainty, total uncertainty, and value difference between options, respectively. We also found a positive correlation between pupil size and total uncertainty. Furthermore, augmenting the choice model with subject-specific total uncertainty estimates decoded from the pupil size improved predictions of held-out choices, suggesting that people used the uncertainty estimate encoded in pupil size to decide which option to explore. Together, the data shed light on the computations underlying uncertainty-driven exploration. Under the assumption that pupil size reflects locus coeruleus-norepinephrine neuromodulatory activity, these results also extend the theory of the locus coeruleus-norepinephrine function in exploration, highlighting its selective role in driving uncertainty-guided random exploration.


Asunto(s)
Nivel de Alerta , Pupila , Humanos , Incertidumbre , Recompensa , Norepinefrina
16.
PLoS One ; 18(5): e0286067, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-37200364

RESUMEN

Correlation does not imply causation, but this does not necessarily stop people from drawing causal inferences from correlational statements. We show that people do in fact infer causality from statements of association, under minimal conditions. In Study 1, participants interpreted statements of the form "X is associated with Y" to imply that Y causes X. In Studies 2 and 3, participants interpreted statements of the form "X is associated with an increased risk of Y" to imply that X causes Y. Thus, even the most orthodox correlational language can give rise to causal inferences.


Asunto(s)
Lenguaje , Humanos , Causalidad
17.
J Exp Psychol Gen ; 152(6): 1754-1767, 2023 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-37199971

RESUMEN

Value-based decisions are often guided by past experience. If a choice led to a good outcome, we are more likely to repeat it. This basic idea is well-captured by reinforcement-learning models. However, open questions remain about how we assign value to options we did not choose and which we therefore never had the chance to learn about directly. One solution to this problem is proposed by policy gradient reinforcement-learning models; these do not require direct learning of value, instead optimizing choices according to a behavioral policy. For example, a logistic policy predicts that if a chosen option was rewarded, the unchosen option would be deemed less desirable. Here, we test the relevance of these models to human behavior and explore the role of memory in this phenomenon. We hypothesize that a policy may emerge from an associative memory trace formed during deliberation between choice options. In a preregistered study (n = 315) we show that people tend to invert the value of unchosen options relative to the outcome of chosen options, a phenomenon we term inverse decision bias. The inverse decision bias is correlated with memory for the association between choice options; moreover, it is reduced when memory formation is experimentally interfered with. Finally, we present a new memory-based policy gradient model that predicts both the inverse decision bias and its dependence on memory. Our findings point to a significant role of associative memory in valuation of unchosen options and introduce a new perspective on the interaction between decision-making, memory, and counterfactual reasoning. (PsycInfo Database Record (c) 2023 APA, all rights reserved).


Asunto(s)
Conducta de Elección , Toma de Decisiones , Humanos , Aprendizaje , Refuerzo en Psicología , Recompensa
18.
Proc Natl Acad Sci U S A ; 120(22): e2215015120, 2023 05 30.
Artículo en Inglés | MEDLINE | ID: mdl-37216526

RESUMEN

Teaching enables humans to impart vast stores of culturally specific knowledge and skills. However, little is known about the neural computations that guide teachers' decisions about what information to communicate. Participants (N = 28) played the role of teachers while being scanned using fMRI; their task was to select examples that would teach learners how to answer abstract multiple-choice questions. Participants' examples were best described by a model that selects evidence that maximizes the learner's belief in the correct answer. Consistent with this idea, participants' predictions about how well learners would do closely tracked the performance of an independent sample of learners (N = 140) who were tested on the examples they had provided. In addition, regions that play specialized roles in processing social information, namely the bilateral temporoparietal junction and middle and dorsal medial prefrontal cortex, tracked learners' posterior belief in the correct answer. Our results shed light on the computational and neural architectures that support our extraordinary abilities as teachers.


Asunto(s)
Aprendizaje , Mentalización , Enseñanza , Humanos , Encéfalo/diagnóstico por imagen
19.
bioRxiv ; 2023 Apr 04.
Artículo en Inglés | MEDLINE | ID: mdl-37066383

RESUMEN

To behave adaptively, animals must learn to predict future reward, or value. To do this, animals are thought to learn reward predictions using reinforcement learning. However, in contrast to classical models, animals must learn to estimate value using only incomplete state information. Previous work suggests that animals estimate value in partially observable tasks by first forming "beliefs"-optimal Bayesian estimates of the hidden states in the task. Although this is one way to solve the problem of partial observability, it is not the only way, nor is it the most computationally scalable solution in complex, real-world environments. Here we show that a recurrent neural network (RNN) can learn to estimate value directly from observations, generating reward prediction errors that resemble those observed experimentally, without any explicit objective of estimating beliefs. We integrate statistical, functional, and dynamical systems perspectives on beliefs to show that the RNN's learned representation encodes belief information, but only when the RNN's capacity is sufficiently large. These results illustrate how animals can estimate value in tasks without explicitly estimating beliefs, yielding a representation useful for systems with limited capacity. Author Summary: Natural environments are full of uncertainty. For example, just because my fridge had food in it yesterday does not mean it will have food today. Despite such uncertainty, animals can estimate which states and actions are the most valuable. Previous work suggests that animals estimate value using a brain area called the basal ganglia, using a process resembling a reinforcement learning algorithm called TD learning. However, traditional reinforcement learning algorithms cannot accurately estimate value in environments with state uncertainty (e.g., when my fridge's contents are unknown). One way around this problem is if agents form "beliefs," a probabilistic estimate of how likely each state is, given any observations so far. However, estimating beliefs is a demanding process that may not be possible for animals in more complex environments. Here we show that an artificial recurrent neural network (RNN) trained with TD learning can estimate value from observations, without explicitly estimating beliefs. The trained RNN's error signals resembled the neural activity of dopamine neurons measured during the same task. Importantly, the RNN's activity resembled beliefs, but only when the RNN had enough capacity. This work illustrates how animals could estimate value in uncertain environments without needing to first form beliefs, which may be useful in environments where computing the true beliefs is too costly.

20.
J Exp Psychol Gen ; 152(6): 1565-1579, 2023 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-36877460

RESUMEN

By collaborating with others, humans can pool their limited knowledge, skills, and resources to achieve goals that outstrip the abilities of any one person. What cognitive capacities make human collaboration possible? Here, we propose that collaboration is grounded in an intuitive understanding of how others think and of what they can do-in other words, of their mental states and competence. We present a belief-desire-competence framework that formalizes this proposal by extending existing models of commonsense psychological reasoning. Our framework predicts that agents recursively reason how much effort they and their partner will allocate to a task, based on the rewards at stake and on their own and their collaborator's competence. Across three experiments (N = 249), we show that the belief-desire-competence framework captures human judgments in a variety of contexts that are critical to collaboration, including predicting whether a joint activity will succeed (Experiment 1), selecting incentives for collaborators (Experiment 2), and choosing which individuals to recruit for a collaborative task (Experiment 3). Our work provides a theoretical framework for understanding how commonsense psychological reasoning contributes to collaborative achievements. (PsycInfo Database Record (c) 2023 APA, all rights reserved).


Asunto(s)
Juicio , Motivación , Humanos , Solución de Problemas , Recompensa , Toma de Decisiones
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA
...