Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 41
Filtrar
Mais filtros










Base de dados
Intervalo de ano de publicação
1.
Front Robot AI ; 10: 1256763, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-37929074

RESUMO

In recent years, soft robots gain increasing attention as a result of their compliance when operating in unstructured environments, and their flexibility that ensures safety when interacting with humans. However, challenges lie on the difficulty to develop control algorithms due to various limitations induced by their soft structure. In this paper, we introduce a novel technique that aims to perform motion control of a modular bio-inspired soft-robotic arm, with the main focus lying on facilitating the qualitative reproduction of well-specified periodic trajectories. The introduced method combines the notion behind two previously developed methodologies both based on the Movement Primitive (MP) theory, by exploiting their capabilities while coping with their main drawbacks. Concretely, the requested actuation is initially computed using a Probabilistic MP (ProMP)-based method that considers the trajectory as a combination of simple movements previously learned and stored as a MP library. Subsequently, the key components of the resulting actuation are extracted and filtered in the frequency domain. These are eventually used as input to a Central Pattern Generator (CPG)-based model that takes over the generation of rhythmic patterns at the motor level. The proposed methodology is evaluated on a two-module soft arm. Results show that the first algorithmic component (ProMP) provides an immediate estimation of the requested actuation by avoiding time-consuming training, while the latter (CPG) further simplifies the execution by allowing its control through a low-dimensional parameterization. Altogether, these results open new avenues for the rapid acquisition of periodic movements in soft robots, and their compression into CPG parameters for long-term storage and execution.

3.
Elife ; 122023 05 30.
Artigo em Inglês | MEDLINE | ID: mdl-37252780

RESUMO

Our brain is continuously challenged by daily experiences. Thus, how to avoid systematic erasing of previously encoded memories? While it has been proposed that a dual-learning system with 'slow' learning in the cortex and 'fast' learning in the hippocampus could protect previous knowledge from interference, this has never been observed in the living organism. Here, we report that increasing plasticity via the viral-induced overexpression of RGS14414 in the prelimbic cortex leads to better one-trial memory, but that this comes at the price of increased interference in semantic-like memory. Indeed, electrophysiological recordings showed that this manipulation also resulted in shorter NonREM-sleep bouts, smaller delta-waves and decreased neuronal firing rates. In contrast, hippocampal-cortical interactions in form of theta coherence during wake and REM-sleep as well as oscillatory coupling during NonREM-sleep were enhanced. Thus, we provide the first experimental evidence for the long-standing and unproven fundamental idea that high thresholds for plasticity in the cortex protect preexisting memories and modulating these thresholds affects both memory encoding and consolidation mechanisms.


Assuntos
Hipocampo , Memória , Córtex Cerebral/fisiologia , Hipocampo/fisiologia , Memória/fisiologia , Sono/fisiologia , Sono REM , Humanos
4.
Cogn Res Princ Implic ; 8(1): 23, 2023 04 20.
Artigo em Inglês | MEDLINE | ID: mdl-37081292

RESUMO

Healthcare professionals' statistical illiteracy can impair medical decision quality and compromise patient safety. Previous studies have documented clinicians' insufficient proficiency in statistics and a tendency in overconfidence. However, an underexplored aspect is clinicians' awareness of their lack of statistical knowledge that precludes any corrective intervention attempt. Here, we investigated physicians', residents' and medical students' alignment between subjective confidence judgments and objective accuracy in basic medical statistics. We also examined how gender, profile of experience and practice of research activity affect this alignment, and the influence of problem framing (conditional probabilities, CP vs. natural frequencies, NF). Eight hundred ninety-eight clinicians completed an online survey assessing skill and confidence on three topics: vaccine efficacy, p value and diagnostic test results interpretation. Results evidenced an overall consistent poor proficiency in statistics often combined with high confidence, even in incorrect answers. We also demonstrate that despite overconfidence bias, clinicians show a degree of metacognitive sensitivity, as their confidence judgments discriminate between their correct and incorrect answers. Finally, we confirm the positive impact of the more intuitive NF framing on accuracy. Together, our results pave the way for the development of teaching recommendations and pedagogical interventions such as promoting metacognition on basic knowledge and statistical reasoning as well as the use of NF to tackle statistical illiteracy in the medical context.


Assuntos
Ilusões , Metacognição , Médicos , Humanos , Julgamento , Pessoal de Saúde , Médicos/psicologia
5.
J Neurosci ; 43(3): 458-471, 2023 01 18.
Artigo em Inglês | MEDLINE | ID: mdl-36216504

RESUMO

Model-free and model-based computations are argued to distinctly update action values that guide decision-making processes. It is not known, however, if these model-free and model-based reinforcement learning mechanisms recruited in operationally based instrumental tasks parallel those engaged by pavlovian-based behavioral procedures. Recently, computational work has suggested that individual differences in the attribution of incentive salience to reward predictive cues, that is, sign- and goal-tracking behaviors, are also governed by variations in model-free and model-based value representations that guide behavior. Moreover, it is not appreciated if these systems that are characterized computationally using model-free and model-based algorithms are conserved across tasks for individual animals. In the current study, we used a within-subject design to assess sign-tracking and goal-tracking behaviors using a pavlovian conditioned approach task and then characterized behavior using an instrumental multistage decision-making (MSDM) task in male rats. We hypothesized that both pavlovian and instrumental learning processes may be driven by common reinforcement-learning mechanisms. Our data confirm that sign-tracking behavior was associated with greater reward-mediated, model-free reinforcement learning and that it was also linked to model-free reinforcement learning in the MSDM task. Computational analyses revealed that pavlovian model-free updating was correlated with model-free reinforcement learning in the MSDM task. These data provide key insights into the computational mechanisms mediating associative learning that could have important implications for normal and abnormal states.SIGNIFICANCE STATEMENT Model-free and model-based computations that guide instrumental decision-making processes may also be recruited in pavlovian-based behavioral procedures. Here, we used a within-subject design to test the hypothesis that both pavlovian and instrumental learning processes were driven by common reinforcement-learning mechanisms. Sign-tracking and goal-tracking behaviors were assessed in rats using a pavlovian conditioned approach task, and then instrumental behavior was characterized using an MSDM task. We report that sign-tracking behavior was associated with greater model-free, but not model-based, learning in the MSDM task. These data suggest that pavlovian and instrumental behaviors may be driven by conserved reinforcement-learning mechanisms.


Assuntos
Reforço Psicológico , Recompensa , Ratos , Masculino , Animais , Aprendizagem , Motivação , Condicionamento Operante , Sinais (Psicologia)
7.
Front Neurorobot ; 16: 864380, 2022.
Artigo em Inglês | MEDLINE | ID: mdl-35812782

RESUMO

Experience replay is widely used in AI to bootstrap reinforcement learning (RL) by enabling an agent to remember and reuse past experiences. Classical techniques include shuffled-, reversed-ordered- and prioritized-memory buffers, which have different properties and advantages depending on the nature of the data and problem. Interestingly, recent computational neuroscience work has shown that these techniques are relevant to model hippocampal reactivations recorded during rodent navigation. Nevertheless, the brain mechanisms for orchestrating hippocampal replay are still unclear. In this paper, we present recent neurorobotics research aiming to endow a navigating robot with a neuro-inspired RL architecture (including different learning strategies, such as model-based (MB) and model-free (MF), and different replay techniques). We illustrate through a series of numerical simulations how the specificities of robotic experimentation (e.g., autonomous state decomposition by the robot, noisy perception, state transition uncertainty, non-stationarity) can shed new lights on which replay techniques turn out to be more efficient in different situations. Finally, we close the loop by raising new hypotheses for neuroscience from such robotic models of hippocampal replay.

8.
Sci Adv ; 7(38): eabh2392, 2021 Sep 17.
Artigo em Inglês | MEDLINE | ID: mdl-34524842

RESUMO

The ability to attribute thoughts to others, also called theory of mind (TOM), has been extensively studied in humans; however, its evolutionary origins have been challenged. Computationally, the basis of TOM has been interpreted within the predictive coding framework and associated with activity in the temporoparietal junction (TPJ). Here, we revealed, using a nonlinguistic task and functional magnetic resonance imaging, that activity in a region of the macaque middle superior temporal cortex was specifically modulated by the predictability of social situations. As in human TPJ, this region could be distinguished from other temporal regions involved in face processing. Our result suggests the existence of a precursor for the TOM ability in the last common ancestor of human and Old World monkeys.

9.
Behav Neurosci ; 135(2): 108, 2021 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-34060868

RESUMO

This special issue, commissioned after the 4th Quadrennial Meeting on Orbitofrontal Cortex Function held in Paris in November of 2019 (https://ofc2019.sciencesconf.org/), is intended to provide a snapshot of this ongoing transformation; we hope that the ideas presented herein will provide a foundation for the next stage in the evolution of our understanding of this magical brain region. (PsycInfo Database Record (c) 2021 APA, all rights reserved).


Assuntos
Córtex Pré-Frontal
10.
Behav Neurosci ; 135(2): 226-244, 2021 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-34060876

RESUMO

Our understanding of orbitofrontal cortex (OFC) function has progressed remarkably over the past decades in part due to theoretical advances in associative and reinforcement learning theories. These theoretical accounts of OFC function have implicated the region in progressively more psychologically refined processes from the value and sensory-specific properties of expected outcomes to the representation and inference over latent state representations in cognitive maps of task space. While these accounts have been successful at modeling many of the effects of causal manipulation of OFC function in both rodents and primates, recent findings suggest that further refinement of our current models are still required. Here, we briefly review how our understanding of OFC function has developed to understand two cardinal deficits following OFC dysfunction: Reversal learning and outcome devaluation. We then consider recent findings that OFC dysfunction also significantly affects initial acquisition learning, often assumed to be intact. To account for these findings, we consider a possible role for the OFC in the arbitration and exploration between model-free (MF) and model-based (MB) learning systems, offline updating of MB representations. While the function of the OFC as a whole is still likely to be integral to the formation and use of a cognitive map of task space, these refinements suggest a way in which distinct orbital subregions, such as the rodent lateral OFC, might contribute to this overall function. (PsycInfo Database Record (c) 2021 APA, all rights reserved).


Assuntos
Córtex Pré-Frontal , Roedores , Animais , Reforço Psicológico , Reversão de Aprendizagem
11.
Nat Commun ; 11(1): 3771, 2020 07 28.
Artigo em Inglês | MEDLINE | ID: mdl-32724052

RESUMO

People and other animals learn the values of choices by observing the contingencies between them and their outcomes. However, decisions are not guided by choice-linked reward associations alone; macaques also maintain a memory of the general, average reward rate - the global reward state - in an environment. Remarkably, global reward state affects the way that each choice outcome is valued and influences future decisions so that the impact of both choice success and failure is different in rich and poor environments. Successful choices are more likely to be repeated but this is especially the case in rich environments. Unsuccessful choices are more likely to be abandoned but this is especially likely in poor environments. Functional magnetic resonance imaging (fMRI) revealed two distinct patterns of activity, one in anterior insula and one in the dorsal raphe nucleus, that track global reward state as well as specific outcome events.


Assuntos
Córtex Cerebral/fisiologia , Comportamento de Escolha/fisiologia , Modelos Neurológicos , Núcleos da Rafe/fisiologia , Recompensa , Animais , Comportamento Animal , Córtex Cerebral/diagnóstico por imagem , Processamento de Imagem Assistida por Computador , Macaca mulatta , Imageamento por Ressonância Magnética , Masculino , Modelos Animais , Núcleos da Rafe/diagnóstico por imagem
12.
Neurobiol Learn Mem ; 173: 107265, 2020 09.
Artigo em Inglês | MEDLINE | ID: mdl-32531423

RESUMO

Kleefstra syndrome is a disorder caused by a mutation in the EHMT1 gene characterized in humans by general developmental delay, mild to severe intellectual disability and autism. Here, we characterized cumulative memory in the Ehmt1+/- mouse model using the Object Space Task. We combined conventional behavioral analysis with automated analysis by deep-learning networks, a session-based computational learning model, and a trial-based classifier. Ehmt1+/- mice showed more anxiety-like features and generally explored objects less, but the difference decreased over time. Interestingly, when analyzing memory-specific exploration, Ehmt1+/- show increased expression of cumulative memory, but a deficit in a more simple, control memory condition. Using our automatic classifier to differentiate between genotypes, we found that cumulative memory features are better suited for classification than general exploration differences. Thus, detailed behavioral classification with the Object Space Task produced a more detailed behavioral phenotype of the Ehmt1+/- mouse model.


Assuntos
Comportamento Animal/fisiologia , Anormalidades Craniofaciais/fisiopatologia , Comportamento Exploratório/fisiologia , Cardiopatias Congênitas/fisiopatologia , Deficiência Intelectual/fisiopatologia , Memória/fisiologia , Animais , Deleção Cromossômica , Cromossomos Humanos Par 9/genética , Anormalidades Craniofaciais/genética , Aprendizado Profundo , Modelos Animais de Doenças , Cardiopatias Congênitas/genética , Histona-Lisina N-Metiltransferase/genética , Deficiência Intelectual/genética , Masculino , Camundongos
13.
Biol Cybern ; 114(2): 231-248, 2020 04.
Artigo em Inglês | MEDLINE | ID: mdl-32065253

RESUMO

Hippocampal offline reactivations during reward-based learning, usually categorized as replay events, have been found to be important for performance improvement over time and for memory consolidation. Recent computational work has linked these phenomena to the need to transform reward information into state-action values for decision making and to propagate it to all relevant states of the environment. Nevertheless, it is still unclear whether an integrated reinforcement learning mechanism could account for the variety of awake hippocampal reactivations, including variety in order (forward and reverse reactivated trajectories) and variety in the location where they occur (reward site or decision-point). Here, we present a model-based bidirectional search model which accounts for a variety of hippocampal reactivations. The model combines forward trajectory sampling from current position and backward sampling through prioritized sweeping from states associated with large reward prediction errors until the two trajectories connect. This is repeated until stabilization of state-action values (convergence), which could explain why hippocampal reactivations drastically diminish when the animal's performance stabilizes. Simulations in a multiple T-maze task show that forward reactivations are prominently found at decision-points while backward reactivations are exclusively generated at reward sites. Finally, the model can generate imaginary trajectories that are not allowed to the agent during task performance. We raise some experimental predictions and implications for future studies of the role of the hippocampo-prefronto-striatal network in learning.


Assuntos
Hipocampo/fisiologia , Modelos Neurológicos , Vigília , Algoritmos , Animais , Simulação por Computador , Aprendizagem , Aprendizagem em Labirinto , Células de Lugar/fisiologia , Reforço Psicológico , Recompensa , Roedores
14.
Psychopharmacology (Berl) ; 236(8): 2373-2388, 2019 Aug.
Artigo em Inglês | MEDLINE | ID: mdl-31367850

RESUMO

In the context of Pavlovian conditioning, two types of behaviour may emerge within the population (Flagel et al. Nature, 469(7328): 53-57, 2011). Animals may choose to engage either with the conditioned stimulus (CS), a behaviour known as sign-tracking (ST) which is sensitive to dopamine inhibition for its acquisition, or with the food cup in which the reward or unconditioned stimulus (US) will eventually be delivered, a behaviour known as goal-tracking (GT) which is dependent on dopamine for its expression only. Previous work by Lesaint et al. (PLoS Comput Biol, 10(2), 2014) offered a computational explanation for these phenomena and led to the prediction that varying the duration of the inter-trial interval (ITI) would change the relative ST-GT proportion in the population as well as phasic dopamine responses. A recent study verified this prediction, but also found a rich variance of ST and GT behaviours within the trial which goes beyond the original computational model. In this paper, we provide a computational perspective on these novel results.


Assuntos
Simulação por Computador , Condicionamento Clássico/fisiologia , Condicionamento Operante/fisiologia , Objetivos , Animais , Dopamina/metabolismo , Masculino , Motivação , Recompensa , Fatores de Tempo
15.
PLoS Biol ; 17(6): e3000322, 2019 06.
Artigo em Inglês | MEDLINE | ID: mdl-31206519

RESUMO

Declarative memory encompasses representations of specific events as well as knowledge extracted by accumulation over multiple episodes. To investigate how these different sorts of memories are created, we developed a new behavioral task in rodents. The task consists of 3 distinct conditions (stable, overlapping, and random). Rodents are exposed to multiple sample trials, in which they explore objects in specific spatial arrangements, with object identity changing from trial to trial. In the stable condition, the locations are constant during all sample trials even though the objects themselves change; in the test trial, 1 object's location is changed. In the random condition, object locations are presented in the sample phase without a specific spatial pattern. In the overlapping condition, 1 location is shared (overlapping) between all trials, while the other location changes during sample trials. We show that in the overlapping condition, instead of only remembering the last sample trial, rodents form a cumulative memory of the sample trials. Here, we could show that both mice and rats can accumulate information across multiple trials and express a long-term abstracted memory.


Assuntos
Memória de Longo Prazo/fisiologia , Memória/fisiologia , Memória Espacial/fisiologia , Animais , Comportamento Animal/fisiologia , Masculino , Camundongos , Camundongos Endogâmicos C57BL , Ratos , Ratos Wistar , Reconhecimento Psicológico/fisiologia , Percepção Espacial/fisiologia
16.
Sci Rep ; 9(1): 6770, 2019 05 01.
Artigo em Inglês | MEDLINE | ID: mdl-31043685

RESUMO

In a volatile environment where rewards are uncertain, successful performance requires a delicate balance between exploitation of the best option and exploration of alternative choices. It has theoretically been proposed that dopamine contributes to the control of this exploration-exploitation trade-off, specifically that the higher the level of tonic dopamine, the more exploitation is favored. We demonstrate here that there is a formal relationship between the rescaling of dopamine positive reward prediction errors and the exploration-exploitation trade-off in simple non-stationary multi-armed bandit tasks. We further show in rats performing such a task that systemically antagonizing dopamine receptors greatly increases the number of random choices without affecting learning capacities. Simulations and comparison of a set of different computational models (an extended Q-learning model, a directed exploration model, and a meta-learning model) fitted on each individual confirm that, independently of the model, decreasing dopaminergic activity does not affect learning rate but is equivalent to an increase in random exploration rate. This study shows that dopamine could adapt the exploration-exploitation trade-off in decision-making when facing changing environmental contingencies.


Assuntos
Tomada de Decisões , Antagonistas de Dopamina/farmacologia , Dopamina/química , Comportamento Exploratório/fisiologia , Modelos Teóricos , Recompensa , Animais , Dopamina/metabolismo , Comportamento Exploratório/efeitos dos fármacos , Masculino , Aprendizagem por Probabilidade , Ratos , Ratos Long-Evans
17.
J Neurophysiol ; 120(6): 2877-2896, 2018 12 01.
Artigo em Inglês | MEDLINE | ID: mdl-30303758

RESUMO

Multiple in vivo studies have shown that place cells from the hippocampus replay previously experienced trajectories. These replays are commonly considered to mainly reflect memory consolidation processes. Some data, however, have highlighted a functional link between replays and reinforcement learning (RL). This theory, extensively used in machine learning, has introduced efficient algorithms and can explain various behavioral and physiological measures from different brain regions. RL algorithms could constitute a mechanistic description of replays and explain how replays can reduce the number of iterations required to explore the environment during learning. We review the main findings concerning the different hippocampal replay types and the possible associated RL models (either model-based, model-free, or hybrid model types). We conclude by tying these frameworks together. We illustrate the link between data and RL through a series of model simulations. This review, at the frontier between informatics and biology, paves the way for future work on replays.


Assuntos
Hipocampo/fisiologia , Reforço Psicológico , Animais , Conectoma , Humanos , Modelos Neurológicos , Sono
18.
Nat Commun ; 9(1): 4503, 2018 10 29.
Artigo em Inglês | MEDLINE | ID: mdl-30374019

RESUMO

In economics and perceptual decision-making contextual effects are well documented, where decision weights are adjusted as a function of the distribution of stimuli. Yet, in reinforcement learning literature whether and how contextual information pertaining to decision states is integrated in learning algorithms has received comparably little attention. Here, we investigate reinforcement learning behavior and its computational substrates in a task where we orthogonally manipulate outcome valence and magnitude, resulting in systematic variations in state-values. Model comparison indicates that subjects' behavior is best accounted for by an algorithm which includes both reference point-dependence and range-adaptation-two crucial features of state-dependent valuation. In addition, we find that state-dependent outcome valuation progressively emerges, is favored by increasing outcome information and correlated with explicit understanding of the task structure. Finally, our data clearly show that, while being locally adaptive (for instance in negative valence and small magnitude contexts), state-dependent valuation comes at the cost of seemingly irrational choices, when options are extrapolated out from their original contexts.


Assuntos
Aprendizagem/fisiologia , Valores de Referência , Reforço Psicológico , Adolescente , Adulto , Algoritmos , Atenção , Comportamento/fisiologia , Simulação por Computador , Tomada de Decisões/fisiologia , Feminino , Humanos , Masculino , Modelos Neurológicos , Recompensa , Adulto Jovem
19.
PLoS Biol ; 16(9): e2004015, 2018 09.
Artigo em Inglês | MEDLINE | ID: mdl-30256785

RESUMO

Recent computational models of sign tracking (ST) and goal tracking (GT) have accounted for observations that dopamine (DA) is not necessary for all forms of learning and have provided a set of predictions to further their validity. Among these, a central prediction is that manipulating the intertrial interval (ITI) during autoshaping should change the relative ST-GT proportion as well as DA phasic responses. Here, we tested these predictions and found that lengthening the ITI increased ST, i.e., behavioral engagement with conditioned stimuli (CS) and cue-induced phasic DA release. Importantly, DA release was also present at the time of reward delivery, even after learning, and DA release was correlated with time spent in the food cup during the ITI. During conditioning with shorter ITIs, GT was prominent (i.e., engagement with food cup), and DA release responded to the CS while being absent at the time of reward delivery after learning. Hence, shorter ITIs restored the classical DA reward prediction error (RPE) pattern. These results validate the computational hypotheses, opening new perspectives on the understanding of individual differences in Pavlovian conditioning and DA signaling.


Assuntos
Dopamina/metabolismo , Modelos Biológicos , Recompensa , Animais , Condicionamento Clássico , Objetivos , Masculino , Ratos Sprague-Dawley
20.
PLoS Comput Biol ; 14(4): e1006092, 2018 04.
Artigo em Inglês | MEDLINE | ID: mdl-29630600

RESUMO

We present a computational model of spatial navigation comprising different learning mechanisms in mammals, i.e., associative, cognitive mapping and parallel systems. This model is able to reproduce a large number of experimental results in different variants of the Morris water maze task, including standard associative phenomena (spatial generalization gradient and blocking), as well as navigation based on cognitive mapping. Furthermore, we show that competitive and cooperative patterns between different navigation strategies in the model allow to explain previous apparently contradictory results supporting either associative or cognitive mechanisms for spatial learning. The key computational mechanism to reconcile experimental results showing different influences of distal and proximal cues on the behavior, different learning times, and different abilities of individuals to alternatively perform spatial and response strategies, relies in the dynamic coordination of navigation strategies, whose performance is evaluated online with a common currency through a modular approach. We provide a set of concrete experimental predictions to further test the computational model. Overall, this computational work sheds new light on inter-individual differences in navigation learning, and provides a formal and mechanistic approach to test various theories of spatial cognition in mammals.


Assuntos
Modelos Psicológicos , Navegação Espacial/fisiologia , Animais , Comportamento Animal/fisiologia , Cognição/fisiologia , Biologia Computacional , Simulação por Computador , Sinais (Psicologia) , Mamíferos , Aprendizagem em Labirinto/fisiologia , Memória/fisiologia , Reforço Psicológico
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...