Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 10 de 10
Filtrar
Mais filtros










Base de dados
Intervalo de ano de publicação
2.
Elife ; 122024 Jun 28.
Artigo em Inglês | MEDLINE | ID: mdl-38941238

RESUMO

How human prefrontal and insular regions interact while maximizing rewards and minimizing punishments is unknown. Capitalizing on human intracranial recordings, we demonstrate that the functional specificity toward reward or punishment learning is better disentangled by interactions compared to local representations. Prefrontal and insular cortices display non-selective neural populations to rewards and punishments. Non-selective responses, however, give rise to context-specific interareal interactions. We identify a reward subsystem with redundant interactions between the orbitofrontal and ventromedial prefrontal cortices, with a driving role of the latter. In addition, we find a punishment subsystem with redundant interactions between the insular and dorsolateral cortices, with a driving role of the insula. Finally, switching between reward and punishment learning is mediated by synergistic interactions between the two subsystems. These results provide a unifying explanation of distributed cortical representations and interactions supporting reward and punishment learning.


Assuntos
Aprendizagem , Córtex Pré-Frontal , Punição , Recompensa , Humanos , Masculino , Adulto , Feminino , Córtex Pré-Frontal/fisiologia , Aprendizagem/fisiologia , Adulto Jovem , Córtex Insular/fisiologia , Lobo Frontal/fisiologia
3.
Nat Hum Behav ; 2024 Jun 14.
Artigo em Inglês | MEDLINE | ID: mdl-38877287

RESUMO

Recent evidence indicates that reward value encoding in humans is highly context dependent, leading to suboptimal decisions in some cases, but whether this computational constraint on valuation is a shared feature of human cognition remains unknown. Here we studied the behaviour of n = 561 individuals from 11 countries of markedly different socioeconomic and cultural makeup. Our findings show that context sensitivity was present in all 11 countries. Suboptimal decisions generated by context manipulation were not explained by risk aversion, as estimated through a separate description-based choice task (that is, lotteries) consisting of matched decision offers. Conversely, risk aversion significantly differed across countries. Overall, our findings suggest that context-dependent reward value encoding is a feature of human cognition that remains consistently present across different countries, as opposed to description-based decision-making, which is more permeable to cultural factors.

4.
Biol Psychiatry ; 95(10): 974-984, 2024 May 15.
Artigo em Inglês | MEDLINE | ID: mdl-38101503

RESUMO

BACKGROUND: Drugs like opioids are potent reinforcers thought to co-opt value-based decisions by overshadowing other rewarding outcomes, but how this happens at a neurocomputational level remains elusive. Range adaptation is a canonical process of fine-tuning representations of value based on reward context. Here, we tested whether recent opioid exposure impacts range adaptation in opioid use disorder, potentially explaining why shifting decision making away from drug taking during this vulnerable period is so difficult. METHODS: Participants who had recently (<90 days) used opioids (n = 34) or who had abstained from opioid use for ≥ 90 days (n = 20) and comparison control participants (n = 44) completed a reinforcement learning task designed to induce robust contextual modulation of value. Two models were used to assess the latent process that participants engaged while making their decisions: 1) a Range model that dynamically tracks context and 2) a standard Absolute model that assumes stationary, objective encoding of value. RESULTS: Control participants and ≥90-days-abstinent participants with opioid use disorder exhibited choice patterns consistent with range-adapted valuation. In contrast, participants with recent opioid use were more prone to learn and encode value on an absolute scale. Computational modeling confirmed the behavior of most control participants and ≥90-days-abstinent participants with opioid use disorder (75%), but a minority in the recent use group (38%), was better fit by the Range model than the Absolute model. Furthermore, the degree to which participants relied on range adaptation correlated with duration of continuous abstinence and subjective craving/withdrawal. CONCLUSIONS: Reduced context adaptation to available rewards could explain difficulty deciding about smaller (typically nondrug) rewards in the aftermath of drug exposure.


Assuntos
Transtornos Relacionados ao Uso de Opioides , Reforço Psicológico , Humanos , Masculino , Adulto , Feminino , Recompensa , Adulto Jovem , Tomada de Decisões/efeitos dos fármacos , Tomada de Decisões/fisiologia , Analgésicos Opioides/administração & dosagem , Analgésicos Opioides/farmacologia , Comportamento de Escolha/efeitos dos fármacos , Comportamento de Escolha/fisiologia , Adaptação Psicológica/efeitos dos fármacos , Adaptação Psicológica/fisiologia
5.
Nat Commun ; 14(1): 6534, 2023 10 17.
Artigo em Inglês | MEDLINE | ID: mdl-37848435

RESUMO

Reinforcement-based adaptive decision-making is believed to recruit fronto-striatal circuits. A critical node of the fronto-striatal circuit is the thalamus. However, direct evidence of its involvement in human reinforcement learning is lacking. We address this gap by analyzing intra-thalamic electrophysiological recordings from eight participants while they performed a reinforcement learning task. We found that in both the anterior thalamus (ATN) and dorsomedial thalamus (DMTN), low frequency oscillations (LFO, 4-12 Hz) correlated positively with expected value estimated from computational modeling during reward-based learning (after outcome delivery) or punishment-based learning (during the choice process). Furthermore, LFO recorded from ATN/DMTN were also negatively correlated with outcomes so that both components of reward prediction errors were signaled in the human thalamus. The observed differences in the prediction signals between rewarding and punishing conditions shed light on the neural mechanisms underlying action inhibition in punishment avoidance learning. Our results provide insight into the role of thalamus in reinforcement-based decision-making in humans.


Assuntos
Reforço Psicológico , Recompensa , Humanos , Aprendizagem da Esquiva/fisiologia , Punição , Tálamo
6.
J Affect Disord ; 340: 694-702, 2023 11 01.
Artigo em Inglês | MEDLINE | ID: mdl-37591352

RESUMO

BACKGROUND: Reward sensitivity is an essential dimension related to mood fluctuations in bipolar disorder (BD), but there is currently a debate around hypersensitivity or hyposensitivity hypotheses to reward in BD during remission, probably related to a heterogeneous population within the BD spectrum and a lack of reward bias evaluation. Here, we examine reward maximization vs. punishment avoidance learning within the BD spectrum during remission. METHODS: Patients with BD-I (n = 45), BD-II (n = 34) and matched (n = 30) healthy controls (HC) were included. They performed an instrumental learning task designed to dissociate reward-based from punishment-based reinforcement learning. Computational modeling was used to identify the mechanisms underlying reinforcement learning performances. RESULTS: Behavioral results showed a significant reward learning deficit across BD subtypes compared to HC, captured at the computational level by a lower sensitivity to rewards compared to punishments in both BD subtypes. Computational modeling also revealed a higher choice randomness in BD-II compared to BD-I that reflected a tendency of BD-I to perform better during punishment avoidance learning than BD-II. LIMITATIONS: Our patients were not naive to antipsychotic treatment and were not euthymic (but in syndromic remission) according to the International Society for Bipolar Disorder definition. CONCLUSIONS: Our results are consistent with the reward hyposensitivity theory in BD. Computational modeling suggests distinct underlying mechanisms that produce similar observable behaviors, making it a useful tool for distinguishing how symptoms interact in BD versus other disorders. In the long run, a better understanding of these processes could contribute to better prevention and management of BD.


Assuntos
Transtorno Bipolar , Punição , Humanos , Recompensa , Reforço Psicológico , Aprendizagem da Esquiva
7.
Res Sq ; 2023 Mar 02.
Artigo em Inglês | MEDLINE | ID: mdl-36909645

RESUMO

Recent evidence indicates that reward value encoding in humans is highly context-dependent, leading to suboptimal decisions in some cases. But whether this computational constraint on valuation is a shared feature of human cognition remains unknown. To address this question, we studied the behavior of individuals from across 11 countries of markedly different socioeconomic and cultural makeup using an experimental approach that reliably captures context effects in reinforcement learning. Our findings show that all samples presented evidence of similar sensitivity to context. Crucially, suboptimal decisions generated by context manipulation were not explained by risk aversion, as estimated through a separate description-based choice task (i.e., lotteries) consisting of matched decision offers. Conversely, risk aversion significantly differed across countries. Overall, our findings suggest that context-dependent reward value encoding is a hardcoded feature of human cognition, while description-based decision-making is significantly sensitive to cultural factors.

8.
Curr Opin Behav Sci ; 38: 40-48, 2021 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-34423103

RESUMO

Computational psychiatry provides a powerful new approach for linking the behavioral manifestations of addiction to their precise cognitive and neurobiological substrates. However, this emerging area of research is still limited in important ways. While research has identified features of reinforcement learning and decision-making in substance users that differ from health, less emphasis has been placed on capturing addiction cycles/states dynamically, within-person. In addition, the focus on few behavioral variables at a time has precluded more detailed consideration of related processes and heterogeneous clinical profiles. We propose that a longitudinal and multidimensional examination of value-based processes, a type of dynamic "computational fingerprint", will provide a more complete understanding of addiction as well as aid in developing better tailored and timed interventions.

9.
Nat Commun ; 12(1): 3344, 2021 06 07.
Artigo em Inglês | MEDLINE | ID: mdl-34099678

RESUMO

Whether maximizing rewards and minimizing punishments rely on distinct brain systems remains debated, given inconsistent results coming from human neuroimaging and animal electrophysiology studies. Bridging the gap across techniques, we recorded intracerebral activity from twenty participants while they performed an instrumental learning task. We found that both reward and punishment prediction errors (PE), estimated from computational modeling of choice behavior, correlate positively with broadband gamma activity (BGA) in several brain regions. In all cases, BGA scaled positively with the outcome (reward or punishment versus nothing) and negatively with the expectation (predictability of reward or punishment). However, reward PE were better signaled in some regions (such as the ventromedial prefrontal and lateral orbitofrontal cortex), and punishment PE in other regions (such as the anterior insula and dorsolateral prefrontal cortex). These regions might therefore belong to brain systems that differentially contribute to the repetition of rewarded choices and the avoidance of punished choices.


Assuntos
Córtex Cerebral/anatomia & histologia , Córtex Cerebral/patologia , Punição , Recompensa , Adulto , Animais , Encéfalo/anatomia & histologia , Encéfalo/fisiologia , Mapeamento Encefálico , Neurociência Cognitiva , Condicionamento Operante , Feminino , Humanos , Imageamento por Ressonância Magnética , Masculino , Pessoa de Meia-Idade , Neuroimagem , Córtex Pré-Frontal/fisiologia
10.
Cereb Cortex ; 27(2): 1545-1557, 2017 02 01.
Artigo em Inglês | MEDLINE | ID: mdl-26796212

RESUMO

The ability to monitor our own errors is mediated by a network that includes dorsomedial prefrontal cortex (dmPFC) and anterior insula (AI). However, the dynamics of the underlying neurophysiological processes remain unclear. In particular, whether AI is on the receiving or driving end of the error-monitoring network is unresolved. Here, we recorded intracerebral electroencephalography signals simultaneously from AI and dmPFC in epileptic patients while they performed a stop-signal task. We found that errors selectively modulated broadband neural activity in human AI. Granger causality estimates revealed that errors were immediately followed by a feedforward influence from AI onto anterior cingulate cortex and, subsequently, onto presupplementary motor area. The reverse pattern of information flow was observed on correct responses. Our findings provide the first direct electrophysiological evidence indicating that the anterior insula rapidly detects and conveys error signals to dmPFC, while the latter might use this input to adapt behavior following inappropriate actions.


Assuntos
Mapeamento Encefálico , Giro do Cíngulo/fisiologia , Córtex Motor/fisiologia , Desempenho Psicomotor/fisiologia , Adulto , Eletroencefalografia/métodos , Feminino , Humanos , Imageamento por Ressonância Magnética/métodos , Masculino , Tempo de Reação
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...