Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 28
Filter
1.
Diagnostics (Basel) ; 14(9)2024 Apr 23.
Article in English | MEDLINE | ID: mdl-38732275

ABSTRACT

Injury to the recurrent laryngeal nerve (RLN) can be a devastating complication of thyroid and parathyroid surgery. Intraoperative neuromonitoring (IONM) has been proposed as a method to reduce the number of RLN injuries but the data are inconsistent. We performed a meta-analysis to critically assess the data. After applying inclusion and exclusion criteria, 60 studies, including five randomized trials and eight non-randomized prospective trials, were included. A meta-analysis of all studies demonstrated an odds ratio (OR) of 0.66 (95% CI [0.56, 0.79], p < 0.00001) favoring IONM compared to the visual identification of the RLN in limiting permanent RLN injuries. A meta-analysis of studies employing contemporaneous controls and routine postoperative laryngoscopy to diagnose RLN injuries (considered to be the most reliable design) demonstrated an OR of 0.69 (95% CI [0.56, 0.84], p = 0.0003), favoring IONM. Strong consideration should be given to employing IONM when performing thyroid and parathyroid surgery.

2.
Nat Neurosci ; 26(8): 1438-1448, 2023 08.
Article in English | MEDLINE | ID: mdl-37474639

ABSTRACT

Memorization and generalization are complementary cognitive processes that jointly promote adaptive behavior. For example, animals should memorize safe routes to specific water sources and generalize from these memories to discover environmental features that predict new ones. These functions depend on systems consolidation mechanisms that construct neocortical memory traces from hippocampal precursors, but why systems consolidation only applies to a subset of hippocampal memories is unclear. Here we introduce a new neural network formalization of systems consolidation that reveals an overlooked tension-unregulated neocortical memory transfer can cause overfitting and harm generalization in an unpredictable world. We resolve this tension by postulating that memories only consolidate when it aids generalization. This framework accounts for partial hippocampal-cortical memory transfer and provides a normative principle for reconceptualizing numerous observations in the field. Generalization-optimized systems consolidation thus provides new insight into how adaptive behavior benefits from complementary learning systems specialized for memorization and generalization.


Subject(s)
Learning , Memory Consolidation , Animals , Generalization, Psychological , Hippocampus
3.
Neuron ; 111(12): 1966-1978.e8, 2023 06 21.
Article in English | MEDLINE | ID: mdl-37119818

ABSTRACT

Mammals form mental maps of the environments by exploring their surroundings. Here, we investigate which elements of exploration are important for this process. We studied mouse escape behavior, in which mice are known to memorize subgoal locations-obstacle edges-to execute efficient escape routes to shelter. To test the role of exploratory actions, we developed closed-loop neural-stimulation protocols for interrupting various actions while mice explored. We found that blocking running movements directed at obstacle edges prevented subgoal learning; however, blocking several control movements had no effect. Reinforcement learning simulations and analysis of spatial data show that artificial agents can match these results if they have a region-level spatial representation and explore with object-directed movements. We conclude that mice employ an action-driven process for integrating subgoals into a hierarchical cognitive map. These findings broaden our understanding of the cognitive toolkit that mammals use to acquire spatial knowledge.


Subject(s)
Learning , Reinforcement, Psychology , Mice , Animals , Mammals
4.
Neuron ; 111(9): 1504-1516.e9, 2023 05 03.
Article in English | MEDLINE | ID: mdl-36898375

ABSTRACT

Human understanding of the world can change rapidly when new information comes to light, such as when a plot twist occurs in a work of fiction. This flexible "knowledge assembly" requires few-shot reorganization of neural codes for relations among objects and events. However, existing computational theories are largely silent about how this could occur. Here, participants learned a transitive ordering among novel objects within two distinct contexts before exposure to new knowledge that revealed how they were linked. Blood-oxygen-level-dependent (BOLD) signals in dorsal frontoparietal cortical areas revealed that objects were rapidly and dramatically rearranged on the neural manifold after minimal exposure to linking information. We then adapt online stochastic gradient descent to permit similar rapid knowledge assembly in a neural network model.


Subject(s)
Learning , Neural Networks, Computer , Humans , Frontal Lobe
5.
Elife ; 122023 02 14.
Article in English | MEDLINE | ID: mdl-36786427

ABSTRACT

Making optimal decisions in the face of noise requires balancing short-term speed and accuracy. But a theory of optimality should account for the fact that short-term speed can influence long-term accuracy through learning. Here, we demonstrate that long-term learning is an important dynamical dimension of the speed-accuracy trade-off. We study learning trajectories in rats and formally characterize these dynamics in a theory expressed as both a recurrent neural network and an analytical extension of the drift-diffusion model that learns over time. The model reveals that choosing suboptimal response times to learn faster sacrifices immediate reward, but can lead to greater total reward. We empirically verify predictions of the theory, including a relationship between stimulus exposure and learning speed, and a modulation of reaction time by future learning prospects. We find that rats' strategies approximately maximize total reward over the full learning epoch, suggesting cognitive control over the learning process.


Subject(s)
Decision Making , Learning , Animals , Rats , Decision Making/physiology , Reaction Time/physiology , Reward , Neural Networks, Computer
6.
Trends Neurosci ; 46(3): 199-210, 2023 03.
Article in English | MEDLINE | ID: mdl-36682991

ABSTRACT

How do humans and other animals learn new tasks? A wave of brain recording studies has investigated how neural representations change during task learning, with a focus on how tasks can be acquired and coded in ways that minimise mutual interference. We review recent work that has explored the geometry and dimensionality of neural task representations in neocortex, and computational models that have exploited these findings to understand how the brain may partition knowledge between tasks. We discuss how ideas from machine learning, including those that combine supervised and unsupervised learning, are helping neuroscientists understand how natural tasks are learned and coded in biological brains.


Subject(s)
Machine Learning , Neural Networks, Computer , Animals , Humans , Brain
7.
PLoS Comput Biol ; 19(1): e1010808, 2023 Jan.
Article in English | MEDLINE | ID: mdl-36656823

ABSTRACT

Humans can learn several tasks in succession with minimal mutual interference but perform more poorly when trained on multiple tasks at once. The opposite is true for standard deep neural networks. Here, we propose novel computational constraints for artificial neural networks, inspired by earlier work on gating in the primate prefrontal cortex, that capture the cost of interleaved training and allow the network to learn two tasks in sequence without forgetting. We augment standard stochastic gradient descent with two algorithmic motifs, so-called "sluggish" task units and a Hebbian training step that strengthens connections between task units and hidden units that encode task-relevant information. We found that the "sluggish" units introduce a switch-cost during training, which biases representations under interleaved training towards a joint representation that ignores the contextual cue, while the Hebbian step promotes the formation of a gating scheme from task units to the hidden layer that produces orthogonal representations which are perfectly guarded against interference. Validating the model on previously published human behavioural data revealed that it matches performance of participants who had been trained on blocked or interleaved curricula, and that these performance differences were driven by misestimation of the true category boundary.


Subject(s)
Learning , Neural Networks, Computer , Animals , Humans , Machine Learning , Prefrontal Cortex , Curriculum
8.
J Stat Mech ; 2023(11): 114004, 2023 Nov 01.
Article in English | MEDLINE | ID: mdl-38524253

ABSTRACT

Learning in deep neural networks is known to depend critically on the knowledge embedded in the initial network weights. However, few theoretical results have precisely linked prior knowledge to learning dynamics. Here we derive exact solutions to the dynamics of learning with rich prior knowledge in deep linear networks by generalising Fukumizu's matrix Riccati solution (Fukumizu 1998 Gen 1 1E-03). We obtain explicit expressions for the evolving network function, hidden representational similarity, and neural tangent kernel over training for a broad class of initialisations and tasks. The expressions reveal a class of task-independent initialisations that radically alter learning dynamics from slow non-linear dynamics to fast exponential trajectories while converging to a global optimum with identical representational similarity, dissociating learning trajectories from the structure of initial internal representations. We characterise how network weights dynamically align with task structure, rigorously justifying why previous solutions successfully described learning from small initial weights without incorporating their fine-scale structure. Finally, we discuss the implications of these findings for continual learning, reversal learning and learning of structured knowledge. Taken together, our results provide a mathematical toolkit for understanding the impact of prior knowledge on deep learning.

10.
Neuron ; 110(7): 1258-1270.e11, 2022 04 06.
Article in English | MEDLINE | ID: mdl-35085492

ABSTRACT

How do neural populations code for multiple, potentially conflicting tasks? Here we used computational simulations involving neural networks to define "lazy" and "rich" coding solutions to this context-dependent decision-making problem, which trade off learning speed for robustness. During lazy learning the input dimensionality is expanded by random projections to the network hidden layer, whereas in rich learning hidden units acquire structured representations that privilege relevant over irrelevant features. For context-dependent decision-making, one rich solution is to project task representations onto low-dimensional and orthogonal manifolds. Using behavioral testing and neuroimaging in humans and analysis of neural signals from macaque prefrontal cortex, we report evidence for neural coding patterns in biological brains whose dimensionality and neural geometry are consistent with the rich learning regime.


Subject(s)
Neural Networks, Computer , Task Performance and Analysis , Brain , Learning , Prefrontal Cortex
11.
J Stat Mech ; 2022(11): 114014, 2022 Nov 01.
Article in English | MEDLINE | ID: mdl-37817944

ABSTRACT

In animals and humans, curriculum learning-presenting data in a curated order-is critical to rapid learning and effective pedagogy. A long history of experiments has demonstrated the impact of curricula in a variety of animals but, despite its ubiquitous presence, a theoretical understanding of the phenomenon is still lacking. Surprisingly, in contrast to animal learning, curricula strategies are not widely used in machine learning and recent simulation studies reach the conclusion that curricula are moderately effective or even ineffective in most cases. This stark difference in the importance of curriculum raises a fundamental theoretical question: when and why does curriculum learning help? In this work, we analyse a prototypical neural network model of curriculum learning in the high-dimensional limit, employing statistical physics methods. We study a task in which a sparse set of informative features are embedded amidst a large set of noisy features. We analytically derive average learning trajectories for simple neural networks on this task, which establish a clear speed benefit for curriculum learning in the online setting. However, when training experiences can be stored and replayed (for instance, during sleep), the advantage of curriculum in standard neural networks disappears, in line with observations from the deep learning literature. Inspired by synaptic consolidation techniques developed to combat catastrophic forgetting, we propose curriculum-aware algorithms that consolidate synapses at curriculum change points and investigate whether this can boost the benefits of curricula. We derive generalisation performance as a function of consolidation strength (implemented as an L 2 regularisation/elastic coupling connecting learning phases), and show that curriculum-aware algorithms can yield a large improvement in test performance. Our reduced analytical descriptions help reconcile apparently conflicting empirical results, trace regimes where curriculum learning yields the largest gains, and provide experimentally-accessible predictions for the impact of task parameters on curriculum benefits. More broadly, our results suggest that fully exploiting a curriculum may require explicit adjustments in the loss.

12.
Nat Rev Neurosci ; 22(1): 55-67, 2021 01.
Article in English | MEDLINE | ID: mdl-33199854

ABSTRACT

Neuroscience research is undergoing a minor revolution. Recent advances in machine learning and artificial intelligence research have opened up new ways of thinking about neural computation. Many researchers are excited by the possibility that deep neural networks may offer theories of perception, cognition and action for biological brains. This approach has the potential to radically reshape our approach to understanding neural systems, because the computations performed by deep networks are learned from experience, and not endowed by the researcher. If so, how can neuroscientists use deep networks to model and understand biological brains? What is the outlook for neuroscientists who seek to characterize computations or neural codes, or who wish to understand perception, attention, memory and executive functions? In this Perspective, our goal is to offer a road map for systems neuroscience research in the age of deep learning. We discuss the conceptual and methodological challenges of comparing behaviour, learning dynamics and neural representations in artificial and biological systems, and we highlight new research questions that have emerged for neuroscience as a direct consequence of recent advances in machine learning.


Subject(s)
Brain , Deep Learning , Neural Networks, Computer , Humans , Neurosciences
13.
Neural Netw ; 132: 428-446, 2020 Dec.
Article in English | MEDLINE | ID: mdl-33022471

ABSTRACT

We perform an analysis of the average generalization dynamics of large neural networks trained using gradient descent. We study the practically-relevant "high-dimensional" regime where the number of free parameters in the network is on the order of or even larger than the number of examples in the dataset. Using random matrix theory and exact solutions in linear models, we derive the generalization error and training error dynamics of learning and analyze how they depend on the dimensionality of data and signal to noise ratio of the learning problem. We find that the dynamics of gradient descent learning naturally protect against overtraining and overfitting in large networks. Overtraining is worst at intermediate network sizes, when the effective number of free parameters equals the number of samples, and thus can be reduced by making a network smaller or larger. Additionally, in the high-dimensional regime, low generalization error requires starting with small initial weights. We then turn to non-linear neural networks, and show that making networks very large does not harm their generalization performance. On the contrary, it can in fact reduce overtraining, even without early stopping or regularization of any sort. We identify two novel phenomena underlying this behavior in overcomplete models: first, there is a frozen subspace of the weights in which no learning occurs under gradient descent; and second, the statistical properties of the high-dimensional regime yield better-conditioned input correlations which protect against overtraining. We demonstrate that standard application of theories such as Rademacher complexity are inaccurate in predicting the generalization performance of deep neural networks, and derive an alternative bound which incorporates the frozen subspace and conditioning effects and qualitatively matches the behavior observed in simulation.


Subject(s)
Deep Learning
14.
J Stat Mech ; 2020(12): 124010, 2020 Dec.
Article in English | MEDLINE | ID: mdl-34262607

ABSTRACT

Deep neural networks achieve stellar generalisation even when they have enough parameters to easily fit all their training data. We study this phenomenon by analysing the dynamics and the performance of over-parameterised two-layer neural networks in the teacher-student setup, where one network, the student, is trained on data generated by another network, called the teacher. We show how the dynamics of stochastic gradient descent (SGD) is captured by a set of differential equations and prove that this description is asymptotically exact in the limit of large inputs. Using this framework, we calculate the final generalisation error of student networks that have more parameters than their teachers. We find that the final generalisation error of the student increases with network size when training only the first layer, but stays constant or even decreases with size when training both layers. We show that these different behaviours have their root in the different solutions SGD finds for different activation functions. Our results indicate that achieving good generalisation in neural networks goes beyond the properties of SGD alone and depends on the interplay of at least the algorithm, the model architecture, and the data set.

15.
Nat Neurosci ; 22(11): 1761-1770, 2019 11.
Article in English | MEDLINE | ID: mdl-31659335

ABSTRACT

Systems neuroscience seeks explanations for how the brain implements a wide variety of perceptual, cognitive and motor tasks. Conversely, artificial intelligence attempts to design computational systems based on the tasks they will have to solve. In artificial neural networks, the three components specified by design are the objective functions, the learning rules and the architectures. With the growing success of deep learning, which utilizes brain-inspired architectures, these three designed components have increasingly become central to how we model, engineer and optimize complex artificial learning systems. Here we argue that a greater focus on these components would also benefit systems neuroscience. We give examples of how this optimization-based framework can drive theoretical and experimental progress in neuroscience. We contend that this principled perspective on systems neuroscience will help to generate more rapid progress.


Subject(s)
Artificial Intelligence , Deep Learning , Neural Networks, Computer , Animals , Brain/physiology , Humans
16.
Proc Natl Acad Sci U S A ; 116(23): 11537-11546, 2019 06 04.
Article in English | MEDLINE | ID: mdl-31101713

ABSTRACT

An extensive body of empirical research has revealed remarkable regularities in the acquisition, organization, deployment, and neural representation of human semantic knowledge, thereby raising a fundamental conceptual question: What are the theoretical principles governing the ability of neural networks to acquire, organize, and deploy abstract knowledge by integrating across many individual experiences? We address this question by mathematically analyzing the nonlinear dynamics of learning in deep linear networks. We find exact solutions to this learning dynamics that yield a conceptual explanation for the prevalence of many disparate phenomena in semantic cognition, including the hierarchical differentiation of concepts through rapid developmental transitions, the ubiquity of semantic illusions between such transitions, the emergence of item typicality and category coherence as factors controlling the speed of semantic processing, changing patterns of inductive projection over development, and the conservation of semantic similarity in neural representations across species. Thus, surprisingly, our simple neural model qualitatively recapitulates many diverse regularities underlying semantic development, while providing analytic insight into how the statistical structure of an environment can interact with nonlinear deep-learning dynamics to give rise to these regularities.

17.
Med Educ Online ; 18: 20598, 2013 Jul 22.
Article in English | MEDLINE | ID: mdl-23880149

ABSTRACT

INTRODUCTION: We operationalized the taxonomy developed by Hauer and colleagues describing common clinical performance problems. Faculty raters pilot tested the resulting worksheet by observing recordings of problematic simulated clinical encounters involving third-year medical students. This approach provided a framework for structured feedback to guide learner improvement and curricular enhancement. METHODS: Eighty-two problematic clinical encounters from M3 students who failed their clinical competency examination were independently rated by paired clinical faculty members to identify common problems related to the medical interview, physical examination, and professionalism. RESULTS: Eleven out of 26 target performance problems were present in 25% or more encounters. Overall, 37% had unsatisfactory medical interviews, with 'inadequate history to rule out other diagnoses' most prevalent (60%). Seventy percent failed because of physical examination deficiencies, with missing elements (69%) and inadequate data gathering (69%) most common. One-third of the students did not introduce themselves to their patients. Among students failing based on standardized patient (SP) ratings, 93% also failed to demonstrate competency based on the faculty ratings. CONCLUSIONS: Our review form allowed clinical faculty to validate pass/fail decisions based on standardized patient ratings. Detailed information about performance problems contributes to learner feedback and curricular enhancement to guide remediation planning and faculty development.


Subject(s)
Clinical Competence/standards , Curriculum , Documentation , Feedback , Students, Medical , Checklist , Education, Medical, Undergraduate , Faculty, Medical , Humans , Michigan , Pilot Projects
18.
Atten Percept Psychophys ; 73(2): 640-57, 2011 Feb.
Article in English | MEDLINE | ID: mdl-21264716

ABSTRACT

Speed-accuracy trade-offs strongly influence the rate of reward that can be earned in many decision-making tasks. Previous reports suggest that human participants often adopt suboptimal speed-accuracy trade-offs in single session, two-alternative forced-choice tasks. We investigated whether humans acquired optimal speed-accuracy trade-offs when extensively trained with multiple signal qualities. When performance was characterized in terms of decision time and accuracy, our participants eventually performed nearly optimally in the case of higher signal qualities. Rather than adopting decision criteria that were individually optimal for each signal quality, participants adopted a single threshold that was nearly optimal for most signal qualities. However, setting a single threshold for different coherence conditions resulted in only negligible decrements in the maximum possible reward rate. Finally, we tested two hypotheses regarding the possible sources of suboptimal performance: (1) favoring accuracy over reward rate and (2) misestimating the reward rate due to timing uncertainty. Our findings provide support for both hypotheses, but also for the hypothesis that participants can learn to approach optimality. We find specifically that an accuracy bias dominates early performance, but diminishes greatly with practice. The residual discrepancy between optimal and observed performance can be explained by an adaptive response to uncertainty in time estimation.


Subject(s)
Attention , Decision Making , Discrimination, Psychological , Motion Perception , Motivation , Pattern Recognition, Visual , Reaction Time , Reward , Adolescent , Adult , Female , Humans , Male , Practice, Psychological , Probability Learning , Psychophysics , Uncertainty , Young Adult
20.
J Am Coll Surg ; 208(4): 517-9, 2009 Apr.
Article in English | MEDLINE | ID: mdl-19476784

ABSTRACT

BACKGROUND: Two uncommon but serious complications after subclavian central venous port (SCVP) placement are pneumothorax (PNX) and malposition of the catheter. Chest x-rays (CXR) are commonly obtained after SCVP placement to identify these complications, but their use is controversial. STUDY DESIGN: We performed a retrospective review of SCVP placements to establish the incidence of PNX or catheter malposition identified exclusively by postprocedure CXR. RESULTS: Between July 1, 2001, and June 30, 2006, 205 patients underwent elective SCVP placement. Although 4 patients (2%) sustained a PNX, none was identified by routine postprocedure CXR. Postprocedure clinical symptoms (3 to 72 hours later) prompted repeat CXR, which identified the PNX. Five patients (2.4%) had catheter malposition recognized by intraoperative fluoroscopy and corrected intraoperatively. No malpositioned catheters were identified on postprocedure CXR. CONCLUSIONS: In our study, incidence of PNX after SCVP placement was low, and PNX was not detected by intraoperative fluoroscopy or by routine postprocedure CXR. We conclude that the practice of routine postprocedure CXR after SCVP placement is not necessary and should be replaced with diagnostic chest radiography only if symptoms develop.


Subject(s)
Catheterization, Central Venous , Diagnostic Tests, Routine/statistics & numerical data , Radiography, Thoracic/statistics & numerical data , Unnecessary Procedures/statistics & numerical data , Catheterization, Central Venous/adverse effects , Equipment Failure , Fluoroscopy , Humans , Intraoperative Period , Pneumothorax/diagnostic imaging , Pneumothorax/etiology , Retrospective Studies
SELECTION OF CITATIONS
SEARCH DETAIL
...