Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 13 de 13
Filter
Add more filters










Publication year range
1.
Entropy (Basel) ; 26(3)2024 Mar 12.
Article in English | MEDLINE | ID: mdl-38539763

ABSTRACT

Deep neural networks excel in supervised learning tasks but are constrained by the need for extensive labeled data. Self-supervised learning emerges as a promising alternative, allowing models to learn without explicit labels. Information theory has shaped deep neural networks, particularly the information bottleneck principle. This principle optimizes the trade-off between compression and preserving relevant information, providing a foundation for efficient network design in supervised contexts. However, its precise role and adaptation in self-supervised learning remain unclear. In this work, we scrutinize various self-supervised learning approaches from an information-theoretic perspective, introducing a unified framework that encapsulates the self-supervised information-theoretic learning problem. This framework includes multiple encoders and decoders, suggesting that all existing work on self-supervised learning can be seen as specific instances. We aim to unify these approaches to understand their underlying principles better and address the main challenge: many works present different frameworks with differing theories that may seem contradictory. By weaving existing research into a cohesive narrative, we delve into contemporary self-supervised methodologies, spotlight potential research areas, and highlight inherent challenges. Moreover, we discuss how to estimate information-theoretic quantities and their associated empirical problems. Overall, this paper provides a comprehensive review of the intersection of information theory, self-supervised learning, and deep neural networks, aiming for a better understanding through our proposed unified approach.

2.
Nat Commun ; 14(1): 1597, 2023 03 22.
Article in English | MEDLINE | ID: mdl-36949048

ABSTRACT

Neuroscience has long been an essential driver of progress in artificial intelligence (AI). We propose that to accelerate progress in AI, we must invest in fundamental research in NeuroAI. A core component of this is the embodied Turing test, which challenges AI animal models to interact with the sensorimotor world at skill levels akin to their living counterparts. The embodied Turing test shifts the focus from those capabilities like game playing and language that are especially well-developed or uniquely human to those capabilities - inherited from over 500 million years of evolution - that are shared with all animals. Building models that can pass the embodied Turing test will provide a roadmap for the next generation of AI.


Subject(s)
Artificial Intelligence , Neurosciences , Animals , Humans
3.
Neural Netw ; 152: 267-275, 2022 Aug.
Article in English | MEDLINE | ID: mdl-35569196

ABSTRACT

Deep learning (DL) and reinforcement learning (RL) methods seem to be a part of indispensable factors to achieve human-level or super-human AI systems. On the other hand, both DL and RL have strong connections with our brain functions and with neuroscientific findings. In this review, we summarize talks and discussions in the "Deep Learning and Reinforcement Learning" session of the symposium, International Symposium on Artificial Intelligence and Brain Science. In this session, we discussed whether we can achieve comprehensive understanding of human intelligence based on the recent advances of deep learning and reinforcement learning algorithms. Speakers contributed to provide talks about their recent studies that can be key technologies to achieve human-level intelligence.


Subject(s)
Artificial Intelligence , Deep Learning , Algorithms , Humans , Reinforcement, Psychology
4.
IEEE Trans Image Process ; 30: 4036-4045, 2021.
Article in English | MEDLINE | ID: mdl-33735083

ABSTRACT

The task of image generation started receiving some attention from artists and designers, providing inspiration for new creations. However, exploiting the results of deep generative models such as Generative Adversarial Networks can be long and tedious given the lack of existing tools. In this work, we propose a simple strategy to inspire creators with new generations learned from a dataset of their choice, while providing some control over the output. We design a simple optimization method to find the optimal latent parameters corresponding to the closest generation to any input inspirational image. Specifically, we allow the generation given an inspirational image of the user's choosing by performing several optimization steps to recover optimal parameters from the model's latent space. We tested several exploration methods from classical gradient descents to gradient-free optimizers. Many gradient-free optimizers just need comparisons (better/worse than another image), so they can even be used without numerical criterion nor inspirational image, only with human preferences. Thus, by iterating on one's preferences we can make robust facial composite or fashion generation algorithms. Our results on four datasets of faces, fashion images, and textures show that satisfactory images are effectively retrieved in most cases.

5.
Cell ; 182(6): 1372-1376, 2020 09 17.
Article in English | MEDLINE | ID: mdl-32946777

ABSTRACT

Large scientific projects in genomics and astronomy are influential not because they answer any single question but because they enable investigation of continuously arising new questions from the same data-rich sources. Advances in automated mapping of the brain's synaptic connections (connectomics) suggest that the complicated circuits underlying brain function are ripe for analysis. We discuss benefits of mapping a mouse brain at the level of synapses.


Subject(s)
Brain/physiology , Connectome/methods , Nerve Net/physiology , Neurons/physiology , Synapses/physiology , Animals , Mice
6.
PLoS One ; 14(12): e0226222, 2019.
Article in English | MEDLINE | ID: mdl-31856228

ABSTRACT

Failing to distinguish between a sheepdog and a skyscraper should be worse and penalized more than failing to distinguish between a sheepdog and a poodle; after all, sheepdogs and poodles are both breeds of dogs. However, existing metrics of failure (so-called "loss" or "win") used in textual or visual classification/recognition via neural networks seldom leverage a-priori information, such as a sheepdog being more similar to a poodle than to a skyscraper. We define a metric that, inter alia, can penalize failure to distinguish between a sheepdog and a skyscraper more than failure to distinguish between a sheepdog and a poodle. Unlike previously employed possibilities, this metric is based on an ultrametric tree associated with any given tree organization into a semantically meaningful hierarchy of a classifier's classes. An ultrametric tree is a tree with a so-called ultrametric distance metric such that all leaves are at the same distance from the root. Unfortunately, extensive numerical experiments indicate that the standard practice of training neural networks via stochastic gradient descent with random starting points often drives down the hierarchical loss nearly as much when minimizing the standard cross-entropy loss as when trying to minimize the hierarchical loss directly. Thus, this hierarchical loss is unreliable as an objective for plain, randomly started stochastic gradient descent to minimize; the main value of the hierarchical loss may be merely as a meaningful metric of success of a classifier.


Subject(s)
Neural Networks, Computer
7.
Neural Comput ; 28(5): 815-25, 2016 05.
Article in English | MEDLINE | ID: mdl-26890348

ABSTRACT

A complex-valued convolutional network (convnet) implements the repeated application of the following composition of three operations, recursively applying the composition to an input vector of nonnegative real numbers: (1) convolution with complex-valued vectors, followed by (2) taking the absolute value of every entry of the resulting vectors, followed by (3) local averaging. For processing real-valued random vectors, complex-valued convnets can be viewed as data-driven multiscale windowed power spectra, data-driven multiscale windowed absolute spectra, data-driven multiwavelet absolute values, or (in their most general configuration) data-driven nonlinear multiwavelet packets. Indeed, complex-valued convnets can calculate multiscale windowed spectra when the convnet filters are windowed complex-valued exponentials. Standard real-valued convnets, using rectified linear units (ReLUs), sigmoidal (e.g., logistic or tanh) nonlinearities, or max pooling, for example, do not obviously exhibit the same exact correspondence with data-driven wavelets (whereas for complex-valued convnets, the correspondence is much more than just a vague analogy). Courtesy of the exact correspondence, the remarkably rich and rigorous body of mathematical analysis for wavelets applies directly to (complex-valued) convnets.

8.
Nature ; 521(7553): 436-44, 2015 May 28.
Article in English | MEDLINE | ID: mdl-26017442

ABSTRACT

Deep learning allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. These methods have dramatically improved the state-of-the-art in speech recognition, visual object recognition, object detection and many other domains such as drug discovery and genomics. Deep learning discovers intricate structure in large data sets by using the backpropagation algorithm to indicate how a machine should change its internal parameters that are used to compute the representation in each layer from the representation in the previous layer. Deep convolutional nets have brought about breakthroughs in processing images, video, speech and audio, whereas recurrent nets have shone light on sequential data such as text and speech.


Subject(s)
Artificial Intelligence , Algorithms , Artificial Intelligence/trends , Computers , Language , Neural Networks, Computer
9.
IEEE Trans Pattern Anal Mach Intell ; 35(8): 1915-29, 2013 Aug.
Article in English | MEDLINE | ID: mdl-23787344

ABSTRACT

Scene labeling consists of labeling each pixel in an image with the category of the object it belongs to. We propose a method that uses a multiscale convolutional network trained from raw pixels to extract dense feature vectors that encode regions of multiple sizes centered on each pixel. The method alleviates the need for engineered features, and produces a powerful representation that captures texture, shape, and contextual information. We report results using multiple postprocessing methods to produce the final labeling. Among those, we propose a technique to automatically retrieve, from a pool of segmentation components, an optimal set of components that best explain the scene; these components are arbitrary, for example, they can be taken from a segmentation tree or from any family of oversegmentations. The system yields record accuracies on the SIFT Flow dataset (33 classes) and the Barcelona dataset (170 classes) and near-record accuracy on Stanford background dataset (eight classes), while being an order of magnitude faster than competing approaches, producing a $(320\times 240)$ image labeling in less than a second, including feature extraction.

10.
Front Neurosci ; 6: 32, 2012.
Article in English | MEDLINE | ID: mdl-22518097

ABSTRACT

Most scene segmentation and categorization architectures for the extraction of features in images and patches make exhaustive use of 2D convolution operations for template matching, template search, and denoising. Convolutional Neural Networks (ConvNets) are one example of such architectures that can implement general-purpose bio-inspired vision systems. In standard digital computers 2D convolutions are usually expensive in terms of resource consumption and impose severe limitations for efficient real-time applications. Nevertheless, neuro-cortex inspired solutions, like dedicated Frame-Based or Frame-Free Spiking ConvNet Convolution Processors, are advancing real-time visual processing. These two approaches share the neural inspiration, but each of them solves the problem in different ways. Frame-Based ConvNets process frame by frame video information in a very robust and fast way that requires to use and share the available hardware resources (such as: multipliers, adders). Hardware resources are fixed- and time-multiplexed by fetching data in and out. Thus memory bandwidth and size is important for good performance. On the other hand, spike-based convolution processors are a frame-free alternative that is able to perform convolution of a spike-based source of visual information with very low latency, which makes ideal for very high-speed applications. However, hardware resources need to be available all the time and cannot be time-multiplexed. Thus, hardware should be modular, reconfigurable, and expansible. Hardware implementations in both VLSI custom integrated circuits (digital and analog) and FPGA have been already used to demonstrate the performance of these systems. In this paper we present a comparison study of these two neuro-inspired solutions. A brief description of both systems is presented and also discussions about their differences, pros and cons.

11.
Genome Biol ; 11(12): R123, 2010.
Article in English | MEDLINE | ID: mdl-21182762

ABSTRACT

BACKGROUND: Nitrate, acting as both a nitrogen source and a signaling molecule, controls many aspects of plant development. However, gene networks involved in plant adaptation to fluctuating nitrate environments have not yet been identified. RESULTS: Here we use time-series transcriptome data to decipher gene relationships and consequently to build core regulatory networks involved in Arabidopsis root adaptation to nitrate provision. The experimental approach has been to monitor genome-wide responses to nitrate at 3, 6, 9, 12, 15 and 20 minutes using Affymetrix ATH1 gene chips. This high-resolution time course analysis demonstrated that the previously known primary nitrate response is actually preceded by a very fast gene expression modulation, involving genes and functions needed to prepare plants to use or reduce nitrate. A state-space model inferred from this microarray time-series data successfully predicts gene behavior in unlearnt conditions. CONCLUSIONS: The experiments and methods allow us to propose a temporal working model for nitrate-driven gene networks. This network model is tested both in silico and experimentally. For example, the over-expression of a predicted gene hub encoding a transcription factor induced early in the cascade indeed leads to the modification of the kinetic nitrate response of sentinel genes such as NIR, NIA2, and NRT1.1, and several other transcription factors. The potential nitrate/hormone connections implicated by this time-series data are also evaluated.


Subject(s)
Arabidopsis/genetics , Arabidopsis/metabolism , Gene Expression Profiling , Nitrates/metabolism , Adaptation, Physiological , Cluster Analysis , Gene Expression Regulation, Plant , Gene Regulatory Networks , Genes, Plant , Models, Genetic , Nitrogen/metabolism , Oligonucleotide Array Sequence Analysis , Plant Roots/genetics , Plant Roots/metabolism , RNA, Plant/genetics , Systems Biology , Transcription Factors/metabolism
12.
Clin Neurophysiol ; 120(11): 1927-1940, 2009 Nov.
Article in English | MEDLINE | ID: mdl-19837629

ABSTRACT

OBJECTIVE: Research in seizure prediction from intracranial EEG has highlighted the usefulness of bivariate measures of brainwave synchronization. Spatio-temporal bivariate features are very high-dimensional and cannot be analyzed with conventional statistical methods. Hence, we propose state-of-the-art machine learning methods that handle high-dimensional inputs. METHODS: We computed bivariate features of EEG synchronization (cross-correlation, nonlinear interdependence, dynamical entrainment or wavelet synchrony) on the 21-patient Freiburg dataset. Features from all channel pairs and frequencies were aggregated over consecutive time points, to form patterns. Patient-specific machine learning-based classifiers (support vector machines, logistic regression or convolutional neural networks) were trained to discriminate interictal from preictal patterns of features. In this explorative study, we evaluated out-of-sample seizure prediction performance, and compared each combination of feature type and classifier. RESULTS: Among the evaluated methods, convolutional networks combined with wavelet coherence successfully predicted all out-of-sample seizures, without false alarms, on 15 patients, yielding 71% sensitivity and 0 false positives. CONCLUSIONS: Our best machine learning technique applied to spatio-temporal patterns of EEG synchronization outperformed previous seizure prediction methods on the Freiburg dataset. SIGNIFICANCE: By learning spatio-temporal dynamics of EEG synchronization, pattern recognition could capture patient-specific seizure precursors. Further investigation on additional datasets should include the seizure prediction horizon.


Subject(s)
Electroencephalography/classification , Neural Networks, Computer , Seizures/classification , Seizures/diagnosis , Humans , Predictive Value of Tests , Seizures/physiopathology
13.
IEEE Trans Image Process ; 14(9): 1360-71, 2005 Sep.
Article in English | MEDLINE | ID: mdl-16190471

ABSTRACT

We describe a trainable system for analyzing videos of developing C. elegans embryos. The system automatically detects, segments, and locates cells and nuclei in microscopic images. The system was designed as the central component of a fully automated phenotyping system. The system contains three modules 1) a convolutional network trained to classify each pixel into five categories: cell wall, cytoplasm, nucleus membrane, nucleus, outside medium; 2) an energy-based model, which cleans up the output of the convolutional network by learning local consistency constraints that must be satisfied by label images; 3) a set of elastic models of the embryo at various stages of development that are matched to the label images.


Subject(s)
Artificial Intelligence , Caenorhabditis elegans/anatomy & histology , Caenorhabditis elegans/embryology , Image Interpretation, Computer-Assisted/methods , Microscopy, Phase-Contrast/methods , Microscopy, Video/methods , Pattern Recognition, Automated/methods , Phenotype , Algorithms , Animals , Caenorhabditis elegans/classification , Caenorhabditis elegans/growth & development , Embryo, Nonmammalian/cytology , Fetal Development/physiology , Image Enhancement/methods , Reproducibility of Results , Sensitivity and Specificity
SELECTION OF CITATIONS
SEARCH DETAIL
...