Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 15 de 15
Filter
Add more filters










Publication year range
1.
Neural Netw ; 165: 654-661, 2023 Aug.
Article in English | MEDLINE | ID: mdl-37364474

ABSTRACT

We employ properties of high-dimensional geometry to obtain some insights into capabilities of deep perceptron networks to classify large data sets. We derive conditions on network depths, types of activation functions, and numbers of parameters that imply that approximation errors behave almost deterministically. We illustrate general results by concrete cases of popular activation functions: Heaviside, ramp sigmoid, rectified linear, and rectified power. Our probabilistic bounds on approximation errors are derived using concentration of measure type inequalities (method of bounded differences) and concepts from statistical learning theory.


Subject(s)
Neural Networks, Computer
2.
IEEE Trans Neural Netw Learn Syst ; 32(11): 5072-5081, 2021 Nov.
Article in English | MEDLINE | ID: mdl-33044935

ABSTRACT

Suitability of shallow (one-hidden-layer) networks with translation-invariant kernel units for function approximation and classification tasks is investigated. It is shown that a critical property influencing the capabilities of kernel networks is how the Fourier transforms of kernels converge to zero. The Fourier transforms of kernels suitable for multivariable approximation can have negative values but must be almost everywhere nonzero. In contrast, the Fourier transforms of kernels suitable for maximal margin classification must be everywhere nonnegative but can have large sets where they are equal to zero (e.g., they can be compactly supported). The behavior of the Fourier transforms of multivariable kernels is analyzed using the Hankel transform. The general results are illustrated by examples of both univariable and multivariable kernels (such as Gaussian, Laplace, rectangle, sinc, and cut power kernels).

3.
Cas Lek Cesk ; 159(6): 217-225, 2020.
Article in English | MEDLINE | ID: mdl-33297697

ABSTRACT

Seroprevalence studies represent a very important tool to find out what fraction of population has already met with the new type of coronavirus (e.g. SARS-CoV-2). Without these data, it is almost impossible for the state authorities to manage the epidemic and adopt rational measures. This article brings the results of a medium-sized seroprevalence study which was carried out in the spring of 2020 in South Bohemia. In the Strakonice and Písek regions, the ELISA method was used to test the prevalence of IgA and IgG antibodies in 2011 subjects, volunteers from general public and selected professions working in areas with a higher exposure to the infection. The study showed that already in May 2020, 2.9% of inhabitants of the Strakonice region and 1.9% of inhabitants of the Písek region had antibodies against the coronavirus. These numbers imply that for each PCR positive person, there were at least fifty others who had probably already undergone the infection. The article points out three types of problems that might occur in such a study. First, the study must be planned correctly, and possible outcomes must be pre-assessed. Second, an appropriate test must be selected with known parameters. This enables us to correctly estimate the share of false positive and false negative results. Third, the data must be evaluated in a reasonable way and correct inference must be performed. We offer a set of recommendations how to manage these issues and how to solve problems that inevitably arise in such a large-scale testing.


Subject(s)
COVID-19 , Czech Republic , Seroepidemiologic Studies , COVID-19/diagnosis , COVID-19/epidemiology , Czech Republic/epidemiology , Enzyme-Linked Immunosorbent Assay , Humans , Pandemics , SARS-CoV-2
5.
IEEE Trans Neural Netw Learn Syst ; 30(9): 2746-2754, 2019 09.
Article in English | MEDLINE | ID: mdl-30640635

ABSTRACT

The choice of dictionaries of computational units suitable for efficient computation of binary classification tasks is investigated. To deal with exponentially growing sets of tasks with increasingly large domains, a probabilistic model is introduced. The relevance of tasks for a given application area is modeled by a product probability distribution on the set of all binary-valued functions. Approximate measures of network sparsity are studied in terms of variational norms tailored to dictionaries of computational units. Bounds on these norms are proven using the Chernoff-Hoeffding bound on sums of independent random variables that need not be identically distributed. Consequences of the probabilistic results for the choice of dictionaries of computational units are derived. It is shown that when a priori knowledge of a type of classification tasks is limited, then the sparsity may be achieved only at the expense of large sizes of dictionaries.

6.
Neural Netw ; 91: 34-41, 2017 Jul.
Article in English | MEDLINE | ID: mdl-28482227

ABSTRACT

Limitations of approximation capabilities of shallow perceptron networks are investigated. Lower bounds on approximation errors are derived for binary-valued functions on finite domains. It is proven that unless the number of network units is sufficiently large (larger than any polynomial of the logarithm of the size of the domain) a good approximation cannot be achieved for almost any uniformly randomly chosen function on a given domain. The results are obtained by combining probabilistic Chernoff-Hoeffding bounds with estimates of the sizes of sets of functions exactly computable by shallow networks with increasing numbers of units.


Subject(s)
Neural Networks, Computer
7.
Neural Netw ; 57: 23-8, 2014 Sep.
Article in English | MEDLINE | ID: mdl-24892273

ABSTRACT

The role of width of Gaussians in two types of computational models is investigated: Gaussian radial-basis-functions (RBFs) where both widths and centers vary and Gaussian kernel networks which have fixed widths but varying centers. The effect of width on functional equivalence, universal approximation property, and form of norms in reproducing kernel Hilbert spaces (RKHS) is explored. It is proven that if two Gaussian RBF networks have the same input-output functions, then they must have the same numbers of units with the same centers and widths. Further, it is shown that while sets of input-output functions of Gaussian kernel networks with two different widths are disjoint, each such set is large enough to be a universal approximator. Embedding of RKHSs induced by "flatter" Gaussians into RKHSs induced by "sharper" Gaussians is described and growth of the ratios of norms on these spaces with increasing input dimension is estimated. Finally, large sets of argminima of error functionals in sets of input-output functions of Gaussian RBFs are described.


Subject(s)
Algorithms , Computer Simulation , Neural Networks, Computer , Normal Distribution
8.
Neural Netw ; 33: 160-7, 2012 Sep.
Article in English | MEDLINE | ID: mdl-22684117

ABSTRACT

Integral transforms with kernels corresponding to computational units are exploited to derive estimates of network complexity. The estimates are obtained by combining tools from nonlinear approximation theory and functional analysis together with representations of functions in the form of infinite neural networks. The results are applied to perceptron networks.


Subject(s)
Computational Biology/methods , Neural Networks, Computer , Nonlinear Dynamics
9.
Neural Netw ; 24(8): 881-7, 2011 Oct.
Article in English | MEDLINE | ID: mdl-21704495

ABSTRACT

Approximation capabilities of two types of computational models are explored: dictionary-based models (i.e., linear combinations of n-tuples of basis functions computable by units belonging to a set called "dictionary") and linear ones (i.e., linear combinations of n fixed basis functions). The two models are compared in terms of approximation rates, i.e., speeds of decrease of approximation errors for a growing number n of basis functions. Proofs of upper bounds on approximation rates by dictionary-based models are inspected, to show that for individual functions they do not imply estimates for dictionary-based models that do not hold also for some linear models. Instead, the possibility of getting faster approximation rates by dictionary-based models is demonstrated for worst-case errors in approximation of suitable sets of functions. For such sets, even geometric upper bounds hold.


Subject(s)
Computer Simulation , Linear Models , Algorithms , Neural Networks, Computer , Reproducibility of Results
10.
Neural Netw ; 24(2): 171-82, 2011 Mar.
Article in English | MEDLINE | ID: mdl-21094023

ABSTRACT

Neural networks provide a more flexible approximation of functions than traditional linear regression. In the latter, one can only adjust the coefficients in linear combinations of fixed sets of functions, such as orthogonal polynomials or Hermite functions, while for neural networks, one may also adjust the parameters of the functions which are being combined. However, some useful properties of linear approximators (such as uniqueness, homogeneity, and continuity of best approximation operators) are not satisfied by neural networks. Moreover, optimization of parameters in neural networks becomes more difficult than in linear regression. Experimental results suggest that these drawbacks of neural networks are offset by substantially lower model complexity, allowing accuracy of approximation even in high-dimensional cases. We give some theoretical results comparing requirements on model complexity for two types of approximators, the traditional linear ones and so called variable-basis types, which include neural networks, radial, and kernel models. We compare upper bounds on worst-case errors in variable-basis approximation with lower bounds on such errors for any linear approximator. Using methods from nonlinear approximation and integral representations tailored to computational units, we describe some cases where neural networks outperform any linear approximator.


Subject(s)
Dictionaries as Topic , Linear Models , Neural Networks, Computer , Computational Biology , Models, Neurological , Statistics, Nonparametric
12.
Neural Comput ; 21(10): 2970-89, 2009 Oct.
Article in English | MEDLINE | ID: mdl-19635020

ABSTRACT

Complexity of one-hidden-layer networks is studied using tools from nonlinear approximation and integration theory. For functions with suitable integral representations in the form of networks with infinitely many hidden units, upper bounds are derived on the speed of decrease of approximation error as the number of network units increases. These bounds are obtained for various norms using the framework of Bochner integration. Results are applied to perceptron networks.


Subject(s)
Computer Simulation , Neural Networks, Computer , Action Potentials/physiology , Algorithms , Brain/physiology , Nerve Net/physiology , Neurons/physiology
13.
Neural Comput ; 20(1): 252-70, 2008 Jan.
Article in English | MEDLINE | ID: mdl-18045008

ABSTRACT

Supervised learning of perceptron networks is investigated as an optimization problem. It is shown that both the theoretical and the empirical error functionals achieve minima over sets of functions computable by networks with a given number n of perceptrons. Upper bounds on rates of convergence of these minima with n increasing are derived. The bounds depend on a certain regularity of training data expressed in terms of variational norms of functions interpolating the data (in the case of the empirical error) and the regression function (in the case of the expected error). Dependence of this type of regularity on dimensionality and on magnitudes of partial derivatives is investigated. Conditions on the data, which guarantee that a good approximation of global minima of error functionals can be achieved using networks with a limited complexity, are derived. The conditions are in terms of oscillatory behavior of the data measured by the product of a function of the number of variables d, which is decreasing exponentially fast, and the maximum of the magnitudes of the squares of the L(1)-norms of the iterated partial derivatives of the order d of the regression function or some function, which interpolates the sample of the data. The results are illustrated by examples of data with small and high regularity constructed using Boolean functions and the gaussian function.


Subject(s)
Artificial Intelligence , Neural Networks, Computer , Algorithms , Biological Clocks/physiology , Computer Simulation , Logic , Normal Distribution
14.
Neural Netw ; 10(6): 1061-1068, 1997 Aug.
Article in English | MEDLINE | ID: mdl-12662499

ABSTRACT

We estimate variation with respect to half-spaces in terms of "flows through hyperplanes". Our estimate is derived from an integral representation for smooth compactly supported multivariable functions proved using properties of the Heaviside and delta distributions. Consequently we obtain conditions which guarantee approximation error rate of order O by one-hidden-layer networks with n sigmoidal perceptrons. Copyright 1997 Elsevier Science Ltd.

15.
Neural Comput ; 3(4): 617-622, 1991.
Article in English | MEDLINE | ID: mdl-31167327

ABSTRACT

We show that Kolmogorov's theorem on representations of continuous functions of n-variables by sums and superpositions of continuous functions of one variable is relevant in the context of neural networks. We give a version of this theorem with all of the one-variable functions approximated arbitrarily well by linear combinations of compositions of affine functions with some given sigmoidal function. We derive an upper estimate of the number of hidden units.

SELECTION OF CITATIONS
SEARCH DETAIL
...