Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 15 de 15
Filter
1.
PLoS One ; 14(9): e0223319, 2019.
Article in English | MEDLINE | ID: mdl-31568495

ABSTRACT

To identify pathways between stress indicators and adverse pregnancy outcomes, we applied a nonparametric graph-learning algorithm, PC-KCI, to data from an observational prospective cohort study. The Measurement of Maternal Stress study (MOMS) followed 744 women with a singleton intrauterine pregnancy recruited between June 2013 and May 2015. Infant adverse pregnancy outcomes were prematurity (<37 weeks' gestation), infant days spent in hospital after birth, and being small for gestational age (percentile gestational weight at birth). Maternal adverse pregnancy outcomes were pre-eclampsia, gestational diabetes, and gestational hypertension. PC-KCI replicated well-established pathways, such as the relationship between gestational weeks and preterm premature rupture of membranes. PC-KCI also identified previously unobserved pathways to adverse pregnancy outcomes, including 1) a link between hair cortisol levels (at 12-21 weeks of pregnancy) and pre-eclampsia; 2) two pathways to preterm birth depending on race, with one linking Hispanic race, pre-gestational diabetes and gestational weeks, and a second pathway linking black race, hair cortisol, preeclampsia, and gestational weeks; and 3) a relationship between maternal childhood trauma, perceived social stress in adulthood, and low weight for gestational age. Our approach confirmed previous findings and identified previously unobserved pathways to adverse pregnancy outcomes. It presents a method for a global assessment of a clinical problem for further study of possible causal pathways.


Subject(s)
Abortion, Spontaneous/epidemiology , Algorithms , Diabetes, Gestational/epidemiology , Hypertension, Pregnancy-Induced/epidemiology , Pre-Eclampsia/epidemiology , Stress, Psychological/epidemiology , Abortion, Spontaneous/diagnosis , Abortion, Spontaneous/metabolism , Adult , Biomarkers/metabolism , Delivery, Obstetric , Diabetes, Gestational/diagnosis , Diabetes, Gestational/metabolism , Female , Gestational Age , Hair/chemistry , Hair/metabolism , Humans , Hydrocortisone/metabolism , Hypertension, Pregnancy-Induced/diagnosis , Hypertension, Pregnancy-Induced/metabolism , Infant, Low Birth Weight , Infant, Newborn , Infant, Premature , Live Birth , Pre-Eclampsia/diagnosis , Pre-Eclampsia/metabolism , Pregnancy , Prospective Studies , Statistics, Nonparametric , Stillbirth , Stress, Psychological/diagnosis , Stress, Psychological/metabolism , United States/epidemiology
2.
Ann Appl Stat ; 9(4): 1997-2022, 2015 Dec.
Article in English | MEDLINE | ID: mdl-34326914

ABSTRACT

Functional neuroimaging measures how the brain responds to complex stimuli. However, sample sizes are modest, noise is substantial, and stimuli are high dimensional. Hence, direct estimates are inherently imprecise and call for regularization. We compare a suite of approaches which regularize via shrinkage: ridge regression, the elastic net (a generalization of ridge regression and the lasso), and a hierarchical Bayesian model based on small area estimation (SAE). We contrast regularization with spatial smoothing and combinations of smoothing and shrinkage. All methods are tested on functional magnetic resonance imaging (fMRI) data from multiple subjects participating in two different experiments related to reading, for both predicting neural response to stimuli and decoding stimuli from responses. Interestingly, when the regularization parameters are chosen by cross-validation independently for every voxel, low/high regularization is chosen in voxels where the classification accuracy is high/low, indicating that the regularization intensity is a good tool for identification of relevant voxels for the cognitive task. Surprisingly, all the regularization methods work about equally well, suggesting that beating basic smoothing and shrinkage will take not only clever methods, but also careful modeling.

3.
J Stat Mech ; 2014(5)2014 May.
Article in English | MEDLINE | ID: mdl-26167197

ABSTRACT

The proliferation of models for networks raises challenging problems of model selection: the data are sparse and globally dependent, and models are typically high-dimensional and have large numbers of latent variables. Together, these issues mean that the usual model-selection criteria do not work properly for networks. We illustrate these challenges, and show one way to resolve them, by considering the key network-analysis problem of dividing a graph into communities or blocks of nodes with homogeneous patterns of links to the rest of the network. The standard tool for undertaking this is the stochastic block model, under which the probability of a link between two nodes is a function solely of the blocks to which they belong. This imposes a homogeneous degree distribution within each block; this can be unrealistic, so degree-corrected block models add a parameter for each node, modulating its overall degree. The choice between ordinary and degree-corrected block models matters because they make very different inferences about communities. We present the first principled and tractable approach to model selection between standard and degree-corrected block models, based on new large-graph asymptotics for the distribution of log-likelihood ratios under the stochastic block model, finding substantial departures from classical results for sparse graphs. We also develop linear-time approximations for log-likelihoods under both the stochastic block model and the degree-corrected model, using belief propagation. Applications to simulated and real networks show excellent agreement with our approximations. Our results thus both solve the practical problem of deciding on degree correction and point to a general approach to model selection in network analysis.

5.
Br J Math Stat Psychol ; 66(1): 8-38, 2013 Feb.
Article in English | MEDLINE | ID: mdl-22364575

ABSTRACT

A substantial school in the philosophy of science identifies Bayesian inference with inductive inference and even rationality as such, and seems to be strengthened by the rise and practical success of Bayesian statistics. We argue that the most successful forms of Bayesian statistics do not actually support that particular philosophy but rather accord much better with sophisticated forms of hypothetico-deductivism. We examine the actual role played by prior distributions in Bayesian models, and the crucial aspects of model checking and model revision, which fall outside the scope of Bayesian confirmation theory. We draw on the literature on the consistency of Bayesian updating and also on our experience of applied work in social science. Clarity about these matters should benefit not just philosophy of science, but also statistical practice. At best, the inductivist view has encouraged researchers to fit and compare models without checking them; at worst, theorists have actively discouraged practitioners from performing model checking because it does not fit into their framework.


Subject(s)
Bayes Theorem , Philosophy , Humans , Psychometrics/statistics & numerical data , Research/statistics & numerical data , Social Sciences/statistics & numerical data
6.
Article in English | MEDLINE | ID: mdl-26321855

ABSTRACT

We informally call a stochastic process learnable if it admits a generalization error approaching zero in probability for any concept class with finite VC-dimension (IID processes are the simplest example). A mixture of learnable processes need not be learnable itself, and certainly its generalization error need not decay at the same rate. In this paper, we argue that it is natural in predictive PAC to condition not on the past observations but on the mixture component of the sample path. This definition not only matches what a realistic learner might demand, but also allows us to sidestep several otherwise grave problems in learning from dependent data. In particular, we give a novel PAC generalization bound for mixtures of learnable processes with a generalization error that is not worse than that of each mixture component. We also provide a characterization of mixtures of absolutely regular (ß-mixing) processes, of independent probability-theoretic interest.

7.
Ann Stat ; 41(2): 508-535, 2013 Apr.
Article in English | MEDLINE | ID: mdl-26166910

ABSTRACT

The growing availability of network data and of scientific interest in distributed systems has led to the rapid development of statistical models of network structure. Typically, however, these are models for the entire network, while the data consists only of a sampled sub-network. Parameters for the whole network, which is what is of interest, are estimated by applying the model to the sub-network. This assumes that the model is consistent under sampling, or, in terms of the theory of stochastic processes, that it defines a projective family. Focusing on the popular class of exponential random graph models (ERGMs), we show that this apparently trivial condition is in fact violated by many popular and scientifically appealing models, and that satisfying it drastically limits ERGM's expressive power. These results are actually special cases of more general results about exponential families of dependent random variables, which we also prove. Using such results, we offer easily checked conditions for the consistency of maximum likelihood estimation in ERGMs, and discuss some possible constructive responses.

8.
JMLR Workshop Conf Proc ; 31: 289-297, 2013.
Article in English | MEDLINE | ID: mdl-26279743

ABSTRACT

We introduce mixed LICORS, an algorithm for learning nonlinear, high-dimensional dynamics from spatio-temporal data, suitable for both prediction and simulation. Mixed LICORS extends the recent LICORS algorithm (Goerg and Shalizi, 2012) from hard clustering of predictive distributions to a non-parametric, EM-like soft clustering. This retains the asymptotic predictive optimality of LICORS, but, as we show in simulations, greatly improves out-of-sample forecasts with limited data. The new method is implemented in the publicly-available R package LICORS.

9.
Stat Politics Policy ; 3(1): 5, 2012 Feb.
Article in English | MEDLINE | ID: mdl-26504738

ABSTRACT

VanderWeele et al.'s paper is a useful contribution to the on-going scientific conversation about the detection of contagion from purely observational data. It is especially helpful as a corrective to some of the more extreme statements of Lyons (2011). Unfortunately, this paper, too, goes too far in some places, and so needs some correction itself.

10.
JMLR Workshop Conf Proc ; 15: 516-524, 2011.
Article in English | MEDLINE | ID: mdl-26279742

ABSTRACT

The literature on statistical learning for time series assumes the asymptotic independence or "mixing" of the data-generating process. These mixing assumptions are never tested, and there are no methods for estimating mixing rates from data. We give an estimator for the beta-mixing rate based on a single stationary sample path and show it is L1-risk consistent.

11.
Sociol Methods Res ; 40(2): 211-239, 2011 May.
Article in English | MEDLINE | ID: mdl-22523436

ABSTRACT

The authors consider processes on social networks that can potentially involve three factors: homophily, or the formation of social ties due to matching individual traits; social contagion, also known as social influence; and the causal effect of an individual's covariates on his or her behavior or other measurable responses. The authors show that generically, all of these are confounded with each other. Distinguishing them from one another requires strong assumptions on the parametrization of the social process or on the adequacy of the covariates used (or both). In particular the authors demonstrate, with simple examples, that asymmetries in regression coefficients cannot identify causal effects and that very simple models of imitation (a form of social contagion) can produce substantial correlations between an individual's enduring traits and his or her choices, even when there is no intrinsic affinity between them. The authors also suggest some possible constructive responses to these results.

12.
J Am Stat Assoc ; 105(489): 170-180, 2010 03.
Article in English | MEDLINE | ID: mdl-21753862

ABSTRACT

State-space models provide an important body of techniques for analyzing time-series, but their use requires estimating unobserved states. The optimal estimate of the state is its conditional expectation given the observation histories, and computing this expectation is hard when there are nonlinearities. Existing filtering methods, including sequential Monte Carlo, tend to be either inaccurate or slow. In this paper, we study a nonlinear filter for nonlinear/non-Gaussian state-space models, which uses Laplace's method, an asymptotic series expansion, to approximate the state's conditional mean and variance, together with a Gaussian conditional distribution. This Laplace-Gaussian filter (LGF) gives fast, recursive, deterministic state estimates, with an error which is set by the stochastic characteristics of the model and is, we show, stable over time. We illustrate the estimation ability of the LGF by applying it to the problem of neural decoding and compare it to sequential Monte Carlo both in simulations and with real data. We find that the LGF can deliver superior results in a small fraction of the computing time.

13.
Neural Comput ; 22(1): 121-57, 2010 Jan.
Article in English | MEDLINE | ID: mdl-19764880

ABSTRACT

Neurons perform computations, and convey the results of those computations through the statistical structure of their output spike trains. Here we present a practical method, grounded in the information-theoretic analysis of prediction, for inferring a minimal representation of that structure and for characterizing its complexity. Starting from spike trains, our approach finds their causal state models (CSMs), the minimal hidden Markov models or stochastic automata capable of generating statistically identical time series. We then use these CSMs to objectively quantify both the generalizable structure and the idiosyncratic randomness of the spike train. Specifically, we show that the expected algorithmic information content (the information needed to describe the spike train exactly) can be split into three parts describing (1) the time-invariant structure (complexity) of the minimal spike-generating process, which describes the spike train statistically; (2) the randomness (internal entropy rate) of the minimal spike-generating process; and (3) a residual pure noise term not described by the minimal spike-generating process. We use CSMs to approximate each of these quantities. The CSMs are inferred nonparametrically from the data, making only mild regularity assumptions, via the causal state splitting reconstruction algorithm. The methods presented here complement more traditional spike train analyses by describing not only spiking probability and spike train entropy, but also the complexity of a spike train's structure. We demonstrate our approach using both simulated spike trains and experimental data recorded in rat barrel cortex during vibrissa stimulation.


Subject(s)
Action Potentials/physiology , Brain/physiology , Nerve Net/physiology , Neural Networks, Computer , Neural Pathways/physiology , Neurons/physiology , Algorithms , Animals , Computer Simulation , Entropy , Markov Chains , Normal Distribution , Rats , Sensory Receptor Cells/physiology , Signal Processing, Computer-Assisted , Somatosensory Cortex/physiology , Stochastic Processes , Synaptic Transmission/physiology
14.
Phys Rev E Stat Nonlin Soft Matter Phys ; 73(3 Pt 2): 036104, 2006 Mar.
Article in English | MEDLINE | ID: mdl-16605595

ABSTRACT

Most current methods for identifying coherent structures in spatially extended systems rely on prior information about the form which those structures take. Here we present two approaches to automatically filter the changing configurations of spatial dynamical systems and extract coherent structures. One, local sensitivity filtering, is a modification of the local Lyapunov exponent approach suitable to cellular automata and other discrete spatial systems. The other, local statistical complexity filtering, calculates the amount of information needed for optimal prediction of the system's behavior in the vicinity of a given point. By examining the changing spatiotemporal distributions of these quantities, we can find the coherent structures in a variety of pattern-forming cellular automata, without needing to guess or postulate the form of that structure. We apply both filters to elementary and cyclical cellular automata (ECA and CCA) and find that they readily identify particles, domains, and other more complicated structures. We compare the results from ECA with earlier ones based upon the theory of formal languages and the results from CCA with a more traditional approach based on an order parameter and free energy. While sensitivity and statistical complexity are equally adept at uncovering structure, they are based on different system properties (dynamical and probabilistic, respectively) and provide complementary information.


Subject(s)
Biophysics/methods , Computational Biology/methods , Automation , Models, Statistical , Programming Languages , Sensitivity and Specificity , Software , Thermodynamics , Time Factors
15.
Phys Rev Lett ; 93(11): 118701, 2004 Sep 10.
Article in English | MEDLINE | ID: mdl-15447385

ABSTRACT

Despite broad interest in self-organizing systems, there are few quantitative, experimentally applicable criteria for self-organization. The existing criteria all give counter-intuitive results for important cases. In this Letter, we propose a new criterion, namely, an internally generated increase in the statistical complexity, the amount of information required for optimal prediction of the system's dynamics. We precisely define this complexity for spatially extended dynamical systems, using the probabilistic ideas of mutual information and minimal sufficient statistics. This leads to a general method for predicting such systems and a simple algorithm for estimating statistical complexity. The results of applying this algorithm to a class of models of excitable media (cyclic cellular automata) strongly support our proposal.


Subject(s)
Algorithms , Cell Communication/physiology , Models, Biological , Models, Statistical , Nonlinear Dynamics , Population Dynamics , Computer Simulation
SELECTION OF CITATIONS
SEARCH DETAIL
...