Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 13 de 13
Filtrar
Más filtros











Base de datos
Intervalo de año de publicación
1.
J Clin Med ; 13(15)2024 Jul 23.
Artículo en Inglés | MEDLINE | ID: mdl-39124559

RESUMEN

Background/Objectives: This retrospective case series analyzed visual outcomes in patients with a prior history of implantable collamer lens (ICL) implantation who underwent cataract extraction (CE). A secondary aim was to investigate the relationship between vault height and the rate of cataract development. Methods: Visual acuity and refraction measurements were collected after CE at one week, one month and six months. Vault height measurements were correlated to the time until symptomatic cataracts were removed. Results: A total of 44 eyes were analyzed at six months after CE with efficacy and safety indexes of 1.20 ± 1.11 and 1.50 ± 1.06, respectively. In addition, 70% of eyes had a post-operative uncorrected distance visual acuity (UDVA) within one line of pre-operative corrected distance visual acuity (CDVA). Refractive predictability at six months demonstrated that 43% and 69% of eyes were within ±0.25 D and ±0.50 D of SEQ target, respectively. Astigmatism measured by refractive cylinder was ≤0.25 D in 17% and ≤0.50 D in 34% of eyes pre-operatively compared to 40% and 60% of eyes, respectively, at six months post-operatively. Vault heights one week after ICL (p < 0.0081) and one week before CE (p < 0.0154) demonstrated a positive linear regression with the time until CE. Conclusions: This sample population achieved favorable visual outcomes six months after CE, similar to six months after ICL implantation. Patients with a history of ICL implantation will similarly have a good visual prognosis after CE.

2.
Phys Med Biol ; 68(8)2023 04 03.
Artículo en Inglés | MEDLINE | ID: mdl-36808921

RESUMEN

Objective. To investigate quantitative imaging markers based on parameters from two diffusion-weighted imaging (DWI) models, continuous-time random-walk (CTRW) and intravoxel incoherent motion (IVIM) models, for characterizing malignant and benign breast lesions by using a machine learning algorithm.Approach. With IRB approval, 40 women with histologically confirmed breast lesions (16 benign, 24 malignant) underwent DWI with 11b-values (50 to 3000 s/mm2) at 3T. Three CTRW parameters,Dm,α, andßand three IVIM parametersDdiff,Dperf, andfwere estimated from the lesions. A histogram was generated and histogram features of skewness, variance, mean, median, interquartile range; and the value of the 10%, 25% and 75% quantiles were extracted for each parameter from the regions-of-interest. Iterative feature selection was performed using the Boruta algorithm that uses the Benjamin Hochberg False Discover Rate to first determine significant features and then to apply the Bonferroni correction to further control for false positives across multiple comparisons during the iterative procedure. Predictive performance of the significant features was evaluated using Support Vector Machine, Random Forest, Naïve Bayes, Gradient Boosted Classifier (GB), Decision Trees, AdaBoost and Gaussian Process machine learning classifiers.Main Results. The 75% quantile, and median ofDm; 75% quantile off;mean, median, and skewness ofß;kurtosis ofDperf; and 75% quantile ofDdiffwere the most significant features. The GB differentiated malignant and benign lesions with an accuracy of 0.833, an area-under-the-curve of 0.942, and an F1 score of 0.87 providing the best statistical performance (p-value < 0.05) compared to the other classifiers.Significance. Our study has demonstrated that GB with a set of histogram features from the CTRW and IVIM model parameters can effectively differentiate malignant and benign breast lesions.


Asunto(s)
Neoplasias de la Mama , Mama , Femenino , Humanos , Teorema de Bayes , Mama/diagnóstico por imagen , Mama/patología , Neoplasias de la Mama/diagnóstico por imagen , Neoplasias de la Mama/patología , Imagen de Difusión por Resonancia Magnética/métodos , Aprendizaje Automático , Movimiento (Física) , Reproducibilidad de los Resultados
3.
Biometrics ; 79(4): 3332-3344, 2023 12.
Artículo en Inglés | MEDLINE | ID: mdl-36807124

RESUMEN

We consider inference problems for high-dimensional (HD) functional data with a dense number of T repeated measurements taken for a large number of p variables from a small number of n experimental units. The spatial and temporal dependence, high dimensionality, and dense number of repeated measurements pose theoretical and computational challenges. This paper has two aims; our first aim is to solve the theoretical and computational challenges in testing equivalence among covariance matrices from HD functional data. The second aim is to provide computationally efficient and tuning-free tools with guaranteed stochastic error control. The weak convergence of the stochastic process formed by the test statistics is established under the "large p, large T, and small n" setting. If the null is rejected, we further show that the locations of the change points can be estimated consistently. The estimator's rate of convergence is shown to depend on the data dimension, sample size, number of repeated measurements, and signal-to-noise ratio. We also show that our proposed computation algorithms can significantly reduce the computation time and are applicable to real-world data with a large number of HD-repeated measurements (e.g., functional magnetic resonance imaging (fMRI) data). Simulation results demonstrate both the finite sample performance and computational effectiveness of our proposed procedures. We observe that the empirical size of the test is well controlled at the nominal level, and the locations of multiple change points can be accurately identified. An application to fMRI data demonstrates that our proposed methods can identify event boundaries in the preface of the television series Sherlock. Code to implement the procedures is available in an R package named TechPhD.


Asunto(s)
Algoritmos , Imagen por Resonancia Magnética , Simulación por Computador , Imagen por Resonancia Magnética/métodos , Tamaño de la Muestra
4.
Biometrika ; 106(3): 619-634, 2019 Sep.
Artículo en Inglés | MEDLINE | ID: mdl-31427823

RESUMEN

This paper deals with the detection and identification of changepoints among covariances of high-dimensional longitudinal data, where the number of features is greater than both the sample size and the number of repeated measurements. The proposed methods are applicable under general temporal-spatial dependence. A new test statistic is introduced for changepoint detection, and its asymptotic distribution is established. If a changepoint is detected, an estimate of the location is provided. The rate of convergence of the estimator is shown to depend on the data dimension, sample size, and signal-to-noise ratio. Binary segmentation is used to estimate the locations of possibly multiple changepoints, and the corresponding estimator is shown to be consistent under mild conditions. Simulation studies provide the empirical size and power of the proposed test and the accuracy of the changepoint estimator. An application to a time-course microarray dataset identifies gene sets with significant gene interaction changes over time.

5.
Genet Epidemiol ; 43(2): 137-149, 2019 03.
Artículo en Inglés | MEDLINE | ID: mdl-30456931

RESUMEN

Single-variant-based genome-wide association studies have successfully detected many genetic variants that are associated with a number of complex traits. However, their power is limited due to weak marginal signals and ignoring potential complex interactions among genetic variants. The set-based strategy was proposed to provide a remedy where multiple genetic variants in a given set (e.g., gene or pathway) are jointly evaluated, so that the systematic effect of the set is considered. Among many, the kernel-based testing (KBT) framework is one of the most popular and powerful methods in set-based association studies. Given a set of candidate kernels, the method has been proposed to choose the one with the smallest p-value. Such a method, however, can yield inflated Type 1 error, especially when the number of variants in a set is large. Alternatively one can get p values by permutations which, however, could be very time-consuming. In this study, we proposed an efficient testing procedure that cannot only control Type 1 error rate but also have power close to the one obtained under the optimal kernel in the candidate kernel set, for quantitative trait association studies. Our method, a maximum kernel-based U-statistic method, is built upon the KBT framework and is based on asymptotic results under a high-dimensional setting. Hence it can efficiently deal with the case where the number of variants in a set is much larger than the sample size. Both simulation and real data analysis demonstrate the advantages of the method compared with its counterparts.


Asunto(s)
Algoritmos , Estudios de Asociación Genética/métodos , Estadística como Asunto , Simulación por Computador , Estudio de Asociación del Genoma Completo , Humanos , Recién Nacido , Modelos Genéticos
6.
Stat Appl Genet Mol Biol ; 17(2)2018 02 08.
Artículo en Inglés | MEDLINE | ID: mdl-29420308

RESUMEN

Gene-environment (G×E) interaction plays a pivotal role in understanding the genetic basis of complex disease. When environmental factors are measured continuously, one can assess the genetic sensitivity over different environmental conditions on a disease trait. Motivated by the increasing awareness of gene set based association analysis over single variant based approaches, we proposed an additive varying-coefficient model to jointly model variants in a genetic system. The model allows us to examine how variants in a gene set are moderated by an environment factor to affect a disease phenotype. We approached the problem from a variable selection perspective. In particular, we select variants with varying, constant and zero coefficients, which correspond to cases of G×E interaction, no G×E interaction and no genetic effect, respectively. The procedure was implemented through a two-stage iterative estimation algorithm via the smoothly clipped absolute deviation penalty function. Under certain regularity conditions, we established the consistency property in variable selection as well as effect separation of the two stage iterative estimators, and showed the optimal convergence rates of the estimates for varying effects. In addition, we showed that the estimate of non-zero constant coefficients enjoy the oracle property. The utility of our procedure was demonstrated through simulation studies and real data analysis.


Asunto(s)
Peso al Nacer/genética , Interacción Gen-Ambiente , Modelos Genéticos , Modelos Estadísticos , Polimorfismo de Nucleótido Simple , Algoritmos , Índice de Masa Corporal , Edad Gestacional , Humanos , Recién Nacido , Madres , Fenotipo
7.
Biometrics ; 73(3): 972-980, 2017 09.
Artículo en Inglés | MEDLINE | ID: mdl-28182830

RESUMEN

Missing values appear very often in many applications, but the problem of missing values has not received much attention in testing order-restricted alternatives. Under the missing at random (MAR) assumption, we impute the missing values nonparametrically using kernel regression. For data with imputation, the classical likelihood ratio test designed for testing the order-restricted means is no longer applicable since the likelihood does not exist. This article proposes a novel method for constructing test statistics for assessing means with an increasing order or a decreasing order based on jackknife empirical likelihood (JEL) ratio. It is shown that the JEL ratio statistic evaluated under the null hypothesis converges to a chi-bar-square distribution, whose weights depend on missing probabilities and nonparametric imputation. Simulation study shows that the proposed test performs well under various missing scenarios and is robust for normally and nonnormally distributed data. The proposed method is applied to an Alzheimer's disease neuroimaging initiative data set for finding a biomarker for the diagnosis of the Alzheimer's disease.


Asunto(s)
Interpretación Estadística de Datos , Distribución de Chi-Cuadrado , Proyectos de Investigación
8.
Scand Stat Theory Appl ; 43(3): 886-903, 2016 Sep.
Artículo en Inglés | MEDLINE | ID: mdl-27795610

RESUMEN

Functional data analysis has become an important area of research due to its ability of handling high dimensional and complex data structures. However, the development is limited in the context of linear mixed effect models, and in particular, for small area estimation. The linear mixed effect models are the backbone of small area estimation. In this article, we consider area level data, and fit a varying coefficient linear mixed effect model where the varying coefficients are semi-parametrically modeled via B-splines. We propose a method of estimating the fixed effect parameters and consider prediction of random effects that can be implemented using a standard software. For measuring prediction uncertainties, we derive an analytical expression for the mean squared errors, and propose a method of estimating the mean squared errors. The procedure is illustrated via a real data example, and operating characteristics of the method are judged using finite sample simulation studies.

9.
J Econom ; 195(1): 154-168, 2016 Nov.
Artículo en Inglés | MEDLINE | ID: mdl-28663668

RESUMEN

In linear regression models with high dimensional data, the classical z-test (or t-test) for testing the significance of each single regression coefficient is no longer applicable. This is mainly because the number of covariates exceeds the sample size. In this paper, we propose a simple and novel alternative by introducing the Correlated Predictors Screening (CPS) method to control for predictors that are highly correlated with the target covariate. Accordingly, the classical ordinary least squares approach can be employed to estimate the regression coefficient associated with the target covariate. In addition, we demonstrate that the resulting estimator is consistent and asymptotically normal even if the random errors are heteroscedastic. This enables us to apply the z-test to assess the significance of each covariate. Based on the p-value obtained from testing the significance of each covariate, we further conduct multiple hypothesis testing by controlling the false discovery rate at the nominal level. Then, we show that the multiple hypothesis testing achieves consistent model selection. Simulation studies and empirical examples are presented to illustrate the finite sample performance and the usefulness of the proposed method, respectively.

10.
BMC Bioinformatics ; 16: 61, 2015 Feb 22.
Artículo en Inglés | MEDLINE | ID: mdl-25887316

RESUMEN

BACKGROUND: Allelic specific expression (ASE) increases our understanding of the genetic control of gene expression and its links to phenotypic variation. ASE testing is implemented through binomial or beta-binomial tests of sequence read counts of alternative alleles at a cSNP of interest in heterozygous individuals. This requires prior ascertainment of the cSNP genotypes for all individuals. To meet the needs, we propose hidden Markov methods to call SNPs from next generation RNA sequence data when ASE possibly exists. RESULTS: We propose two hidden Markov models (HMMs), HMM-ASE and HMM-NASE that consider or do not consider ASE, respectively, in order to improve genotyping accuracy. Both HMMs have the advantages of calling the genotypes of several SNPs simultaneously and allow mapping error which, respectively, utilize the dependence among SNPs and correct the bias due to mapping error. In addition, HMM-ASE exploits ASE information to further improve genotype accuracy when the ASE is likely to be present. Simulation results indicate that the HMMs proposed demonstrate a very good prediction accuracy in terms of controlling both the false discovery rate (FDR) and the false negative rate (FNR). When ASE is present, the HMM-ASE had a lower FNR than HMM-NASE, while both can control the false discovery rate (FDR) at a similar level. By exploiting linkage disequilibrium (LD), a real data application demonstrate that the proposed methods have better sensitivity and similar FDR in calling heterozygous SNPs than the VarScan method. Sensitivity and FDR are similar to that of the BCFtools and Beagle methods. The resulting genotypes show good properties for the estimation of the genetic parameters and ASE ratios. CONCLUSIONS: We introduce HMMs, which are able to exploit LD and account for the ASE and mapping errors, to simultaneously call SNPs from the next generation RNA sequence data. The method introduced can reliably call for cSNP genotypes even in the presence of ASE and under low sequencing coverage. As a byproduct, the proposed method is able to provide predictions of ASE ratios for the heterozygous genotypes, which can then be used for ASE testing.


Asunto(s)
Desequilibrio Alélico , Secuenciación de Nucleótidos de Alto Rendimiento/métodos , Desequilibrio de Ligamiento , Cadenas de Markov , Polimorfismo de Nucleótido Simple/genética , Análisis de Secuencia de ARN/métodos , Algoritmos , Alelos , Biología Computacional/métodos , Genotipo , Humanos , ARN/genética
11.
BMC Proc ; 8(Suppl 1): S61, 2014.
Artículo en Inglés | MEDLINE | ID: mdl-25519336

RESUMEN

The genetic basis of blood pressure often involves multiple genetic factors and their interactions with environmental factors. Gene-environment interaction is assumed to play an important role in determining individual blood pressure variability. Older people are more prone to high blood pressure than younger ones and the risk may not display a linear trend over the life span. However, which gene shows sensitivity to aging in its effect on blood pressure is not clear. In this work, we allowed the genetic effect to vary over time and propose a varying-coefficient model to identify potential genetic players that show nonlinear response across different age stages. We detected 2 novel loci, gene MIR1263 (a microRNA coding gene) on chromosome 3 and gene UNC13B on chromosome 9, that are nonlinearly associated with diastolic blood pressure. Further experimental validation is needed to confirm this finding.

12.
Front Genet ; 5: 395, 2014.
Artículo en Inglés | MEDLINE | ID: mdl-25429300

RESUMEN

Single variant analysis in genome-wide association studies (GWAS) has been proven to be successful in identifying thousands of genetic variants associated with hundreds of complex diseases. However, these identified variants only explain a small fraction of inheritable variability in many diseases, suggesting that other resources, such as multilevel genetic variations, may contribute to disease susceptibility. In this work, we proposed to combine genetic variants that belong to a gene set, such as at gene- and pathway-level to form an integrated signal aimed to identify major players that function in a coordinated manner conferring disease risk. The integrated analysis provides novel insight into disease etiology while individual signals could be easily missed by single variant analysis. We applied our approach to a genome-wide association study of type 2 diabetes (T2D) with male and female data analyzed separately. Novel sex-specific genes and pathways were identified to increase the risk of T2D. We also demonstrated the performance of signal integration through simulation studies.

13.
PLoS One ; 9(3): e91702, 2014.
Artículo en Inglés | MEDLINE | ID: mdl-24643065

RESUMEN

Cytoplasm contains important metabolism reaction organelles such as mitochondria and chloroplast (in plant). In particular, mitochondria contains special DNA information which can be passed to offsprings through maternal gametes, and has been confirmed to play a pivotal role in nuclear activities. Experimental evidences have documented the importance of cyto-nuclear interactions in affecting important biological traits. While studies have also pointed out the role of interaction between imprinting nuclear DNA and cytoplasm, no statistical method has been developed to efficiently model such effect and further quantify its effect size. In this work, we developed an efficient statistical model for genome-wide estimating and testing the cytoplasmic effect, nuclear DNA imprinting effect as well as the interaction between them under reciprocal backcross and F2 designs derived from inbred lines. Parameters are estimated under maximum likelihood framework implemented with the EM algorithm. Extensive simulations show good performance in a variety of scenarios. The utility of the method is demonstrated by analyzing a published data set in an F2 family derived from C3H/HeJBir and C57BL/6 J mouse strains. Important cyto-nuclear interactions were identified. Our approach provides a quantitative framework for identifying and estimating cyto-nuclear interactions subject to genomic imprinting involved in the genetic control of complex traits.


Asunto(s)
Núcleo Celular/genética , Citoplasma/genética , Epistasis Genética , Impresión Genómica , Modelos Genéticos , Algoritmos , Animales , Simulación por Computador , Cruzamientos Genéticos , Femenino , Masculino , Ratones , Ratones Endogámicos C3H , Ratones Endogámicos C57BL , Sitios de Carácter Cuantitativo
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA