Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 16 de 16
Filtrar
1.
BMC Med ; 21(1): 182, 2023 05 15.
Artigo em Inglês | MEDLINE | ID: mdl-37189125

RESUMO

BACKGROUND: In high-dimensional data (HDD) settings, the number of variables associated with each observation is very large. Prominent examples of HDD in biomedical research include omics data with a large number of variables such as many measurements across the genome, proteome, or metabolome, as well as electronic health records data that have large numbers of variables recorded for each patient. The statistical analysis of such data requires knowledge and experience, sometimes of complex methods adapted to the respective research questions. METHODS: Advances in statistical methodology and machine learning methods offer new opportunities for innovative analyses of HDD, but at the same time require a deeper understanding of some fundamental statistical concepts. Topic group TG9 "High-dimensional data" of the STRATOS (STRengthening Analytical Thinking for Observational Studies) initiative provides guidance for the analysis of observational studies, addressing particular statistical challenges and opportunities for the analysis of studies involving HDD. In this overview, we discuss key aspects of HDD analysis to provide a gentle introduction for non-statisticians and for classically trained statisticians with little experience specific to HDD. RESULTS: The paper is organized with respect to subtopics that are most relevant for the analysis of HDD, in particular initial data analysis, exploratory data analysis, multiple testing, and prediction. For each subtopic, main analytical goals in HDD settings are outlined. For each of these goals, basic explanations for some commonly used analysis methods are provided. Situations are identified where traditional statistical methods cannot, or should not, be used in the HDD setting, or where adequate analytic tools are still lacking. Many key references are provided. CONCLUSIONS: This review aims to provide a solid statistical foundation for researchers, including statisticians and non-statisticians, who are new to research with HDD or simply want to better evaluate and understand the results of HDD analyses.


Assuntos
Pesquisa Biomédica , Objetivos , Humanos , Projetos de Pesquisa
2.
BMC Bioinformatics ; 24(1): 96, 2023 Mar 16.
Artigo em Inglês | MEDLINE | ID: mdl-36927444

RESUMO

BACKGROUND: The research of biomarker-treatment interactions is commonly investigated in randomized clinical trials (RCT) for improving medicine precision. The hierarchical interaction constraint states that an interaction should only be in a model if its main effects are also in the model. However, this constraint is not guaranteed in the standard penalized statistical approaches. We aimed to find a compromise for high-dimensional data between the need for sparse model selection and the need for the hierarchical constraint. RESULTS: To favor the property of the hierarchical interaction constraint, we proposed to create groups composed of the biomarker main effect and its interaction with treatment and to perform the bi-level selection on these groups. We proposed two weighting approaches (Single Wald (SW) and likelihood ratio test (LRT)) for the adaptive lasso method. The selection performance of these two approaches is compared to alternative lasso extensions (adaptive lasso with ridge-based weights, composite Minimax Concave Penalty, group exponential lasso and Sparse Group Lasso) through a simulation study. A RCT (NSABP B-31) randomizing 1574 patients (431 events) with early breast cancer aiming to evaluate the effect of adjuvant trastuzumab on distant-recurrence free survival with expression data from 462 genes measured in the tumour will serve for illustration. The simulation study illustrates that the adaptive lasso LRT and SW, and the group exponential lasso favored the hierarchical interaction constraint. Overall, in the alternative scenarios, they had the best balance of false discovery and false negative rates for the main effects of the selected interactions. For NSABP B-31, 12 gene-treatment interactions were identified more than 20% by the different methods. Among them, the adaptive lasso (SW) approach offered the best trade-off between a high number of selected gene-treatment interactions and a high proportion of selection of both the gene-treatment interaction and its main effect. CONCLUSIONS: Adaptive lasso with Single Wald and likelihood ratio test weighting and the group exponential lasso approaches outperformed their competitors in favoring the hierarchical constraint of the biomarker-treatment interaction. However, the performance of the methods tends to decrease in the presence of prognostic biomarkers.


Assuntos
Neoplasias da Mama , Medicina de Precisão , Humanos , Feminino , Ensaios Clínicos Controlados Aleatórios como Assunto , Biomarcadores , Neoplasias da Mama/tratamento farmacológico , Neoplasias da Mama/genética , Simulação por Computador
3.
Biometrics ; 79(1): 319-331, 2023 03.
Artigo em Inglês | MEDLINE | ID: mdl-34510407

RESUMO

Publication bias and p-hacking are two well-known phenomena that strongly affect the scientific literature and cause severe problems in meta-analyses. Due to these phenomena, the assumptions of meta-analyses are seriously violated and the results of the studies cannot be trusted. While publication bias is very often captured well by the weighting function selection model, p-hacking is much harder to model and no definitive solution has been found yet. In this paper, we advocate the selection model approach to model publication bias and propose a mixture model for p-hacking. We derive some properties for these models, and we compare them formally and through simulations. Finally, two real data examples are used to show how the models work in practice.


Assuntos
Viés de Publicação , Viés , Metanálise como Assunto
4.
Lifetime Data Anal ; 29(2): 420-440, 2023 04.
Artigo em Inglês | MEDLINE | ID: mdl-35476164

RESUMO

In this paper we propose a boosting algorithm to extend the applicability of a first hitting time model to high-dimensional frameworks. Based on an underlying stochastic process, first hitting time models do not require the proportional hazards assumption, hardly verifiable in the high-dimensional context, and represent a valid parametric alternative to the Cox model for modelling time-to-event responses. First hitting time models also offer a natural way to integrate low-dimensional clinical and high-dimensional molecular information in a prediction model, that avoids complicated weighting schemes typical of current methods. The performance of our novel boosting algorithm is illustrated in three real data examples.


Assuntos
Algoritmos , Humanos , Análise de Sobrevida , Modelos de Riscos Proporcionais , Processos Estocásticos
5.
ALTEX ; 40(2): 287-298, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-36317504

RESUMO

Quantitative adverse outcome pathway network (qAOPN) is gaining momentum due to its predictive nature, alignment with quantitative risk assessment, and great potential as a computational new approach methodology (NAM) to reduce laboratory animal tests. The present work aimed to demonstrate two advanced modeling approaches, piecewise structural equation modeling (PSEM) and Bayesian network (BN), for de novo qAOPN model construction based on routine ecotoxicological data. A previously published AOP network comprised of four linear AOPs linking excessive reactive oxygen species production to mortality in aquatic organisms was employed as a case study. The demonstrative case study intended to answer: Which linear AOP in the network contributed the most to the AO? Can any of the upstream KEs accurately predict the AO? What are the advantages and limitations of PSEM or BN in qAOPN development? The outcomes from the two approaches showed that both PSEM and BN are suitable for constructing a complex qAOPN based on limited experimental data. Besides quantification of response-response relationships, both approaches could identify the most influencing linear AOP in a complex network and evaluate the predictive ability of the AOP, albeit some discrepancies in predictive ability were identified for the two approaches using this specific dataset. The advantages and limitations of the two approaches for qAOPN construction are discussed in detail, and suggestions on optimal workflows of PSEM and BN are provided to guide future qAOPN development.


Assuntos
Rotas de Resultados Adversos , Animais , Teorema de Bayes , Análise de Classes Latentes , Alternativas aos Testes com Animais , Ecotoxicologia/métodos , Medição de Risco/métodos
6.
Stat Med ; 40(2): 369-381, 2021 01 30.
Artigo em Inglês | MEDLINE | ID: mdl-33089538

RESUMO

Statistical models are often fitted to obtain a concise description of the association of an outcome variable with some covariates. Even if background knowledge is available to guide preselection of covariates, stepwise variable selection is commonly applied to remove irrelevant ones. This practice may introduce additional variability and selection is rarely certain. However, these issues are often ignored and model stability is not questioned. Several resampling-based measures were proposed to describe model stability, including variable inclusion frequencies (VIFs), model selection frequencies, relative conditional bias (RCB), and root mean squared difference ratio (RMSDR). The latter two were recently proposed to assess bias and variance inflation induced by variable selection. Here, we study the consistency and accuracy of resampling estimates of these measures and the optimal choice of the resampling technique. In particular, we compare subsampling and bootstrapping for assessing stability of linear, logistic, and Cox models obtained by backward elimination in a simulation study. Moreover, we exemplify the estimation and interpretation of all suggested measures in a study on cardiovascular risk. The VIF and the model selection frequency are only consistently estimated in the subsampling approach. By contrast, the bootstrap is advantageous in terms of bias and precision for estimating the RCB as well as the RMSDR. Though, unbiased estimation of the latter quantity requires independence of covariates, which is rarely encountered in practice. Our study stresses the importance of addressing model stability after variable selection and shows how to cope with it.


Assuntos
Modelos Estatísticos , Simulação por Computador , Humanos , Modelos de Riscos Proporcionais
7.
Chem Sci ; 11(18): 4584-4601, 2020 May 14.
Artigo em Inglês | MEDLINE | ID: mdl-33224459

RESUMO

Homogeneous catalysis using transition metal complexes is ubiquitously used for organic synthesis, as well as technologically relevant in applications such as water splitting and CO2 reduction. The key steps underlying homogeneous catalysis require a specific combination of electronic and steric effects from the ligands bound to the metal center. Finding the optimal combination of ligands is a challenging task due to the exceedingly large number of possibilities and the non-trivial ligand-ligand interactions. The classic example of Vaska's complex, trans-[Ir(PPh3)2(CO)(Cl)], illustrates this scenario. The ligands of this species activate iridium for the oxidative addition of hydrogen, yielding the dihydride cis-[Ir(H)2(PPh3)2(CO)(Cl)] complex. Despite the simplicity of this system, thousands of derivatives can be formulated for the activation of H2, with a limited number of ligands belonging to the same general categories found in the original complex. In this work, we show how DFT and machine learning (ML) methods can be combined to enable the prediction of reactivity within large chemical spaces containing thousands of complexes. In a space of 2574 species derived from Vaska's complex, data from DFT calculations are used to train and test ML models that predict the H2-activation barrier. In contrast to experiments and calculations requiring several days to be completed, the ML models were trained and used on a laptop on a time-scale of minutes. As a first approach, we combined Bayesian-optimized artificial neural networks (ANN) with features derived from autocorrelation and deltametric functions. The resulting ANNs achieved high accuracies, with mean absolute errors (MAE) between 1 and 2 kcal mol-1, depending on the size of the training set. By using a Gaussian process (GP) model trained with a set of selected features, including fingerprints, accuracy was further enhanced. Remarkably, this GP model minimized the MAE below 1 kcal mol-1, by using only 20% or less of the data available for training. The gradient boosting (GB) method was also used to assess the relevance of the features, which was used for both feature selection and model interpretation purposes. Features accounting for chemical composition, atom size and electronegativity were found to be the most determinant in the predictions. Further, the ligand fragments with the strongest influence on the H2-activation barrier were identified.

8.
PLoS One ; 15(11): e0242334, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-33186404

RESUMO

The time it takes a student to graduate with a university degree is mitigated by a variety of factors such as their background, the academic performance at university, and their integration into the social communities of the university they attend. Different universities have different populations, student services, instruction styles, and degree programs, however, they all collect institutional data. This study presents data for 160,933 students attending a large American research university. The data includes performance, enrollment, demographics, and preparation features. Discrete time hazard models for the time-to-graduation are presented in the context of Tinto's Theory of Drop Out. Additionally, a novel machine learning method: gradient boosted trees, is applied and compared to the typical maximum likelihood method. We demonstrate that enrollment factors (such as changing a major) lead to greater increases in model predictive performance of when a student graduates than performance factors (such as grades) or preparation (such as high school GPA).


Assuntos
Escolaridade , Universidades/estatística & dados numéricos , Adulto , Feminino , Humanos , Funções Verossimilhança , Modelos Logísticos , Masculino , Fatores de Tempo , Estados Unidos , Adulto Jovem
9.
Brief Bioinform ; 21(6): 1904-1919, 2020 12 01.
Artigo em Inglês | MEDLINE | ID: mdl-31750518

RESUMO

Data integration, i.e. the use of different sources of information for data analysis, is becoming one of the most important topics in modern statistics. Especially in, but not limited to, biomedical applications, a relevant issue is the combination of low-dimensional (e.g. clinical data) and high-dimensional (e.g. molecular data such as gene expressions) data sources in a prediction model. Not only the different characteristics of the data, but also the complex correlation structure within and between the two data sources, pose challenging issues. In this paper, we investigate these issues via simulations, providing some useful insight into strategies to combine low- and high-dimensional data in a regression prediction model. In particular, we focus on the effect of the correlation structure on the results, while accounting for the influence of our specific choices in the design of the simulation study.


Assuntos
Biologia Computacional , Simulação por Computador , Modelos Estatísticos
10.
BMC Med Res Methodol ; 19(1): 162, 2019 07 24.
Artigo em Inglês | MEDLINE | ID: mdl-31340753

RESUMO

BACKGROUND: Omics data can be very informative in survival analysis and may improve the prognostic ability of classical models based on clinical risk factors for various diseases, for example breast cancer. Recent research has focused on integrating omics and clinical data, yet has often ignored the need for appropriate model building for clinical variables. Medical literature on classical prognostic scores, as well as biostatistical literature on appropriate model selection strategies for low dimensional (clinical) data, are often ignored in the context of omics research. The goal of this paper is to fill this methodological gap by investigating the added predictive value of gene expression data for models using varying amounts of clinical information. METHODS: We analyze two data sets from the field of survival prognosis of breast cancer patients. First, we construct several proportional hazards prediction models using varying amounts of clinical information based on established medical knowledge. These models are then used as a starting point (i.e. included as a clinical offset) for identifying informative gene expression variables using resampling procedures and penalized regression approaches (model based boosting and the LASSO). In order to assess the added predictive value of the gene signatures, measures of prediction accuracy and separation are examined on a validation data set for the clinical models and the models that combine the two sources of information. RESULTS: For one data set, we do not find any substantial added predictive value of the omics data when compared to clinical models. On the second data set, we identify a noticeable added predictive value, however only for scenarios where little or no clinical information is included in the modeling process. We find that including more clinical information can lead to a smaller number of selected omics predictors. CONCLUSIONS: New research using omics data should include all available established medical knowledge in order to allow an adequate evaluation of the added predictive value of omics data. Including all relevant clinical information in the analysis might also lead to more parsimonious models. The developed procedure to assess the predictive value of the omics data can be readily applied to other scenarios.


Assuntos
Neoplasias da Mama/genética , Neoplasias da Mama/mortalidade , Genômica/estatística & dados numéricos , Modelos Estatísticos , Análise de Sobrevida , Conjuntos de Dados como Assunto , Feminino , Expressão Gênica , Humanos , Fatores de Risco
11.
Obstet Gynecol ; 130(5): 1017-1024, 2017 11.
Artigo em Inglês | MEDLINE | ID: mdl-29016504

RESUMO

OBJECTIVE: To establish the diagnostic test accuracy of evacuation proctography, magnetic resonance imaging (MRI), transperineal ultrasonography, and endovaginal ultrasonography for detecting posterior pelvic floor disorders (rectocele, enterocele, intussusception, and anismus) in women with obstructed defecation syndrome and secondarily to identify the most patient-friendly imaging technique. METHODS: In this prospective cohort study, 131 women with symptoms of obstructed defecation syndrome underwent evacuation proctogram, MRI, and transperineal and endovaginal ultrasonography. Images were analyzed by two blinded observers. In the absence of a reference standard, latent class analysis was used to assess diagnostic test accuracy of multiple tests with area under the curve (AUC) as the primary outcome measure. Secondary outcome measures were interobserver agreement calculated as Cohen's κ and patient acceptability using a visual analog scale. RESULTS: No significant differences in diagnostic accuracy were found among the imaging techniques for all the target conditions. Estimates of diagnostic test accuracy were highest for rectocele using MRI (AUC 0.79) or transperineal ultrasonography (AUC 0.85), for enterocele using transperineal (AUC 0.73) or endovaginal ultrasonography (AUC 0.87), for intussusception using evacuation proctography (AUC 0.76) or endovaginal ultrasonography (AUC 0.77), and for anismus using endovaginal (AUC 0.95) or transperineal ultrasonography (AUC 0.78). Interobserver agreement for the diagnosis of rectocele (κ 0.53-0.72), enterocele (κ 0.54-0.94) and anismus (κ 0.43-0.81) was moderate to excellent, but poor to fair for intussusception (κ -0.03 to 0.37) with all techniques. Patient acceptability was better for transperineal and endovaginal ultrasonography as compared with MRI and evacuation proctography (P<.001). CONCLUSION: Evacuation proctography, MRI, and transperineal and endovaginal ultrasonography were shown to have similar diagnostic test accuracy. Evacuation proctography is not the best available imaging technique. There is no one optimal test for the diagnosis of all posterior pelvic floor disorders. Because transperineal and endovaginal ultrasonography have good test accuracy and patient acceptability, we suggest these could be used for initial assessment of obstructed defecation syndrome. CLINICAL TRIAL REGISTRATION: ClinicalTrials.gov, NCT02239302.


Assuntos
Constipação Intestinal/diagnóstico por imagem , Defecografia/métodos , Endossonografia/métodos , Imageamento por Ressonância Magnética/métodos , Distúrbios do Assoalho Pélvico/diagnóstico por imagem , Adulto , Idoso , Idoso de 80 Anos ou mais , Área Sob a Curva , Constipação Intestinal/etiologia , Defecação , Defecografia/psicologia , Endossonografia/psicologia , Feminino , Humanos , Imageamento por Ressonância Magnética/psicologia , Pessoa de Meia-Idade , Aceitação pelo Paciente de Cuidados de Saúde , Distúrbios do Assoalho Pélvico/complicações , Períneo/diagnóstico por imagem , Estudos Prospectivos , Síndrome , Vagina/diagnóstico por imagem
12.
Comput Math Methods Med ; 2017: 7691937, 2017.
Artigo em Inglês | MEDLINE | ID: mdl-28546826

RESUMO

As modern biotechnologies advance, it has become increasingly frequent that different modalities of high-dimensional molecular data (termed "omics" data in this paper), such as gene expression, methylation, and copy number, are collected from the same patient cohort to predict the clinical outcome. While prediction based on omics data has been widely studied in the last fifteen years, little has been done in the statistical literature on the integration of multiple omics modalities to select a subset of variables for prediction, which is a critical task in personalized medicine. In this paper, we propose a simple penalized regression method to address this problem by assigning different penalty factors to different data modalities for feature selection and prediction. The penalty factors can be chosen in a fully data-driven fashion by cross-validation or by taking practical considerations into account. In simulation studies, we compare the prediction performance of our approach, called IPF-LASSO (Integrative LASSO with Penalty Factors) and implemented in the R package ipflasso, with the standard LASSO and sparse group LASSO. The use of IPF-LASSO is also illustrated through applications to two real-life cancer datasets. All data and codes are available on the companion website to ensure reproducibility.


Assuntos
Algoritmos , Biologia Computacional/métodos , Medicina de Precisão/métodos , Humanos , Neoplasias/diagnóstico , Análise de Regressão , Reprodutibilidade dos Testes , Estatística como Assunto/normas
13.
Biometrics ; 72(1): 272-80, 2016 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-26288150

RESUMO

In recent years, increasing attention has been devoted to the problem of the stability of multivariable regression models, understood as the resistance of the model to small changes in the data on which it has been fitted. Resampling techniques, mainly based on the bootstrap, have been developed to address this issue. In particular, the approaches based on the idea of "inclusion frequency" consider the repeated implementation of a variable selection procedure, for example backward elimination, on several bootstrap samples. The analysis of the variables selected in each iteration provides useful information on the model stability and on the variables' importance. Recent findings, nevertheless, show possible pitfalls in the use of the bootstrap, and alternatives such as subsampling have begun to be taken into consideration in the literature. Using model selection frequencies and variable inclusion frequencies, we empirically compare these two different resampling techniques, investigating the effect of their use in selected classical model selection procedures for multivariable regression. We conduct our investigations by analyzing two real data examples and by performing a simulation study. Our results reveal some advantages in using a subsampling technique rather than the bootstrap in this context.


Assuntos
Algoritmos , Modelos Estatísticos , Análise Multivariada , Análise de Regressão , Tamanho da Amostra , Simulação por Computador , Interpretação Estatística de Dados , Reprodutibilidade dos Testes , Sensibilidade e Especificidade
14.
BMC Med Res Methodol ; 14: 117, 2014 Oct 28.
Artigo em Inglês | MEDLINE | ID: mdl-25352096

RESUMO

BACKGROUND: In the last years, the importance of independent validation of the prediction ability of a new gene signature has been largely recognized. Recently, with the development of gene signatures which integrate rather than replace the clinical predictors in the prediction rule, the focus has been moved to the validation of the added predictive value of a gene signature, i.e. to the verification that the inclusion of the new gene signature in a prediction model is able to improve its prediction ability. METHODS: The high-dimensional nature of the data from which a new signature is derived raises challenging issues and necessitates the modification of classical methods to adapt them to this framework. Here we show how to validate the added predictive value of a signature derived from high-dimensional data and critically discuss the impact of the choice of methods on the results. RESULTS: The analysis of the added predictive value of two gene signatures developed in two recent studies on the survival of leukemia patients allows us to illustrate and empirically compare different validation techniques in the high-dimensional framework. CONCLUSIONS: The issues related to the high-dimensional nature of the omics predictors space affect the validation process. An analysis procedure based on repeated cross-validation is suggested.


Assuntos
Leucemia Linfocítica Crônica de Células B/genética , Leucemia Linfocítica Crônica de Células B/mortalidade , Leucemia Mieloide Aguda/genética , Leucemia Mieloide Aguda/mortalidade , Adolescente , Adulto , Fatores Etários , Idoso , Idoso de 80 Anos ou mais , Feminino , Genômica , Humanos , Região Variável de Imunoglobulina/genética , Leucemia Linfocítica Crônica de Células B/diagnóstico , Leucemia Mieloide Aguda/diagnóstico , Masculino , Pessoa de Meia-Idade , Modelos Estatísticos , Proteínas Nucleares/genética , Nucleofosmina , Prognóstico , Modelos de Riscos Proporcionais , Taxa de Sobrevida , Sequências de Repetição em Tandem/genética , Estudos de Validação como Assunto , Adulto Jovem , Tirosina Quinase 3 Semelhante a fms/genética
15.
Stat Med ; 33(30): 5310-29, 2014 Dec 30.
Artigo em Inglês | MEDLINE | ID: mdl-25042390

RESUMO

In biomedical literature, numerous prediction models for clinical outcomes have been developed based either on clinical data or, more recently, on high-throughput molecular data (omics data). Prediction models based on both types of data, however, are less common, although some recent studies suggest that a suitable combination of clinical and molecular information may lead to models with better predictive abilities. This is probably due to the fact that it is not straightforward to combine data with different characteristics and dimensions (poorly characterized high-dimensional omics data, well-investigated low-dimensional clinical data). In this paper, we analyze two publicly available datasets related to breast cancer and neuroblastoma, respectively, in order to show some possible ways to combine clinical and omics data into a prediction model of time-to-event outcome. Different strategies and statistical methods are exploited. The results are compared and discussed according to different criteria, including the discriminative ability of the models, computed on a validation dataset.


Assuntos
Biometria/métodos , Interpretação Estatística de Dados , Modelos de Riscos Proporcionais , Adulto , Neoplasias da Mama/genética , Pré-Escolar , Bases de Dados Genéticas , Feminino , Humanos , Lactente , Masculino , Pessoa de Meia-Idade , Neuroblastoma/genética , Análise de Regressão , Análise de Sobrevida , Fatores de Tempo
16.
BMC Bioinformatics ; 12: 49, 2011 Feb 08.
Artigo em Inglês | MEDLINE | ID: mdl-21303507

RESUMO

BACKGROUND: Cluster analysis is a crucial tool in several biological and medical studies dealing with microarray data. Such studies pose challenging statistical problems due to dimensionality issues, since the number of variables can be much higher than the number of observations. RESULTS: Here, we present a general framework to deal with the clustering of microarray data, based on a three-step procedure: (i) gene filtering; (ii) dimensionality reduction; (iii) clustering of observations in the reduced space. Via a nonparametric model-based clustering approach we obtain promising results both in simulated and real data. CONCLUSIONS: The proposed algorithm is a simple and effective tool for the clustering of microarray data, in an unsupervised setting.


Assuntos
Algoritmos , Análise por Conglomerados , Biologia Computacional/métodos , Modelos Estatísticos , Análise de Sequência com Séries de Oligonucleotídeos/métodos , Simulação por Computador , Perfilação da Expressão Gênica/métodos , Humanos , Análise de Componente Principal , Software , Estatísticas não Paramétricas
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...