Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 15 de 15
Filter
1.
Signif (Oxf) ; 18(2): 44-45, 2021 Apr.
Article in English | MEDLINE | ID: mdl-34230829

ABSTRACT

Paul Allin and David J. Hand call for official statistics to take centre stage.

2.
Mach Learn ; 110(3): 451-456, 2021.
Article in English | MEDLINE | ID: mdl-33746357

ABSTRACT

The F-measure, also known as the F1-score, is widely used to assess the performance of classification algorithms. However, some researchers find it lacking in intuitive interpretation, questioning the appropriateness of combining two aspects of performance as conceptually distinct as precision and recall, and also questioning whether the harmonic mean is the best way to combine them. To ease this concern, we describe a simple transformation of the F-measure, which we call F ∗ (F-star), which has an immediate practical interpretation.

3.
Patterns (N Y) ; 1(3): 100037, 2020 Jun 12.
Article in English | MEDLINE | ID: mdl-33205105

ABSTRACT

AI systems will only fulfill their promise for society if they can be relied upon. This means that the role and task of the system must be properly formulated; that the system must be bug free, be based on properly representative data, and can cope with anomalies and data quality issues; and that its output is sufficiently accurate for the task.

4.
Big Data ; 6(3): 176-190, 2018 09 01.
Article in English | MEDLINE | ID: mdl-30283727

ABSTRACT

Ready data availability, cheap storage capacity, and powerful tools for extracting information from data have the potential to significantly enhance the human condition. However, as with all advanced technologies, this comes with the potential for misuse. Ethical oversight and constraints are needed to ensure that an appropriate balance is reached. Ethical issues involving data may be more challenging than the ethical challenges of some other advanced technologies partly because data and data science are ubiquitous, having the potential to impact all aspects of life, and partly because of their intrinsic complexity. We explore the nature of data, personal data, data ownership, consent and purpose of use, trustworthiness of data as well as of algorithms and of those using the data, and matters of privacy and confidentiality. A checklist is given of topics that need to be considered.


Subject(s)
Data Collection/ethics , Data Science/ethics , Confidentiality , Ethics , Humans , Informed Consent , Internet , Ownership , Privacy , Trust
5.
Philos Trans A Math Phys Eng Sci ; 373(2039)2015 Apr 13.
Article in English | MEDLINE | ID: mdl-25750151

ABSTRACT

The nature of statistics has changed over time. It was originally concerned with descriptive 'matters of state'--with summarizing population numbers, economic strength and social conditions. But during the course of the twentieth century its aim broadened to include inference--how to use data to shed light on underlying mechanisms, about what might happen in the future, about what would happen if certain actions were taken. Central to this development was Ronald Fisher. Over the course of his life he was responsible for many of the major conceptual advances in statistics. This is particularly illustrated by his 1922 paper, in which he introduced many of the concepts which remain fundamental to our understanding of how to extract meaning from data, right to the present day. It is no exaggeration to say that Fisher's work, as illustrated by the ideas he described and developed in this paper, underlies all modern science, and much more besides. This commentary was written to celebrate the 350th anniversary of the journal Philosophical Transactions of the Royal Society.

6.
Sci Am ; 310(2): 72-5, 2014 Feb.
Article in English | MEDLINE | ID: mdl-24640335
7.
Stat Med ; 29(14): 1502-10, 2010 Jun 30.
Article in English | MEDLINE | ID: mdl-20087877

ABSTRACT

Because accurate diagnosis lies at the heart of medicine, it is important to be able to evaluate the effectiveness of diagnostic tests. A variety of accuracy measures are used. One particularly widely used measure is the AUC, the area under the receiver operating characteristic (ROC) curve. This measure has a well-understood weakness when comparing ROC curves which cross. However, it also has the more fundamental weakness of failing to balance different kinds of misdiagnoses effectively. This is not merely an aspect of the inevitable arbitrariness in choosing a performance measure, but is a core property of the way the AUC is defined. This property is explored, and an alternative, the H measure, is described.


Subject(s)
Area Under Curve , Diagnostic Tests, Routine/standards , ROC Curve , Sensitivity and Specificity , Aged , Aged, 80 and over , Female , Humans , Logistic Models , Middle Aged , Osteoporosis/diagnosis , Surveys and Questionnaires
8.
Stat Appl Genet Mol Biol ; 7(2): Article15, 2008.
Article in English | MEDLINE | ID: mdl-19120032

ABSTRACT

The performance results of a wide range of different classifiers applied to proteomic mass spectra data, in a blind comparative assessment organised by Bart Mertens, are reviewed. The different approaches are summarised, issues of how to evaluate and compare the predictions are described, and the results of the different methods are examined. Although the different methods perform differently, their rank ordering varies according to how one measures performance, so that one cannot draw unequivocal conclusions about which is 'best.' Instead, it is clear that what matters is not the method by itself, but the interaction of method and user - the degree of sophistication of the user with a method. Nevertheless, such competitions do serve the useful role of setting (constantly improving) baselines against which new researchers can pit their wits and methods, as well as providing standards against which new methods should be assessed.


Subject(s)
Breast Neoplasms/diagnosis , Breast Neoplasms/metabolism , Mass Spectrometry/methods , Proteomics/methods , Calibration , Female , Humans , Reproducibility of Results , Sensitivity and Specificity
9.
Drug Saf ; 30(7): 621-2, 2007.
Article in English | MEDLINE | ID: mdl-17604416

ABSTRACT

Data mining is the discovery of interesting, unexpected or valuable structures in large datasets. As such, it has two rather different aspects. One of these concerns large-scale, 'global' structures, and the aim is to model the shapes, or features of the shapes, of distributions. The other concerns small-scale, 'local' structures, and the aim is to detect these anomalies and decide if they are real or chance occurrences. In the context of signal detection in the pharmaceutical sector, most interest lies in the second of the above two aspects; however, signal detection occurs relative to an assumed background model, therefore, some discussion of the first aspect is also necessary. This paper gives a lightning overview of data mining and its relation to statistics, with particular emphasis on tools for the detection of adverse drug reactions.


Subject(s)
Adverse Drug Reaction Reporting Systems/organization & administration , Information Systems/organization & administration , Databases, Factual , Drug Industry , Humans , Product Surveillance, Postmarketing
10.
Lifetime Data Anal ; 11(4): 545-64, 2005 Dec.
Article in English | MEDLINE | ID: mdl-16328576

ABSTRACT

The banks have been accumulating huge data bases for many years and are increasingly turning to statistics to provide insight into customer behaviour, among other things. Credit risk is an important issue and certain stochastic models have been developed in recent years to describe and predict loan default. Two of the major models currently used in the industry are considered here, and various ways of extending their application to the case where a loan is repaid in installments are explored. The aspect of interest is the probability distribution of the total loss due to repayment default at some time. Thus, the loss distribution is determined by the distribution of times to default, here regarded as a discrete-time survival distribution. In particular, the probabilities of large losses are to be assessed for insurance purposes.


Subject(s)
Financial Management/statistics & numerical data , Industry/economics , Models, Statistical , Risk Assessment/statistics & numerical data , Commerce , England , Probability , Stochastic Processes , United States
11.
NMR Biomed ; 18(8): 587-94, 2005 Dec.
Article in English | MEDLINE | ID: mdl-16273507

ABSTRACT

A Bayesian nonlinear hierarchical random coefficients model was used in a reanalysis of a previously published longitudinal study of the extracellular direct current (DC)-potential and apparent diffusion coefficient (ADC) responses to focal ischaemia. The main purpose was to examine the data for evidence of an ADC threshold for anoxic depolarisation. A Markov chain Monte Carlo simulation approach was adopted. The Metropolis algorithm was used to generate three parallel Markov chains and thus obtain a sampled posterior probability distribution for each of the DC-potential and ADC model parameters, together with a number of derived parameters. The latter were used in a subsequent threshold analysis. The analysis provided no evidence indicating a consistent and reproducible ADC threshold for anoxic depolarisation.


Subject(s)
Hypoxia-Ischemia, Brain/metabolism , Markov Chains , Algorithms , Animals , Diffusion , Monte Carlo Method , Rats
12.
Proc Natl Acad Sci U S A ; 102(47): 16939-44, 2005 Nov 22.
Article in English | MEDLINE | ID: mdl-16287981

ABSTRACT

We present a method for Bayesian model-based hierarchical coclustering of gene expression data and use it to study the temporal transcription responses of an Anopheles gambiae cell line upon challenge with multiple microbial elicitors. The method fits statistical regression models to the gene expression time series for each experiment and performs coclustering on the genes by optimizing a joint probability model, characterizing gene coregulation between multiple experiments. We compute the model using a two-stage Expectation-Maximization-type algorithm, first fixing the cross-experiment covariance structure and using efficient Bayesian hierarchical clustering to obtain a locally optimal clustering of the gene expression profiles and then, conditional on that clustering, carrying out Bayesian inference on the cross-experiment covariance using Markov chain Monte Carlo simulation to obtain an expectation. For the problem of model choice, we use a cross-validatory approach to decide between individual experiment modeling and varying levels of coclustering. Our method successfully generates tightly coregulated clusters of genes that are implicated in related processes and therefore can be used for analysis of global transcript responses to various stimuli and prediction of gene functions.


Subject(s)
Anopheles/genetics , Anopheles/immunology , Gene Expression/immunology , Algorithms , Animals , Anopheles/drug effects , Anopheles/microbiology , Bayes Theorem , Cell Line , Cluster Analysis , Gene Expression/drug effects , Gene Expression Profiling , Immunity/genetics , Immunity/physiology , Models, Genetic , Zymosan/pharmacology
13.
J Biomed Biotechnol ; 2005(2): 215-25, 2005 Jun 30.
Article in English | MEDLINE | ID: mdl-16046827

ABSTRACT

The vast potential of the genomic insight offered by microarray technologies has led to their widespread use since they were introduced a decade ago. Application areas include gene function discovery, disease diagnosis, and inferring regulatory networks. Microarray experiments enable large-scale, high-throughput investigations of gene activity and have thus provided the data analyst with a distinctive, high-dimensional field of study. Many questions in this field relate to finding subgroups of data profiles which are very similar. A popular type of exploratory tool for finding subgroups is cluster analysis, and many different flavors of algorithms have been used and indeed tailored for microarray data. Cluster analysis, however, implies a partitioning of the entire data set, and this does not always match the objective. Sometimes pattern discovery or bump hunting tools are more appropriate. This paper reviews these various tools for finding interesting subgroups.

14.
J Cereb Blood Flow Metab ; 23(6): 677-88, 2003 Jun.
Article in English | MEDLINE | ID: mdl-12796716

ABSTRACT

Markov chain Monte Carlo simulation was used in a reanalysis of the longitudinal data obtained by Harris et al. (J Cereb Blood Flow Metab 20:28-36) in a study of the direct current (DC) potential and apparent diffusion coefficient (ADC) responses to focal ischemia. The main purpose was to provide a formal analysis of the temporal relationship between the ADC and DC responses, to explore the possible involvement of a common latent (driving) process. A Bayesian nonlinear hierarchical random coefficients model was adopted. DC and ADC transition parameter posterior probability distributions were generated using three parallel Markov chains created using the Metropolis algorithm. Particular attention was paid to the within-subject differences between the DC and ADC time course characteristics. The results show that the DC response is biphasic, whereas the ADC exhibits monophasic behavior, and that the two DC components are each distinguishable from the ADC response in their time dependencies. The DC and ADC changes are not, therefore, driven by a common latent process. This work demonstrates a general analytical approach to the multivariate, longitudinal data-processing problem that commonly arises in stroke and other biomedical research.


Subject(s)
Cerebrovascular Circulation/physiology , Ischemic Attack, Transient/physiopathology , Markov Chains , Models, Cardiovascular , Animals , Computer Simulation , Longitudinal Studies , Rats
15.
Pain ; 16(4): 375-383, 1983 Aug.
Article in English | MEDLINE | ID: mdl-6622047

ABSTRACT

The response profiles on the McGill Pain Questionnaire (MPQ) were compared with those obtained from a checklist format, consisting of the 78 MPQ words arranged in random order. Both forms were administered to 3 patient groups: (a) primiparae experiencing post-episiotomy pain (n = 60); (b) outpatients attending a rheumatology clinic wisdom tooth extraction (n = 60); and (c) inpatients having undergone wisdom tooth extraction (n = 60). The order of administration was balanced, so that within each patient group 40 patients received either one of the study forms and 20 both, yielding total sample sizes of 120 and 60 for further statistical analyses. Comparison of numbers of words checked in the two formats showed considerable similarity and so for purposes of further comparison, the MPQ structure was imposed on the checklist. This permitted comparison of summary scores, with no significant differences in mean level, with the sole exception of the evaluative subscale. Comparison of individual subgroup profiles on both forms also showed considerable similarity. A second objective was to compare the format in discriminating between patient groups. It was found that the MPQ offered a higher correct classification rate, although there was little in it, with MPQ subgroup scores rather than subscale scores showing marginally better results.


Subject(s)
Pain/physiopathology , Surveys and Questionnaires , Adult , Animals , Arthritis, Rheumatoid/physiopathology , Episiotomy/adverse effects , Female , Humans , Male , Middle Aged , Statistics as Topic , Tooth Extraction/adverse effects
SELECTION OF CITATIONS
SEARCH DETAIL
...