Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 19.242
Filter
1.
BMC Bioinformatics ; 25(1): 180, 2024 May 08.
Article in English | MEDLINE | ID: mdl-38720249

ABSTRACT

BACKGROUND: High-throughput sequencing (HTS) has become the gold standard approach for variant analysis in cancer research. However, somatic variants may occur at low fractions due to contamination from normal cells or tumor heterogeneity; this poses a significant challenge for standard HTS analysis pipelines. The problem is exacerbated in scenarios with minimal tumor DNA, such as circulating tumor DNA in plasma. Assessing sensitivity and detection of HTS approaches in such cases is paramount, but time-consuming and expensive: specialized experimental protocols and a sufficient quantity of samples are required for processing and analysis. To overcome these limitations, we propose a new computational approach specifically designed for the generation of artificial datasets suitable for this task, simulating ultra-deep targeted sequencing data with low-fraction variants and demonstrating their effectiveness in benchmarking low-fraction variant calling. RESULTS: Our approach enables the generation of artificial raw reads that mimic real data without relying on pre-existing data by using NEAT, a fine-grained read simulator that generates artificial datasets using models learned from multiple different datasets. Then, it incorporates low-fraction variants to simulate somatic mutations in samples with minimal tumor DNA content. To prove the suitability of the created artificial datasets for low-fraction variant calling benchmarking, we used them as ground truth to evaluate the performance of widely-used variant calling algorithms: they allowed us to define tuned parameter values of major variant callers, considerably improving their detection of very low-fraction variants. CONCLUSIONS: Our findings highlight both the pivotal role of our approach in creating adequate artificial datasets with low tumor fraction, facilitating rapid prototyping and benchmarking of algorithms for such dataset type, as well as the important need of advancing low-fraction variant calling techniques.


Subject(s)
Benchmarking , High-Throughput Nucleotide Sequencing , Neoplasms , High-Throughput Nucleotide Sequencing/methods , Humans , Neoplasms/genetics , Mutation , Algorithms , DNA, Neoplasm/genetics , Sequence Analysis, DNA/methods , Computational Biology/methods
2.
BMC Genom Data ; 25(1): 45, 2024 May 07.
Article in English | MEDLINE | ID: mdl-38714942

ABSTRACT

OBJECTIVES: Cellular deconvolution is a valuable computational process that can infer the cellular composition of heterogeneous tissue samples from bulk RNA-sequencing data. Benchmark testing is a crucial step in the development and evaluation of new cellular deconvolution algorithms, and also plays a key role in the process of building and optimizing deconvolution pipelines for specific experimental applications. However, few in vivo benchmarking datasets exist, particularly for whole blood, which is the single most profiled human tissue. Here, we describe a unique dataset containing whole blood gene expression profiles and matched circulating leukocyte counts from a large cohort of human donors with utility for benchmarking cellular deconvolution pipelines. DATA DESCRIPTION: To produce this dataset, venous whole blood was sampled from 138 total donors recruited at an academic medical center. Genome-wide expression profiling was subsequently performed via next-generation RNA sequencing, and white blood cell differentials were collected in parallel using flow cytometry. The resultant final dataset contains donor-level expression data for over 45,000 protein coding and non-protein coding genes, as well as matched neutrophil, lymphocyte, monocyte, and eosinophil counts.


Subject(s)
Benchmarking , Humans , Leukocyte Count , Gene Expression Profiling/methods , Transcriptome , Sequence Analysis, RNA/methods , Leukocytes/metabolism , High-Throughput Nucleotide Sequencing , Algorithms
3.
PLoS One ; 19(5): e0301696, 2024.
Article in English | MEDLINE | ID: mdl-38781237

ABSTRACT

In the domain of question subjectivity classification, there exists a need for detailed datasets that can foster advancements in Automatic Subjective Question Answering (ASQA) systems. Addressing the prevailing research gaps, this paper introduces the Fine-Grained Question Subjectivity Dataset (FQSD), which comprises 10,000 questions. The dataset distinguishes between subjective and objective questions and offers additional categorizations such as Subjective-types (Target, Attitude, Reason, Yes/No, None) and Comparison-form (Single, Comparative). Annotation reliability was confirmed via robust evaluation techniques, yielding a Fleiss's Kappa score of 0.76 and Pearson correlation values up to 0.80 among three annotators. We benchmarked FQSD against existing datasets such as (Yu, Zha, and Chua 2012), SubjQA (Bjerva 2020), and ConvEx-DS (Hernandez-Bocanegra 2021). Our dataset excelled in scale, linguistic diversity, and syntactic complexity, establishing a new standard for future research. We employed visual methodologies to provide a nuanced understanding of the dataset and its classes. Utilizing transformer-based models like BERT, XLNET, and RoBERTa for validation, RoBERTa achieved an outstanding F1-score of 97%, confirming the dataset's efficacy for the advanced subjectivity classification task. Furthermore, we utilized Local Interpretable Model-agnostic Explanations (LIME) to elucidate model decision-making, ensuring transparent and reliable model predictions in subjectivity classification tasks.


Subject(s)
Benchmarking , Humans , Benchmarking/methods , Reproducibility of Results
4.
Nat Commun ; 15(1): 4134, 2024 May 16.
Article in English | MEDLINE | ID: mdl-38755121

ABSTRACT

Defining the number and abundance of different cell types in tissues is important for understanding disease mechanisms as well as for diagnostic and prognostic purposes. Typically, this is achieved by immunohistological analyses, cell sorting, or single-cell RNA-sequencing. Alternatively, cell-specific DNA methylome information can be leveraged to deconvolve cell fractions from a bulk DNA mixture. However, comprehensive benchmarking of deconvolution methods and modalities was not yet performed. Here we evaluate 16 deconvolution algorithms, developed either specifically for DNA methylome data or more generically. We assess the performance of these algorithms, and the effect of normalization methods, while modeling variables that impact deconvolution performance, including cell abundance, cell type similarity, reference panel size, method for methylome profiling (array or sequencing), and technical variation. We observe differences in algorithm performance depending on each these variables, emphasizing the need for tailoring deconvolution analyses. The complexity of the reference, marker selection method, number of marker loci and, for sequencing-based assays, sequencing depth have a marked influence on performance. By developing handles to select the optimal analysis configuration, we provide a valuable source of information for studies aiming to deconvolve array- or sequencing-based methylation data.


Subject(s)
Algorithms , Benchmarking , DNA Methylation , Epigenome , Humans , Sequence Analysis, DNA/methods , DNA/genetics , High-Throughput Nucleotide Sequencing/methods
5.
Br J Surg ; 111(5)2024 May 03.
Article in English | MEDLINE | ID: mdl-38747683

ABSTRACT

BACKGROUND: Clinical auditing is a powerful tool to evaluate and improve healthcare. Deviations from the expected quality of care are identified by benchmarking the results of individual hospitals using national averages. This study aimed to evaluate the use of quality indicators for benchmarking hepato-pancreato-biliary (HPB) surgery and when outlier hospitals could be identified. METHODS: A population-based study used data from two nationwide Dutch HPB audits (DHBA and DPCA) from 2014 to 2021. Sample size calculations determined the threshold (in percentage points) to identify centres as statistical outliers, based on current volume requirements (annual minimum of 20 resections) on a two-year period (2020-2021), covering mortality rate, failure to rescue (FTR), major morbidity rate and textbook/ideal outcome (TO) for minor liver resection (LR), major LR, pancreaticoduodenectomy (PD) and distal pancreatectomy (DP). RESULTS: In total, 10 963 and 7365 patients who underwent liver and pancreatic resection respectively were included. Benchmark and corresponding range of mortality rates were 0.6% (0 -3.2%) and 3.3% (0-16.7%) for minor and major LR, and 2.7% (0-7.0%) and 0.6% (0-4.2%) for PD and DP respectively. FTR rates were 5.4% (0-33.3%), 14.2% (0-100%), 7.5% (1.6%-28.5%) and 3.1% (0-14.9%). For major morbidity rate, corresponding rates were 9.8% (0-20.5%), 28.1% (0-47.1%), 36% (15.8%-58.3%) and 22.3% (5.2%-46.1%). For TO, corresponding rates were 73.6% (61.3%-94.4%), 54.1% (35.3-100), 46.8% (25.3%-59.4%) and 63.3% (30.7%-84.6%). Mortality rate thresholds indicating a significant outlier were 8.6% and 15.4% for minor and major LR and 14.2% and 8.6% for PD and DP. For FTR, these thresholds were 17.9%, 31.6%, 22.9% and 15.0%. For major morbidity rate, these thresholds were 26.1%, 49.7%, 57.9% and 52.9% respectively. For TO, lower thresholds were 52.5%, 32.5%, 25.8% and 41.4% respectively. Higher hospital volumes decrease thresholds to detect outliers. CONCLUSION: Current event rates and minimum volume requirements per hospital are too low to detect any meaningful between hospital differences in mortality rate and FTR. Major morbidity rate and TO are better candidates to use for benchmarking.


Subject(s)
Benchmarking , Quality Indicators, Health Care , Humans , Netherlands/epidemiology , Pancreatectomy/standards , Pancreatectomy/mortality , Male , Pancreaticoduodenectomy/standards , Pancreaticoduodenectomy/mortality , Hepatectomy/mortality , Hepatectomy/standards , Female , Middle Aged , Aged , Hospital Mortality
6.
PLoS One ; 19(5): e0302696, 2024.
Article in English | MEDLINE | ID: mdl-38753612

ABSTRACT

Pathway enrichment analysis is a ubiquitous computational biology method to interpret a list of genes (typically derived from the association of large-scale omics data with phenotypes of interest) in terms of higher-level, predefined gene sets that share biological function, chromosomal location, or other common features. Among many tools developed so far, Gene Set Enrichment Analysis (GSEA) stands out as one of the pioneering and most widely used methods. Although originally developed for microarray data, GSEA is nowadays extensively utilized for RNA-seq data analysis. Here, we quantitatively assessed the performance of a variety of GSEA modalities and provide guidance in the practical use of GSEA in RNA-seq experiments. We leveraged harmonized RNA-seq datasets available from The Cancer Genome Atlas (TCGA) in combination with large, curated pathway collections from the Molecular Signatures Database to obtain cancer-type-specific target pathway lists across multiple cancer types. We carried out a detailed analysis of GSEA performance using both gene-set and phenotype permutations combined with four different choices for the Kolmogorov-Smirnov enrichment statistic. Based on our benchmarks, we conclude that the classic/unweighted gene-set permutation approach offered comparable or better sensitivity-vs-specificity tradeoffs across cancer types compared with other, more complex and computationally intensive permutation methods. Finally, we analyzed other large cohorts for thyroid cancer and hepatocellular carcinoma. We utilized a new consensus metric, the Enrichment Evidence Score (EES), which showed a remarkable agreement between pathways identified in TCGA and those from other sources, despite differences in cancer etiology. This finding suggests an EES-based strategy to identify a core set of pathways that may be complemented by an expanded set of pathways for downstream exploratory analysis. This work fills the existing gap in current guidelines and benchmarks for the use of GSEA with RNA-seq data and provides a framework to enable detailed benchmarking of other RNA-seq-based pathway analysis tools.


Subject(s)
Benchmarking , RNA-Seq , Humans , RNA-Seq/methods , Computational Biology/methods , Neoplasms/genetics , Databases, Genetic , Gene Expression Profiling/methods
7.
Sci Rep ; 14(1): 11438, 2024 05 20.
Article in English | MEDLINE | ID: mdl-38763952

ABSTRACT

The utilization of artificial intelligence (AI) in healthcare is on the rise, demanding increased accessibility to (public) medical data for benchmarking. The digitization of healthcare in recent years has facilitated medical data scientists' access to extensive hospital data, fostering AI-based research. A notable addition to this trend is the Salzburg Intensive Care database (SICdb), made publicly available in early 2023. Covering over 27 thousand intensive care admissions at the University Hospital Salzburg from 2013 to 2021, this dataset presents a valuable resource for AI-driven investigations. This article explores the SICdb and conducts a comparative analysis with the widely recognized Medical Information Mart for Intensive Care - version IV (MIMIC-IV) database. The comparison focuses on key aspects, emphasizing the availability and granularity of data provided by the SICdb, particularly vital signs and laboratory measurements. The analysis demonstrates that the SICdb offers more detailed information with higher data availability and temporal resolution for signal data, especially for vital signs, compared to the MIMIC-IV. This is advantageous for longitudinal studies of patients' health conditions in the intensive care unit. The SICdb provides a valuable resource for medical data scientists and researchers. The database offers comprehensive and diverse healthcare data in a European country, making it well suited for benchmarking and enhancing AI-based healthcare research. The importance of ongoing efforts to expand and make public datasets available for advancing AI applications in the healthcare domain is emphasized by the findings.


Subject(s)
Critical Care , Databases, Factual , Intensive Care Units , Humans , Artificial Intelligence , Male , Female , Aged , Middle Aged , Adult , Aged, 80 and over , Benchmarking , Adolescent
8.
Eur J Epidemiol ; 39(4): 349-361, 2024 Apr.
Article in English | MEDLINE | ID: mdl-38717556

ABSTRACT

Prospective benchmarking of an observational analysis against a randomized trial increases confidence in the benchmarking process as it relies exclusively on aligning the protocol of the trial and the observational analysis, while the trials findings are unavailable. The Randomized Evaluation of Decreased Usage of Betablockers After Myocardial Infarction (REDUCE-AMI, ClinicalTrials.gov ID: NCT03278509) trial started recruitment in September 2017 and results are expected in 2024. REDUCE-AMI aimed to estimate the effect of long-term use of beta blockers on the risk of death and myocardial following a myocardial infarction with preserved left ventricular systolic ejection fraction. We specified the protocol of a target trial as similar as possible to that of REDUCE-AMI, then emulated the target trial using observational data from Swedish healthcare registries. Had everyone followed the treatment strategy as specified in the target trial protocol, the observational analysis estimated a reduction in the 5-year risk of death or myocardial infarction of 0.8 percentage points for beta blockers compared with no beta blockers; effects ranging from an absolute reduction of 4.5 percentage points to an increase of 2.8 percentage points in the risk of death or myocardial infarction were compatible with our data under conventional statistical criteria. Once results of REDUCE-AMI are published, we will compare the results of our observational analysis against those from the trial. If this prospective benchmarking is successful, it supports the credibility of additional analyses using these observational data, which can rapidly deliver answers to questions that could not be answered by the initial trial. If benchmarking proves unsuccessful, we will conduct a "postmortem" analysis to identify the reasons for the discrepancy. Prospective benchmarking shifts the investigator focus away from an endeavour to use observational data to obtain similar results as a completed randomized trial, to a systematic attempt to align the design and analysis of the trial and the observational analysis.


Subject(s)
Adrenergic beta-Antagonists , Benchmarking , Myocardial Infarction , Registries , Humans , Sweden , Prospective Studies , Adrenergic beta-Antagonists/therapeutic use , Female , Male , Aged , Randomized Controlled Trials as Topic , Middle Aged
9.
J Environ Manage ; 359: 121017, 2024 May.
Article in English | MEDLINE | ID: mdl-38718602

ABSTRACT

Energy transition currently brings focus on fuel cell micro-combined heat and power (mCHP) systems for residential uses. The two main technologies already commercialized are the Proton Exchange Membrane Fuel Cells (PEMFCs) and Solid Oxide Fuel Cells (SOFCs). The pollutant emissions of one system of each technology have been tested with a portable probe both in laboratory and field-test configurations. In this paper, the nitrogen oxides (NOx), sulphur dioxide (SO2), and carbon monoxide (CO) emission levels are compared to other combustion technologies such as a recent Euro 6 diesel automotive vehicle, a classical gas condensing boiler, and a gas absorption heat pump. At last, a method of converting the concentration of pollutants (in ppm) measured by the sensors into pollutant intensity per unit of energy (in mg/kWh) is documented and reported. This allows for comparing the pollutant emissions levels with relevant literature, especially other studies conducted with other measuring sensors. Both tested residential fuel cell technologies fed by natural gas can be considered clean regarding SO2 and NOx emissions. The CO emissions can be considered quite low for the tested SOFC and even nil for the tested PEMFC. The biggest issue of natural gas fuel cell technologies still lies in the carbon dioxide (CO2) emissions associated with the fossil fuel they consume. The gas absorption heat pump however shows worse NOx and CO levels than the classical gas condensing boiler. At last, this study illustrates that the high level of hybridization between a fuel cell and a gas boiler may be responsible for unexpected ON/OFF cycling behaviours and therefore prevent both sub-systems from operating as optimally and reliably as they would have as standalone units.


Subject(s)
Air Pollutants , Nitrogen Oxides , Air Pollutants/analysis , Nitrogen Oxides/analysis , Carbon Monoxide/analysis , Sulfur Dioxide/analysis , Benchmarking , Vehicle Emissions/analysis , Environmental Monitoring/methods
10.
Elife ; 122024 May 24.
Article in English | MEDLINE | ID: mdl-38787371

ABSTRACT

Spatial transcriptomics (ST) technologies allow the profiling of the transcriptome of cells while keeping their spatial context. Since most commercial untargeted ST technologies do not yet operate at single-cell resolution, computational methods such as deconvolution are often used to infer the cell type composition of each sequenced spot. We benchmarked 11 deconvolution methods using 63 silver standards, 3 gold standards, and 2 case studies on liver and melanoma tissues. We developed a simulation engine called synthspot to generate silver standards from single-cell RNA-sequencing data, while gold standards are generated by pooling single cells from targeted ST data. We evaluated methods based on their performance, stability across different reference datasets, and scalability. We found that cell2location and RCTD are the top-performing methods, but surprisingly, a simple regression model outperforms almost half of the dedicated spatial deconvolution methods. Furthermore, we observe that the performance of all methods significantly decreased in datasets with highly abundant or rare cell types. Our results are reproducible in a Nextflow pipeline, which also allows users to generate synthetic data, run deconvolution methods and optionally benchmark them on their dataset (https://github.com/saeyslab/spotless-benchmark).


Subject(s)
Benchmarking , Gene Expression Profiling , Transcriptome , Humans , Gene Expression Profiling/methods , Single-Cell Analysis/methods , Software , Computational Biology/methods , Sequence Analysis, RNA/methods , Melanoma/genetics , Reproducibility of Results , Liver
11.
Nat Commun ; 15(1): 4376, 2024 May 23.
Article in English | MEDLINE | ID: mdl-38782890

ABSTRACT

Topologically associating domains (TADs), megabase-scale features of chromatin spatial architecture, are organized in a domain-within-domain TAD hierarchy. Within TADs, the inner and smaller subTADs not only manifest cell-to-cell variability, but also precisely regulate transcription and differentiation. Although over 20 TAD callers are able to detect TAD, their usability in biomedicine is confined by a disagreement of outputs and a limit in understanding TAD hierarchy. We compare 13 computational tools across various conditions and develop a metric to evaluate the similarity of TAD hierarchy. Although outputs of TAD hierarchy at each level vary among callers, data resolutions, sequencing depths, and matrices normalization, they are more consistent when they have a higher similarity of larger TADs. We present comprehensive benchmarking of TAD hierarchy callers and operational guidance to researchers of life science researchers. Moreover, by simulating the mixing of different types of cells, we confirm that TAD hierarchy is generated not simply from stacking Hi-C heatmaps of heterogeneous cells. Finally, we propose an air conditioner model to decipher the role of TAD hierarchy in transcription.


Subject(s)
Benchmarking , Chromatin , Chromatin/chemistry , Humans , Computational Biology/methods , Software , Chromatin Assembly and Disassembly
12.
Brief Bioinform ; 25(4)2024 May 23.
Article in English | MEDLINE | ID: mdl-38797968

ABSTRACT

A major challenge of precision oncology is the identification and prioritization of suitable treatment options based on molecular biomarkers of the considered tumor. In pursuit of this goal, large cancer cell line panels have successfully been studied to elucidate the relationship between cellular features and treatment response. Due to the high dimensionality of these datasets, machine learning (ML) is commonly used for their analysis. However, choosing a suitable algorithm and set of input features can be challenging. We performed a comprehensive benchmarking of ML methods and dimension reduction (DR) techniques for predicting drug response metrics. Using the Genomics of Drug Sensitivity in Cancer cell line panel, we trained random forests, neural networks, boosting trees and elastic nets for 179 anti-cancer compounds with feature sets derived from nine DR approaches. We compare the results regarding statistical performance, runtime and interpretability. Additionally, we provide strategies for assessing model performance compared with a simple baseline model and measuring the trade-off between models of different complexity. Lastly, we show that complex ML models benefit from using an optimized DR strategy, and that standard models-even when using considerably fewer features-can still be superior in performance.


Subject(s)
Algorithms , Antineoplastic Agents , Benchmarking , Machine Learning , Humans , Antineoplastic Agents/pharmacology , Antineoplastic Agents/therapeutic use , Neoplasms/drug therapy , Neoplasms/genetics , Neural Networks, Computer , Cell Line, Tumor
13.
Brief Bioinform ; 25(4)2024 May 23.
Article in English | MEDLINE | ID: mdl-38796691

ABSTRACT

Limited gene capture efficiency and spot size of spatial transcriptome (ST) data pose significant challenges in cell-type characterization. The heterogeneity and complexity of cell composition in the mammalian brain make it more challenging to accurately annotate ST data from brain. Many algorithms attempt to characterize subtypes of neuron by integrating ST data with single-nucleus RNA sequencing (snRNA-seq) or single-cell RNA sequencing. However, assessing the accuracy of these algorithms on Stereo-seq ST data remains unresolved. Here, we benchmarked 9 mapping algorithms using 10 ST datasets from four mouse brain regions in two different resolutions and 24 pseudo-ST datasets from snRNA-seq. Both actual ST data and pseudo-ST data were mapped using snRNA-seq datasets from the corresponding brain regions as reference data. After comparing the performance across different areas and resolutions of the mouse brain, we have reached the conclusion that both robust cell-type decomposition and SpatialDWLS demonstrated superior robustness and accuracy in cell-type annotation. Testing with publicly available snRNA-seq data from another sequencing platform in the cortex region further validated our conclusions. Altogether, we developed a workflow for assessing suitability of mapping algorithm that fits for ST datasets, which can improve the efficiency and accuracy of spatial data annotation.


Subject(s)
Algorithms , Benchmarking , Brain , Single-Cell Analysis , Animals , Mice , Brain/metabolism , Single-Cell Analysis/methods , RNA-Seq/methods , Transcriptome , Sequence Analysis, RNA/methods , Neurons/metabolism , Gene Expression Profiling/methods
14.
J Healthc Manag ; 69(3): 178-189, 2024.
Article in English | MEDLINE | ID: mdl-38728544

ABSTRACT

GOAL: A lack of improvement in productivity in recent years may be the result of suboptimal measurement of productivity. Hospitals and clinics benefit from external benchmarks that allow assessment of clinical productivity. Work relative value units have long served as a common currency for this purpose. Productivity is determined by comparing work relative value units to full-time equivalents (FTEs), but FTEs do not have a universal or standardized definition, which could cause problems. We propose a new clinical labor input measure-"clinic time"-as a substitute for using the reported measure of FTEs. METHODS: In this observational validation study, we used data from a cluster randomized trial to compare FTE with clinic time. We compared these two productivity measures graphically. For validation, we estimated two separate ordinary least squares (OLS) regression models. To validate and simultaneously adjust for endogeneity, we used instrumental variables (IV) regression with the proportion of days in a pay period that were federal holidays as an instrument. We used productivity data collected between 2018 and 2020 from Veterans Health Administration (VA) cardiology and orthopedics providers as part of a 2-year cluster randomized trial of medical scribes mandated by the VA Maintaining Internal Systems and Strengthening Integrated Outside Networks (MISSION) Act of 2018. PRINCIPAL FINDINGS: Our cohort included 654 unique providers. For both productivity variables, the values for patients per clinic day were consistently higher than those for patients per day per FTE. To validate these measures, we estimated separate OLS and IV regression models, predicting wait times from the two productivity measures. The slopes from the two productivity measures were positive and small in magnitude with OLS, but negative and large in magnitude with IV regression. The magnitude of the slope for patients per clinic day was much larger than the slope for patients per day per FTE. Current metrics that rely on FTE data may suffer from self-report bias and low reporting frequency. Using clinic time as an alternative is an effective way to mitigate these biases. PRACTICAL APPLICATIONS: Measuring productivity accurately is essential because provider productivity plays an important role in facilitating clinic operations outcomes. Most importantly, tracking a more valid productivity metric is a concrete, cost-effective management tactic to improve the provision of care in the long term.


Subject(s)
Efficiency, Organizational , Humans , United States , Efficiency , United States Department of Veterans Affairs , Benchmarking , Female , Relative Value Scales , Male
15.
J Healthc Manag ; 69(3): 219-230, 2024.
Article in English | MEDLINE | ID: mdl-38728547

ABSTRACT

GOAL: Boarding emergency department (ED) patients is associated with reductions in quality of care, patient safety and experience, and ED operational efficiency. However, ED boarding is ultimately reflective of inefficiencies in hospital capacity management. The ability of a hospital to accommodate variability in patient flow presumably affects its financial performance, but this relationship is not well studied. We investigated the relationship between ED boarding and hospital financial performance measures. Our objective was to see if there was an association between key financial measures of business performance and limitations in patient progression efficiency, as evidenced by ED boarding. METHODS: Cross-sectional ED operational data were collected from the Emergency Department Benchmarking Alliance, a voluntarily self-reporting operational database that includes 54% of EDs in the United States. Freestanding EDs, pediatric EDs and EDs with missing boarding data were excluded. The key operational outcome variable was boarding time. We reviewed the financial information of these nonprofit institutions by accessing their Internal Revenue Service Form 990. We examined standard measures of financial performance, including return on equity, total margin, total asset turnover, and equity multiplier (EM). We studied these associations using quantile regressions of added ED volume, ED admission percentage, urban versus nonurban ED site location, trauma status, and percentage of the population receiving Medicare and Medicaid as covariates in the regression models. PRINCIPAL FINDINGS: Operational data were available for 892 EDs from 31 states. Of those, 127 reported a Form 990 in the year corresponding to the ED boarding measures. Median boarding time across EDs was 148 min (interquartile range [IQR]: 100-216). A significant relationship exists between boarding and the EM, along with a negative association with the hospital's total profit margin in the highest-performing hospitals (by profit margin percentage). After adjusting for the covariates in the regression model, we found that for every 10 min above 90 min of boarding, the mean EM for the top quartile increased from 245.8% to 249.5% (p < .001). In hospitals in the top 90th percentile of total margin, every 10 min beyond the median ED boarding interval led to a decrease in total margin of 0.24%. PRACTICAL APPLICATIONS: Using the largest available national registry of ED operational data and concordant nonprofit financial reports, higher boarding among the highest-profitability hospitals (i.e., top 10%) is associated with a drag on profit margin, while hospitals with the highest boarding are associated with the highest leverage (i.e., indicated by the EM). These relationships suggest an association between a key ED indicator of hospital capacity management and overall institutional financial performance.


Subject(s)
Efficiency, Organizational , Emergency Service, Hospital , Emergency Service, Hospital/statistics & numerical data , Emergency Service, Hospital/economics , Cross-Sectional Studies , United States , Humans , Efficiency, Organizational/economics , Benchmarking
16.
PLoS One ; 19(5): e0301720, 2024.
Article in English | MEDLINE | ID: mdl-38739583

ABSTRACT

A key benefit of the Open Computing Language (OpenCL) software framework is its capability to operate across diverse architectures. Field programmable gate arrays (FPGAs) are a high-speed computing architecture used for computation acceleration. This study investigates the impact of memory access time on overall performance in general FPGA computing environments through the creation of eight benchmarks within the OpenCL framework. The developed benchmarks capture a range of memory access behaviors, and they play a crucial role in assessing the performance of spinning and sleeping on FPGA-based architectures. The results obtained guide the formulation of new implementations and contribute to defining an abstraction of FPGAs. This abstraction is then utilized to create tailored implementations of primitives that are well-suited for this platform. While other research endeavors concentrate on creating benchmarks with the Compute Unified Device Architecture (CUDA) to scrutinize the memory systems across diverse GPU architectures and propose recommendations for future generations of GPU computation platforms, this study delves into the memory system analysis for the broader FPGA computing platform. It achieves this by employing the highly abstracted OpenCL framework, exploring various data workload characteristics, and experimentally delineating the appropriate implementation of primitives that can seamlessly integrate into a design tailored for the FPGA computing platform. Additionally, the results underscore the efficacy of employing a task-parallel model to mitigate the need for high-cost synchronization mechanisms in designs constructed on general FPGA computing platforms.


Subject(s)
Benchmarking , Software , Humans , Programming Languages
17.
BMC Public Health ; 24(1): 1245, 2024 May 06.
Article in English | MEDLINE | ID: mdl-38711106

ABSTRACT

OBJECTIVE: To benchmark the university food environment and explore students' experiences with food insecurity and healthy eating in order to inform interventions to improve access and affordability of healthy foods for university students. DESIGN: A food environment audit was conducted on the university campus using the Uni-Food tool from April to May 2022 and was comprised of three main components, university systems and governance, campus facilities and environment, and food retail outlets. A qualitative study design was also used to conduct focus groups and semi-structured interviews with students to explore key themes regarding their experiences with food insecurity and healthy eating. SETTING: Macquarie University, Australia. PARTICIPANTS: For the food environment audit 24 retail outlets on campus and for the qualitative component 29 domestic and international students enrolled at Macquarie University. RESULTS: The university only scored 27% in total for all components in the food environment audit. The results showed the need for better governance and leadership of the food environment. The qualitative component suggested that the main barriers to accessing healthy foods were related to availability, pricing, and knowledge of healthy foods. Future intervention ideas included free fruits and vegetables, food relief, discounts, improved self-catering facilities, education, and increased healthy food outlets. CONCLUSIONS: Improving governance measures related to healthy eating on campus are a core priority to strengthen the food environment and students identified pricing and availability as key issues. These findings will inform effective and feasible interventions to improve food security and healthy eating on campus.


Subject(s)
Benchmarking , Diet, Healthy , Food Insecurity , Qualitative Research , Students , Humans , Universities , Students/psychology , Students/statistics & numerical data , Diet, Healthy/psychology , Female , Male , Australia , Young Adult , Focus Groups , Adult , Organizational Case Studies , Food Supply/statistics & numerical data
18.
Nat Commun ; 15(1): 4055, 2024 May 14.
Article in English | MEDLINE | ID: mdl-38744843

ABSTRACT

We introduce GRouNdGAN, a gene regulatory network (GRN)-guided reference-based causal implicit generative model for simulating single-cell RNA-seq data, in silico perturbation experiments, and benchmarking GRN inference methods. Through the imposition of a user-defined GRN in its architecture, GRouNdGAN simulates steady-state and transient-state single-cell datasets where genes are causally expressed under the control of their regulating transcription factors (TFs). Training on six experimental reference datasets, we show that our model captures non-linear TF-gene dependencies and preserves gene identities, cell trajectories, pseudo-time ordering, and technical and biological noise, with no user manipulation and only implicit parameterization. GRouNdGAN can synthesize cells under new conditions to perform in silico TF knockout experiments. Benchmarking various GRN inference algorithms reveals that GRouNdGAN effectively bridges the existing gap between simulated and biological data benchmarks of GRN inference algorithms, providing gold standard ground truth GRNs and realistic cells corresponding to the biological system of interest.


Subject(s)
Algorithms , Computer Simulation , Gene Regulatory Networks , RNA-Seq , Single-Cell Analysis , Single-Cell Analysis/methods , RNA-Seq/methods , Humans , Transcription Factors/metabolism , Transcription Factors/genetics , Computational Biology/methods , Benchmarking , Sequence Analysis, RNA/methods , Single-Cell Gene Expression Analysis
19.
Genome Biol ; 25(1): 97, 2024 Apr 15.
Article in English | MEDLINE | ID: mdl-38622738

ABSTRACT

BACKGROUND: As most viruses remain uncultivated, metagenomics is currently the main method for virus discovery. Detecting viruses in metagenomic data is not trivial. In the past few years, many bioinformatic virus identification tools have been developed for this task, making it challenging to choose the right tools, parameters, and cutoffs. As all these tools measure different biological signals, and use different algorithms and training and reference databases, it is imperative to conduct an independent benchmarking to give users objective guidance. RESULTS: We compare the performance of nine state-of-the-art virus identification tools in thirteen modes on eight paired viral and microbial datasets from three distinct biomes, including a new complex dataset from Antarctic coastal waters. The tools have highly variable true positive rates (0-97%) and false positive rates (0-30%). PPR-Meta best distinguishes viral from microbial contigs, followed by DeepVirFinder, VirSorter2, and VIBRANT. Different tools identify different subsets of the benchmarking data and all tools, except for Sourmash, find unique viral contigs. Performance of tools improved with adjusted parameter cutoffs, indicating that adjustment of parameter cutoffs before usage should be considered. CONCLUSIONS: Together, our independent benchmarking facilitates selecting choices of bioinformatic virus identification tools and gives suggestions for parameter adjustments to viromics researchers.


Subject(s)
Benchmarking , Viruses , Metagenome , Ecosystem , Metagenomics/methods , Computational Biology/methods , Databases, Genetic , Viruses/genetics
20.
PLoS One ; 19(4): e0290706, 2024.
Article in English | MEDLINE | ID: mdl-38625859

ABSTRACT

In many applications, artificial neural networks are best trained for a task by following a curriculum, in which simpler concepts are learned before more complex ones. This curriculum can be hand-crafted by the engineer or optimised like other hyperparameters, by evaluating many curricula. However, this is computationally intensive and the hyperparameters are unlikely to generalise to new datasets. An attractive alternative, demonstrated in influential prior works, is that the network could choose its own curriculum by monitoring its learning. This would be particularly beneficial for continual learning, in which the network must learn from an environment that is changing over time, relevant both to practical applications and in the modelling of human development. In this paper we test the generality of this approach using a proof-of-principle model, training a network on two sequential tasks under static and continual conditions, and investigating both the benefits of a curriculum and the handicap induced by continuous learning. Additionally, we test a variety of prior task-switching metrics, and find that in some cases even in this simple scenario the a network is often unable to choose the optimal curriculum, as the benefits are sometimes only apparent with hindsight, at the end of training. We discuss the implications of the results for network engineering and models of human development.


Subject(s)
Curriculum , Neural Networks, Computer , Humans , Upper Extremity , Education, Continuing , Benchmarking
SELECTION OF CITATIONS
SEARCH DETAIL
...