Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 15 de 15
Filter
1.
JMIR Med Inform ; 11: e40964, 2023 Feb 24.
Article in English | MEDLINE | ID: mdl-36826984

ABSTRACT

BACKGROUND: Management of abdominal aortic aneurysms (AAAs) requires serial imaging surveillance to evaluate the aneurysm dimension. Natural language processing (NLP) has been previously developed to retrospectively identify patients with AAA from electronic health records (EHRs). However, there are no reported studies that use NLP to identify patients with AAA in near real-time from radiology reports. OBJECTIVE: This study aims to develop and validate a rule-based NLP algorithm for near real-time automatic extraction of AAA diagnosis from radiology reports for case identification. METHODS: The AAA-NLP algorithm was developed and deployed to an EHR big data infrastructure for near real-time processing of radiology reports from May 1, 2019, to September 2020. NLP extracted named entities for AAA case identification and classified subjects as cases and controls. The reference standard to assess algorithm performance was a manual review of processed radiology reports by trained physicians following standardized criteria. Reviewers were blinded to the diagnosis of each subject. The AAA-NLP algorithm was refined in 3 successive iterations. For each iteration, the AAA-NLP algorithm was modified based on performance compared to the reference standard. RESULTS: A total of 360 reports were reviewed, of which 120 radiology reports were randomly selected for each iteration. At each iteration, the AAA-NLP algorithm performance improved. The algorithm identified AAA cases in near real-time with high positive predictive value (0.98), sensitivity (0.95), specificity (0.98), F1 score (0.97), and accuracy (0.97). CONCLUSIONS: Implementation of NLP for accurate identification of AAA cases from radiology reports with high performance in near real time is feasible. This NLP technique will support automated input for patient care and clinical decision support tools for the management of patients with AAA. .

2.
BMC Med Inform Decis Mak ; 22(1): 272, 2022 10 18.
Article in English | MEDLINE | ID: mdl-36258218

ABSTRACT

BACKGROUND: Cardiac magnetic resonance (CMR) imaging is important for diagnosis and risk stratification of hypertrophic cardiomyopathy (HCM) patients. However, collection of information from large numbers of CMR reports by manual review is time-consuming, error-prone and costly. Natural language processing (NLP) is an artificial intelligence method for automated extraction of information from narrative text including text in CMR reports in electronic health records (EHR). Our objective was to assess whether NLP can accurately extract diagnosis of HCM from CMR reports. METHODS: An NLP system with two tiers was developed for information extraction from narrative text in CMR reports; the first tier extracted information regarding HCM diagnosis while the second extracted categorical and numeric concepts for HCM classification. We randomly allocated 200 HCM patients with CMR reports from 2004 to 2018 into training (100 patients with 185 CMR reports) and testing sets (100 patients with 206 reports). RESULTS: NLP algorithms demonstrated very high performance compared to manual annotation. The algorithm to extract HCM diagnosis had accuracy of 0.99. The accuracy for categorical concepts included HCM morphologic subtype 0.99, systolic anterior motion of the mitral valve 0.96, mitral regurgitation 0.93, left ventricular (LV) obstruction 0.94, location of obstruction 0.92, apical pouch 0.98, LV delayed enhancement 0.93, left atrial enlargement 0.99 and right atrial enlargement 0.98. Accuracy for numeric concepts included maximal LV wall thickness 0.96, LV mass 0.99, LV mass index 0.98, LV ejection fraction 0.98 and right ventricular ejection fraction 0.99. CONCLUSIONS: NLP identified and classified HCM from CMR narrative text reports with very high performance.


Subject(s)
Cardiomyopathy, Hypertrophic , Natural Language Processing , Humans , Stroke Volume , Artificial Intelligence , Ventricular Function, Right , Cardiomyopathy, Hypertrophic/diagnostic imaging , Cardiomyopathy, Hypertrophic/pathology , Magnetic Resonance Imaging , Magnetic Resonance Spectroscopy
3.
J Med Internet Res ; 24(8): e27333, 2022 08 22.
Article in English | MEDLINE | ID: mdl-35994324

ABSTRACT

BACKGROUND: Clinical practice guidelines recommend antiplatelet and statin therapies as well as blood pressure control and tobacco cessation for secondary prevention in patients with established atherosclerotic cardiovascular diseases (ASCVDs). However, these strategies for risk modification are underused, especially in rural communities. Moreover, resources to support the delivery of preventive care to rural patients are fewer than those for their urban counterparts. Transformative interventions for the delivery of tailored preventive cardiovascular care to rural patients are needed. OBJECTIVE: A multidisciplinary team developed a rural-specific, team-based model of care intervention assisted by clinical decision support (CDS) technology using participatory design in a sociotechnical conceptual framework. The model of care intervention included redesigned workflows and a novel CDS technology for the coordination and delivery of guideline recommendations by primary care teams in a rural clinic. METHODS: The design of the model of care intervention comprised 3 phases: problem identification, experimentation, and testing. Input from team members (n=35) required 150 hours, including observations of clinical encounters, provider workshops, and interviews with patients and health care professionals. The intervention was prototyped, iteratively refined, and tested with user feedback. In a 3-month pilot trial, 369 patients with ASCVDs were randomized into the control or intervention arm. RESULTS: New workflows and a novel CDS tool were created to identify patients with ASCVDs who had gaps in preventive care and assign the right care team member for delivery of tailored recommendations. During the pilot, the intervention prototype was iteratively refined and tested. The pilot demonstrated feasibility for successful implementation of the sociotechnical intervention as the proportion of patients who had encounters with advanced practice providers (nurse practitioners and physician assistants), pharmacists, or tobacco cessation coaches for the delivery of guideline recommendations in the intervention arm was greater than that in the control arm. CONCLUSIONS: Participatory design and a sociotechnical conceptual framework enabled the development of a rural-specific, team-based model of care intervention assisted by CDS technology for the transformation of preventive health care delivery for ASCVDs.


Subject(s)
Decision Support Systems, Clinical , Rural Population , Ambulatory Care Facilities , Blood Pressure , Humans , Preventive Health Services
4.
Mayo Clin Proc Innov Qual Outcomes ; 5(1): 94-102, 2021 Feb.
Article in English | MEDLINE | ID: mdl-33718788

ABSTRACT

OBJECTIVE: To evaluate usability of a quality improvement tool that promotes guideline-based care for patients with peripheral arterial disease (PAD). PATIENTS AND METHODS: The study was conducted from July 19, 2018, to August 21, 2019. We compared the usability of a PAD cohort knowledge solution (CKS) with standard management supported by an electronic health record (EHR). Two scenarios were developed for usability evaluation; the first for the PAD-CKS while the second evaluated standard EHR workflow. Providers were asked to provide opinions about the PAD-CKS tool and to generate a System Usability Scale (SUS) score. Metrics analyzed included time required, number of mouse clicks, and number of keystrokes. RESULTS: Usability evaluations were completed by 11 providers. SUS for the PAD-CKS was excellent at 89.6. Time required to complete 21 tasks in the CKS was 4 minutes compared with 12 minutes for standard EHR workflow (median, P = .002). Completion of CKS tasks required 34 clicks compared with 148 clicks for the EHR (median, P = .002). Keystrokes for CKS task completion was 8 compared with 72 for EHR (median, P = .004). Providers indicated that overall they found the tool easy to use and the PAD mortality risk score useful. CONCLUSIONS: Usability evaluation of the PAD-CKS tool demonstrated time savings, a high SUS score, and a reduction of mouse clicks and keystrokes for task completion compared to standard workflow using the EHR. Provider feedback regarding the strengths and weaknesses also created opportunities for iterative improvement of the PAD-CKS tool.

5.
J Biomed Inform ; 113: 103660, 2021 01.
Article in English | MEDLINE | ID: mdl-33321199

ABSTRACT

Coronavirus Disease 2019 has emerged as a significant global concern, triggering harsh public health restrictions in a successful bid to curb its exponential growth. As discussion shifts towards relaxation of these restrictions, there is significant concern of second-wave resurgence. The key to managing these outbreaks is early detection and intervention, and yet there is a significant lag time associated with usage of laboratory confirmed cases for surveillance purposes. To address this, syndromic surveillance can be considered to provide a timelier alternative for first-line screening. Existing syndromic surveillance solutions are however typically focused around a known disease and have limited capability to distinguish between outbreaks of individual diseases sharing similar syndromes. This poses a challenge for surveillance of COVID-19 as its active periods tend to overlap temporally with other influenza-like illnesses. In this study we explore performing sentinel syndromic surveillance for COVID-19 and other influenza-like illnesses using a deep learning-based approach. Our methods are based on aberration detection utilizing autoencoders that leverages symptom prevalence distributions to distinguish outbreaks of two ongoing diseases that share similar syndromes, even if they occur concurrently. We first demonstrate that this approach works for detection of outbreaks of influenza, which has known temporal boundaries. We then demonstrate that the autoencoder can be trained to not alert on known and well-managed influenza-like illnesses such as the common cold and influenza. Finally, we applied our approach to 2019-2020 data in the context of a COVID-19 syndromic surveillance task to demonstrate how implementation of such a system could have provided early warning of an outbreak of a novel influenza-like illness that did not match the symptom prevalence profile of influenza and other known influenza-like illnesses.


Subject(s)
COVID-19/epidemiology , Influenza, Human/epidemiology , Sentinel Surveillance , COVID-19/virology , Deep Learning , Disease Outbreaks , Humans , SARS-CoV-2/isolation & purification
6.
medRxiv ; 2020 Jun 09.
Article in English | MEDLINE | ID: mdl-32577704

ABSTRACT

Coronavirus Disease 2019 (COVID-19) has emerged as a significant global concern, triggering harsh public health restrictions in a successful bid to curb its exponential growth. As discussion shifts towards relaxation of these restrictions, there is significant concern of second-wave resurgence. The key to managing these outbreaks is early detection and intervention, and yet there is significant lag time associated with usage of laboratory confirmed cases for surveillance purposes. To address this, syndromic surveillance can be considered to provide a timelier alternative for first-line screening. Existing syndromic surveillance solutions are however typically focused around a known disease and have limited capability to distinguish between outbreaks of individual diseases sharing similar syndromes. This poses a challenge for surveillance of COVID-19 as its active periods are tend to overlap temporally with other influenza-like illnesses. In this study we explore performing sentinel syndromic surveillance for COVID-19 and other influenza-like illnesses using a deep learning-based approach. Our methods are based on aberration detection utilizing autoencoders that leverages symptom prevalence distributions to distinguish outbreaks of two ongoing diseases that share similar syndromes, even if they occur concurrently. We first demonstrate that this approach works for detection of outbreaks of influenza, which has known temporal boundaries. We then demonstrate that the autoencoder can be trained to not alert on known and well-managed influenza-like illnesses such as the common cold and influenza. Finally, we applied our approach to 2019-2020 data in the context of a COVID-19 syndromic surveillance task to demonstrate how implementation of such a system could have provided early warning of an outbreak of a novel influenza-like illness that did not match the symptom prevalence profile of influenza and other known influenza-like illnesses.

7.
NPJ Digit Med ; 2: 130, 2019.
Article in English | MEDLINE | ID: mdl-31872069

ABSTRACT

Data is foundational to high-quality artificial intelligence (AI). Given that a substantial amount of clinically relevant information is embedded in unstructured data, natural language processing (NLP) plays an essential role in extracting valuable information that can benefit decision making, administration reporting, and research. Here, we share several desiderata pertaining to development and usage of NLP systems, derived from two decades of experience implementing clinical NLP at the Mayo Clinic, to inform the healthcare AI community. Using a framework, we developed as an example implementation, the desiderata emphasize the importance of a user-friendly platform, efficient collection of domain expert inputs, seamless integration with clinical data, and a highly scalable computing infrastructure.

8.
Proc IEEE Int Conf Big Data ; 2019: 2800-2807, 2019 Dec.
Article in English | MEDLINE | ID: mdl-38213777

ABSTRACT

The big data revolution has created a hefty demand for searching large-scale electronic health records (EHRs) to support clinical practice, research, and administration. Despite the volume of data involved, fast and accurate identification of clinical narratives pertinent to a clinical case being seen by any given provider is crucial for decision-making at the point of care. In the general domain, this capability is accomplished through a combination of the inverted index data structure, horizontal scaling, and information retrieval (IR) scoring algorithms. These technologies are also being used in the clinical domain, but have met limited success, particularly as clinical cases become more complex. One barrier affecting clinical performance is that contextual information, such as negation, temporality, and the subject of clinical mentions, impact clinical relevance but is not considered in general IR methodologies. In this study, we implemented a solution by identifying and incorporating the aforementioned semantic contexts as part of IR indexing/scoring with Elasticsearch. Experiments were conducted in comparison to baseline approaches with respect to: 1) evaluation of the impact on the quality (relevance) of the returned results, and 2) evaluation of the impact on execution time and storage requirements. The results showed a 5.1-23.1% improvement in retrieval quality, along with achieving 35% faster query execution time. Cost-wise, the solution required 1.5-2 times larger space and about 3 times increase in indexing time. The higher relevance demonstrated the merit of incorporating contextual information into clinical IR, and the near-constant increase in time and space suggested promising scalability.

9.
J Am Heart Assoc ; 7(23): e009680, 2018 12 04.
Article in English | MEDLINE | ID: mdl-30571601

ABSTRACT

Background Automated individualized risk prediction tools linked to electronic health records ( EHR s) are not available for management of patients with peripheral arterial disease. The goal of this study was to create a prognostic tool for patients with peripheral arterial disease using data elements automatically extracted from an EHR to enable real-time and individualized risk prediction at the point of care. Methods and Results A previously validated phenotyping algorithm was deployed to an EHR linked to the Rochester Epidemiology Project to identify peripheral arterial disease cases from Olmsted County, MN, for the years 1998 to 2011. The study cohort was composed of 1676 patients: 593 patients died over 5-year follow-up. The c-statistic for survival in the overall data set was 0.76 (95% confidence interval [CI], 0.74-0.78), and the c-statistic across 10 cross-validation data sets was 0.75 (95% CI, 0.73-0.77). Stratification of cases demonstrated increasing mortality risk by subgroup (low: hazard ratio, 0.35 [95% CI, 0.21-0.58]; intermediate-high: hazard ratio, 2.98 [95% CI, 2.37-3.74]; high: hazard ratio, 8.44 [95% CI, 6.66-10.70], all P<0.0001 versus the reference subgroup). An equation for risk calculation was derived from Cox model parameters and ß estimates. Big data infrastructure enabled deployment of the real-time risk calculator to the point of care via the EHR . Conclusions This study demonstrates that electronic tools can be deployed to EHR s to create automated real-time risk calculators to predict survival of patients with peripheral arterial disease. Moreover, the prognostic model developed may be translated to patient care as an automated and individualized real-time risk calculator deployed at the point of care.


Subject(s)
Diagnosis, Computer-Assisted/methods , Electronic Health Records , Peripheral Arterial Disease/diagnosis , Aged , Algorithms , Female , Humans , Male , Peripheral Arterial Disease/mortality , Prognosis , Proportional Hazards Models , Risk Assessment , Risk Factors
10.
JAMA Netw Open ; 1(8): e185547, 2018 12 07.
Article in English | MEDLINE | ID: mdl-30646276

ABSTRACT

Importance: The prevalence and morbidity of peripheral artery disease (PAD) are high, with limb outcomes including revascularization and amputation. In community-dwelling patients with PAD, the role of noninvasive evaluation for risk assessment and rates of limb outcomes have not been established to date. Objective: To evaluate whether ankle-brachial indices are associated with limb outcomes in community-dwelling patients with PAD. Design, Setting, and Participants: A population-based, observational, test-based cohort study of patients was performed from January 1, 1998, to December 31, 2014. Data analysis was conducted from July 15 to December 15, 2017. Participants included a community-based cohort of 1413 patients with PAD from Olmsted County, Minnesota, identified by validated algorithms deployed to electronic health records. Automated algorithms identified limb outcomes used to build Cox proportional hazards regression models. Ankle-brachial indices and presence of poorly compressible arteries were electronically identified from digital data sets. Guideline-recommended management strategies within 6 months of diagnosis were also electronically retrieved, including therapy with statins, antiplatelet agents, angiotensin-converting enzyme inhibitors or angiotensin-receptor blockers, and smoking abstention. Main Outcomes and Measures: Ankle-brachial index (index ≤0.9 indicates PAD; <.05, severe PAD; and ≥1.40, poorly compressible arteries) and limb revascularization or amputation. Results: Of 1413 patients, 633 (44.8%) were women; mean (SD) age was 70.8 (13.3) years. A total of 283 patients (20.0%) had severe PAD (ankle-brachial indices <0.5) and 350 (24.8%) had poorly compressible arteries (ankle-brachial indices ≥1.4); 780 (55.2%) individuals with less than severe disease formed the reference group. Only 32 of 283 patients (11.3%) with severe disease and 68 of 350 patients (19.4%) with poorly compressible arteries were receiving 4 guideline-recommended management strategies. In the severe disease subgroup, the 1-year event rate for revascularization was 32.4% (90 events); in individuals with poorly compressible arteries, the 1-year amputation rate was 13.9% (47 events). In models adjusted for age, sex, and critical limb ischemia, poorly compressible arteries were associated with amputation (hazard ratio [HR], 3.12; 95% CI, 2.16-4.50; P < .001) but not revascularization (HR, 0.91; 95% CI, 0.69-1.20; P = .49). In contrast, severe disease was associated with revascularization (HR, 2.69; 95% CI, 2.15-3.37; P < .001) but not amputation (HR, 1.30; 95% CI, 0.82-2.07; P = .27). Conclusions and Relevance: Community-dwelling patients with severe PAD or poorly compressible arteries have high rates of revascularization or limb loss, respectively. Guideline-recommended management strategies for secondary risk prevention are underused in the community.


Subject(s)
Amputation, Surgical/statistics & numerical data , Ankle Brachial Index/statistics & numerical data , Peripheral Arterial Disease , Vascular Surgical Procedures/statistics & numerical data , Aged , Aged, 80 and over , Female , Humans , Lower Extremity/blood supply , Lower Extremity/surgery , Male , Middle Aged , Peripheral Arterial Disease/epidemiology , Peripheral Arterial Disease/physiopathology , Peripheral Arterial Disease/surgery , Treatment Outcome
11.
Biomed Inform Insights ; 8(Suppl 1): 13-22, 2016.
Article in English | MEDLINE | ID: mdl-27385912

ABSTRACT

The concept of optimizing health care by understanding and generating knowledge from previous evidence, ie, the Learning Health-care System (LHS), has gained momentum and now has national prominence. Meanwhile, the rapid adoption of electronic health records (EHRs) enables the data collection required to form the basis for facilitating LHS. A prerequisite for using EHR data within the LHS is an infrastructure that enables access to EHR data longitudinally for health-care analytics and real time for knowledge delivery. Additionally, significant clinical information is embedded in the free text, making natural language processing (NLP) an essential component in implementing an LHS. Herein, we share our institutional implementation of a big data-empowered clinical NLP infrastructure, which not only enables health-care analytics but also has real-time NLP processing capability. The infrastructure has been utilized for multiple institutional projects including the MayoExpertAdvisor, an individualized care recommendation solution for clinical care. We compared the advantages of big data over two other environments. Big data infrastructure significantly outperformed other infrastructure in terms of computing speed, demonstrating its value in making the LHS a possibility in the near future.

12.
J Am Med Inform Assoc ; 23(2): 248-56, 2016 Mar.
Article in English | MEDLINE | ID: mdl-26568604

ABSTRACT

OBJECTIVE: The objective of the Strategic Health IT Advanced Research Project area four (SHARPn) was to develop open-source tools that could be used for the normalization of electronic health record (EHR) data for secondary use--specifically, for high throughput phenotyping. We describe the role of Intermountain Healthcare's Clinical Element Models ([CEMs] Intermountain Healthcare Health Services, Inc, Salt Lake City, Utah) as normalization "targets" within the project. MATERIALS AND METHODS: Intermountain's CEMs were either repurposed or created for the SHARPn project. A CEM describes "valid" structure and semantics for a particular kind of clinical data. CEMs are expressed in a computable syntax that can be compiled into implementation artifacts. The modeling team and SHARPn colleagues agilely gathered requirements and developed and refined models. RESULTS: Twenty-eight "statement" models (analogous to "classes") and numerous "component" CEMs and their associated terminology were repurposed or developed to satisfy SHARPn high throughput phenotyping requirements. Model (structural) mappings and terminology (semantic) mappings were also created. Source data instances were normalized to CEM-conformant data and stored in CEM instance databases. A model browser and request site were built to facilitate the development. DISCUSSION: The modeling efforts demonstrated the need to address context differences and granularity choices and highlighted the inevitability of iso-semantic models. The need for content expertise and "intelligent" content tooling was also underscored. We discuss scalability and sustainability expectations for a CEM-based approach and describe the place of CEMs relative to other current efforts. CONCLUSIONS: The SHARPn effort demonstrated the normalization and secondary use of EHR data. CEMs proved capable of capturing data originating from a variety of sources within the normalization pipeline and serving as suitable normalization targets.


Subject(s)
Electronic Health Records/standards , Information Storage and Retrieval , Medical Record Linkage/methods , Health Information Systems/standards , Semantics , Utah , Vocabulary, Controlled
13.
J Am Med Inform Assoc ; 20(e2): e341-8, 2013 Dec.
Article in English | MEDLINE | ID: mdl-24190931

ABSTRACT

RESEARCH OBJECTIVE: To develop scalable informatics infrastructure for normalization of both structured and unstructured electronic health record (EHR) data into a unified, concept-based model for high-throughput phenotype extraction. MATERIALS AND METHODS: Software tools and applications were developed to extract information from EHRs. Representative and convenience samples of both structured and unstructured data from two EHR systems-Mayo Clinic and Intermountain Healthcare-were used for development and validation. Extracted information was standardized and normalized to meaningful use (MU) conformant terminology and value set standards using Clinical Element Models (CEMs). These resources were used to demonstrate semi-automatic execution of MU clinical-quality measures modeled using the Quality Data Model (QDM) and an open-source rules engine. RESULTS: Using CEMs and open-source natural language processing and terminology services engines-namely, Apache clinical Text Analysis and Knowledge Extraction System (cTAKES) and Common Terminology Services (CTS2)-we developed a data-normalization platform that ensures data security, end-to-end connectivity, and reliable data flow within and across institutions. We demonstrated the applicability of this platform by executing a QDM-based MU quality measure that determines the percentage of patients between 18 and 75 years with diabetes whose most recent low-density lipoprotein cholesterol test result during the measurement year was <100 mg/dL on a randomly selected cohort of 273 Mayo Clinic patients. The platform identified 21 and 18 patients for the denominator and numerator of the quality measure, respectively. Validation results indicate that all identified patients meet the QDM-based criteria. CONCLUSIONS: End-to-end automated systems for extracting clinical information from diverse EHR systems require extensive use of standardized vocabularies and terminologies, as well as robust information models for storing, discovering, and processing that information. This study demonstrates the application of modular and open-source resources for enabling secondary use of EHR data through normalization into standards-based, comparable, and consistent format for high-throughput phenotyping to identify patient cohorts.


Subject(s)
Data Mining , Electronic Health Records/standards , Medical Informatics Applications , Natural Language Processing , Phenotype , Algorithms , Biomedical Research , Computer Security , Humans , Software , Vocabulary, Controlled
14.
Am J Public Health ; 103(3): 448-9, 2013 Mar.
Article in English | MEDLINE | ID: mdl-23327237

ABSTRACT

Health disparities and solutions are heterogeneous within and among racial and ethnic groups, yet existing administrative databases lack the granularity to reflect important sociocultural distinctions. We measured the efficacy of a natural-language-processing algorithm to identify a specific immigrant group. The algorithm demonstrated accuracy and precision in identifying Somali patients from the electronic medical records at a single institution. This technology holds promise to identify and track immigrants and refugees in the United States in local health care settings.


Subject(s)
Health Status Disparities , Natural Language Processing , Algorithms , Electronic Health Records/statistics & numerical data , Humans , Minnesota/epidemiology , Refugees/statistics & numerical data , Somalia/ethnology
15.
J Biomed Semantics ; 4(1): 1, 2013 Jan 03.
Article in English | MEDLINE | ID: mdl-23286462

ABSTRACT

BACKGROUND: One challenge in reusing clinical data stored in electronic medical records is that these data are heterogenous. Clinical Natural Language Processing (NLP) plays an important role in transforming information in clinical text to a standard representation that is comparable and interoperable. Information may be processed and shared when a type system specifies the allowable data structures. Therefore, we aim to define a common type system for clinical NLP that enables interoperability between structured and unstructured data generated in different clinical settings. RESULTS: We describe a common type system for clinical NLP that has an end target of deep semantics based on Clinical Element Models (CEMs), thus interoperating with structured data and accommodating diverse NLP approaches. The type system has been implemented in UIMA (Unstructured Information Management Architecture) and is fully functional in a popular open-source clinical NLP system, cTAKES (clinical Text Analysis and Knowledge Extraction System) versions 2.0 and later. CONCLUSIONS: We have created a type system that targets deep semantics, thereby allowing for NLP systems to encapsulate knowledge from text and share it alongside heterogenous clinical data sources. Rather than surface semantics that are typically the end product of NLP algorithms, CEM-based semantics explicitly build in deep clinical semantics as the point of interoperability with more structured data types.

SELECTION OF CITATIONS
SEARCH DETAIL
...