Efficiently treating cardiac patients before the onset of a heart attack relies on the precise prediction of heart disease. Identifying and detecting the risk factors for heart disease such as diabetes mellitus, Coron...
详细信息
Efficiently treating cardiac patients before the onset of a heart attack relies on the precise prediction of heart disease. Identifying and detecting the risk factors for heart disease such as diabetes mellitus, Coronary Artery Disease (CAD), hyperlipidemia, hypertension, smoking, familial CAD history, obesity, and medications is critical for developing effective preventative and management measures. Although Electronic Health Records (EHRs) have emerged as valuable resources for identifying these risk factors, their unstructured format poses challenges for cardiologists in retrieving relevant information. This research proposed employing transfer learning techniques to automatically extract heart disease risk factors from EHRs. Leveraging transfer learning, a deep learning technique has demonstrated a significant performance in various clinical natural language processing (NLP) applications, particularly in heart disease risk prediction. This study explored the application of transformer-based language models, specifically utilizing pre-trained architectures like BERT (bidirectional encoder representations from transformers), RoBERTa, BioClinicalBERT, XLNet, and BioBERT for heart disease detection and extraction of related risk factors from clinical notes, using the i2b2 dataset. These transformer models are pre-trained on an extensive corpus of medical literature and clinical records to gain a deep understanding of contextualized language representations. Adapted models are then fine-tuned using annotated datasets specific to heart disease, such as the i2b2 dataset, enabling them to learn patterns and relationships within the domain. These models have demonstrated superior performance in extracting semantic information from EHRs, automating high-performance heart disease risk factor identification, and performing downstream NLP tasks within the clinical domain. This study proposed fine-tuned five widely used transformer-based models, namely BERT, RoBERTa, BioClini
The rapid expansion of online content and big data has precipitated an urgent need for efficient summarization techniques to swiftly comprehend vast textual documents without compromising their original *** approaches...
详细信息
The rapid expansion of online content and big data has precipitated an urgent need for efficient summarization techniques to swiftly comprehend vast textual documents without compromising their original *** approaches in Extractive Text Summarization(ETS)leverage the modeling of inter-sentence relationships,a task of paramount importance in producing coherent *** study introduces an innovative model that integrates Graph Attention Networks(GATs)with Transformer-based bidirectionalencoder Representa-tions fromtransformers(BERT)and Latent Dirichlet Allocation(LDA),further enhanced by Term Frequency-Inverse Document Frequency(TF-IDF)values,to improve sentence selection by capturing comprehensive topical *** approach constructs a graph with nodes representing sentences,words,and topics,thereby elevating the interconnectivity and enabling a more refined understanding of text *** model is stretched to Multi-Document Summarization(MDS)from Single-Document Summarization,offering significant improvements over existing models such as THGS-GMM and Topic-GraphSum,as demonstrated by empirical evaluations on benchmark news datasets like Cable News Network(CNN)/Daily Mail(DM)and *** results consistently demonstrate superior performance,showcasing the model’s robustness in handling complex summarization tasks across single and multi-document *** research not only advances the integration of BERT and LDA within a GATs but also emphasizes our model’s capacity to effectively manage global information and adapt to diverse summarization challenges.
This article combined BERT (bidirectionalencoder Representation fromtransformers), Bi-LSTM (bidirectional Long Short-Term Memory), and CRF (Conditional Random Field) models to transform unstructured legal text into ...
详细信息
This article combined BERT (bidirectionalencoder Representation fromtransformers), Bi-LSTM (bidirectional Long Short-Term Memory), and CRF (Conditional Random Field) models to transform unstructured legal text into structured data through information extraction, improving the effectiveness of legal information extraction. The BERT model can be used for deep semantic embedding of legal texts, generating context-sensitive representations for each word. The Bi-LSTM network can capture long-distance dependencies in the text, extract sequence features, and apply CRF layers to globally optimize sequence labels to ensure accurate annotation of entity boundaries and relationships. In the dataset for extracting legal entity relationships related to prostitution constructed in this article, the accuracy, precision, recall rate, and F1 score of entity classification reached 93.6%, 92.7%, 92.1%, and 92.4%, respectively. All 153 samples in the Engage_in_ prostitution relationship were correctly classified. In order to analyze the stability of legal information extraction and classification, the model proposed by this article was tested on five datasets: CAIL2019, CJRC (Chinese Judicial Reading Comprehension), LexGLUE (Legal General Language Understanding Evaluation), COLIEE (Competition on Legal Information Extraction/Appointment), and ECHR (European Court of Human Rights). The accuracy of the article's model fluctuated only 1.2% on different datasets, while the precision remained stable and the recall fluctuated by 0.7%. This article provided reliable technical support for legal intelligence research by combining BERT, Bi-LSTM, and CRF to accurately extract and classify legal information.
Traditional research on preschool language development often fails to capture the complex nonlinear relationships and high-dimensional characteristics of language growth, leading to low prediction accuracy and poor cr...
详细信息
Traditional research on preschool language development often fails to capture the complex nonlinear relationships and high-dimensional characteristics of language growth, leading to low prediction accuracy and poor cross-cultural applicability. This paper introduces a novel BERT (bidirectional encoder representations from transformers)-based model to predict preschool language development and evaluate its cross-cultural effectiveness. Text data from preschool children's language datasets across multiple cultural backgrounds is collected, cleaned, and preprocessed to create suitable training samples. Special attention is given to the unique grammatical structures and cultural expressions in each language to ensure compatibility with the model. The BERT model is used to encode the processed text, leveraging its bidirectional self-attention mechanism to extract contextual information and generate deep feature representations essential for understanding preschool language development. The model combines both grammatical and semantic features for meaningful representations in subsequent predictions. Fine-tuning the pre-trained BERT model using the Adam optimizer enhances prediction accuracy, while cross-validation and hyperparameter tuning further improve its performance. Culturally specific annotations and vocabularies are incorporated to ensure the model's effective prediction of language development across different regions. Experimental results show that the BERT model achieves an MAE (Mean Absolute Error) between 0.20 and 0.25, an MSE (Mean Squared Error) between 0.05 and 0.08, and an average R2 value of 0.84 across English, Chinese, Spanish, and Japanese. These results demonstrate the model's high accuracy and strong cross-cultural stability in predicting preschool language development.
Event-based surveillance is crucial for the early detection and rapid response to potential public health risks. In recent years, social networking services (SNS) have been recognized for their potential role in this ...
详细信息
Event-based surveillance is crucial for the early detection and rapid response to potential public health risks. In recent years, social networking services (SNS) have been recognized for their potential role in this domain. Previous studies have demonstrated the capacity of SNS posts for the early detection of health crises and affected individuals, including those related to infectious diseases. However, the reliability of such posts, being subjective and not clinically diagnosed, remains a challenge. In this study, we address this issue by assessing the classification performance of transformer-based pretrained language models to accurately classify Japanese tweets related to heat stroke, a significant health effect of climate change, as true or false. We also evaluated the efficacy of combining SNS and artificial intelligence for event-based public health surveillance by visualizing the data on correctly classified tweets and heat stroke emergency medical evacuees in time-space and animated video, respectively. The transformer-based pretrained language models exhibited good performance in classifying the tweets. Spatiotemporal and animated video visualizations revealed a reasonable correlation. This study demonstrates the potential of using Japanese tweets and deep learning algorithms based on transformer networks for event-based surveillance at high spatiotemporal levels to enable early detection of heat stroke risks.
Background: Artificial intelligence (AI), more specifically large language models (LLMs), holds significant potential in revolutionizing emergency care delivery by optimizing clinical workflows and enhancing the quali...
详细信息
Background: Artificial intelligence (AI), more specifically large language models (LLMs), holds significant potential in revolutionizing emergency care delivery by optimizing clinical workflows and enhancing the quality of decision-making. Although enthusiasm for integrating LLMs into emergency medicine (EM) is growing, the existing literature is characterized by a disparate collection of individual studies, conceptual analyses, and preliminary implementations. Given these complexities and gaps in understanding, a cohesive framework is needed to comprehend the existing body of knowledge on the application of LLMs in Objective: Given the absence of a comprehensive framework for exploring the roles of LLMs in EM, this scoping review aims to systematically map the existing literature on LLMs' potential applications within EM and identify directions for future research. Addressing this gap will allow for informed advancements in the field. Methods: Using PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) criteria, we searched Ovid MEDLINE, Embase, Web of Science, and Google Scholar for papers published between January 2018 and August 2023 that discussed LLMs' use in EM. We excluded other forms of AI. A total of 1994 unique titles and abstracts were screened, and each full-text paper was independently reviewed by 2 authors. Data were abstracted independently, and 5 authors performed a collaborative quantitative and qualitative synthesis of the data. Results: A total of 43 papers were included. Studies were predominantly from 2022 to 2023 and conducted in the United States and China. We uncovered four major themes: (1) clinical decision-making and support was highlighted as a pivotal area, with LLMs playing a substantial role in enhancing patient care, notably through their application in real-time triage, allowing early recognition of patient urgency;(2) efficiency, workflow, and information management demonstrat
Background: Obesity is a chronic, multifactorial, and relapsing disease, affecting people of all ages worldwide, and is directly related to multiple complications. Understanding public attitudes and perceptions toward...
详细信息
Background: Obesity is a chronic, multifactorial, and relapsing disease, affecting people of all ages worldwide, and is directly related to multiple complications. Understanding public attitudes and perceptions toward obesity is essential for developing effective health policies, prevention strategies, and treatment approaches. Objective: This study investigated the sentiments of the general public, celebrities, and important organizations regarding obesity using social media data, specifically from Twitter (subsequently rebranded as X).Methods: The study analyzes a dataset of 53,414 tweets related to obesity posted on Twitter during the COVID-19 pandemic, from April 2019 to December 2022. Sentiment analysis was performed using the XLM-RoBERTa-base model, and topic modeling was conducted using the BERTopic library. Results: The analysis revealed that tweets regarding obesity were predominantly negative. Spikes in Twitter activity correlated with significant political events, such as the exchange of obesity-related comments between US politicians and criticism of theUnited Kingdom's obesity campaign. Topic modeling identified 243 clusters representing various obesity-related topics, such as childhood obesity;the US President's obesity struggle;COVID-19 vaccinations;the UK government's obesity campaign;body shaming;racism and high obesity rates among Black American people;smoking, substance abuse, and alcohol consumption among people with obesity;environmental risk factors;and surgical treatments. Conclusions: Twitter serves as a valuable source for understanding obesity-related sentiments and attitudes among the public, celebrities, and influential organizations. Sentiments regarding obesity were predominantly negative. Negative portrayals of obesity by influential politicians and celebrities were shown to contribute to negative public sentiments, which can have adverse effects on public health. It is essential for public figures to be mindful of their impact on publ
This research proposes a novel method for enhancing the prediction of microbe-disease associations and the identification of possible drug treatments through the utilization of advanced computational strategies. The f...
详细信息
This research proposes a novel method for enhancing the prediction of microbe-disease associations and the identification of possible drug treatments through the utilization of advanced computational strategies. The foundation of this approach is the integration of data from the Human Microbe-Disease Association Database with advanced techniques such as Conditional Tabular Generative Adversarial Networks for augmenting data and a combination of Graph Neural Networks and Graph Attention Networks for modeling the intricate relationships between microbes and diseases. Additionally, this study leverages the bidirectional encoder representations from transformers for the encoding of diseases and drugs, thereby improving the prediction capabilities for disease-drug associations. A thorough experimental analysis validates the efficacy of these approaches in decoding complex biological datasets, offering valuable insights into disease etiology and potential treatment pathways. The results highlight the significance of merging generative models, graph-based deep learning, and transformer-based natural language processing models to push forward the boundaries of biomedical research and its applications.
Sentiment or opinion largely relies on public commentary, where reflections are either positive or negative. Sentiment classification automates the process of determining the orientation of a subject based on text, wh...
详细信息
Sentiment or opinion largely relies on public commentary, where reflections are either positive or negative. Sentiment classification automates the process of determining the orientation of a subject based on text, which aims to classify documents by expressed views. This task faces significant challenges due to issues, like negation, ambiguity, and complex language structures. In this work, a new Squirrel Search Mayfly Algorithm_Hierarchical Deep Learning for Text (SSMA_HDLTex) is established for sentiment classification using aspect term extraction and optimized deep learning. Primarily, the document on Amazon review is taken as input and later it is applied to the process of tokenization. In this process, bidirectional encoder representations from transformers is employed to partition the sentence into tokens. Later, Aspect Term Extraction is effectuated. At last, sentiment classification is done by employing HDLTex which is trained by utilizing the presented SSMA method. The novel SSMA method is devised by integrating the Squirrel Search Algorithm as well as the Mayfly Algorithm. The proposed SSMA_HDLTex has attained maximal and precision, F-measure, and recall, of 0.936, 0.937, and 0.941correspondingly. This innovative approach significantly enhances the accuracy and reliability of sentiment classification tasks.
The cross-domain aspect detection and categorization is a vital task in natural language processing, enabling the automated identification and categorization of aspects in textual data spanning diverse domains. Tradit...
详细信息
The cross-domain aspect detection and categorization is a vital task in natural language processing, enabling the automated identification and categorization of aspects in textual data spanning diverse domains. Traditional methods face several complexities such as scalability, limited contextual understanding of words, data sparsity, and adaptation difficulty. So, a novel method named gated bidirectional recurrent encoder-based adaptive honey badger (GBRE-AHB) algorithm is proposed for cross-domain aspect detection and categorization. In this study, the bidirectionalencoderrepresentations for transformers (BERT) is utilized to capture contextual information from text and enable better aspect identification and categorization by understanding the context. The local optimization problems are identified and solved by determining an adaptive strategy. Also, the gated recurrent unit (GRU) is employed to sequence the text data and allow aspect detection by considering the sequence in which aspects appear within a document. The study is validated on the datasets, namely the IMDB dataset of 50 K movie reviews and the cell phone reviews sentiment analysis-body dataset. The efficiency is validated by various metrics that attained the ranges as F1-score (97.85%), specificity (97.83%), recall (97.82%), precision (97.94%), and accuracy (98.67%), respectively. The experimental results revealed that the proposed method for cross-domain aspect detection and categorization as well as improved the reliability as well as longevity of the model and determined the impacts applied in the categorization process.
暂无评论