Background: Eating disorders affect an increasing number of people. Social networks provide information that can help. Objective: We aimed to find machine learning models capable of efficiently categorizing tweets abo...
详细信息
Diversity in use of Question and Answering (Q/A) is evolving as a popular application in the area of Natural Language Processing (NLP). The alive unsupervised word embedding approaches are efficient to collect Latent-...
详细信息
Diversity in use of Question and Answering (Q/A) is evolving as a popular application in the area of Natural Language Processing (NLP). The alive unsupervised word embedding approaches are efficient to collect Latent-Semantic data on number of tasks. But certain methods are still unable to tackle issues such as polysemous-unaware with task-unaware phenomena in NLP tasks. GloVe understands word embedding by availing information statistics from word co-occurrence matrices. Nevertheless, word-pairs in the matrices are taken from a pre-established window of local context, which may result in constrained word-pairs and also probably semantic inappropriate word-pairs. SemGloVe employed in this paper, refines semantic co-occurrences from BERT into static GloVe word-embedding with bidirectional-Long-Short-Term-Memory (BERT- Bi-LSTM) model for text categorization in Q/A. This method utilizes the CR23K and CR1000k datasets for the effective text classification of NLP. The proposed model, with SemGloVe Embedding on BERT combined with Bi-LSTM, produced better results on metrics like accuracy, precision, recall, and F1 Score as 0.92, 0.79, 0.85, and 0.73, respectively, when compared to existing methods of Text2GraphQL, GPT-2, BERT and SPARQL. The BERT model with Bi-LSTM is better in every way for responding to different kinds of questions.
In recent decades, abstractive text summarization using multimodal input has attracted many researchers due to the capability of gathering information from various sources to create a concise summary. However, the exi...
详细信息
In recent decades, abstractive text summarization using multimodal input has attracted many researchers due to the capability of gathering information from various sources to create a concise summary. However, the existing methodologies based on multimodal summarization provide only a summary for the short videos and poor results for the lengthy videos. To address the aforementioned issues, this research presented the Multimodal Abstractive Summarization using bidirectional encoder representations from transformers (MAS-BERT) with an attention mechanism. The purpose of the video summarization is to increase the speed of searching for a large collection of videos so that the users can quickly decide whether the video is relevant or not by reading the summary. Initially, the data is obtained from the publicly available How2 dataset and is encoded using the bidirectional Gated Recurrent Unit (Bi-GRU) encoder and the Long Short Term Memory (LSTM) encoder. The textual data which is embedded in the embedding layer is encoded using a bidirectional GRU encoder and the features with audio and video data are encoded with LSTM encoder. After this, BERT based attention mechanism is used to combine the modalities and finally, the BI-GRU based decoder is used for summarizing the multimodalities. The results obtained through the experiments that show the proposed MAS-BERT has achieved a better result of 60.2 for Rouge-1 whereas, the existing Decoder-only Multimodal transformer (DMmT) and the Factorized Multimodal transformer based Decoder Only Language model (FLORAL) has achieved 49.58 and 56.89 respectively. Our work facilitates users by providing better contextual information and user experience and would help video-sharing platforms for customer retention by allowing users to search for relevant videos by looking at its summary.
Product life cycles continue to shorten with the rapid pace of technological innovation and intensifying global competition. This escalation in speed, quality, and cost demands for new product development can be met b...
详细信息
Product life cycles continue to shorten with the rapid pace of technological innovation and intensifying global competition. This escalation in speed, quality, and cost demands for new product development can be met by quickly identifying and reflecting customer requirements (CRs) based on quality function deployment (QFD). This study presents a new approach to overcome the limitations of the traditional qualitative methods in QFD by utilizing large-scale online product review data, considering the importance, topics, and context to apply optimal natural language processing techniques. Extracting CRs using term frequency-inverse document frequency (TF-IDF), topic modeling, and bidirectional encoder representations from transformers (BERT), followed by summarizing the extracted CRs into sentences using generative artificial intelligence, enables a more precise analysis of online product review data without human intervention. This approach allows for the swift and accurate incorporation of CRs into product development. Implementing a review-specialized BERT, which understands the characteristics of review language, showed superior multi-class classification performance by at least 1% across all aspects-precision, recall, F1-score, and accuracy-compared to the base BERT.
This paper introduces the concept of a smart menstrual cup, incorporating advanced technology to address challenges associated with traditional menstrual health management methods. The proposed method, Hybrid Bidirect...
详细信息
This paper introduces the concept of a smart menstrual cup, incorporating advanced technology to address challenges associated with traditional menstrual health management methods. The proposed method, Hybrid bidirectionalencoder Generative transformer based Bees Search (Hybrid BEGT-BS), aims to enhance the longevity and reliability of smart menstrual cups. The Generative Pre-trained transformer (GPT) is employed to extract key information and determine emotional tones related to menstruation, while the bidirectional encoder representations from transformer (BERT) classifies menstrual-related data such as cycle tracking, hygiene, and symptoms. Further different validation metrics such as F1-score, specificity, Area Under the Curve-Receiver Operating Characteristic (AUC-ROC), Mean Squared Error (MSE), recall, energy consumption, accuracy, and precision are employed to assess the method's effectiveness. from the comparative results it demonstrates the superior performance of the Hybrid BEGT-BS method in enhancing the performance of smart menstrual cups.
Background: A challenge in updating systematic reviews is the workload in screening the articles. Many screening models using natural language processing technology have been implemented to scrutinize articles based o...
详细信息
Background: A challenge in updating systematic reviews is the workload in screening the articles. Many screening models using natural language processing technology have been implemented to scrutinize articles based on titles and abstracts. While these approaches show promise, traditional models typically treat abstracts as uniform text. We hypothesize that selective training on specific abstract components could enhance model performance for systematic review screening. Objective: We evaluated the efficacy of a novel screening model that selects specific components from abstracts to improve performance and developed an automatic systematic review update model using an abstract component classifier to categorize abstracts based on their components. Methods: A screening model was created based on the included and excluded articles in the existing systematic review used as the scheme for the automatic update of the systematic review. A prior publication was selected for the systematic review, and articlesincluded or excluded in the articles screening process were used as training data. Thetitles and abstracts were classified into 5 categories (Title, Introduction, Methods, Results, and Conclusion). Thirty-one component-composition datasets created by combining 5 component datasets. We implemented 31 screening models using the component-composition datasets and compared their performances. Comparisons were conducted using 3 pretrained models: bidirectional encoder representations from transformer (BERT), BioLinkBERT, and BioM-Efficiently Learning an encoder that Classifies Token Replacements Accurately (ELECTRA). Moreover, to automate the component selection of abstracts, we developed the Abstract Component Classifier Model and created component datasets using this classifier model classification. Using the component datasets classified using the Abstract Component Classifier Model, we created 10 component-composition datasets used by the top 10 screening models with t
The widespread use of social media and its development have offered a medium for the propagation of fake contents quickly among the masses. Fake contents frequently misguide individuals and lead to erroneous social ju...
详细信息
The widespread use of social media and its development have offered a medium for the propagation of fake contents quickly among the masses. Fake contents frequently misguide individuals and lead to erroneous social judgments. Individuals and society have been harmed by the dissemination of low-quality news content on social media. In this paper, we have worked on a benchmark dataset of news content and proposed an approach comprising basic natural language processing techniques with different deep learning models for categorising content as real or fake. Different deep learning models employed are LSTM, bi-LSTM, LSTM and bi-LSTM with an attention mechanism. We compared the outcomes by using one hot word embedding and pre-trained GloVe technique. On benchmark LIAR dataset, the LSTM achieved a better accuracy of 67.2%, while the bi-LSTM with GloVe word embedding reached an accuracy of 67%. An accuracy of 98.22% is achieved using bi-LSTM and 97.98% using LSTM on Real-Fake dataset. Fake news can be a menace to society, so if it is detected early, harmony can be maintained in society and individuals can avoid being misled.
Text categorization (TC) is one of the most useful automatic tools in today's world to organize huge text data automatically. It is widely used by practitioners to classify texts automatically for different purpos...
详细信息
Text categorization (TC) is one of the most useful automatic tools in today's world to organize huge text data automatically. It is widely used by practitioners to classify texts automatically for different purposes, including sentiment analysis, authorship detection, spam detection, and so on. However, studying TC task for different fields can be challenging since it is required to train a separate model on a labeled and large data set specific to that field. This is very time-consuming, and creating a domain-specific large and labeled data is often very hard. In order to overcome this problem, language models are recently employed to transfer learned information from a large data to another downstream task. bidirectional encoder representations from transformer (BERT) is one of the most popular language models and has been shown to provide very good results for TC tasks. Hence, in this study, we use four pretrained BERT models trained on formal text data as well as our own BERT models trained on Facebook messages. We then fine-tuned BERT models on different downstream data sets collected from different domains such as Twitter, Instagram, and so on. We aim to investigate whether fine-tuned BERT models can provide satisfying results on different downstream tasks of different domains via transfer learning. The results of our extensive experiments show that BERT models provide very satisfying results and selecting both the BERT model and downstream tasks' data from the same or similar domain is akin to improve the performance in a further direction. This shows that a well-trained language model can remove the need for a separate training process for each different downstream TC task within the OSN domain.
Modern home energy management systems (HEMSs) have great flexibility of energy consumption for customers, but at the same time, bear a range of problems, such as the high system complexity, uncertainty and time-varyin...
详细信息
Modern home energy management systems (HEMSs) have great flexibility of energy consumption for customers, but at the same time, bear a range of problems, such as the high system complexity, uncertainty and time-varying nature of load consumptions, and renewable sources generation. This has brought great challenges for the real-time control. To solve these problems, we propose an HEMS that integrates a kernel-based real-time adaptive dynamic programming (K-RT-ADP) with a new preprocessing short-term prediction technique. For the preprocessing short-term prediction, we propose a gated recurrent unit-bidirectionalencoderrepresentationsfrom the transformer (GRU-BERT) model to improve the forecasting accuracy of electrical loads and renewable energy generation. In particular, we classify household appliances into the temperature-sensitive loads, human activity sensitive loads, and insensitive/constant loads. The GRU-BERT model can incorporate weather and human activity information to predict load consumption and solar generation. For real-time control, we propose and employ the K-RT-ADP HEMS based on the GRU-BERT prediction algorithm. The objective of the K-RT-ADP HEMS is to minimize the electricity cost and maximize the solar energy utilization. To enhance the nonlinear approximation ability and generalization ability of the adaptive dynamic programming (ADP) algorithm, the K-RT-ADP algorithm leverages kernel mapping instead of neural networks. Hardware-in-the-loop experiments demonstrate the superiority of the proposed K-RT-ADP HEMS over the traditional ADP control through comparison.
Dialogue system is one of the research area coming into picture because of advancement in natural language processing and deep learning methods. Dialogue systems are designed for communication between humans and machi...
详细信息
ISBN:
(纸本)9783031837920;9783031837937
Dialogue system is one of the research area coming into picture because of advancement in natural language processing and deep learning methods. Dialogue systems are designed for communication between humans and machine. When humans communicate with each other they use their own intelligence to carry conversation but this intelligence is missing in machines. Researchers have attempted to accommodate external knowledge with machines to generate knowledge-enhanced responses. Knowledge graph is one of the structured ways of providing an abstraction of the real world knowledge to the machine, and machine in turn can use this knowledge to improve the quality of response generated by dialogue systems. Generating knowledge grounded response is a challenging task. Recently most of the architectures are end-to-end dialogue system, in contrast to them this paper proposes three step architecture which extracts entity from input using inside outside beginning 2 tagging and bidirectional encoder representations from transformers, secondly entity related sub-graph is extracted using laplacian matrix method then knowledge grounded response are generated using extracted subgraph and Gated recurrent unit encoder-decoder model. This architecture has independent fact retrieval system which is detached from two tunable NER model and response generation model which makes the model training easy as compared to end-to-end trainable system and also improves the overall performance of the system. Proposed model is tested on standard benchmark dataset In-car and shows performance comparable with existing models.
暂无评论