This work presents a novel systematic methodology to analyse the capabilities and limitations of Large language Models (LLMs) with feedback from a formal inference engine, on logic theory induction. The analysis is co...
详细信息
Recent progress in the interdisciplinary studies of computer vision (CV) and naturallanguageprocessing (NLP) has enabled the development of intelligent systems that can describe what they see and answer questions ac...
详细信息
ISBN:
(数字)9781665469463
ISBN:
(纸本)9781665469463
Recent progress in the interdisciplinary studies of computer vision (CV) and naturallanguageprocessing (NLP) has enabled the development of intelligent systems that can describe what they see and answer questions accordingly. However, despite showing usefulness in performing these vision-language tasks, existing methods still struggle in understanding real-life problems (i.e., how to do something) and suggesting step-by-step guidance to solve them. With an overarching goal of developing intelligent systems to assist humans in various daily activities, we propose VisualHow, a free form and open-ended research that focuses on understanding a real-life problem and deriving its solution by incorporating key components across multiple modalities. We develop a new dataset with 20,028 real-life problems and 102,933 steps that constitute their solutions, where each step consists of both a visual illustration and a textual description that guide the problem solving. To establish better understanding of problems and solutions, we also provide annotations of multimodal attention that localizes important components across modalities and solution graphs that encapsulate different steps in structured representations. These data and annotations enable a family of new vision-language tasks that solve real-life problems. Through extensive experiments with representative models, we demonstrate their effectiveness on training and testing models for the new tasks, and there is significant scope for improvement by learning effective attention mechanisms.
Alzheimer's disease (AD) is a progressive neurodegenerative disorder that affects millions worldwide, leading to cognitive decline and placing a substantial burden on families and healthcare systems. Early detecti...
详细信息
Part-of-speech tagging is a fundamental task that provides the elemental structure and content information for additional naturallanguageprocessing. Although Part-of-speech tagging problems have traditionally been f...
详细信息
ISBN:
(纸本)9781665421973
Part-of-speech tagging is a fundamental task that provides the elemental structure and content information for additional naturallanguageprocessing. Although Part-of-speech tagging problems have traditionally been formulated as sequential labeling tasks, none of the proposed ensemble approaches have focused on sequence alignment during post-processing. Herein, we present a weighted ensemble technique using a sequence alignment approach for a Part-of-speech tagger. Through this technique, we introduce a simple but powerful post-processor, which is a sub-sequence selector using a similarity score calculated through sequence alignment methods. Such methods are based on an existing DNA alignment approach applied toward naturallanguage. Experiments were conducted using an ensemble of sequence alignment methods with three different sub-sequence units, i.e., the sequence, word, and character span. Experiments on English and Korean datasets show that our sequence alignment ensemble technique outperforms a basic hard voting method. Most of the results of the ensemble sequence alignment approach with various sub-sequence units showed an increase in F1-score over hard voting. F1-score increased up to 0.36 for the general hard voting method on the test dataset.
Math Word Problems (MWPs) represent a critical area of research in naturallanguageprocessing and artificial intelligence, with the goal of translating naturallanguage descriptions into mathematical expressions and ...
详细信息
ISBN:
(数字)9798331519827
ISBN:
(纸本)9798331519834
Math Word Problems (MWPs) represent a critical area of research in naturallanguageprocessing and artificial intelligence, with the goal of translating naturallanguage descriptions into mathematical expressions and deriving solutions. Despite recent progress with graph-based and tree-based neural models, many approaches still struggle to capture the implicit mathematical knowledge embedded in these problems. This limitation impedes their ability to accurately represent the problem-solving process, especially for complex MWPs that require an understanding of implicit mathematical relationships. To tackle this challenge, we propose integrating mathematical knowledge directly into graph neural network models to enhance their effectiveness in solving intricate MWPs. Specifically, we introduce a formula graph where each formula is abstracted into a graph structure and integrated into an enhanced model, thereby facilitating efficient resolution of MWPs involving implicit knowledge. Our approach aims to improve model performance on complex MWPs, as demonstrated through empirical evaluation on the Ape210K dataset.
Text summarization task aims to condense lengthy documents into shorter texts while preserving the essential information. However, related work mostly focused on preserving semantic information of documents and ignore...
详细信息
ISBN:
(纸本)9789819796700;9789819796717
Text summarization task aims to condense lengthy documents into shorter texts while preserving the essential information. However, related work mostly focused on preserving semantic information of documents and ignored the human values alignment between documents and summaries. What's more, summaries generated by LLMs are often found to contain hallucinations and lack factuality. In this paper, we introduce a two-stage extract-generate summarization framework to address the above issues. In the first stage, sentences containing both semantic and human values are extracted through a diffusion model. Then in the second stage, overall summaries are generated via LLMs. By extracting important sentences before summarization, we can shift LLMs' attention to preserve more human values information from the documents, hence achieving the goal of human values alignment. Specifically, we follow Schwartz's Theory of Human Values for human values definition due to its wide adoption as a value framework in the realm of NLP(naturallanguageprocessing). To verify the performance of our framework, we designed several LLM-only methods including in-context methods as baselines to compare with. Experiments conducted on CNN/DM and Reddit summarization datasets demonstrate improvements on both human values alignment and factuality achieved by our framework, indicating the importance of an extractor before leveraging LLM for summarization.
To alleviate the teen vaping crisis, this study analyzed the 'QuitVaping' subreddit. We identified posts sharing experience and/or advice using DistilBERT and investigated distinctive features in the data with...
详细信息
The rapid proliferation of open-source language models significantly increases the risks of downstream backdoor attacks. These backdoors can introduce dangerous behaviours during model deployment and can evade detecti...
Large-scale text-to-image diffusion models have demonstrated impressive capabilities for downstream tasks by leveraging strong vision-language alignment from generative pre-training. Recently, a number of works have e...
详细信息
Contextualized language models have emerged as a de facto standard in naturallanguageprocessing due to the vast amount of knowledge they acquire during pretraining. Nonetheless, their ability to solve tasks that req...
详细信息
暂无评论