Large language models (LLMs) have significantly improved naturallanguageprocessing, holding the potential to support health workers and their clients directly. Unfortunately, there is a substantial and variable drop...
详细信息
ISBN:
(纸本)9798350383744;9798350383737
Large language models (LLMs) have significantly improved naturallanguageprocessing, holding the potential to support health workers and their clients directly. Unfortunately, there is a substantial and variable drop in performance for low-resource languages. This paper presents an exploratory case study in Malawi, aiming to enhance the performance of LLMs in Chichewa through innovative prompt engineering techniques. By focusing on practical evaluations over traditional metrics, we assess the subjective utility of LLM outputs, prioritizing end-user satisfaction. Our findings suggest that tailored prompt engineering may improve LLM utility in underserved linguistic contexts, offering a promising avenue to bridge the language inclusivity gap in digital health interventions.
We present a novel multimodal interpretable VQA model that can answer the question more accurately and generate diverse explanations. Although researchers have proposed several methods that can generate human-readable...
详细信息
ISBN:
(纸本)9781728198354
We present a novel multimodal interpretable VQA model that can answer the question more accurately and generate diverse explanations. Although researchers have proposed several methods that can generate human-readable and fine-grained naturallanguage sentences to explain a model's decision, these methods have focused solely on the information in the image. Ideally, the model should refer to various information inside and outside the image to correctly generate explanations, just as we use background knowledge daily. The proposed method incorporates information from outside knowledge and multiple image captions to increase the diversity of information available to the model. The contribution of this paper is to construct an interpretable visual question answering model using multimodal inputs to improve the rationality of generated results. Experimental results show that our model can outperform state-of-the-art methods regarding answer accuracy and explanation rationality.
In-context learning (ICL) ability has emerged with the increasing scale of large language models (LLMs), enabling them to learn input-label mappings from demonstrations and perform well on downstream tasks. However, u...
详细信息
ISBN:
(纸本)9798350344868;9798350344851
In-context learning (ICL) ability has emerged with the increasing scale of large language models (LLMs), enabling them to learn input-label mappings from demonstrations and perform well on downstream tasks. However, under the standard ICL setting, LLMs may sometimes neglect query-related information in demonstrations, leading to incorrect predictions. To address this limitation, we propose a new paradigm called Hint-enhanced In-Context Learning (HICL) to explore the power of ICL in open-domain question answering, an important form in knowledge-intensive tasks. HICL leverages LLMs' reasoning ability to extract query-related knowledge from demonstrations, then concatenates the knowledge to prompt LLMs in a more explicit way. Furthermore, we track the source of this knowledge to identify specific examples, and introduce a Hint-related Example Retriever (HER) to select informative examples for enhanced demonstrations. We evaluate HICL with HER on 3 open-domain QA benchmarks, and observe average performance gains of 2.89 EM score and 2.52 F1 score on gpt-3.5-turbo, 7.62 EM score and 7.27 F1 score on LLaMA-2-Chat-7B compared with standard setting.
In view of the limitation of existing research, which primarily relies on similarity as input features for the model without considering the extraction of features from the intrinsic information of metabolite or disea...
详细信息
ISBN:
(纸本)9798350386783;9798350386776
In view of the limitation of existing research, which primarily relies on similarity as input features for the model without considering the extraction of features from the intrinsic information of metabolite or disease, and only focuses on the existing associations between metabolite and disease while neglecting their potential associations, this study proposes the GCDNLP model, a graph attention network that combines community detection and naturallanguageprocessing to predict metabolite-disease associations. Firstly, the Mol2Vec algorithm is employed to extract the ontological features of metabolite from the perspective of their molecular structures, which are then combined with the metabolite similarity features. Secondly, the Doc2Vec+ TF-IDF algorithm is introduced to compute the ontological features of disease. Finally, the prior probability matrix construction algorithm PPMCD based on community detection is proposed to uncover the unconfirmed potential associations between metabolite and disease within the network. The computed results, represented as prior probabilities, are then integrated with the graph attention network for further processing. Experimental results demonstrate the superiority of this model over existing algorithms in multiple evaluation metrics.
"Glad you introduced a traceability process in your company. Now you are able to relate the various types of artifacts in order to build up a large traceability graph" - what are the next steps? How do you d...
详细信息
ISBN:
(纸本)9798350326918
"Glad you introduced a traceability process in your company. Now you are able to relate the various types of artifacts in order to build up a large traceability graph" - what are the next steps? How do you deal with the enormous amount of information? We provide use cases to create, analyse and explore traceability graphs, based on our experiences from itemis ANALYZE, our professional traceability management system, that connect your whole development toolchain, including modelling tools and code. Furthermore, we fine-grain the 'requirements' in traceability by emerging Machine learning (ML) techniques that aid in analysing their similarities thereby proposing a tailored solution which forms the basis for requirements -specific knowledge graph. Finally, we use graph-based algorithms to analyse different use-case of created knowledge graphs.
naturallanguageprocessing (NLP) has witnessed significant advancements in recent years, particularly with the emergence of large language models. These models, such as GPT-3.5 and its variants, have revolutionized v...
详细信息
Machine learning-based approaches have been widely used to address naturallanguageprocessing (NLP) problems. Considering the similarity between naturallanguage text and source code, researchers have been working on...
详细信息
ISBN:
(纸本)9798350322637
Machine learning-based approaches have been widely used to address naturallanguageprocessing (NLP) problems. Considering the similarity between naturallanguage text and source code, researchers have been working on applying techniques from NLP to deal with code. On the other hand, source code and naturallanguage are by nature different. For example, code is highly structured and executable. Thus, directly applying the NLP techniques may not be optimal, and how to effectively optimize these NLP techniques to adapt to software engineering (SE) tasks remains a challenge. Therefore, to tackle the challenge, in this dissertation, we focus on two research directions: 1) distributed code representations, and 2) logging statements, which are two important intersections between the naturallanguage and source code. For distributed code representations, we first discuss the limitations of existing code embedding techniques, and then, we propose a novel approach to learn more generalizable code embeddings in a task-agnostic manner. For logging statements, we first propose an automated deep learning-based approach to automatically generate accurate logging texts by translating the related source code into short textual descriptions. Then, we make the first attempt to comprehensively study the temporal relations between logging and its corresponding source code, which is later used to detect issues in logging statements. We anticipate that our study can provide useful suggestions and support to developers in utilizing NLP techniques to assist in SE tasks.
The rapid growth of data has led to a significant increase in unstructured data, such as text, audio, and images, which dominate modern information processing. However, the complexity of unstructured data presents cha...
详细信息
With the advent of the digital age, the explosive growth of data has brought unprecedented opportunities and challenges to societal development. The rapid evolution of data science is driving innovation and progress a...
详细信息
language models that can learn a task at inference time, called in-context learning (ICL), show increasing promise in naturallanguage inference tasks. In ICL, a model user constructs a prompt to describe a task with ...
详细信息
ISBN:
(纸本)9798350395129;9798350395112
language models that can learn a task at inference time, called in-context learning (ICL), show increasing promise in naturallanguage inference tasks. In ICL, a model user constructs a prompt to describe a task with a naturallanguage instruction and zero or more examples, called demonstrations. The prompt is then input to the language model to generate a completion. In this paper, we apply ICL to the design and evaluation of satisfaction arguments, which describe how a requirement is satisfied by a system specification and associated domain knowledge. The approach builds on three prompt design patterns, including augmented generation, prompt tuning, and chain-of-thought prompting, and is evaluated on a privacy problem to check whether a mobile app scenario and associated design description satisfies eight consent requirements from the EU General Data Protection Regulation (GDPR). The overall results show that GPT-4 can be used to verify requirements satisfaction with 96.7% accuracy and dissatisfaction with 93.2% accuracy. Inverting the requirement improves verification of dissatisfaction to 97.2%. Chain-of-thought prompting improves overall GPT3.5 performance by 9.0% accuracy. We discuss the trade-offs among templates, models and prompt strategies and provide a detailed analysis of the generated specifications to inform how the approach can be applied in practice.
暂无评论