Software documents are commonly processed by naturallanguageprocessing (NLP) libraries to extract information. The libraries provide similar functional APIs to achieve NLP tasks, numerous toolkits result in a proble...
详细信息
Software documents are commonly processed by naturallanguageprocessing (NLP) libraries to extract information. The libraries provide similar functional APIs to achieve NLP tasks, numerous toolkits result in a problem of selection. In this work, we propose a method to combine the strengths of different NLP libraries to avoid the subjective selection of a specific NLP library. The combined usage is conducted through two steps, i.e. document-level selection of primary NLP library and sentence-level overwriting. The primary NLP library is determined according to the overlap degree of the results. The highest overlap degree indicated the most effective NLP library on a specific NLP task. Through sentence-level overwriting, the possible fine-gained improvements from other libraries are extracted to overwrite the outputs of primary library. We evaluate the combined method with six widely used NLP libraries and 200 documents from three different sources. The results show that the combined method can generally outperform all the studied NLP libraries in terms of accuracy. The finding means that our combined method can be used instead of individual NLP library for more effective results.
Large language Models (LLMs) excel in numerous naturallanguageprocessing (NLP) tasks but encounter significant challenges in practical applications, including hallucinations, outdated information, and a lack of doma...
详细信息
Large language Models (LLMs) have demonstrated remarkable abilities in text comprehension and logical reasoning, indicating that the text representations learned by LLMs can facilitate their languageprocessing capabi...
With the emergence of large-scale language models (LLM), the powerful capabilities of LLM in naturallanguageprocessing have attracted attention. Based on programming language LLM (Programming language Model, PLM), w...
详细信息
ISBN:
(纸本)9798350371000;9798350370997
With the emergence of large-scale language models (LLM), the powerful capabilities of LLM in naturallanguageprocessing have attracted attention. Based on programming language LLM (Programming language Model, PLM), we use prompt templates to explore its potential in the field of automatic vulnerability repair, and combine it with a special workflow to improve its efficiency in automatic vulnerability repair tasks. Specifically, we design four prompt templates for handling vulnerable code, and design an iterative reasoning method to improve the efficiency of vulnerability fixing. We selected multiple typical LLMs for evaluation on multiple data sets. The results show that reasonable prompt templates can effectively improve the efficiency of automatic vulnerability repair, which is significantly improved compared with neural machine translation technology. In addition, we also discussed previous bug fixing related work and our work, and pointed out some of our shortcomings and directions for future improvements.
Large language Models (LLMs) have attracted a lot of attention due to their success in naturallanguageprocessing tasks. This paper provides a thorough overview by examining the architecture, applications, problems, ...
详细信息
The article discusses the development of a system that uses artificial intelligence (AI) to generate individualized mathematics assignments for bilingual students in Tatarstan, Russia. The goal is to enhance learning ...
详细信息
This paper delves into the critical task of measuring semantic similarity in text documents, a fundamental need in today's data-rich landscape. Efficiently gauging semantic con-nections is vital for applications s...
详细信息
Sign language is an essential communication medium for individuals with hearing impairments. It enables them to convey messages, disseminate knowledge, and transfer ideas within the deaf community. However, not everyo...
详细信息
Large language models (LLMs) have shown exceptional performance in the domain of composite artificial intelligence tasks, offering a preliminary insight into the potential of general artificial intelligence. The fine-...
详细信息
ISBN:
(纸本)9789819794331;9789819794348
Large language models (LLMs) have shown exceptional performance in the domain of composite artificial intelligence tasks, offering a preliminary insight into the potential of general artificial intelligence. The fine-tuning process for LLMs necessitates significant computational resources, often surpassing those available from standard consumer-grade GPUs. To this end, we introduce the Adaptive Quantization Low-Rank Adaptation fine-tuning (AQLoRA), a method that reduces memory demands during fine-tuning by utilizing quantization coupled with pruning techniques. This dual strategy not only reduces memory usage but also preserves accuracy. AQLoRA refines the original Low-Rank Adaptation fine-tuning (LoRA) method by efficiently quantizing LLMs weights, prioritizing computational resource allocation based on weight importance, and effectively integrating the quantized model with auxiliary weights post fine-tuning. Applying AQLoRA to the ChatGLM2-6B model, we demonstrate its effectiveness in both naturallanguage generation (NLG) and naturallanguage understanding (NLU) across diverse fine-tuning datasets and scenarios. Our findings reveal that AQLoRA achieves balance between performance and memory efficiency, reducing memory consumption by 25% in NLG tasks. For NLU tasks, it enhances performance by 10% and reduces memory consumption by 10% compared to state-of-the-art methods.
With the advent of large language models (LLMs), requirements engineers have gained a powerful naturallanguageprocessing tool to analyze, query, and validate a wide variety of textual artifacts, thus potentially sup...
详细信息
ISBN:
(纸本)9798350395129;9798350395112
With the advent of large language models (LLMs), requirements engineers have gained a powerful naturallanguageprocessing tool to analyze, query, and validate a wide variety of textual artifacts, thus potentially supporting the whole requirements engineering process from requirements elicitation to management. However, the input for the requirements engineering process often encompasses a variety of potential information sources in various formats, especially graphical models such as process models. Hence, this work aims to contribute to the state of the art by assessing the feasibility of utilizing graphical process models and their textual representations in the requirements engineering process. In particular, we focus on the extraction of textual process descriptions from process models as i) input for the requirements engineering process and ii) documentation as the result of process-oriented requirements engineering. To this end, we explore, quantify, and compare traditional deterministic and LLM-based extraction methods where the latter includes GPT3, GPT3.5, GPT4, and LLAMA. The evaluation assesses output quality and information loss based on one data set. The results indicate that LLMs produce human-like process descriptions based on the predefined patterns, but apparently lack true comprehension of the process models.
暂无评论