The practical application of artificial intelligence in the field of naturallanguageprocessing is becoming more and more extensive, and the technology is changing day by day. With the rapid update and development of...
详细信息
People with Visual Impairments are vulnerable to various challenges in their everyday lives ranging from Access to Information, Travelling to places, Recognizing their surroundings, Isolated Lifestyle and much more. T...
详细信息
Retrieval-augmented generation (RAG) enhances large language models (LLMs) by integrating external knowledge sources, enabling more useful information and generating accurate responses. This paper explores RAG's a...
详细信息
ISBN:
(纸本)9798331540364
Retrieval-augmented generation (RAG) enhances large language models (LLMs) by integrating external knowledge sources, enabling more useful information and generating accurate responses. This paper explores RAG's architecture and applications, combining generator and retriever models to access and utilize vast external data repositories. While RAG holds significant promise for various naturallanguageprocessing (NLP) processes like dialogue generation, summarization, and question answering, it also presents unique security challenges that must be addressed to ensure system integrity and reliability. RAG systems face several security threats, including data poisoning, model manipulation, privacy leakage, biased information retrieval, and harmful outputs generation. Generally, in the traditional RAG application, security threat is one of the major concerns. To tighten the security system and enhance the efficiency of the model on processing more complex data this paper outlines key strategies for securing RAG-based applications to mitigate these risks paper outlines key strategies for securing RAG-based applications to mitigate these risks. Ensuring data security through filtering, sanitization, and provenance tracking can prevent data poisoning and enhance the quality of external knowledge sources. Strengthening model security via adversarial training, input validation, and anomaly detection improves resilience against manipulative attacks. Implementing output monitoring and filtering techniques, such as factual verification, language moderation, and bias detection, ensures the accuracy and safety of generated responses. Additionally, robust infrastructure and access control measures, including secure data storage, secure APIs, and regulated model access, protect against unauthorized access and manipulation. Moreover, this study analyzes various use cases for LLMs enhanced by RAG, including personalized recommendations, customer support automation, content creation,
Entity relation extraction aims to extract knowledge triples from unstructured or semi-structured text data and can be applied to various fields, including medicine, finance knowledge graph construction and intelligen...
详细信息
ISBN:
(纸本)9798400708305
Entity relation extraction aims to extract knowledge triples from unstructured or semi-structured text data and can be applied to various fields, including medicine, finance knowledge graph construction and intelligent question-answering. Traditional entity relation extraction requires a large amount of labeled data, consumes a lot of labor and time, and the trained model lacks generalization ability, which is difficult to migrate to other fields. Zero-shot entity relation extraction relieves the dependence on labeled data in traditional method. Based on unlabeled text data, zero-shot entity relation extraction has strong domain adaptability, which is a very challenging and practical task. Recent work on large language models shows that large models can effectively complete downstream tasks through naturallanguage instructions and have good generalization ability. Inspired by this, we explore the use of large models for information extraction. Due to the randomness of large language model generation, we introduce in-context learning in entity relation extraction task to guide large language model to output data in a specified format to help obtain structured data. At the same time, we propose a three-stage extraction framework for decomposing entity relation extraction tasks, and each stage is conducted in the form of question and answer to reduce the complexity of extraction. We evaluated the knowledge triples extraction performance of the model on three self-built test datasets in different fields, and the experimental result showed that our proposed method achieved impressive performance in the zero-shot entity relation extraction task, surpassing the comparison model on multiple metrics, proving the effectiveness and domain adaptability of the proposed method.
Definitions are a fundamental building block in lexicography, linguistics and computational semantics. In NLP, they have been used for retrofitting word embeddings or augmenting contextual representations in language ...
详细信息
Quantifying the likeness between words, sentences, paragraphs, and documents plays a crucial role in various applications of naturallanguageprocessing (NLP). As Bert, Elmo, and Roberta exemplified, contemporary meth...
详细信息
ISBN:
(纸本)9798350370027;9798350370034
Quantifying the likeness between words, sentences, paragraphs, and documents plays a crucial role in various applications of naturallanguageprocessing (NLP). As Bert, Elmo, and Roberta exemplified, contemporary methodologies leverage neural networks to generate embeddings, necessitating substantial data and training time for cutting-edge performance. Alternatively, semantic similarity metrics are based on knowledge bases like WordNet, using approaches such as the shortest path between words. MinHashing, a nimble technique, quickly approximates Jaccard similarity scores for document pairs. In this study, we propose employing MinHashing to gauge semantic scores by enhancing original documents with information from semantic networks, incorporating relationships such as synonyms, antonyms, hyponyms, and hypernyms. This augmentation improves lexical similarity based on semantic insights. The MinHash algorithm calculates compact signatures for extended vectors, mitigating dimensionality concerns. The similarity of these signatures reflects the semantic score between the documents. Our method achieves approximately 64% accuracy in the MRPC and SICK data sets.
In order to detect plagiarism in academic and professional contexts, this work presents an intelligent naturallanguageprocessing (NLP) method. The program outperforms conventional exact match algorithms by utilizing...
详细信息
Temporal knowledge Graph (TKG) is an extension of regular knowledge graph by attaching the time scope. Existing temporal knowledge graph question answering (TKGQA) models solely approach simple questions, owing to the...
详细信息
ISBN:
(纸本)9798350344868;9798350344851
Temporal knowledge Graph (TKG) is an extension of regular knowledge graph by attaching the time scope. Existing temporal knowledge graph question answering (TKGQA) models solely approach simple questions, owing to the prior assumption that each question only contains a single temporal fact with explicit/implicit temporal constraints. Hence, they perform poorly on questions which own multiple temporal facts. In this paper, we propose Joint Multi Facts Reasoning Network (JMFRN), to jointly reasoning multiple temporal facts for accurately answering complex temporal questions. Specifically, JMFRN first retrieves question-related temporal facts from TKG for each entity of the given complex question. For joint reasoning, we design two different attention (i.e., entity-aware and time-aware) modules, which are suitable for universal settings, to aggregate entities and timestamps information of retrieved facts. Moreover, to filter incorrect type answers, we introduce an additional answer type discrimination task. Extensive experiments demonstrate our proposed method significantly outperforms the state-of-art on the well-known complex temporal question benchmark TimeQuestions.
Pre-trained language Models have been shown to be able to emulate deductive reasoning in naturallanguage. However, PLMs are easily affected by irrelevant information (e.g., entity) in instance-level proofs when learn...
详细信息
Building a quality management system (QMS) relies heavily on two key steps: mapping and modeling business processes. The process approach, a core principle of QMS, emphasizes identifying and managing processes and the...
详细信息
ISBN:
(纸本)9783031850660;9783031850677
Building a quality management system (QMS) relies heavily on two key steps: mapping and modeling business processes. The process approach, a core principle of QMS, emphasizes identifying and managing processes and their interactions systematically. This ensures achieving desired outcomes aligned with the organization's policies and strategic goals. Following this approach, organizations need to establish, implement, maintain, and continuously improve their quality management system (SQM), encompassing all relevant processes and their connections, to meet regulatory requirements. This article aims to present our approach for extending the BPMN formalism by a quality dimension based on the MDA approach and on an ontological approach for the definition of a meta-model for the quality management which covers the ISO 9001 Norm.
暂无评论