Goal of this paper is to determine actual trends in geocontext extraction methods and to understand which types of geocontext information are the most interesting for users. For this purposes comparison of recent rese...
详细信息
ISBN:
(纸本)9785757704890
Goal of this paper is to determine actual trends in geocontext extraction methods and to understand which types of geocontext information are the most interesting for users. For this purposes comparison of recent researches about geocontext analysis was done. Researches were compared by the type of achieved result, used formalism, source data and limitations. As the main result of comparison new approach for automatic semantic places recognition was proposed. This approach is based on geotags markup with semantic user-defined tags. The solution allows extracting information (coordinates and a set of corresponding semantic tags on the naturallanguage) about locations which are interesting for the location-based services users. The main advantage of the approach is its simplicity - the method does not rely on any syntax analysis algorithms during the semantic labeling stage. For illustrating the approach an example of the general purpose accidents monitoring service for the Geo2Tag platform was described.
Text sentiment analysis in naturallanguageprocessing is an important means of understanding and interpreting human emotions in texts. It plays an important role in various applications, such as customer feedback ana...
详细信息
naturallanguageprocessing (NLP) stands at the forefront of the rapidly evolving landscape of Machine Learning, witnessing the emergence and evolution of diverse methodologies over the past decade. This study delves ...
详细信息
ISBN:
(纸本)9783031683220;9783031683237
naturallanguageprocessing (NLP) stands at the forefront of the rapidly evolving landscape of Machine Learning, witnessing the emergence and evolution of diverse methodologies over the past decade. This study delves into the dynamic trends within the NLP domain, specifically spanning the years 2010 to 2022, through an empirical analysis of papers presented at conferences hosted by the Association for Computational Linguistics (ACL). We utilize ChatGPT in order to extract meaningful information from the data before performing an in depth analysis. Our investigation encompasses an exploration of several key aspects, namely computational trends, research trends and geographic trends. We further investigate the entry cost into NLP, the longevity of hardware and the environmental impact of NLP. The code to run our system is publicly available at https://***/ieeta-pt/nlp-trends.
Mitigating bias in language models (LMs) has become a critical problem due to the widespread deployment of LMs in the industry and customer-facing applications. Numerous approaches revolve around data pre-processing a...
详细信息
The use of naturallanguageprocessing Algorithms (NLP) for automation purposes in various applications is frequently encountered recently. Some research managed to identify the dominant emotion from a text using neur...
详细信息
Prompt-based Large language Models(LLMs) are surprisingly powerful in generating naturallanguage reasoning steps or Chains of Thoughts(CoT) for multi-hop question answering(QA). However, LLMs struggle when they lack ...
详细信息
ISBN:
(纸本)9798350344868;9798350344851
Prompt-based Large language Models(LLMs) are surprisingly powerful in generating naturallanguage reasoning steps or Chains of Thoughts(CoT) for multi-hop question answering(QA). However, LLMs struggle when they lack access to necessary knowledge or when the knowledge within their parameters is outdated. Additionally, LLMs that rely solely on CoT tend to generate hallucinations during the reasoning process. To address these dilemmas, we propose the Chain of Question (CoQ) framework, a novel multi-hop QA approach. This approach decomposes a complex original question into multiple sub-questions according to a CoT to retrieve knowledge from an external knowledge base. It then answers the question process based on the retrieved knowledge in accordance with a CoT. We design that each point of thought generated during the reasoning process be supported by the knowledge retrieved in the external knowledge base. Experiments show that CoQ is effective in reducing model hallucinations, leading to higher factual accuracy than *** average, it reduces factual errors by 31% over CoT, and even by 38% on the two most commonly used models today.
In this work, we study whether multilingual language models (MultiLMs) can transfer logical reasoning abilities to other languages when they are fine-tuned for reasoning in a different language. We evaluate the cross-...
详细信息
ISBN:
(纸本)9798891760615
In this work, we study whether multilingual language models (MultiLMs) can transfer logical reasoning abilities to other languages when they are fine-tuned for reasoning in a different language. We evaluate the cross-lingual reasoning abilities of MultiLMs in two schemes: (1) where the language of the context and the question remain the same in the new languages that are tested (i.e., the reasoning is still monolingual, but the model must transfer the learned reasoning ability across languages), and (2) where the language of the context and the question is different (which we term code-switched reasoning). On two logical reasoning datasets, RuleTaker and LeapOfThought, we demonstrate that although MultiLMs can transfer reasoning ability across languages in a monolingual setting, they struggle to transfer reasoning abilities in a code-switched setting. Following this observation, we propose a novel attention mechanism that uses a dedicated set of parameters to encourage cross-lingual attention in code-switched sequences, which improves the reasoning performance by up to 14% and 4% on the RuleTaker and LeapOfThought datasets,
Deep learning is the new frontier of machine learning research, which has led to many recent breakthroughs in English naturallanguageprocessing. However, there are inherent differences between Chinese and English, a...
详细信息
ISBN:
(纸本)9783642416439;9783642416446
Deep learning is the new frontier of machine learning research, which has led to many recent breakthroughs in English naturallanguageprocessing. However, there are inherent differences between Chinese and English, and little work has been done to apply deep learning techniques to Chinese naturallanguageprocessing. In this paper, we propose a deep neural network model: text window denoising autoencoder, as well as a complete pre-training solution as a new way to solve classical Chinese naturallanguageprocessing problems. This method does not require any linguistic knowledge or manual feature design, and can be applied to various Chinese naturallanguageprocessing tasks, such as Chinese word segmentation. On the PKU dataset of Chinese word segmentation bakeoff 2005, applying this method decreases the F1 error rate by 11.9% for deep neural network based models. We are the first to apply deep learning methods to Chinese word segmentation to our best knowledge.
Character language models have access to surface morphological patterns, but it is not clear whether or how they learn abstract morphological regularities. We instrument a character language model with several probes,...
详细信息
Ontology, the shared formal conceptualization of domain information, has been shown to have multiple applications in modeling, processing and understanding naturallanguage text. In this work, we use distributed word ...
详细信息
ISBN:
(纸本)9781509061679
Ontology, the shared formal conceptualization of domain information, has been shown to have multiple applications in modeling, processing and understanding naturallanguage text. In this work, we use distributed word vectors out of various recent language models from Deep Learning for semi- automated domain ontology creation for closed domains. We cover all major aspects of Domain Ontology Induction or Learning like concept identification, attribute identification, taxonomical and nontaxonomical relationship identification using the distributed word vectors. Preliminary results show that simple clustering based methods using distributed word vectors from these language models outperforms methods using models like LSI in ontology learning for closed domains.
暂无评论