With the advent of transfer learning approaches, naturallanguageprocessing (NLP) problems have experienced tremendous progress, as demonstrated by models such as Generative Pre-trained Transformers (GPT) and Bidirec...
详细信息
ISBN:
(纸本)9798350388800
With the advent of transfer learning approaches, naturallanguageprocessing (NLP) problems have experienced tremendous progress, as demonstrated by models such as Generative Pre-trained Transformers (GPT) and Bidirectional Encoder Representations from Transformers (BERT). The usefulness of such transfer learning strategies across a range of NLP tasks and domains is investigated in this work. The study uses a methodical methodology to assess BERT and GPT's performance on a wide range of tasks. In addition, the study evaluates the generalizability and flexibility of these models across a broad variety of disciplines, including social media, finance, legal, and biological literature. The study's methodology entails rigorous assessment utilizing task-specific standard metrics after pre-trained BERT and GPT models have been fine-tuned using task-specific datasets. To determine the relative benefits and drawbacks of transfer learning strategies in various contexts, comparative studies are carried out against baseline models and other cutting-edge methodologies. Additionally, the study looks at how the performance of BERT and GPT is affected by variables including task difficulty, dataset size, and domain specificity. The results provide a comprehensive understanding of the benefits and drawbacks of transfer learning strategies in a variety of NLP tasks and domains. While BERT performs admirably on tests requiring semantic comprehension and contextual knowledge, GPT is superior at producing text that is both cohesive and appropriate to the situation. Both models, however, show sensitivity to dataset features and idiosyncrasies unique to the domain, indicating the necessity for customized fine-tuning techniques for best results. All things considered, this study advances our knowledge of the usefulness and efficiency of transfer learning strategies and provides insightful information for academics and practitioners who want to use BERT, GPT, and related models in a variety
Communication is a critical aspect of every individual's interaction, and individuals typically exchange information in a variety of languages. However, individuals with hearing and speech impairments may encounte...
详细信息
This paper explores the field of Spoken language Understanding (SLU), a subdomain of naturallanguageprocessing (NLP). SLU focuses on the processing and comprehension of spoken or voice input, aiming to extract meani...
详细信息
Biomedical Named Entity Linking (NEL) plays a crucial role in extracting valuable information from text for downstream tasks in bioinformatics. However, challenges like entity ambiguity and complex relationships hinde...
详细信息
Semantic parsing, in a sense, is a formal language process that translates naturallanguage into computer recognizable language. Therefore, semantic parsing can be treated as a special machine translation problem. Mod...
详细信息
This paper presents a translation-based knowledge geraph embedding method via efficient relation rotation (TransERR), a straightforward yet effective alternative to traditional translation-based knowledge graph embedd...
详细信息
Large language models (LLMs) have shown significant achievements in solving a wide range of tasks. Recently, LLMs' capability to store, retrieve and infer with symbolic knowledge has drawn a great deal of attentio...
详细信息
Text-to-Music Retrieval, finding music based on a given naturallanguage query, plays a pivotal role in content discovery within extensive music databases. To address this challenge, prior research has predominantly f...
详细信息
ISBN:
(纸本)9798350344868;9798350344851
Text-to-Music Retrieval, finding music based on a given naturallanguage query, plays a pivotal role in content discovery within extensive music databases. To address this challenge, prior research has predominantly focused on a joint embedding of music audio and text, utilizing it to retrieve music tracks that exactly match descriptive queries related to musical attributes (i.e. genre, instrument) and contextual elements (i.e. mood, theme). However, users also articulate a need to explore music that shares similarities with their favorite tracks or artists, such as I need a similar track to Superstition by Stevie Wonder. To address these concerns, this paper proposes an improved Text-to-Music Retrieval model, denoted as TTMR++, which utilizes rich text descriptions generated with a finetuned large language model and metadata. To accomplish this, we obtained various types of seed text from several existing music tag and caption datasets and a knowledge graph dataset of artists and tracks. The experimental results show the effectiveness of TTMR++ in comparison to state-of-the-art music-text joint embedding models through a comprehensive evaluation involving various musical text queries. (1)
The paper Augmenting Content Retrieval Using NLP in AIML describes using naturallanguageprocessing (NLP) with Artificial Intelligence Markup language (AIML) to enhance content retrieval. Among Markup languages, AIML...
详细信息
While intelligent virtual assistants like Siri, Alexa, and Google Assistant have become ubiquitous in modern life, they still face limitations in their ability to follow multi-step instructions and accomplish complex ...
详细信息
ISBN:
(纸本)9798400704901
While intelligent virtual assistants like Siri, Alexa, and Google Assistant have become ubiquitous in modern life, they still face limitations in their ability to follow multi-step instructions and accomplish complex goals articulated in naturallanguage. However, recent breakthroughs in large language models (LLMs) show promise for overcoming existing barriers by enhancing naturallanguageprocessing and reasoning capabilities. Though promising, applying LLMs to create more advanced virtual assistants still faces challenges like ensuring robust performance and handling variability in real-world user commands. This paper proposes a novel LLM-based virtual assistant that can automatically perform multi-step operations within mobile apps based on high-level user requests. The system represents an advance in assistants by providing an end-to-end solution for parsing instructions, reasoning about goals, and executing actions. LLM-based Process Automation (LLMPA) has modules for decomposing instructions, generating descriptions, detecting interface elements, predicting next actions, and error checking. Experiments demonstrate the system completing complex mobile operation tasks in Alipay based on naturallanguage instructions. This showcases how large language models can enable automated assistants to accomplish real-world tasks. The main contributions are the novel LLMPA architecture optimized for app process automation, the methodology for applying LLMs to mobile apps, and demonstrations of multi-step task completion in a real-world environment. Notably, this work represents the first real-world deployment and extensive evaluation of a large language model-based virtual assistant in a widely used mobile application with an enormous user base numbering in the hundreds of millions.
暂无评论