Metaphor detection, a critical task in naturallanguageprocessing, involves identifying whether a particular word in a sentence is used metaphorically. Traditional approaches often rely on supervised learning models ...
详细信息
With the increasing amount of problematic peer reviews in top AI conferences, the community is urgently in need of automatic quality control measures. In this paper, we restrict our attention to substantiation ' o...
详细信息
The paper proposes an algorithm of text data representation for time-series econometric modeling. We link two areas in empirical econometrics, text mining and time-series modeling, and provide a multi-step methodology...
详细信息
ISBN:
(纸本)9798350398342
The paper proposes an algorithm of text data representation for time-series econometric modeling. We link two areas in empirical econometrics, text mining and time-series modeling, and provide a multi-step methodology for processing qualitative text data for quantitative time-series models. We also present an empirical example of text data processing of research article meta-data in economics and show its applicability to sentiment analysis using annual data from 1933 - 2015
language-based colorization generates realistic and aesthetically appealing colors by leveraging the guidance of intuitive and user-friendly naturallanguage descriptions. Previous methods in language-based image colo...
详细信息
Large language models (LLMs) trained on general domain corpora showed remarkable results on naturallanguageprocessing (NLP) tasks. However, previous research demonstrated LLMs trained using domain-focused corpora pe...
详细信息
Large language models (LLMs) are susceptible to red teaming attacks, which can induce LLMs to generate harmful content. Previous research constructs attack prompts via manual or automatic methods, which have their own...
详细信息
ISBN:
(纸本)9798891760615
Large language models (LLMs) are susceptible to red teaming attacks, which can induce LLMs to generate harmful content. Previous research constructs attack prompts via manual or automatic methods, which have their own limitations on construction cost and quality. To address these issues, we propose an integrated approach that combines manual and automatic methods to economically generate high-quality attack prompts. Specifically, considering the impressive capabilities of newly emerged LLMs, we propose an attack framework to instruct LLMs to mimic human-generated prompts through in-context learning. Furthermore, we propose a defense framework that fine-tunes victim LLMs through iterative interactions with the attack framework to enhance their safety against red teaming attacks. Extensive experiments on different LLMs validate the effectiveness of our proposed attack and defense frameworks. Additionally, we release a series of attack prompts datasets named SAP with varying sizes, facilitating the safety evaluation and enhancement of more LLMs. Our code and dataset is available on https://***/Aatrox103/SAP.
Public scarce resource allocation plays a crucial role in economics as it directly influences the efficiency and equity in society. Traditional studies including theoretical model-based, empirical study-based and simu...
详细信息
Contrastive language-Audio Pretraining (CLAP) is pre-trained to associate audio features with human language, making it a natural zero-shot classifier to recognize unseen sound categories. To adapt CLAP to downstream ...
详细信息
ISBN:
(纸本)9798350344868;9798350344851
Contrastive language-Audio Pretraining (CLAP) is pre-trained to associate audio features with human language, making it a natural zero-shot classifier to recognize unseen sound categories. To adapt CLAP to downstream tasks, prior works inevitably require labeled domain audios, which limits their scalability under data scarcity and deprives them of the capability to detect novel classes as the original CLAP. In this work, by leveraging the modality alignment in CLAP, we propose an efficient audio-free prompt tuning scheme aimed at optimizing a few prompt tokens from texts instead of audios, which regularizes the model space to avoid overfitting the seen classes as well. Based on this, a multi-grained prompt design is further explored to fuse global and local information. Experiments on several tasks demonstrate that our approach can boost the CLAP and outperform other training methods on model performance and training efficiency. While conducting zero-shot inference on unseen categories, it still shows better transferability than the vanilla CLAP. Moreover, our method is flexible enough even if only knowing the downstream class names. The code is available at https://***/Ming-er/Audio-Free-P-Tuning.
Large vision-language models (LVLMs) suffer from hallucination, resulting in misalignment between the output textual response and the input visual content. Recent research indicates that the over-reliance on the Large...
详细信息
作者:
Wang, LikunYunnan Univ
Sch Informat Sci & Engn Kunming 650500 Yunnan Peoples R China
With the booming development of social networks, a massive amount of short texts emerge every day, containing valuable information such as user interests and intentions. Therefore, the mining and classification of sho...
详细信息
ISBN:
(纸本)9798400709760
With the booming development of social networks, a massive amount of short texts emerge every day, containing valuable information such as user interests and intentions. Therefore, the mining and classification of short text information is particularly important. However, the inherent characteristics of sparse features and high noise in short texts limit the performance of traditional machine learning methods in short text classification. Meanwhile, many neural network models often rely on a large amount of annotated data during the training process, but obtaining sufficient annotated data is a challenging task in practical situations. Taking inspiration from recent large-scale language models, this article proposes an efficient fine-tuning method for short text classification based on the LLaMA large-scale language model. Utilizing the powerful learning ability of large language models to expand text information, fine-tuning the freezing model and instruction learning through LoRA can more fully classify downstream specific tasks. From the experimental results obtained from real datasets, it can be seen that the method proposed in this paper has achieved an improvement in the accuracy of short text classification.
暂无评论