Owing to the increased video content consumption in recent years, the need for advanced contextual advertising methods that leverage increasing user engagement and relevance on advertisement-based video-on-demand plat...
详细信息
Owing to the increased video content consumption in recent years, the need for advanced contextual advertising methods that leverage increasing user engagement and relevance on advertisement-based video-on-demand platforms has increased. Traditional behavior-based advertisement targeting is waning, particularly owing to the recent strict privacy policies that favor user consent and privacy. This study proposes an innovative approach for integrating advanced naturallanguageprocessing with multimodal analysis for video contextual advertising. To this end, transformer-based architectures, specifically BERTopic, computer vision techniques, and large language models were used to extract sets of topics from visual and textual video data automatically and systematically. The proposed framework decodes the taxonomy of content efficiently through videos in different levels of noise and languages. empirical analysis of the YouTube-8M dataset shows the potential for the approach to change the paradigm in video advertising. Built to be scalable and easily adaptable, this solution can handle multifarious and complex user-generated content well, suited for a wide range of applications across various media platforms.
We investigate a surprising limitation of LLMs: their inability to consistently generate text in a user's desired language. We create the language Confusion Benchmark (LCB) to evaluate such failures, covering 15 t...
详细信息
We show that LLMs hallucinate because their output is not constrained to be synonymous with claims for which they have evidence: a condition that we call evidential closure. Information about the truth or falsity of s...
详细信息
ISBN:
(纸本)9798891760608
We show that LLMs hallucinate because their output is not constrained to be synonymous with claims for which they have evidence: a condition that we call evidential closure. Information about the truth or falsity of sentences is not statistically identified in the standard neural language generation setup, and so cannot be conditioned on to generate new strings. We then show how to constrain LLMs to produce output that satisfies evidential closure. A multimodal LLM must learn about the external world (perceptual learning);it must learn a mapping from strings to states of the world (extensional learning);and, to achieve fluency when generalizing beyond a body of evidence, it must learn mappings from strings to their synonyms (intensional learning). The output of a unimodal LLM must be synonymous with strings in a validated evidence set. Finally, we present a heuristic procedure, Learn-Babble-Prune, that yields faithful output from an LLM by rejecting output that is not synonymous with claims for which the LLM has evidence.
Public scarce resource allocation plays a crucial role in economics as it directly influences the efficiency and equity in society. Traditional studies including theoretical model-based, empirical study-based and simu...
详细信息
We systematically study how three large language models with code capabilities - CodeT5, Codex, and ChatGPT - generalize to out-of-domain data. We consider two fundamental applications - code summarization, and code g...
详细信息
ISBN:
(纸本)9798891760608
We systematically study how three large language models with code capabilities - CodeT5, Codex, and ChatGPT - generalize to out-of-domain data. We consider two fundamental applications - code summarization, and code generation. We split data into domains following its natural boundaries - by an organization, by a project, and by a module within the software project. We establish that samples from each new domain present all the models with a significant challenge of distribution shift. We study how established methods adapt models to better generalize to new domains. Our experiments show that while multitask learning alone is a reasonable baseline, combining it with few-shot finetuning on examples retrieved from training data can achieve very strong performance. Moreover, this solution can outperform direct finetuning for very low-data scenarios. Finally, we consider variations of this approach to create a more broadly applicable method to adapt to multiple domains at once. We find that for code generation, a model adapted to multiple domains simultaneously performs on par with those adapted to a single domain(1).
Cross-domain Named Entity Recognition (CDNER) is crucial for Knowledge Graph (KG) construction and naturallanguageprocessing (NLP), enabling learning from source to target domains with limited data. Previous studies...
详细信息
Bilingual Lexicon Induction (BLI) is a core task in multilingual NLP that still, to a large extent, relies on calculating cross-lingual word representations. Inspired by the global paradigm shift in NLP towards Large ...
详细信息
ISBN:
(纸本)9798891760608
Bilingual Lexicon Induction (BLI) is a core task in multilingual NLP that still, to a large extent, relies on calculating cross-lingual word representations. Inspired by the global paradigm shift in NLP towards Large language Models (LLMs), we examine the potential of the latest generation of LLMs for the development of bilingual lexicons. We ask the following research question: Is it possible to prompt and fine-tune multilingual LLMs (mLLMs) for BLI, and how does this approach compare against and complement current BLI approaches? To this end, we systematically study 1) zero-shot prompting for unsupervised BLI and 2) few-shot in-context prompting with a set of seed translation pairs, both without any LLM finetuning, as well as 3) standard BLI-oriented finetuning of smaller LLMs. We experiment with 18 open-source text-to-text mLLMs of different sizes (from 0.3B to 13B parameters) on two standard BLI benchmarks covering a range of typologically diverse languages. Our work is the first to demonstrate strong BLI capabilities of text-to-text mLLMs. The results reveal that few-shot prompting with in-context examples from nearest neighbours achieves the best performance, establishing new state-of-the-art BLI scores for many language pairs. We also conduct a series of in-depth analyses and ablation studies, providing more insights on BLI with (m)LLMs, also along with their limitations.
Prompting is now a dominant method for evaluating the linguistic knowledge of large language models (LLMs). While other methods directly read out models' probability distributions over strings, prompting requires ...
详细信息
ISBN:
(纸本)9798891760608
Prompting is now a dominant method for evaluating the linguistic knowledge of large language models (LLMs). While other methods directly read out models' probability distributions over strings, prompting requires models to access this internal information by processing linguistic input, thereby implicitly testing a new type of emergent ability: metalinguistic judgment. In this study, we compare metalinguistic prompting and direct probability measurements as ways of measuring models' linguistic knowledge. Broadly, we find that LLMs' metalinguistic judgments are inferior to quantities directly derived from representations. Furthermore, consistency gets worse as the prompt query diverges from direct measurements of next-word probabilities. Our findings suggest that negative results relying on metalinguistic prompts cannot be taken as conclusive evidence that an LLM lacks a particular linguistic generalization. Our results also highlight the value that is lost with the move to closed APIs where access to probability distributions is limited.
There is a mismatch between psychological and computational studies on emotions. Psychological research aims at explaining and documenting internal mechanisms of these phenomena, while computational work often simplif...
详细信息
Chinese Spelling Check (CSC) is a meaningful task in the area of naturallanguageprocessing (NLP) which aims at detecting spelling errors in Chinese texts and then correcting these errors. However, CSC models are bas...
详细信息
ISBN:
(纸本)9798350344868;9798350344851
Chinese Spelling Check (CSC) is a meaningful task in the area of naturallanguageprocessing (NLP) which aims at detecting spelling errors in Chinese texts and then correcting these errors. However, CSC models are based on pretrained language models, which are trained on a general corpus. Consequently, their performance may drop when confronted with downstream tasks involving domain-specific terms. In this paper, we conduct a thorough evaluation about the domain adaption ability of various typical CSC models by building three new datasets encompassing rich domain-specific terms from the financial, medical, and legal domains. Then we conduct empirical investigations in the corresponding domain-specific test datasets to ascertain the cross-domain adaptation ability of several typical CSC models. We also test the performance of the popular large language model ChatGPT. As shown in our experiments, the performances of the CSC models drop significantly in the new domains.
暂无评论