Mixture of experts (MoE) has become the standard for constructing production-level large language models (LLMs) due to its promise to boost model capacity without causing significant overheads. Nevertheless, existing ...
详细信息
With the recent advancement of Large language Models (LLMs), efforts have been made to leverage LLMs in crucial social science study methods, including predicting human features of social life such as presidential vot...
详细信息
Perceiving and understanding non-speech sounds and non-verbal speech is essential to making decisions that help us interact with our surroundings. In this paper, we propose GAMA, a novel General-purpose Large AudioLan...
详细信息
This position paper proposes a data-centric viewpoint of AI research, focusing on large language models (LLMs). We start by making a key observation that data is instrumental in the developmental (e.g., pretraining an...
Few studies on legitimation of new technologies were able to provide insights into the longitudinal changes in legitimacy outcomes and the social dynamics that underpin such outcomes. Using a novel mixed-methods appro...
详细信息
Few studies on legitimation of new technologies were able to provide insights into the longitudinal changes in legitimacy outcomes and the social dynamics that underpin such outcomes. Using a novel mixed-methods approach, combining naturallanguageprocessing with a qualitative text analysis, and drawing on the concept of social cohesion to investigate the social relations among actors, the study offers new insights into the legitimation of cultured meat in Germany. Using 424 newspaper articles, we identify four topics in the public discourse related to cultured meat and positive average sentiment on each topic over the period 2011 -2021. Furthermore, we find the actors, groups, and social relations that shape the observed legitimacy outcomes. The empirical findings are used to develop propositions about the role of social cohesion in legitimacy creation. The study paves the way for future studies on social cohesion dynamics in socio-technical change.
Multimodal Large language Models (MLLMs) extend the capacity of LLMs to understand multimodal information comprehensively, achieving remarkable performance in many vision-centric tasks. Despite that, recent studies ha...
Recent approaches to zero-shot commonsense reasoning have enabled Pre-trained language Models (PLMs) to learn a broad range of commonsense knowledge without being tailored to specific situations. However, they often s...
详细信息
The global escalation in emergency department patient visits poses significant challenges to efficient clinical management, particularly in clinical triage. Traditionally managed by human professionals, clinical triag...
详细信息
Recently, mobile AI agents based on VLMs have gained increasing *** works typically utilize VLM pre-trained on general-domain data as a foundation, fine-tuning it on instruction-based mobile ***, the proportion of mob...
详细信息
The rise of voice interface applications has renewed interest in improving the robustness of spoken language understanding(SLU). Many advances have come from end-to-end speech-language joint training, such as inferrin...
详细信息
ISBN:
(纸本)9789819794362;9789819794379
The rise of voice interface applications has renewed interest in improving the robustness of spoken language understanding(SLU). Many advances have come from end-to-end speech-language joint training, such as inferring semantics directly from speech signals and post-editing automatic speech recognition (ASR) output. Despite their performance achievements, these methods either suffer from the unavailability of a large number of paired error-prone ASR transcriptions and ground-truth annotations or are computationally costly. To mitigate these issues, we propose an ASR-robust pre-trained language model (ASRLM), which involves a generator generating simulated ASR transcriptions from ground-truth annotations and a sample-efficient discriminator distinguishing reasonable ASR errors from unrealistic ones. Experimental results demonstrate that ASRLM improves performance on a wide range of SLU tasks in the presence of ASR errors while saving 27% of the computation cost compared to baselines. Analysis also shows that our proposed generator is better than other simulation methods, including both BERT and GPT4-based, at simulating real-world ASR error situations.
暂无评论