Existing speculative decoding methods typically require additional model structure and training processes to assist the model for draft token generation. This makes the migration of acceleration methods to the new mod...
详细信息
Large language Models (LLMs) have demonstrated remarkable capability in a variety of NLP tasks. However, LLMs are also prone to generate nonfactual content. Uncertainty Quantification (UQ) is pivotal in enhancing our ...
详细信息
In this paper, we study the problem of generating structured objects that conform to a complex schema, with intricate dependencies between the different components (facets) of the object. The facets of the object (att...
详细信息
The inability to utilise future contexts and the pre-determined left-to-right generation order are major limitations of unidirectional language models. Bidirectionality has been introduced to address those deficiencie...
详细信息
Fine-tuning large language models (LLMs) has achieved remarkable performance across various naturallanguageprocessing tasks, yet it demands more and more memory as model sizes keep growing. To address this issue, th...
详细信息
Text written by humans makes up the vast majority of the data used to pre-train and fine-tune large language models (LLMs).Many sources of this data-like code, forum posts, personal websites, and books-are easily attr...
详细信息
The advent of large language models (LLMs) has dramatically advanced the state-of-the-art in numerous naturallanguage generation tasks. For LLMs to be applied reliably, it is essential to have an accurate measure of ...
详细信息
naturallanguageprocessing (NLP) is poised to substantially influence the world. However, significant progress comes hand-in-hand with substantial risks. Addressing them requires broad engagement with various fields ...
详细信息
ISBN:
(纸本)9798891760608
naturallanguageprocessing (NLP) is poised to substantially influence the world. However, significant progress comes hand-in-hand with substantial risks. Addressing them requires broad engagement with various fields of study. Yet, little empirical work examines the state of such engagement (past or current). In this paper, we quantify the degree of influence between 23 fields of study and NLP (on each other). We analyzed similar to 77k NLP papers, similar to 3.1m citations from NLP papers to other papers, and similar to 1.8m citations from other papers to NLP papers. We show that, unlike most fields, the cross-field engagement of NLP, measured by our proposed Citation Field Diversity Index (CFDI), has declined from 0.58 in 1980 to 0.31 in 2022 (an all-time low). In addition, we find that NLP has grown more insular-citing increasingly more NLP papers and having fewer papers that act as bridges between fields. NLP citations are dominated by computer science;Less than 8% of NLP citations are to linguistics, and less than 3% are to math and psychology. These findings underscore NLP's urgent need to reflect on its engagement with various fields.
Active Learning (AL) addresses the high costs of collecting human annotations by strategically annotating the most informative ***, for subjective NLP tasks, incorporating a wide range of perspectives in the annotatio...
详细信息
language models can be manipulated by adversarial attacks, which introduce subtle perturbations to input data. While recent attack methods can achieve a relatively high attack success rate (ASR), we've observed th...
详细信息
暂无评论