Recent advances in instruction-tuned Large Vision-language Models (LVLMs) have imbued the models with the ability to generate high-level, image-grounded explanations with ease. While such capability is largely attribu...
详细信息
In this work, we show that contemporary language models have a previously unknown skill - the capacity for electronic circuit design from high-level textual descriptions, akin to code generation. We introduce two benc...
详细信息
ISBN:
(纸本)9798891760615
In this work, we show that contemporary language models have a previously unknown skill - the capacity for electronic circuit design from high-level textual descriptions, akin to code generation. We introduce two benchmarks: PINS100, assessing model knowledge of electrical components, and MICRO25, evaluating a model's capability to design common micro-controller circuits and code in the ARDUINO ecosystem that involve input, output, sensors, motors, protocols, and logic - with models such as GPT-4 and Claude-V1 achieving between 60% to 96% PASS@ 1 on generating full devices. We include six case studies of using language models as a design assistant for moderately complex devices, such as a radiation-powered random number generator, an emoji keyboard, a visible spectrometer, and several assistive devices, while offering a qualitative analysis performance, outlining evaluation challenges, and suggesting areas of development to improve complex circuit design and practical utility. With this work, we aim to spur research at the juncture of naturallanguageprocessing and electronic design.
In many scientific fields, large language models (LLMs) have revolutionized the way text and other modalities of data (e.g., molecules and proteins) are handled, achieving superior performance in various applications ...
详细信息
Improving user experience and providing personalized search results in E-commerce services heavily rely on understanding purchase intention. However, existing methods for acquiring large-scale intentions bank on disti...
详细信息
Both humans and large language models are able to learn language without explicit structural supervision. What inductive biases make this learning possible? We address this fundamental cognitive question by leveraging...
详细信息
ISBN:
(纸本)9798891760615
Both humans and large language models are able to learn language without explicit structural supervision. What inductive biases make this learning possible? We address this fundamental cognitive question by leveraging transformer language models: we inject inductive bias into language models by pretraining on formally-structured data, and then evaluate the biased learners' ability to learn typologicallydiverse naturallanguages. Our experimental setup creates a testbed for hypotheses about inductive bias in human language learning. We investigate the effect of injecting models with three types of inductive bias: 1) recursive, hierarchical processing, 2) crossing token-token relationships that can't be modeled by contextfree grammars, and 3) a Zipfian power-law vocabulary distribution. We show that noncontext-free relationships form the best inductive biases. Our study leverages the capabilities of transformer models to run controlled language learning experiments that are not possible to run on humans, and surfaces hypotheses about the structures that facilitate language learning in both humans and machines.
Recent language Models (LMs) have shown impressive capabilities in generating texts with the knowledge internalized in parameters. Yet, LMs often generate the factually incorrect responses to the given queries, since ...
详细信息
ISBN:
(纸本)9798891760608
Recent language Models (LMs) have shown impressive capabilities in generating texts with the knowledge internalized in parameters. Yet, LMs often generate the factually incorrect responses to the given queries, since their knowledge may be inaccurate, incomplete, and outdated. To address this problem, previous works propose to augment LMs with the knowledge retrieved from an external knowledge source. However, such approaches often show suboptimal text generation performance due to two reasons: 1) the model may fail to retrieve the knowledge relevant to the given query, or 2) the model may not faithfully reflect the retrieved knowledge in the generated text. To overcome these, we propose to verify the output and the knowledge of the knowledge-augmented LMs with a separate verifier, which is a small LM that is trained to detect those two types of errors through instruction-finetuning. Then, when the verifier recognizes an error, we can rectify it by either retrieving new knowledge or generating new text. Further, we use an ensemble of the outputs from different instructions with a single verifier to enhance the reliability of the verification processes. We validate the effectiveness of the proposed verification steps on multiple question answering benchmarks, whose results show that the proposed verifier effectively identifies retrieval and generation errors, allowing LMs to provide more factually correct outputs. Our code is available at https://***/JinheonBaek/KALMV.
With the prevalence of the Internet and various types of social media, our daily life is surrounded by a huge amount of text information, which can provide us with the convenience of accessing information and communic...
详细信息
ISBN:
(纸本)9798400709760
With the prevalence of the Internet and various types of social media, our daily life is surrounded by a huge amount of text information, which can provide us with the convenience of accessing information and communication, but also brings challenges in management and maintenance. In the face of the current practical needs, traditional machine learning methods, such as TF-IDF, SVM, Naive Bayes need to rely on manual completion of the feature design, and can not solve the problem of data sparsity and high-dimensional feature vectors, so that the accuracy of text classification is insufficient. In this regard, based on the deep learning algorithm, this paper will propose a set of Chinese text classification model based on Word2vec and TextCNN to realize the automatic classification function of a large number of texts. Practice has proved that the test precision of this classification model is 94.12%, the recall is 92.19%, and the F1 value is 93.11%. Compared with the traditional TF-IDF, SVM and Naive Bayes methods, all the evaluation indexes are obviously improved, and the actual application effect is better than that of the original TextCNN. The overall design is in line with expectations and has certain application and promotion value.
This work focuses on the task of query-based meeting summarization, in which the summary of a context (meeting transcript) is generated in response to a specific query. When using Large language Models (LLMs) for this...
详细信息
As a promising paradigm to collaboratively train models with decentralized data, Federated Learning (FL) can be exploited to fine-tune Large language Models (LLMs). While LLMs correspond to huge size, the scale of the...
详细信息
Classification is a core NLP task architecture with many potential applications. While large language models (LLMs) have brought substantial advancements in text generation, their potential for enhancing classificatio...
详细信息
暂无评论