Automated decision tools, such as advanced energy management systems, are required to involve the electrical grid users in energy flexibility services. This paper focuses on the prediction models as a substantial part...
详细信息
Covering option discovery has been developed to improve the exploration of reinforcement learning in single-agent scenarios, where only sparse reward signals are available. It aims to connect the most distant states i...
详细信息
We perform factor analysis on the raw data of the four major neighborhood and shortest paths-based centrality metrics (Degree, Eigenvector, Betweenness and Closeness) and propose a novel quantitative measure called th...
详细信息
We characterize the statistical efficiency of knowledge transfer through n samples from a teacher to a probabilistic student classifier with input space S over labels A. We show that privileged information at three pr...
详细信息
A novel reconfigurable diplexer with independently controllable center frequency and insertion phase is proposed in this brief. It simply consists of six coupled resonators and two non-resonating-nodes (NRNs). These s...
详细信息
We investigate the use of the Neighborhood-based Bridge Node Centrality (NBNC) tuple to choose nodes for preferential vaccination so that such vaccinated nodes could provide herd immunity and reduce the spreading rate...
详细信息
Predicting financial markets and stock price movements requires analyzing a company's performance, historic price movements, industry-specific events alongside the influence of human factors such as social media a...
详细信息
Adversarial imitation learning (AIL) has stood out as a dominant framework across various imitation learning (IL) applications, with Discriminator Actor Critic (DAC) (Kostrikov et al., 2019) demonstrating the effectiv...
We present a novel approach for test-time adaptation via online self-training, consisting of two components. First, we introduce a statistical framework that detects distribution shifts in the classifier's entropy...
The transformer architecture has prevailed in various deep learning settings due to its exceptional capabilities to select and compose structural information. Motivated by these capabilities, Sanford et al. (2023) pro...
详细信息
The transformer architecture has prevailed in various deep learning settings due to its exceptional capabilities to select and compose structural information. Motivated by these capabilities, Sanford et al. (2023) proposed the sparse token selection task, in which transformers excel while fully-connected networks (FCNs) fail in the worst case. Building upon that, we strengthen the FCN lower bound to an average-case setting and establish an algorithmic separation of transformers over FCNs. Specifically, a one-layer transformer trained with gradient descent provably learns the sparse token selection task and, surprisingly, exhibits strong out-of-distribution length generalization. We provide empirical simulations to justify our theoretical findings. Copyright 2024 by the author(s)
暂无评论