Deep learning has brought significant breakthroughs in sequential recommendation (SR) for capturing dynamic user interests. A series of recent research revealed that models with more parameters usually achieve optimal...
详细信息
Deep learning has brought significant breakthroughs in sequential recommendation (SR) for capturing dynamic user interests. A series of recent research revealed that models with more parameters usually achieve optimal performance for SR tasks, inevitably resulting in great challenges for deploying them in real systems. Following the simple assumption that light networks might already suffice for certain users, in this work, we propose CANet, a conceptually simple yet very scalable framework for assigning adaptive network architecture in an input-dependent manner to reduce unnecessary computation. The core idea of CANet is to route the input user behaviors with a light-weighted router module. Specifically, we first construct the routing space with various submodels parameterized in terms of multiple model dimensions such as the number of layers, hidden size and embedding size. To avoid extra storage overhead of the routing space, we employ a weight-slicing schema to maintain all the submodels in exactly one network. Furthermore, we leverage several solutions to solve the discrete optimization issues caused by the router module. Thanks to them, CANet could adaptively adjust its network architecture for each input in an end-to-end manner, in which the user preference can be effectively captured. To evaluate our work, we conduct extensive experiments on benchmark datasets. Experimental results show that CANet reduces computation by 55 ~ 65% while preserving the accuracy of the original model. Our codes are available at https://***/CANet.
The enormous data distributed at the network edge and ubiquitous connectivity have led to the emergence of the new paradigm of distributed machine learning and large-scale data analytics. Distributed principal compone...
详细信息
The enormous data distributed at the network edge and ubiquitous connectivity have led to the emergence of the new paradigm of distributed machine learning and large-scale data analytics. Distributed principal component analysis (PCA) concerns finding a low-dimensional subspace that contains the most important information of high-dimensional data distributed over the network edge. The subspace is useful for distributed data compression and feature extraction. This work advocates the application of over-the-air federated learning to efficient implementation of distributed PCA in a wireless network under a data-privacy constraint, termed AirPCA. The design features the exploitation of the waveform-superposition property of a multi-access channel to realize over-the-air aggregation of local subspace updates computed and simultaneously transmitted by devices to a server, thereby reducing the multi-access latency. The original drawback of this class of techniques, namely channel-noise perturbation to uncoded analog modulated signals, is turned into a mechanism for escaping from saddle points during stochastic gradient descent (SGD) in the AirPCA algorithm. As a result, the convergence of the AirPCA algorithm is accelerated. To materialize the idea, descent speeds in different types of descent regions are analyzed mathematically using martingale theory by accounting for wireless propagation and techniques including broadband transmission, over-the-air aggregation, channel fading and noise. The results reveal the accelerating effect of noise in saddle regions and the opposite effect in other types of regions. The insight and results are applied to designing an online scheme for adapting receive signal power to the type of current descent region. Specifically, the scheme amplifies the noise effect in saddle regions by reducing signal power and applies the power savings to suppressing the effect in other regions. From experiments using real datasets, such power control is fo
Adversarial attacks have shown the vulnerability of machine learning models, however, it is non-trivial to conduct textual adversarial attacks on natural language processing tasks due to the discreteness of data. Most...
详细信息
This paper studies the algorithm of Gray Level Co-occurrence Matrix (GLCM), explains the concrete meaning of 14 texture features based on GLCM, and points out the redundancy among texture features. Through the calcula...
This paper studies the algorithm of Gray Level Co-occurrence Matrix (GLCM), explains the concrete meaning of 14 texture features based on GLCM, and points out the redundancy among texture features. Through the calculation and analysis of the gray level co-occurrence matrix of texture image and the experiments of texture feature extraction, it is shown that the gray level co-occurrence matrix can reflect the features of the image, and corresponds to the features of the image described by the texture features. At the same time, there is a certain degree of redundancy among the 14 texture features of the image. In practice, we can choose several prominent texture features to classify the image according to the difference of texture features. The analysis of texture features and experimental results are of great significance to the application of image texture features.
Adapting a medical image segmentation model to a new domain is important for improving its cross-domain transferability, and due to the expensive annotation process, Unsupervised Domain Adaptation (UDA) is appealing w...
详细信息
Designing an incentive-compatible auction mechanism that maximizes the auctioneer's revenue while minimizes the bidders' ex-post regret is an important yet intricate problem in economics. Remarkable progress h...
ISBN:
(纸本)9781713871088
Designing an incentive-compatible auction mechanism that maximizes the auctioneer's revenue while minimizes the bidders' ex-post regret is an important yet intricate problem in economics. Remarkable progress has been achieved through learning the optimal auction mechanism by neural networks. In this paper, we consider the popular additive valuation and symmetric valuation setting; i.e., the valuation for a set of items is defined as the sum of all items' valuations in the set, and the valuation distribution is invariant when the bidders and/or the items are permutated. We prove that permutation-equivariant neural networks have significant advantages: the permutation-equivariance decreases the expected ex-post regret, improves the model generalizability, while maintains the expected revenue invariant. This implies that the permutation-equivariance helps approach the theoretically optimal dominant strategy incentive compatible condition, and reduces the required sample complexity for desired generalization. Extensive experiments fully support our theory. To our best knowledge, this is the first work towards understanding the benefits of permutation-equivariance in auction mechanisms.
Trajectory tracking of zebrafish is an important requirement in studying neurological disorders and developing new psychoactive drug. However, many challenges emerge for stable tracking, since zebrafish are similar in...
Trajectory tracking of zebrafish is an important requirement in studying neurological disorders and developing new psychoactive drug. However, many challenges emerge for stable tracking, since zebrafish are similar in appearance, occlusion, agile, non-linear in moving and easy to swarm, all of which will lead to mistrack for multiple fish. And there is no precedent for tracking zebrafish through embedded edge artificial intelligence device. To overcome these difficulties, we present a tracking system for zebrafish based on RK3588-S. First, We construct an embedded edge AI hardware system consisting of two cameras driven by the RK3588-S, one for the front view and the other for top view of the fish. Then, we develop a 2D tracking algorithm based on YOLOv5 and the Observation-Centric Simple online and real-time tracking (OC-SORT) algorithm, which are transplanted to the RK3588-S for tracking the top and front views of the fish at the edge device. Compared with the previous methods, our method has fewer ID exchanges and highly real-time. Moreover, we apply the MQTT mechanism to establish communication links between the edge and the cloud to reliably transmit data to the cloud. The correspondence between cloud server and embedded AI is 1: N. Finally, we design a multi-view data fusion association algorithm to fuse the data of the two views in the cloud, which are further utilized to build the 3D tracklets of the zebrafish.
Binary Particle Swarm Optimization has too strong global search ability and lacks local search ability in the later stage, because it has a unreasonable transfer function. According to the analysis of the transfer fun...
详细信息
Aerodynamic optimal design is crucial for enhancing performance of aircrafts, while calculating multi-target functionals through solving dual equations with arbitrary right-hand sides remains challenging. In this pape...
详细信息
The chemistry and localized structure of materials greatly affect their efficiency in gas evolution reactions leading to cell resistance loses. This work focuses on the role of morphological changes in the nature of t...
详细信息
暂无评论