Federated learning (FL), as a safe distributed training mode, provides strong support for the edge intelligence of the Internet of Vehicles (IoV) to realize efficient collaborative control and safe data sharing. Howev...
详细信息
The recent emergence of time series contrastive clustering methods can be categorized into two classes. The first class uses contrastive learning for universal representations, which can be effective in various downst...
详细信息
The graph with complex annotations is the most potent data type, whose constantly evolving motivates further exploration of the unsupervised dynamic graph representation. One of the representative paradigms is graph c...
The graph with complex annotations is the most potent data type, whose constantly evolving motivates further exploration of the unsupervised dynamic graph representation. One of the representative paradigms is graph contrastive learning. It constructs self-supervised signals by maximizing the mutual information between the statistic graph’s augmentation views. However, the semantics and labels may change within the augmentation process, causing a significant performance drop in downstream tasks. This drawback becomes greatly magnified on dynamic graphs. To address this problem, we designed a simple yet effective framework named CLDG. Firstly, we elaborate that dynamic graphs have temporal translation invariance at different levels. Then, we proposed a sampling layer to extract the temporally-persistent signals. It will encourage the node to maintain consistent local and global representations, i.e., temporal translation invariance under the timespan views. The extensive experiments demonstrate the effectiveness and efficiency of the method on seven datasets by outperforming eight unsupervised state-of-the-art baselines and showing competitiveness against four semi-supervised methods. Compared with the existing dynamic graph method, the number of model parameters and training time is reduced by an average of 2,001.86 times and 130.31 times on seven datasets, respectively. The code and data are available at: https://***/yimingxu24/CLDG.
In this study, a multi-degree-of-freedom (Multi-DOF) robot (MDR) system based on a LightGBM-driven electroencephalogram (EEG) decoding model is designed and developed to assist subjects with hand motor dysfunction in ...
In this study, a multi-degree-of-freedom (Multi-DOF) robot (MDR) system based on a LightGBM-driven electroencephalogram (EEG) decoding model is designed and developed to assist subjects with hand motor dysfunction in their daily activities and neurorehabilitation. The system mainly consists of a motor imagery electroencephalogram (MI-EEG) evoking layer, an intention decoding layer, and an interaction executive layer. The MI-EEG evoking layer initially displays a virtual reality (VR) motion imagination scenario, instructing the subjects to imagine a real hand gripping movement, simultaneously collecting the electrical EEG signals and preprocessing the EEG signals. Secondly, in the intention decoding layer, a network combining temporal-spectral feature fusion and LightGBM (TSFF-LightGBM) for MI-BCI classification is used to more effectively boost brain decoding accuracy and decrease decoding time. Finally, in the interaction executive layer, the Multi-DOF wearable robot is developed to offer hand grasp motion kinesthetic feedback and visual feedback synced with MI. The following are the key benefits of the proposed MDR system: (1) We propose a new lightweight network structure more suitable for brain computer interface (BCI) interaction systems, achieving more accurate decoding and shorter identification time in different data sets, which helps to improve the practicability of the system and promote the practical clinical application of BCI rehabilitation technology. (2) We integrate BCI, VR, a wearable Multi-DOF robot, motion kinesthetic feedback, and visual feedback to improve human-machine interaction. Compared to the most recent investigations, the average accuracy of the MDR system on the publicly accessible datasets BCI IV 2a and HGD reached 75.89% and 93.53%, respectively.
In recent years, the demand for facial expression recognition applications has increased rapidly, and its research has received extensive attention from researchers. However, the current recognition methods based on d...
In recent years, the demand for facial expression recognition applications has increased rapidly, and its research has received extensive attention from researchers. However, the current recognition methods based on deep learning ignore the idea of multiple head attention and semantic consistency, resulting in the model only paying attention to the local area of the feature map. In addition, the model's inconsistent attention to the images before and after the flip, resulting the poor robustness, poor interpretation, and other shortcomings in the model. To address the above problems, we propose an Affinity Separation Loss (ASLoss), which improves the separability of samples through clustering. Moreover, a Separate Multi-head Attention block (SMA), and a Zonal Loss (ZLoss) are also designed to decentralize the model's attention. Experimental results demonstrate that our proposed MACNet method achieves competitive recognition performance on two public datasets RAF-DB and FERPlus.
Building upon the impressive success of CLIP (Contrastive Language-Image Pretraining), recent pioneer works have proposed to adapt the powerful CLIP to video data, leading to efficient and effective video learners for...
详细信息
Short carrier lifetimes is a key challenge limiting the open-circuit voltage (VOC) and power conversion efficiency (PCE) of kesterite Cu2ZnSn(S,Se)4 (CZTSSe) solar cells. In this work, for the first time, lanthanide e...
详细信息
Unmanned Aerial Vehicles (UAVs) possess high mobility and flexible deployment capabilities, prompting the development of UAVs for various application scenarios within the Internet of Things (IoT). The unique capabilit...
详细信息
Classifying large-scale data is a challenging task in machine learning. Feature selection and feature construction can improve the classification performances of classifiers. However, existing feature selections strug...
详细信息
High-fidelity kinship face synthesis has many potential applications, such as kinship verification, missing child identification, and social media analysis. However, it is challenging to synthesize high-quality descen...
High-fidelity kinship face synthesis has many potential applications, such as kinship verification, missing child identification, and social media analysis. However, it is challenging to synthesize high-quality descendant faces with genetic relations due to the lack of large-scale, high-quality annotated kinship data. This paper proposes RFG (Region-level Facial Gene) extraction framework to address this issue. We propose to use IGE (Image-based Gene Encoder), LGE (Latent-based Gene Encoder) and Gene Decoder to learn the RFGs of a given face image, and the relationships between RFGs and the latent space of Style-GAN2. As cycle-like losses are designed to measure the $\mathcal{L}_data$ distances between the output of Gene Decoder and image encoder, and that between the output of LGE and IGE, only face images are required to train our framework, i.e. no paired kinship face data is required. Based upon the proposed RFGs, a crossover and mutation module is further designed to inherit the facial parts of parents. A Gene Pool has also been used to introduce the variations into the mutation of RFGs. The diversity of the faces of descendants can thus be significantly increased. Qualitative, quantitative, and subjective experiments on FIW, TSKinFace, and FF-databases clearly show that the quality and diversity of kinship faces generated by our approach are much better than the existing state-of-the-art methods.
暂无评论