The operation conditions of the rotating machinery are always complex and variable, which makes it difficult to automatically and effectively capture the useful fault features from the measured vibration signals, and ...
详细信息
The operation conditions of the rotating machinery are always complex and variable, which makes it difficult to automatically and effectively capture the useful fault features from the measured vibration signals, and it is a great challenge for rotating machinery fault diagnosis. In this paper, a novel deep autoencoder feature learning method is developed to diagnose rotating machinery fault. Firstly, the maximum correntropy is adopted to design the new deep autoencoder loss function for the enhancement of feature learning from the measured vibration signals. Secondly, artificial fish swarm algorithm is used to optimize the key parameters of the deep autoencoder to adapt to the signal features. The proposed method is applied to the fault diagnosis of gearbox and electrical locomotive roller bearing. The results confirm that the proposed method is more effective and robust than other methods. (C) 2017 Elsevier Ltd. All rights reserved.
Network embedding aims to learn a low-dimensional vector for each node in networks, which is effective in a variety of applications such as network reconstruction and community detection. However, the majority of the ...
详细信息
Network embedding aims to learn a low-dimensional vector for each node in networks, which is effective in a variety of applications such as network reconstruction and community detection. However, the majority of the existing network embedding methods merely exploit the network structure and ignore the rich node attributes, which tend to generate sub-optimal network representation. To learn more desired network representation, diverse information of networks should be exploited. In this paper, we develop a novel deep autoencoder framework to fuse topological structure and node attributes named FSADA. We firstly design a multi-layer autoencoder which consists of multiple non-linear functions to capture and preserve the highly non-linear network structure and node attribute information. Particularly, we adopt a pre-processing procedure to pre-process the original information, which can bet -ter facilitate to extract the intrinsic correlations between topological structure and node attributes. In addition, we design an enhancement module that combines topology and node attribute similarity to construct pairwise constraints on nodes, and then a graph regularization is introduced into the frame-work to enhance the representation in the latent space. Our extensive experimental evaluations demon-strate the superior performance of the proposed method. (c) 2021 Elsevier B.V. All rights reserved.
Natural gas is widely used for domestic and industrial purposes, and whether it is being leaked into the air cannot be directly known. The current problem is that gas leakage is not only economically harmful but also ...
详细信息
Natural gas is widely used for domestic and industrial purposes, and whether it is being leaked into the air cannot be directly known. The current problem is that gas leakage is not only economically harmful but also detrimental to health. Therefore, much research has been done on gas damage and leakage risks, but research on predicting gas leakages is just beginning. In this study, we propose a method based on deep learning to predict gas leakage from environmental data. Our proposed method has successfully improved the performance of machine learning classification algorithms by efficiently preparing training data using a deep autoencoder model. The proposed method was evaluated on an open dataset containing natural gas and environmental information and compared with extreme gradient boost (XGBoost), K-nearest neighbors (KNN), decision tree (DT), random forest (RF), and naive Bayes (NB) algorithms. The proposed method is evaluated using accuracy, F1-score, mean square error (MSE), mean intersection over union (mIoU), and area under the ROC curve (AUC). The presented method in this study outperformed all compared methods. Moreover, the deep autoencoder and ordinal encoder-based XGBoost (DA-MA-XGBoost) showed the best performance by giving 99.51% accuracy, an F1-score of 99.53%, an MSE of 0.003, mIoU of 99.40 and an AUC of 99.62%.
The technology of the brain-computer interface (BCI) employs electroencephalogram (EEG) signals to establish direct interaction between the human body and its surroundings with promising applications in medical rehabi...
详细信息
The technology of the brain-computer interface (BCI) employs electroencephalogram (EEG) signals to establish direct interaction between the human body and its surroundings with promising applications in medical rehabilitative services and cognitive science. deep learning approaches, particularly the detection and analysis of motor imagery signals using convolutional neural network (CNN) frameworks have produced outstanding results in the BCI system in recent years. The complex process of data representation, on the other hand, limits practical applications, and the end-to-end approach reduces the accuracy of recognition. Moreover, since noise and other signal sources can interfere with brain electrical capacitance, EEG classifiers are difficult to improve and have limited generalisation ability. To address these issues, this paper proposes a new approach for EEG motor imagery signal classification by using a variational autoencoder to remove noise from the signals, followed by a combination of deep autoencoder (DAE) and a CNN architecture to classify EEG motor imagery signals which is capable of training a deep neural network to replicate its input to output using encoding and decoding operations. Experimental results show that the proposed approach for motor imagery EEG signal classification is feasible and that it outperforms current CNN-based approaches and several traditional machine learning approaches.
Network embedding has recently attracted lots of attention due to its wide applications on graph tasks such as link prediction, network reconstruction, node stabilization, and community stabilization, which aims to le...
详细信息
Network embedding has recently attracted lots of attention due to its wide applications on graph tasks such as link prediction, network reconstruction, node stabilization, and community stabilization, which aims to learn the low-dimensional representations of nodes with essential features. Most existing network embedding methods mainly focus on static or continuous evolution patterns of microscopic node and link structures in networks, while neglecting the dynamics of macroscopic community structures. In this paper, we propose a Community-aware Dynamic Network Embedding method (short for CDNE) which considers the dynamics of macroscopic community structures. First, we model the problem of dynamic network embedding as a minimization of an overall loss function, which tries to maximally preserve the global node structures, local link structures, and continuous community dynamics. Then, we adopt a stacked deep autoencoder algorithm to solve this minimization problem, obtaining the low-dimensional representations of nodes. Extensive experiments on both synthetic networks and real networks demonstrate the superiority of CDNE over the existing methods on tackling various graph tasks. (C) 2020 Elsevier Inc. All rights reserved.
Increased accuracy and affordability of depth sensors such as Kinect has created a great depth-data source for various 3D oriented applications. Specifically, 3D model retrieval is attracting attention in the field of...
详细信息
Increased accuracy and affordability of depth sensors such as Kinect has created a great depth-data source for various 3D oriented applications. Specifically, 3D model retrieval is attracting attention in the field of computer vision and pattern recognition due to its numerous applications. A cross-domain retrieval approach such as depth image based 3D model retrieval has the challenges of occlusion, noise and view variability present in both query and training data. In this paper, we propose a new supervised deep autoencoder approach followed by semantic modeling to retrieve 3D shapes based on depth images. The key novelty is the two-fold feature abstraction to cope with the incompleteness and ambiguity present in the depth images. First, we develop a supervised autoencoder to extract robust features from both real depth images and synthetic ones rendered from 3D models, which are intended to balance reconstruction and classification capabilities of mix-domain data. Then semantic modeling of the supervised autoencoder features offers the next level of abstraction to cope with the incompleteness and ambiguity of the depth data. It is interesting that unlike any other pairwise model structures, we argue that cross-domain retrieval is still possible using only one single deep network trained on real and synthetic data. The experimental results on the NYUD2 and ModelNet10 datasets demonstrate that the proposed supervised method outperforms the recent approaches for cross-modal 3D model retrieval. (C) 2019 Elsevier B.V. All rights reserved.
Evidential clustering is a promising clustering framework using Dempster-Shafer belief function theory to model uncertain data. However, evidential clustering needs to estimate more parameters compared with other clus...
详细信息
Evidential clustering is a promising clustering framework using Dempster-Shafer belief function theory to model uncertain data. However, evidential clustering needs to estimate more parameters compared with other clustering algorithms, and thus the clustering performance of evidential clustering will be greatly affected if data is insufficient or contaminated. In addition, the existing evidential clustering algorithms can not well deal with high-dimensional data such as texts and images. To solve the above problems, an evidential clustering algorithm based on transfer learning and deep autoencoder (TDEC) is proposed. The TDEC utilizes deep autoencoder to obtain evidential clustering-friendly representations of the original data, and applies the maximum mean discrepancy (MMD) constraint between the source network and the target network, so that the network can learn domain-invariant features. The algorithm jointly trains the deep evidential clustering networks in the source domain and the target domain, and realizes the deep feature representations of high-dimensional data in the target domain for evidential clustering by minimizing reconstruction loss, entropy-based evidential clustering loss, MMD loss and the regular penalty term of the network parameters. In addition, an iterative optimization method to solve the TDEC objective function is proposed. Extensive experiments were conducted to evaluate the clustering performance of the proposed TDEC algorithm compared with the existing shallow transfer clustering algorithms and deep clustering algorithms. For both image and text clustering tasks, the proposed TDEC achieved approximately 5% performance improvement over the comparison algorithms on average. In addition, the practical application value of the proposed TDEC algorithm was demonstrated in unsupervised remote sensing image scene classification.
Nonlinear processing of high-dimensional data is quite common in image filtering algorithms. Bilateral, joint bilateral, and non-local means filters are the examples of the same. Real-time implementation of high-dimen...
详细信息
Nonlinear processing of high-dimensional data is quite common in image filtering algorithms. Bilateral, joint bilateral, and non-local means filters are the examples of the same. Real-time implementation of high-dimensional filters has always been a research challenge due to its computational complexity. In this paper, we have proposed a solution utilizing both color sparseness and color dominance in an image which ensures a faster algorithm for generic high-dimensional filtering. The solution speeds up the filtering algorithm further by psycho-visual saliency-based deep encoded dominant color gamut, learned for different subject classes of images. The proposed bilateral filter has been proved to be efficient both in terms of psycho-visual quality and performance for edge-preserving smoothing and denoising of color images. The results demonstrate competitiveness of our proposed solution with the existing fast bilateral algorithms in terms of the CTQ (critical to quality) parameters.
Background Drug repositioning has caught the attention of scholars at home and abroad due to its effective reduction of the development cost and time of new drugs. However, existing drug repositioning methods that are...
详细信息
Background Drug repositioning has caught the attention of scholars at home and abroad due to its effective reduction of the development cost and time of new drugs. However, existing drug repositioning methods that are based on computational analysis are limited by sparse data and classic fusion methods;thus, we use autoencoders and adaptive fusion methods to calculate drug repositioning. Results In this study, a drug repositioning algorithm based on a deep autoencoder and adaptive fusion was proposed to mitigate the problems of decreased precision and low-efficiency multisource data fusion caused by data sparseness. Specifically, a drug is repositioned by fusing drug-disease associations, drug target proteins, drug chemical structures and drug side effects. First, drug feature data integrated by drug target proteins and chemical structures were processed with dimension reduction via a deep autoencoder to characterize feature representations more densely and abstractly. Then, disease similarity was computed using drug-disease association data, while drug similarity was calculated with drug feature and drug-side effect data. Predictions of drug-disease associations were also calculated using a top-k neighbor method that is commonly used in predictive drug repositioning studies. Finally, a predicted matrix for drug-disease associations was acquired after fusing a wide variety of data via adaptive fusion. Based on experimental results, the proposed algorithm achieves a higher precision and recall rate than the DRCFFS, SLAMS and BADR algorithms with the same dataset. Conclusion The proposed algorithm contributes to investigating the novel uses of drugs, as shown in a case study of Alzheimer's disease. Therefore, the proposed algorithm can provide an auxiliary effect for clinical trials of drug repositioning.
In this paper, we propose a novel image encryption method based on logistic chaotic systems and deep autoencoder. In the encryption phase, first, the plaintext image is randomly scrambled by a logistic chaotic system....
详细信息
In this paper, we propose a novel image encryption method based on logistic chaotic systems and deep autoencoder. In the encryption phase, first, the plaintext image is randomly scrambled by a logistic chaotic system. Then, the random scrambled image is encoded by a deep autoencoder to generate the ciphertext image. In order to obtain the ciphertext image with uniform distribution, we incorporated the uniform distribution constraint into the training of the deep autoencoder. The resulting ciphertext image contains high randomness, which is critical for an excellent image encryption algorithm. Histogram analysis, information entropy analysis, key space analysis, key sensitivity analysis, correlation analysis, and ablation experiments show that the proposed encryption algorithm can effectively resist attacks and has excellent encryption performance while providing high security. (c) 2021 Published by Elsevier B.V.
暂无评论