Egocentric activity recognition has recently generated great popularity in computer vision due to its widespread applications in egocentric video analysis. However, it poses new challenges comparing to the conventiona...
详细信息
Egocentric activity recognition has recently generated great popularity in computer vision due to its widespread applications in egocentric video analysis. However, it poses new challenges comparing to the conventional third-person activity recognition tasks, which are caused by significant body shaking, varied lengths, and poor recoding quality, etc. To handle these challenges, in this paper, we propose deep appearance and motion learning (DAML) for egocentric activity recognition, which leverages the great strength of deep learning networks in feature learning. In contrast to hand- crafted visual features or pre-trained convolutional neural network (CNN) features with limited generality to new egocentric videos, the proposed DAML is built on the deep autoencoder (DAE), and directly extracts appearance and motion feature, the main cue of activities, from egocentric videos. The DAML takes advantages of the great effectiveness and efficiency of the DAE in unsupervised feature learning, which provides a new representation learning framework of egocentric videos. The learned appearance and motion features by the DAML are seamlessly fused to accomplish a rich informative egocentric activity representation which can be readily fed into any supervised learning models for activity recognition. Experimental results on two challenging benchmark datasets show that the DAML achieves high performance on both short- and long-term egocentric activity recognition tasks, which is comparable to or even better than the state-of-the-art counterparts. (C) 2017 Elsevier B.V. All rights reserved.
Deep learning, in particular the deep convolutional neural networks, has received increasing interests in face recognition recently, and a number of deep learning methods have been proposed. This paper summarizes abou...
详细信息
Deep learning, in particular the deep convolutional neural networks, has received increasing interests in face recognition recently, and a number of deep learning methods have been proposed. This paper summarizes about 330 contributions in this area. It reviews major deep learning concepts pertinent to face image analysis and face recognition, and provides a concise overview of studies on specific face recognition problems, such as handling variations in pose, age, illumination, expression, and heterogeneous face matching. A summary of databases used for deep face recognition is given as well. Finally, some open challenges and directions are discussed for future research.
The purpose of this paper is to remove the exponentially decaying DC offset in fault current waveforms using a deep neural network (DNN), even under harmonics and noise distortion. The DNN is implemented using the Ten...
详细信息
The purpose of this paper is to remove the exponentially decaying DC offset in fault current waveforms using a deep neural network (DNN), even under harmonics and noise distortion. The DNN is implemented using the TensorFlow library based on Python. autoencoders are utilized to determine the number of neurons in each hidden layer. Then, the number of hidden layers is experimentally decided by comparing the performance of DNNs with different numbers of hidden layers. Once the optimal DNN size has been determined, intensive training is performed using both the supervised and unsupervised training methodologies. Through various case studies, it was verified that the DNN is immune to harmonics, noise distortion, and variation of the time constant of the DC offset. In addition, it was found that the DNN can be applied to power systems with different voltage levels.
Hyperspectral unmixing is an important and challenging task in the field of remote sensing which arises when the spatial resolution of sensors is insufficient for the separation of spectrally distinct materials. Hyper...
详细信息
Hyperspectral unmixing is an important and challenging task in the field of remote sensing which arises when the spatial resolution of sensors is insufficient for the separation of spectrally distinct materials. Hyperspectral images, like other natural images, have highly correlated pixels and it is very desirable to make use of this spatial information. In this paper, a deep learning based method for blind hyperspectral unmixing is presented. The method uses multitask learning through multiple parallel autoencoders to unmix a neighborhood of pixels simultaneously. Operating on image patches instead of single pixels enables the method to take advantage of spatial information in the hyperspectral image. The method is the first in its class to directly utilize the spatial structure of hyperspectral images (HSIs) for the estimation of the spectral signatures of endmembers in the data cube. We evaluate the proposed method using two real HSIs and compare it to seven state-of-the-art methods that either rely only on spectral or both on spectral and spatial information in the HSIs. The proposed method outperforms all the baseline unmixing methods in experiments.
Online social networks, World Wide Web, media, and technological networks, and other types of so-called Information networks are ubiquitous nowadays. These information networks are inherently heterogeneous and dynamic...
详细信息
Online social networks, World Wide Web, media, and technological networks, and other types of so-called Information networks are ubiquitous nowadays. These information networks are inherently heterogeneous and dynamic. They are heterogeneous as they consist of multi-typed objects and relations, and they are dynamic as they are constantly evolving over time. One of the challenging issues in such heterogeneous and dynamic environments is to forecast those relationships in the network that will appear in the future. In this article, we try to solve the problem of continuous-time relationship prediction in dynamic and heterogeneous information networks. This implies predicting the time it takes for a relationship to appear in the future, given its features that have been extracted by considering both heterogeneity and temporal dynamics of the underlying network. To this end, we first introduce a feature extraction framework that combines the power of meta-path-based modeling and recurrent neural networks to effectively extract features suitable for relationship prediction regarding heterogeneity and dynamicity of the networks. Next, we propose a supervised non-parametric approach, called Non-Parametric Generalized Linear Model (NIP-GLM), which infers the hidden underlying probability distribution of the relationship building time given its features. We then present a learning algorithm to train NP-GLM and an inference method to answer time-related queries. Extensive experiments conducted on synthetic data and three real-world datasets, namely Delicious, MovieLens, and DBLP, demonstrate the effectiveness of NP-am in solving continuous-time relationship prediction problem vis-a-vis competitive baselines.
PurposeConvolutional neural networks have become rapidly popular for image recognition and image analysis because of its powerful potential. In this paper, we developed a method for classifying subtypes of lung adenoc...
详细信息
PurposeConvolutional neural networks have become rapidly popular for image recognition and image analysis because of its powerful potential. In this paper, we developed a method for classifying subtypes of lung adenocarcinoma from pathological images using neural network whose that can evaluate phenotypic features from wider area to consider cellular *** order to recognize the types of tumors, we need not only to detail features of cells, but also to incorporate statistical distribution of the different types of cells. Variants of autoencoders as building blocks of pre-trained convolutional layers of neural networks are implemented. A sparse deep autoencoder which minimizes local information entropy on the encoding layer is then proposed and applied to images of size . We applied this model for feature extraction from pathological images of lung adenocarcinoma, which is comprised of three transcriptome subtypes previously defined by the Cancer Genome Atlas network. Since the tumor tissue is composed of heterogeneous cell populations, recognition of tumor transcriptome subtypes requires more information than local pattern of cells. The parameters extracted using this approach will then be used in multiple reduction stages to perform classification on larger *** were able to demonstrate that these networks successfully recognize morphological features of lung adenocarcinoma. We also performed classification and reconstruction experiments to compare the outputs of the variants. The results showed that the larger input image that covers a certain area of the tissue is required to recognize transcriptome subtypes. The sparse autoencoder network with input provides a 98.9% classification *** study shows the potential of autoencoders as a feature extraction paradigm and paves the way for a whole slide image analysis tool to predict molecular subtypes of tumors from pathological features.
This paper presents a novel automatic facial expressions recognition system (AFERS) using the deep network framework. The proposed AFERS consists of four steps: 1) geometric features extraction;2) regional local binar...
详细信息
This paper presents a novel automatic facial expressions recognition system (AFERS) using the deep network framework. The proposed AFERS consists of four steps: 1) geometric features extraction;2) regional local binary pattern (LBP) features extraction;3) fusion of both the features using autoencoders;and 4) classification using Kohonen self-organizing map (SOM)-based classifier. This paper makes three distinct contributions. The proposed deep network consisting of autoencoders and the SOM-based classifier is computationally more efficient and performance wise more accurate. The fusion of geometric features with LBP features using autoencoders provides better representation of facial expression. The SOM-based classifier proposed in this paper has been improved by making use of a soft-threshold logic and a better learning algorithm. The performance of the proposed approach is validated on two widely used databases (DBs): 1) MMI and 2) extended Cohn-Kanade (CK+). An average recognition accuracy of 97.55% in MMI DB and 98.95% in CK+ DB are obtained using the proposed algorithm. The recognition results obtained from fused features are found to be distinctly superior to both recognition using individual features as well as recognition with a direct concatenation of the individual feature vectors. Simulation results validate that the proposed AFERS is more efficient as compared to the existing approaches.
This paper aims to design and implement a system capable of distinguishing between different activities carried out during a tennis match. The goal is to achieve the correct classification of a set of tennis strokes. ...
详细信息
This paper aims to design and implement a system capable of distinguishing between different activities carried out during a tennis match. The goal is to achieve the correct classification of a set of tennis strokes. The system must exhibit robustness to the variability of the height, age or sex of any subject that performs the actions. A new database is developed to meet this objective. The system is based on two sensor nodes using Bluetooth Low Energy (BLE) wireless technology to communicate with a PC that acts as a central device to collect the information received by the sensors. The data provided by these sensors are processed to calculate their spectrograms. Through the application of innovative deep learning techniques with semi-supervised training, it is possible to carry out the extraction of characteristics and the classification of activities. Preliminary results obtained with a data set of eight players, four women and four men have shown that our approach is able to address the problem of the diversity of human constitutions, weight and sex of different players, providing accuracy greater than 96.5% to recognize the tennis strokes of a new player never seen before by the system.
Gear pitting fault is one of the most common faults in mechanical transmission. Acoustic emission (AE) signals have been effective for gear fault detection because they are less affected by ambient noise than traditio...
详细信息
Gear pitting fault is one of the most common faults in mechanical transmission. Acoustic emission (AE) signals have been effective for gear fault detection because they are less affected by ambient noise than traditional vibration signals. To overcome the problem of low gear pitting fault recognition rate using AE signals and convolutional neural networks, this paper proposes a new method named augmented convolution sparse autoencoder (ACSAE) for gear pitting fault diagnosis using raw AE signals. First, the proposed method combines sparse autoencoder and one-dimensional convolutional neural networks for unsupervised learning and then uses the reinforcement theory to enhance the adaptability and robustness of the network. The ACSAE method can automatically extract fault features directly from the original AE signals without time and frequency domain conversion of the AE signals. AE signals collected from gear test experiments are used to validate the ACME method. The analysis result of the gear pitting fault test shows that the proposed method can effectively performing recognition of the gear pitting faults, and the recognition rate reaches above 98%. The comparative analysis shows that in comparison with fully-connected neural networks, convolutional neural networks, and recurrent neural networks, the ACSAE method has achieved a better diagnostic accuracy for gear fitting faults.
Image transmission holds a major share in data communication, and thus secure image transmission is currently a challenging domain of research. A secure image transmission scheme is proposed that physically transmits ...
详细信息
Image transmission holds a major share in data communication, and thus secure image transmission is currently a challenging domain of research. A secure image transmission scheme is proposed that physically transmits the encrypted image employing visual cryptography scheme (VCS). During physical transmission, the meaningless shares may attract curious hackers and if captured and stacked, the secret may be revealed. Moreover, the increase in transmission overhead due to multiple share images resulted from a single secret image after encryption is another concern regarding the physical implementation of VCS. Focusing on both observations, vector quantization (VQ) is used to encode as well as to compress each of the shares before transmission. To utilize VQ, its two parameters, cell width and dimension of grid, are needed to be optimized for various kind of images without compromising the randomness property of the shares. Hence, a particle swarm optimization-guided VQ is proposed, and furthermore, a multilayer perceptron in conjunction with an autoencoder are also trained in synchronism with that to automatically obtain the optimal VQ for each image type during the transmission. The proposed scheme is successfully implemented with different types of images for secure physical transmission with a 62.8% data volume reduction and 98.07% image quality retrieval. (C) 2019 SPIE and IS&T
暂无评论