Multilayer composite films have gained widespread application across a variety of industries and applications due to their unique properties and functionalities. The precise control of the thickness of each layer ensu...
详细信息
Multilayer composite films have gained widespread application across a variety of industries and applications due to their unique properties and functionalities. The precise control of the thickness of each layer ensures that every layer contributes optimally to the desired attributes. In this study, a method using a one-dimensional convolutional neuralnetwork (1D-CNN) was developed to analyze the spectral data, five kinds of plastic films were assembled with different thickness combinations and measured with FTIR, which provided the data sets for 1D-CNN modeling. Compared to partial least squares (PLS) and fully connected neuralnetwork (FCNN) models, the 1D-CNN model performs better in accuracy of the thickness prediction, with respective root mean square error (RMSE) values dropping from < 9.11 mu m to < 0.31 mu m for film thickness of 13-270 mu m. The developed regression model was visualized using the gradient weighted class activation mapping (Grad-CAM) method for detailed analysis and selection of significant wavenumbers in order to build a more compact 1D-CNN model. Altogether, these results show that the proposed spectral 1D-CNN technique can measure the thickness of components in composite plastic films rapidly and accurately, and has the potential to improve efficiency and reduce the cost of instruments and calculations for spectral quantitation applications.
High-impedance fault (HIF) detection has always been difficult in distribution networks due to the lack of field data and the large difference between field and simulation waveforms. Based on the characteristics of ze...
详细信息
High-impedance fault (HIF) detection has always been difficult in distribution networks due to the lack of field data and the large difference between field and simulation waveforms. Based on the characteristics of zero-sequence currents, a novel HIF detection methodology is proposed, which combines time-frequency spectrum (TFS) and transfer convolutional neuralnetwork (TCNN). First, the TFSs are acquired by applying continuous wavelet transform (CWT) to the collected zero-sequence currents. Then, the TFSs of simulated zero-sequence currents are utilized for training source-domain convolutional neuralnetwork (SCNN). Next, the SCNN is transfer learned with very few TFSs of field zero-sequence currents to obtain TCNN. The performance of the proposed method is verified by simulation samples and field samples. The results show that the proposed method can effectively extract fault features from small-scale training samples under different fault circumstances. Besides, TCNN can adaptively extract the effective features of field HIF and detect field HIF more accurately than SCNN. Finally, this article provides a visualization scheme for interpretability of the neuralnetwork, which offers visual explanations for the decision-making basis of the neuralnetwork.
Deep learning has achieved great success in a variety of research fields and industrial ***,when applied to seismic inversion,the shortage of labeled data severely influences the performance of deep learning-based ***...
详细信息
Deep learning has achieved great success in a variety of research fields and industrial ***,when applied to seismic inversion,the shortage of labeled data severely influences the performance of deep learning-based *** order to tackle this problem,we propose a novel seismic impedance inversion method based on a cycle-consistent generative adversarial network(Cycle-GAN).The proposed Cycle-GAN model includes two generative subnets and two discriminative *** kinds of loss,including cycle-consistent loss,adversarial loss,and estimation loss,are adopted to guide the training *** from the proposed structure,the information contained in unlabeled data can be extracted,and adversarial learning further guarantees that the prediction results share similar distributions with the real ***,a neural network visualization method is adopted to show that the proposed CNN model can learn more distinguishable features than the conventional CNN *** robustness experiments on synthetic data sets show that the proposed method can achieve better performances than other methods in most *** the blind-well experiments on real seismic profiles show that the predicted impedance curve of the proposed method maintains a better correlation with the true impedance curve.
Despite their remarkable performance, the explainability of Vision Transformers (ViTs) remains a challenge. While forward attention-based token attribution techniques have become popular in text processing, their suit...
详细信息
ISBN:
(纸本)9798350365474
Despite their remarkable performance, the explainability of Vision Transformers (ViTs) remains a challenge. While forward attention-based token attribution techniques have become popular in text processing, their suitability for ViTs hasn't been extensively explored. In this paper, we compare these methods against state-of-the-art input attribution methods from the Vision literature, revealing their limitations due to improper aggregation of information across layers. To address this, we introduce two general techniques, PLUS and SkipPLUS, that can be composed with any input attribution method to more effectively aggregate information across layers while handling noisy layers. Through comprehensive and quantitative evaluations of faithfulness and human interpretability on a variety of ViT architectures and datasets, we demonstrate the effectiveness of PLUS and SkipPLUS, establishing a new state-of-the-art in white-box token attribution. We conclude with a comparative analysis highlighting the strengths and weaknesses of the best versions of all the studied methods. The code used in this paper is freely available at https://***/NightMachinery/SkipPLUS-CVPR-2024.
The striking results of deep neuralnetworks (DNN) have motivated its wide acceptance to tackle large datasets and complex tasks such as natural language processing, facial recognition, and artificial image generation...
详细信息
The striking results of deep neuralnetworks (DNN) have motivated its wide acceptance to tackle large datasets and complex tasks such as natural language processing, facial recognition, and artificial image generation. However, DNN parameters are often empirically selected on a trial-and-error approach without detailed information on convergence behavior. While some visualization techniques have been proposed to aid the comprehension of general-purpose neuralnetworks, only a few explore the training process, lacking the ability to adequately display how abstract representations are formed and represent the influence of training parameters during this process. This paper describes neuralnetwork training fingerprint (NNTF), a visual analytics approach to investigate the training process of any neuralnetwork performing classification. NNTF allows understanding how classification decisions change along the training process, displaying information about convergence, oscillations, and training rates. We show its usefulness through case studies and demonstrate how it can support the analysis of training parameters.
Objective. Functional specialization is fundamental to neural information processing. Here, we study whether and how functional specialization emerges in artificial deep convolutional neuralnetworks (CNNs) during a b...
详细信息
Objective. Functional specialization is fundamental to neural information processing. Here, we study whether and how functional specialization emerges in artificial deep convolutional neuralnetworks (CNNs) during a brain-computer interfacing (BCI) task. Approach. We trained CNNs to predict hand movement speed from intracranial electroencephalography (iEEG) and delineated how units across the different CNN hidden layers learned to represent the iEEG signal. Main results. We show that distinct, functionally interpretable neural populations emerged as a result of the training process. While some units became sensitive to either iEEG amplitude or phase, others showed bimodal behavior with significant sensitivity to both features. Pruning of highly sensitive units resulted in a steep drop of decoding accuracy not observed for pruning of less sensitive units, highlighting the functional relevance of the amplitude- and phase-specialized populations. Significance. We anticipate that emergent functional specialization as uncovered here will become a key concept in research towards interpretable deep learning for neuroscience and BCI applications.
As the success of deep models has led to their deployment in all areas of computer vision, it is increasingly important to understand how these representations work and what they are capturing. In this paper, we shed ...
详细信息
As the success of deep models has led to their deployment in all areas of computer vision, it is increasingly important to understand how these representations work and what they are capturing. In this paper, we shed light on deep spatiotemporal representations by visualizing the internal representation of models that have been trained to recognize actions in video. We visualize multiple two-stream architectures to show that local detectors for appearance and motion objects arise to form distributed representations for recognizing human actions. Key observations include the following. First, cross-stream fusion enables the learning of true spatiotemporal features rather than simply separate appearance and motion features. Second, the networks can learn local representations that are highly class specific, but also generic representations that can serve a range of classes. Third, throughout the hierarchy of the network, features become more abstract and show increasing invariance to aspects of the data that are unimportant to desired distinctions (e.g. motion patterns across various speeds). Fourth, visualizations can be used not only to shed light on learned representations, but also to reveal idiosyncrasies of training data and to explain failure cases of the system.
Analyzing and understanding how abstract representations of data are formed inside deep neuralnetworks is a complex task. Among the different methods that have been developed to tackle this problem, multidimensional ...
详细信息
ISBN:
(纸本)9789897584022
Analyzing and understanding how abstract representations of data are formed inside deep neuralnetworks is a complex task. Among the different methods that have been developed to tackle this problem, multidimensional projection techniques have attained positive results in displaying the relationships between data instances, network layers or class features. However, these techniques are often static and lack a way to properly keep a stable space between observations and properly convey flow in such space. In this paper, we employ different dimensionality reduction techniques to create a visual space where the flow of information inside hidden layers can come to light. We discuss the application of each used tool and provide experiments that show how they can be combined to highlight new information about neuralnetwork optimization processes.
Deep neuralnetworks are known for impressive results in a wide range of applications, being responsible for many advances in technology over the past few years. However, debugging and understanding neuralnetworks mo...
详细信息
Deep neuralnetworks are known for impressive results in a wide range of applications, being responsible for many advances in technology over the past few years. However, debugging and understanding neuralnetworks models' inner workings is a complex task, as there are several parameters and variables involved in every decision. Multidimensional projection techniques have been successfully adopted to display neuralnetwork hidden layer outputs in an explainable manner, but comparing different outputs often means overlapping projections or observing them side-by-side, presenting hurdles for users in properly conveying data flow. In this paper, we introduce a novel approach for comparing projections obtained from multiple stages in a neuralnetwork model and visualizing differences in data perception. Changes among projections are transformed into trajectories that, in turn, generate vector fields used to represent the general flow of information. This representation can then be used to create layouts that highlight new information about abstract structures identified by neuralnetworks.
Keeping record of daily meal intake is an effective solution for tackling with obesity and overweight. This can be done by developing apps on smartphones that are able to automatically recommend a short list of most p...
详细信息
Keeping record of daily meal intake is an effective solution for tackling with obesity and overweight. This can be done by developing apps on smartphones that are able to automatically recommend a short list of most probable foods by analyzing the photo taken from food. Then, the user chooses the correct answer from the short list. Hence, the automatic food recognition system must be able to recommend an accurate list. In other words, it is not essential for these apps to have a very high top-1 accuracy. Considering that the app will show the list of 5 most probable foods, the food recognition system must have a high top-5 accuracy. A food recognition system is usually developed by adapting knowledge of state-of-the-art networks such as GoogleNet and ResNet to the domain of food. However, these networks have high number of parameters. In this paper, we propose a 23-layer architecture which has 99.14% and 96.63% fewer parameter compared with ResNet and GoogleNet. Our experiment on Food101 and UECFood-256 datasets shows that although our network reduces the number of parameters dramatically, it produces more accurate results than GoogleNet and its accuracy is comparable with ResNet. (c) 2017 Elsevier B.V. All rights reserved.
暂无评论