Melt density is a crucial quality indicator for polymer composites, yet real-time measurement remains challenging due to processing complexities. While existing machine learning methods offer solutions, they often fal...
详细信息
Melt density is a crucial quality indicator for polymer composites, yet real-time measurement remains challenging due to processing complexities. While existing machine learning methods offer solutions, they often fall short in complex compounding scenarios. This study presents a novel multi-source data-driven approach for measuring melt density in polycarbonate/acrylonitrile butadiene styrene blends. By incorporating ultrasonic, near-infrared, and Raman spectra data acquired during melt processing, a deep separable convolutional neural network model is developed to predict melt density accurately. The model effectively fuses multi-source data to establish the mapping relationship between input data and melt density output. Results demonstrate the model's ability to monitor melt density in real-time, achieving a prediction accuracy with RMSE and R2 indexes of 0.005 g/cm3 and 0.9841, respectively. The proposed approach outperforms existing methods, showcasing its effectiveness and superiority in melt density prediction for polymer compounding *** Establishment of the real-time monitoring system for polymer extrusion processes. Conversion of multi-sensor signals into time-frequency images using wavelet decomposition. Fusion of sensor data into a three-channel tensor-image. Development of a data-driven DSCNN model for predicting melt density. Implementation for online monitoring and prediction in PC/ABS compounding system. Sensor data acquisition and DSCNN model monitoring process diagram. image
In this manuscript, we propose a novel method to perform audio inpainting, i.e., the restoration of audio signals presenting multiple missing parts. Audio inpainting can be interpreted in the context of inverse proble...
详细信息
In this manuscript, we propose a novel method to perform audio inpainting, i.e., the restoration of audio signals presenting multiple missing parts. Audio inpainting can be interpreted in the context of inverse problems as the task of reconstructing an audio signal from its corrupted observation. For this reason, our method is based on a deep prior approach, a recently proposed technique that proved to be effective in the solution of many inverse problems, among which image inpainting. deep prior allows one to consider the structure of a neural network as an implicit prior and to adopt it as a regularizer. Differently from the classical deeplearning paradigm, deep prior performs a single-element training and thus it can be applied to corrupted audio signals independently from the available training data sets. In the context of audio inpainting, a network presenting relevant audio priors will possibly generate a restored version of an audio signal, only provided with its corrupted observation. Our method exploits a time-frequency representation of audio signals and makes use of a multi-resolution convolutional autoencoder, that has been enhanced to perform the harmonic convolution operation. Results show that the proposed technique is able to provide a coherent and meaningful reconstruction of the corrupted audio. It is also able to outperform the methods considered for comparison, in its domain of application.
Application of artificial intelligence methods in agriculture is gaining research attention with focus on improving planting, harvesting, post-harvesting, etc. Fruit quality recognition is crucial for farmers during h...
详细信息
Application of artificial intelligence methods in agriculture is gaining research attention with focus on improving planting, harvesting, post-harvesting, etc. Fruit quality recognition is crucial for farmers during harvesting and sorting, for food retailers for quality monitoring, and for consumers for freshness evaluation, etc. However, there is a lack of multi-fruit datasets to support real-time fruit quality evaluation. To address this gap, we present a new dataset of fruit images aimed at evaluating fruit freshness, which addresses the lack of multi-fruit datasets for real-time fruit quality evaluation. The dataset contains images of 11 fruits categorized into three freshness classes, and five well-known deeplearning models (ShuffleNet, SqueezeNet, EfficientNet, ResNet18, and MobileNet-V2) were adopted as baseline models for fruit quality recognition using the dataset. The study provides a benchmark dataset for the classification task, which could improve research endeavors in the field of fruit quality recognition. The dataset is systematically organized and annotated, making it suitable for testing the performance of state-of-the-art methods and new learning classifiers. The research community in the fields of computer vision, machine learning, and pattern recognition could benefit from this dataset by applying it to various research tasks such as fruit classification and fruit quality recognition. The study achieved impressive results with the best classifier being ResNet-18 with an overall best performance of 99.8% for accuracy. The study also identified limitations, such as the small size of the dataset, and proposed future work to improve deeplearning techniques for fruit quality classification tasks.
Computer vision is a promising domain that focuses on emerging approaches, algorithms and technologies to provide computing capability to machine to analysis visual data, such as image files, videos files and real tim...
详细信息
In the field of ship detection in synthetic aperture radar (SAR) images using deeplearning, traditional models often face challenges related to their complex structure and high computational demands. To address these...
详细信息
In the field of ship detection in synthetic aperture radar (SAR) images using deeplearning, traditional models often face challenges related to their complex structure and high computational demands. To address these problems, this paper proposes a lightweight end-to-end convolutional neural network called BGD-Net, which is based on anchor-free methods. This network incorporates a novel feature pyramid called Ghost, Cross-Stage-Partial, and Path Aggregation Network (G-CSP-PAN) to enhance the detection performance across targets of varying scales. Additionally, it introduces an efficient decoupled detection head, termed the efficient decoupled head (ED-Head), to enhance the interaction between regression and classification. Furthermore, an optimized loss function named optimized efficient intersection over union (OEIoU) loss is proposed for edge regression. The proposed method is evaluated on public datasets, SSDD and SAR-Ship-Dataset, demonstrating a balance between detection accuracy and efficiency.
The rolling elements of the induction motor are highly susceptible to faults. The detection and diagnosis of rolling element faults are accurate and reliable only when the extracted features are accurate. The paper pr...
详细信息
The rolling elements of the induction motor are highly susceptible to faults. The detection and diagnosis of rolling element faults are accurate and reliable only when the extracted features are accurate. The paper proposes an approach for bearing and rotor fault diagnosis using deep optimal feature extraction and selection based on vibration signal analysis. The deep feature extraction is done using an ensemble deep models features extraction approach in which features are extracted from seven pretrained models are fused serially using serial-based feature fusion technique. This leads to a solution for a higher efficacy model, but at the cost of high processingtime as the feature data set gets large. A unique approach termed Ensemble Feature Selection has been developed to address this issue and limit the harmful impact of unwanted features in data-driven diagnostics. The processingtime is further reduced using the shallow classifier at the fully connected layer. The proposed model is tested using the data acquired in the laboratory and validated using the available online benchmark data sets.
deeplearning is developing rapidly, and the emergence of many network architectures has brought significant breakthroughs to training recognition models. Due to the maturity of edge computing technology, we can perfo...
详细信息
deeplearning is developing rapidly, and the emergence of many network architectures has brought significant breakthroughs to training recognition models. Due to the maturity of edge computing technology, we can perform regional image training through distributed nodes, which significantly improves the training model's accuracy while performing transfer learning to achieve better performance. In imageprocessing technology, high-precision recognition of non-luminous images can currently be achieved by modeling, if we replace the visual recognition target with a glowing digital panel, the recognition rate cannot be the same as the static text recognition rate. This article uses Keras to build a convolutional neural networks deeplearning model to identify glowing light-emitting diodes (LED) digits, incremental learning to complete transfer learning on edge computing nodes, and an integrated IoT architecture to achieve better recognition results. In the experiment, the verification results obtained from the distributed training nodes were successfully combined to model and retrain the nodes. The proposed distributed learning method can increase the accuracy from 70% to 89%. At the same time, the misclassified images can be retrained by integrating the transfer learning model with the distributed learning results, and the accuracy reaches more than 92%.
In this study, we address the critical task of early bush-fire detection in remote areas through UAV imaging, a vital effort for protecting inhabitants, infrastructure, and ecosystems within the smart cities paradigm....
详细信息
ISBN:
(纸本)9783031663284;9783031663291
In this study, we address the critical task of early bush-fire detection in remote areas through UAV imaging, a vital effort for protecting inhabitants, infrastructure, and ecosystems within the smart cities paradigm. We introduce FireNet, a deeplearning framework that employs a hybrid Inception Residual Neural Network (Inception-ResNet) model, specifically designed for the rapid and precise classification of fire and non-fire regions in real-timeimageprocessing scenarios. FireNet synergises the strengths of both Inception and ResNet architectures, enhancing the model's feature extraction capabilities while significantly reducing computational overhead. A distinctive aspect of FireNet is the adoption of the HardSwish activation function, demonstrating superior performance over the conventional Rectified Linear Unit (ReLU) in our fire detection cases. Through rigorous evaluation of a robust dataset, FireNet achieves an impressive accuracy of 96%, with a corresponding AUC of approximately 96%. These results not only affirm FireNet's efficacy in accurately identifying fire and non-fire areas but also underscore its importance as a crucial tool for real-time fire monitoring and rescue operations, especially in the context of smart cities. With its outstanding efficiency and accuracy, FireNet represents a significant leap forward in the domain of fire detection technologies, highlighting the role of deeplearning in enhancing urban resilience against fires.
Social networks like LinkedIn, Facebook, and Instagram contribute significantly to the rise of image prevalence in daily life, with numerous images posted in everyday. Detecting image similarity is crucial for many ap...
详细信息
Social networks like LinkedIn, Facebook, and Instagram contribute significantly to the rise of image prevalence in daily life, with numerous images posted in everyday. Detecting image similarity is crucial for many applications. While deeplearning methods like Learned Perceptual image Patch Similarity (LPIPS) are popular, they often overlook image structure. An alternative method involves using pre-trained models ($e.g.$, LeNet-$5$ and VGG-$16$) to extract features and employing classifiers. However, deeplearning methods demand substantial computational resources and they also suffer from uncontrolled false alarms. This paper proposes a novel Generalized Likelihood Ratio Test (GLRT) detector based on a hypothesis testing framework to identify the similarity of structural image pairs. The proposed approach minimizes the need for extensive computational resources, and false alarms can be regulated by employing a threshold. The detector is applied to Local Dissimilarity Maps (LDM), with gray-level values modeled by a statistical distribution. Experimental results on simulated and real data confirm its effectiveness for structural similarity detection. Additionally, a Simple Likelihood Ratio Test (SLRT) is tested on simulated data. Comparisons with deeplearning and classical measures like Structural Similarity Index (SSIM) and Feature Similarity Index (FSIM) show the proposed detector performs comparably or better in terms of Area Under the Curve (AUC) with less computing time, especially for structural similarity.
deep neural networks excel at image identification and computer vision applications such as visual product search, facial recognition, medical image analysis, object detection, semantic segmentation,instance segmentat...
详细信息
deep neural networks excel at image identification and computer vision applications such as visual product search, facial recognition, medical image analysis, object detection, semantic segmentation,instance segmentation, and many others. In image and video recognition applications, convolutional neural networks(CNNs) are widely employed. These networks provide better performance but at a higher cost of computation. With the advent of big data, the growing scale of datasets has made processing and model training a time-consuming operation, resulting in longer training times. Moreover, these large scale datasets contain redundant data points that have minimum impact on the final outcome of the model. To address these issues, an accelerated CNN system is proposed for speeding up training by eliminating the noncritical data points during training alongwith a model compression method. Furthermore, the identification of the critical input data is performed by aggregating the data points at two levels of granularity which are used for evaluating the impact on the model *** experiments are conducted using the proposed method on CIFAR-10 dataset on ResNet models giving a 40% reduction in number of FLOPs with a degradation of just 0.11% accuracy.
暂无评论