Detecting oil spills in marine environments is crucial for avoiding environmental damage and facilitating rapid response efforts. In this study, we propose a robust method for oil spill detection leveraging state-of-t...
详细信息
Detecting oil spills in marine environments is crucial for avoiding environmental damage and facilitating rapid response efforts. In this study, we propose a robust method for oil spill detection leveraging state-of-the-art (SOTA) deeplearning techniques. We constructed an extensive dataset comprising images and frames extracted from video sourced from Google, significantly augmenting the dataset through frame extraction techniques. Each image is meticulously labeled to ensure high-quality training data. Utilizing the Yolov8 segmentation model, we trained our oil spill detection model to accurately identify and segment oil spills in ocean environments. K-means and Truncated Linear Stretching algorithms are combined with trained model weight to increase model detection accuracy. The model demonstrated exceptional performance, yielding high detection accuracy and precise segmentation capabilities. Our results indicate that this approach is highly effective for real-time oil spill detection, offering a promising tool for environmental monitoring and disaster management. In training metrics, the model reached over 97% accuracy in 100 epochs. In evaluation, model achieved its best detection rates by 94% accuracy in F1, 93.9% accuracy in Precision, and 95.5% mAP@0.5 accuracy in Recall curves.
The 'Rootkit Detection using deeplearning with Reinforcement learning' task addresses the important venture of detecting and mitigating rootkits, insidious types of malwares that frequently pass neglected wit...
详细信息
This article mainly studies the method of deep recognition analysis of power technology standard images using Convolutional Neural Network (CNN). In this study, an efficient CNN model was constructed to quickly and ac...
详细信息
Coal gangue sorting is an important link in the process of coal mining and processing, which can effectively reduce the difficulty and cost of coal post-processing. Aiming at the problems of complicated sorting proces...
详细信息
The oil and natural gas industry is a crucial sector in our daily lives. Unwanted incidents in this sector can significantly impact the household sector. Therefore, an automatic early warning system is necessary to de...
详细信息
Single-shot three-dimensional (3D) imaging with compact device footprint, high imaging quality, and fast processing speed is challenging in computational imaging. Mask-based lensless imagers, which replace the bulky o...
详细信息
Single-shot three-dimensional (3D) imaging with compact device footprint, high imaging quality, and fast processing speed is challenging in computational imaging. Mask-based lensless imagers, which replace the bulky optics with customized thin optical masks, are portable and lightweight, and can recover 3D object from a snap-shot image. Existing lensless imaging typically requires extensive calibration of its point spread function and heavy computational resources to reconstruct the object. Here we overcome these challenges and demonstrate a compact and learnable lensless 3D camera for real-time photorealistic imaging. We custom designed and fabricated the optical phase mask with an optimized spatial frequency support and axial resolving ability. We developed a simple and robust physics-aware deeplearning model with adversarial learning module for real-time depth-resolved photorealistic reconstructions. Our lensless imager does not require calibrating the point spread function and has the capability to resolve depth and "see-through" opaque obstacles to image features being blocked, enabling broad applications in computational imaging.(c) 2022 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement
Quantitative magnetic resonance imaging (qMRI) measures physical, physiological or biological properties of tissues and thus provides reproducible imaging biomarkers for disease diagnosis and therapy response monitori...
详细信息
Quantitative magnetic resonance imaging (qMRI) measures physical, physiological or biological properties of tissues and thus provides reproducible imaging biomarkers for disease diagnosis and therapy response monitoring. However, long scanning and reconstruction time, low reproducibility, spatial resolution, and volume of coverage limit the clinical translation of *** overall goal of this dissertation is to improve qMRI by exploiting its sparsity by data-driven deeplearning methods. Such methods can provide more accurate and precise tissue parameters from highly undersampled or accelerated scans using a fraction of reconstruction time of conventional methods. Sampling patterns, image reconstruction and parameter estimation could be jointly optimized to directly minimize parameter quantification error under the same scan *** sparsity in dynamic contrast enhanced magnetic resonance imaging (DCE MRI) was exploited by a long short-term memory (LSTM) neural network-based approach to provide more robust tissue parameter estimation. The network was trained on simulated DCE signals and tested on both simulated and real data. Compared to a conventional linear least squares (LLSQ) fitting method, the LSTM-based approach had higher accuracy for the data with temporally subsampling, total acquisition time truncation, or high noise level. Also, the LSTM-based method reduced the inference time by ~14 times compared to the LLSQ fitting. Validation of the method on real data demonstrated its clinical feasibility to provide high-quality tissue parameter *** temporal sparsity, the spatiotemporal sparsity of DCE MRI was further exploited by convolutional recurrent neural network. 2D Cartesian phase encoding k-space subsampling patterns were jointly optimized with image reconstruction to identify the most informative k-space data to acquire beyond the learned population prior knowledge. Both reconstruction image quality and parameter estimation accuracy were use
Human activity recognition (HAR) employs a broad range of sensors that generate massive volumes of data. Traditional server-based and cloud computing methods require all sensor data to be sent to servers or clouds for...
详细信息
Human activity recognition (HAR) employs a broad range of sensors that generate massive volumes of data. Traditional server-based and cloud computing methods require all sensor data to be sent to servers or clouds for processing, which leads to high latency and bandwidth costs. The long-term data transfer between servers and sensors maximizes the cost of latency and bandwidth. real-timeprocessing is, nevertheless, highly required for human action identification. By bringing processing and quick data storage to the sensors instead of depending on a central database, edge computing is rapidly emerging as a solution to this issue. Artificial intelligence is responsible for most HAR, which demands a lot of processing power and calculation. Artificial intelligence (AI) needs more computation which is not allowed by edge computing. So Edge intelligence, which allows AI to operate at the network edge for actual-time applications, has been made possible by the advent of binarized neural networks. To provide less latency and less memory for human activity identification at the edge network, we construct a hybrid deeplearning-based binarized neural network (HDL-Binary Dilated DenseNet) in this research. Fractal HAR optimization algorithms could be applied to these algorithms. For example, fractal-HAR optimization techniques might be used to provide less latency and less memory human activity identification at the edge network. Using three sensors-based human activity detection datasets such as Radar HAR dataset, UCI HAR dataset and UniMib-SHAR dataset, we implemented the Hybrid Binary Dilated Dense Net. It is then assessed using four criteria. Comparatively, the Hybrid Binary Dilated DenseNet performs better with 99.6% radar HAR dataset which is highest than other models like CNN-BiLSTM and GoogLeNet.
Examining otoscopic images for ear diseases is necessary when the clinical diagnosis of ear diseases extracted from the knowledge of otolaryngologists is limited. Improved diagnosis approaches based on otoscopic image...
详细信息
Examining otoscopic images for ear diseases is necessary when the clinical diagnosis of ear diseases extracted from the knowledge of otolaryngologists is limited. Improved diagnosis approaches based on otoscopic imageprocessing are urgently needed. Recently, convolutional neural networks (CNNs) have been carried out for medical diagnosis to obtain higher accuracy than standard machine learning algorithms and specialists' expertise. Therefore, the proposed approach involves using the Bayesian hyperparameter optimization with the CNN architecture for automatic diagnosis of ear imagery database including four classes: normal, myringosclerosis, earwax plug, and chronic otitis media (COM). The suggested approach was trained using 616 otoscopic images, and the performance of this approach was assessed using 264 testing images. In this paper, the performance of ear disease classification was compared in terms of accuracy, sensitivity, specificity, and positive predictive value (PPV). The results produced a classification accuracy of 98.10%, a sensitivity of 98.11%, a specificity of 99.36%, and a PPV of 98.10%. Finally, the suggested approach demonstrates how to locate optimal CNN hyperparameters for accurate diagnosis of ear diseases while taking time into account. As a result, the usefulness and dependability of the suggested approach will lead to the establishment of an automated tool for better categorization and prediction of different ear diseases.
The Big Video Data generated in today's smart cities has raised concerns from its purposeful usage perspective, where surveillance cameras, among many others are the most prominent resources to contribute to the h...
详细信息
The Big Video Data generated in today's smart cities has raised concerns from its purposeful usage perspective, where surveillance cameras, among many others are the most prominent resources to contribute to the huge volumes of data, making its automated analysis a difficult task in terms of computation and preciseness. Violence detection (VD), broadly plunging under action and activity recognition domain, is used to analyze Big Video data for anomalous actions incurred due to humans. The VD literature is traditionally based on manually engineered features, though advancements to deeplearning based standalone models are developed for real-time VD analysis. This paper focuses on overview of deep sequence learning approaches along with localization strategies of the detected violence. This overview also dives into the initial imageprocessing and machine learning-based VD literature and their possible advantages such as efficiency against the current complex models. Furthermore,the datasets are discussed, to provide an analysis of the current models, explaining their pros and cons with future directions in VD domain derived from an in-depth analysis of the previous methods.
暂无评论