Tensile testing (aka tension testing) is a widely employed mechanical testing technique for analyzing materials' properties and behavior under applied stress. Tensile testing plays a pivotal role in helping engine...
详细信息
Tensile testing (aka tension testing) is a widely employed mechanical testing technique for analyzing materials' properties and behavior under applied stress. Tensile testing plays a pivotal role in helping engineers to make informed decision about material selection and usage. Despite its importance, there is a limited numbers of studies that explored the potential of AI techniques for realtime monitoring and material behavior prediction in tensile testing. To this end, this work presents a deeplearning model designed to predict the material's condition throughout tensile testing and provide an early warning prior to fracture. By leveraging a comprehensive dataset of tension test video samples, the proposed model utilizes both convolution and recurrent neural networks to extract pertinent spatial and temporal visual features, thereby predicting the frames at which material deformation and fracture occur. The evaluation results of our research showed that the proposed model achieved a predictive ability with an F1-score of 97%, on average. The implications of our research are significant for industries and researchers in the field of materials science and engineering. By accurately predicting material status, our model enables automounts, realtime analysis of material behavior during tensile testing, leading to better time and cost efficiency in various applications.
PurposeClinical needle insertion into tissue, commonly assisted by 2D ultrasound imaging for real-time navigation, faces the challenge of precise needle and probe alignment to reduce out-of-plane movement. Recent stud...
详细信息
PurposeClinical needle insertion into tissue, commonly assisted by 2D ultrasound imaging for real-time navigation, faces the challenge of precise needle and probe alignment to reduce out-of-plane movement. Recent studies investigate 3D ultrasound imaging together with deeplearning to overcome this problem, focusing on acquiring high-resolution images to create optimal conditions for needle tip detection. However, high-resolution also requires a lot of time for image acquisition and processing, which limits the real-time capability. Therefore, we aim to maximize the US volume rate with the trade-off of low image resolution. We propose a deeplearning approach to directly extract the 3D needle tip position from sparsely sampled US *** design an experimental setup with a robot inserting a needle into water and chicken liver tissue. In contrast to manual annotation, we assess the needle tip position from the known robot pose. During insertion, we acquire a large data set of low-resolution volumes using a 16 x\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document} 16 element matrix transducer with a volume rate of 4 Hz. We compare the performance of our deeplearning approach with conventional needle *** experiments in water and liver show that deeplearning outperforms the conventional approach while achieving sub-millimeter accuracy. We achieve mean position errors of 0.54 mm in water and 1.54 mm in liver for deep *** study underlines the strengths of deeplearning to predict the 3D needle positions from low-resolution ultrasound volumes. This is an important milestone for real-time needle navigation, simplifying the alignment of needle and ultrasound probe and enabling a 3D motion analysis.
This article aims to propose a reliable real-time monitoring system for distribution network defects, improve intelligent monitoring technology by combining deeplearning technology, and analyse the drawbacks of tradi...
详细信息
This review paper presents an in-depth analysis of deeplearning (DL) models applied to traffic scene understanding, a key aspect of modern intelligent transportation systems. It examines fundamental techniques such a...
详细信息
This review paper presents an in-depth analysis of deeplearning (DL) models applied to traffic scene understanding, a key aspect of modern intelligent transportation systems. It examines fundamental techniques such as classification, object detection, and segmentation, and extends to more advanced applications like action recognition, object tracking, path prediction, scene generation and retrieval, anomaly detection, image-to-image Translation (I2IT), and person re-identification (Person Re-ID). The paper synthesizes insights from a broad range of studies, tracing the evolution from traditional imageprocessing methods to sophisticated DL techniques, such as Convolutional Neural Networks (CNNs) and Generative Adversarial Networks (GANs). The review also explores three primary categories of domain adaptation (DA) methods: clustering-based, discrepancy-based, and adversarial-based, highlighting their significance in traffic scene understanding. The significance of Hyperparameter Optimization (HPO) is also discussed, emphasizing its critical role in enhancing model performance and efficiency, particularly in adapting DL models for practical, real-world use. Special focus is given to the integration of these models in real-world applications, including autonomous driving, traffic management, and pedestrian safety. The review also addresses key challenges in traffic scene understanding, such as occlusions, the dynamic nature of urban traffic, and environmental complexities like varying weather and lighting conditions. By critically analyzing current technologies, the paper identifies limitations in existing research and proposes areas for future exploration. It underscores the need for improved interpretability, real-timeprocessing, and the integration of multi-modal data. This review serves as a valuable resource for researchers and practitioners aiming to apply or advance DL techniques in traffic scene understanding.
Visual Odometry (VO) is a crucial process for estimating camera motion in real-time based on visual information captured. The emergence of deeplearning has significantly transformed VO and Explainable Artificial Inte...
详细信息
Visual Odometry (VO) is a crucial process for estimating camera motion in real-time based on visual information captured. The emergence of deeplearning has significantly transformed VO and Explainable Artificial Intelligence (XAI) in deep Vision-Based Odometry. This survey paper explores the latest advancements in VO facilitated by deeplearning methods, focusing on explainability and interpretability. It provides an overview of state-of-the-art deeplearning techniques and dissects each model into its elemental building blocks to understand their explainable and interpretable aspects. The survey also highlights research gaps in optical flow robustness, occlusion and dynamic objects, real-timeprocessing, drift correction, semantic awareness, and sensor integration. The aim is to catalyze future innovations in deeplearningbased VO and stimulate dialogue about potential directions for the next wave of research, emphasizing explainability and interpretability as integral components of advanced systems.
The deeplearning object detection algorithm has been widely applied in the field of synthetic aperture radar (SAR). By utilizing deep convolutional neural networks (CNNs) and other techniques, these algorithms can ef...
详细信息
The deeplearning object detection algorithm has been widely applied in the field of synthetic aperture radar (SAR). By utilizing deep convolutional neural networks (CNNs) and other techniques, these algorithms can effectively identify and locate targets in SAR images, thereby improving the accuracy and efficiency of detection. In recent years, achieving real-time monitoring of regions has become a pressing need, leading to the direct completion of real-time SAR image target detection on airborne or satellite-borne real-timeprocessing platforms. However, current GPU-based real-timeprocessing platforms struggle to meet the power consumption requirements of airborne or satellite applications. To address this issue, a low-power, low-latency deeplearning SAR object detection algorithm accelerator was designed in this study to enable real-time target detection on airborne and satellite SAR platforms. This accelerator proposes a Process Engine (PE) suitable for multidimensional convolution parallel computing, making full use of Field-Programmable Gate Array (FPGA) computing resources to reduce convolution computing time. Furthermore, a unique memory arrangement design based on this PE aims to enhance memory read/write efficiency while applying dataflow patterns suitable for FPGA computing to the accelerator to reduce computation latency. Our experimental results demonstrate that deploying the SAR object detection algorithm based on Yolov5s on this accelerator design, mounted on a Virtex 7 690t chip, consumes only 7 watts of dynamic power, achieving the capability to detect 52.19 512 x 512-sized SAR images per second.
Facial recognition on resource-limited devices such as the Raspberry Pi poses a challenge due to inherent processing limitations. For real-time applications, finding efficient and reliable solutions is critical. This ...
详细信息
Facial recognition on resource-limited devices such as the Raspberry Pi poses a challenge due to inherent processing limitations. For real-time applications, finding efficient and reliable solutions is critical. This study investigated the feasibility of using transfer learning for facial recognition tasks on the Raspberry Pi and evaluated transfer learning that leverages knowledge from previously trained models. We compared two well-known deeplearning (DL) architectures, InceptionV3 and MobileNetV2, adapted to face recognition datasets. MobileNetV2 outperformed InceptionV3, achieving a training accuracy of 98.20% and an F1 score of 98%, compared to InceptionV3's training accuracy of 86.80% and an F1 score of 91%. As a result, MobileNetV2 emerges as a more powerful architecture for facial recognition tasks on the Raspberry Pi when integrated with transfer learning. These results point to a promising direction for deploying efficient DL applications on edge devices, reducing latency, and enabling real-timeprocessing.
image denoising is a crucial algorithm in imageprocessing that aims to enhance image quality. deeplearning-based image denoising methods can be categorized into supervised and unsupervised approaches. Supervised lea...
详细信息
image denoising is a crucial algorithm in imageprocessing that aims to enhance image quality. deeplearning-based image denoising methods can be categorized into supervised and unsupervised approaches. Supervised learning requires pairs of noisy and noise-free training data, which is impractical in real-world scenarios. Unsupervised learning uses pairs of noisy images for training, but it may yield lower accuracy. Additionally, deeplearning-based methods often require a large amount of training data. To overcome these challenges, this research proposes a self-validation Noise2Noise (SV-N2N) framework that generates validation sets using only noisy images without requiring noise-free pairs. The proposed SV-N2N method effectively reduces noise, comparable to supervised and unsupervised methods, without requiring a noise-free ground truth, which is efficient for solving real-world scenarios.
Underground carbon dioxide (CO2) sequestration is widely accepted as a proven and established technology to respond to global warming from greenhouse gas emissions. It is crucial to monitor the CO2 plume effectively t...
详细信息
Underground carbon dioxide (CO2) sequestration is widely accepted as a proven and established technology to respond to global warming from greenhouse gas emissions. It is crucial to monitor the CO2 plume effectively throughout the life cycle of a geologic CO2 sequestration project to ensure safety and storage efficacy. We propose a fast and efficient deeplearning workflow for near real-time data assimilation, forecasting and visualization of CO2 plume evolution in saline aquifer and demonstrate its application at a field site. Our proposed workflow integrates field measurements, including distributed pressure and temperature at wells to visualize spatial and temporal migration of the CO2 plume in the subsurface. The deeplearning framework has two key concepts that enhance the efficiency of the training and robustness of the CO2 plume prediction: first, instead of using multiple saturation images as output of the deep-learning model, a single time of Flight (TOF) image are used to represent CO2 plume propagation;second, we use variational autoencoderdecoder (VAE) for high dimensional image data compression considering uncertainties in the predicted images. time-of-Flight (TOF) is the travel time of a neural particle from the injection point, which is calculated along streamline based on the velocity field by considering reservoir heterogeneity and driving forces. The latent variables of VAE are estimated through a deep neural network model with inputs of distributed pressure and temperature measurements. Subsequently, the decoder part of the VAE expands the estimated latent variables to the original dimension of the TOF images. We demonstrate the efficacy of the proposed method by comparison with the more traditional pareto-based multi-objective Genetic Algorithm (MOGA) for the assimilation of the field measurements and forecasting of the CO2 plume migration. The power and utility of our proposed workflow are demonstrated by application to the Illinois Basin-Decatu
This study examines the use of artificial intelligence-based image segmentation and imageprocessing techniques in disaster management. The aim of the study is to integrate artificial intelligence and imageprocessing...
详细信息
ISBN:
(数字)9798331533137
ISBN:
(纸本)9798331533144
This study examines the use of artificial intelligence-based image segmentation and imageprocessing techniques in disaster management. The aim of the study is to integrate artificial intelligence and imageprocessing techniques for fast and effective response in disaster areas to facilitate disaster management. Our hypothesis is that fast and accurate analysis can be performed with high-performance and fast AI-based segmentation models and imageprocessing techniques on images taken from UAVs and satellites in disaster areas. In this process, important regions in the images are identified by applying segmentation processes and masks are created. Then, using these masks, numerical results can be obtained with imageprocessing techniques with low computational cost. Thus, it is aimed to make fast and accurate decisions in disaster management. In this study, popular segmentation models were compared and analyzed using 2343 images obtained from Floodnet dataset. The results show that the SegFormer model provides detailed damage analysis and contour area or connected component analysis, which can provide both detailed and numerically accurate results for disaster management. This study reveals that the use of imageprocessing and artificial intelligence in disaster management can improve response processes by increasing operational efficiency.
暂无评论