Detecting micro-defects in densely populated printed circuit boards (PCBs) with complex backgrounds is a critical challenge. To address the problem, the DHNet, a small object detection network based on YOLOv8 employin...
详细信息
Detecting micro-defects in densely populated printed circuit boards (PCBs) with complex backgrounds is a critical challenge. To address the problem, the DHNet, a small object detection network based on YOLOv8 employing multi-scale convolutional kernels is proposed for feature extraction and fusion. The lightweight VOVGSHet module is designed for feature fusion and a pyramid structure to efficiently leverage feature map relationships while minimizing model complexity and parameters. Otherwise, to optimize the original extraction structure and enhance multi-scale defect detection, convolutional kernels of varying sizes process the same input channels. Additionally, the incorporation of the Wise-IoU loss function improves small defect detection accuracy and efficiency. Moreover, extensive experiments on a custom PCB dataset demonstrate DHNet's effectiveness, achieving an outstanding mean Average Precision (mAP) of 84.5%, surpassing the original YOLOv8 network by 4.0%, with parameters only of 2.85 M. Model demonstrates a latency of 3.6 ms on NVIDIA 4090. However, YOLOv8n has a latency of 4.4 ms. Validation on public deepPCB and NEU datasets further confirms DHNet's superiority, which can reach 99.1% and 79.9% mAP, respectively. Finally, successful deployment on the NVIDIA Jetson Nano platform validates DHNet's suitability for real-time defect detection in industrial applications.
To monitor tool wear during cutting processing, tool wear is mainly measured indirectly through sensor signals that are most correlated with wear. The direct measurement method of tool wear using image and optical sen...
详细信息
To monitor tool wear during cutting processing, tool wear is mainly measured indirectly through sensor signals that are most correlated with wear. The direct measurement method of tool wear using image and optical sensors is more accurate than indirect measurement, but it is mainly used to measure the amount of wear because it is difficult to apply in realtime. Existing studies have been conducted mainly on flank wear caused by friction with workpiece. On the other hand, crater wear is an important monitoring factor because it is caused by friction with chips generated during processing and causes sudden tool breakage. However, for crater wear, it is difficult to measure the amount of wear because the indicator of the amount of wear is depth. Therefore, although imageprocessing-based studies have been conducted to measure the amount of crater wear, there is a clear limit to accurately measure the depth only with the image on the top of the tool. In this work, we propose a method to extract unique features of crater wear images through autoencoder, a deeplearning technique, and use them as a new measure of wear.
Breast cancer is one of the most common life-threatening diseases that affects women globally. Saudi Arabia is also one of the countries that suffer from a serious number of this disease among women. In terms of diagn...
详细信息
ISBN:
(纸本)9781665422383
Breast cancer is one of the most common life-threatening diseases that affects women globally. Saudi Arabia is also one of the countries that suffer from a serious number of this disease among women. In terms of diagnosis modalities, a mammogram is the first line for detecting breast cancer. In addition, breast cancer can be screened by real-time ultrasound images, which are of relatively less quality and have more impact (noninvasive) images. Therefore, the purpose of this study is to develop image enhancement techniques using deeplearning and imageprocessing techniques. The main goal is to improve the ultrasound images in order to help radiologists screen the disease more accurately. For this study, ninety female patients of ages between 15 - 77 years are considered. These patients were already diagnosed using ultrasound with breast lesions. The images are visually graded and evaluated by two trained radiologists, both pre- and post- enhancement. In particular, two parameters were considered;1) BI-RAD categories, 2) Breast cancer classification. The agreement between radiologists and post-enhancement was assessed using simple kappa and weighted kappa statistics. Moreover, sensitivity and specificity are also calculated.
The position of blind lanes must be correctly determined in order for blind people to travel safely. Aiming at the low accuracy and slow speed of traditional blind lanes image segmentation algorithms, a semantic segme...
详细信息
The position of blind lanes must be correctly determined in order for blind people to travel safely. Aiming at the low accuracy and slow speed of traditional blind lanes image segmentation algorithms, a semantic segmentation method based on SegNet and MobileNetV3 is proposed. The main idea is to replace the coding part of the original SegNet model with the feature extraction part of MobileNetV3 and remove the pooling layer. Blind lanes images were collected through online search and self-shooting, and then the data were manually marked by LabelMe software and trained on TensorFlow deeplearning framework. The experimental results show that the improved model has high segmentation accuracy and recognition speed. The pixel accuracy of blind lanes segmentation is 98.21%, the mean intersection over union is 96.29%, and the average time for processing a 416 x 416 image is 0.057 s, which meets the real-time requirements of the blind guidance system.
The accelerated development of the Industrial Internet of Things (IIoT) is catalyzing the digitalization of industrial production to achieve Industry 4.0. In this article, we propose a novel digital twin (DT) empowere...
详细信息
The accelerated development of the Industrial Internet of Things (IIoT) is catalyzing the digitalization of industrial production to achieve Industry 4.0. In this article, we propose a novel digital twin (DT) empowered IIoT (DTEI) architecture, in which DTs capture the properties of industrial devices for real-timeprocessing and intelligent decision making. To alleviate data transmission burden and privacy leakage, we aim to optimize federated learning (FL) to construct the DTEI model. Specifically, to cope with the heterogeneity of IIoT devices, we develop the DTEI-assisted deep reinforcement learning method for the selection process of IIoT devices in FL, especially for selecting IIoT devices with high utility values. Furthermore, we propose an asynchronous FL scheme to address the discrete effects caused by heterogeneous IIoT devices. Experimental results show that our proposed scheme features faster convergence and higher training accuracy compared to the benchmark.
A railway pantograph supplies a vehicle with the electric power from the OCL (Overhead Contact Line). The contact strip, which directly contacts the OCL, is vulnerable to wear and should be periodically replaced to pr...
详细信息
A railway pantograph supplies a vehicle with the electric power from the OCL (Overhead Contact Line). The contact strip, which directly contacts the OCL, is vulnerable to wear and should be periodically replaced to prevent accidents from power outages. To that end, the pantograph is regularly checked visually during routine maintenance, along with the panhead which includes the horn and the contact strip. However, real-time monitoring is still difficult, and a reliable method needs to be developed. The existing methods use imageprocessing to detect the actual condition of the contact strip. In this paper, we suggest a method for detecting the wear size of the contact strip using imageprocessing and deeplearning. In addition, we demonstrate how to assess the condition of the panhead and the horn in the contact. The monitoring equipment is built on the test bed and is automatically configured to acquire images of the moving pantograph. The tilt of the panhead is estimated from image recognition and the condition of the horn is assessed with deeplearning. In the future, this method can be used not only for efficient pantograph maintenance but also for determining the performance according to the contact condition.
Understanding and monitoring water levels are essential for various applications, including environmental protection, public safety, and resource management. Water level estimation, a critical aspect of hydrological m...
详细信息
Understanding and monitoring water levels are essential for various applications, including environmental protection, public safety, and resource management. Water level estimation, a critical aspect of hydrological monitoring, is often constrained by challenges such as resource scarcity, high costs, and time-intensive processes. This research addresses these limitations by developing a machine learning-based system for automatic and real-time water level control. Specifically, it investigates the effectiveness of a non-contact, image-based water level measurement approach, leveraging recent advancements in mobile imaging technology. images were captured using a standard smartphone equipped with an RGB camera for water level analysis. Through precise image alignment processing under both clear and turbid conditions, the water's edge on a gauge was accurately detected. The study centers on the development and comparison of three computational models: Artificial Neural Networks (ANN), deeplearning (DL), and Convolutional Neural Networks (CNN). These models were trained to estimate water levels from processed image data. Results demonstrated varying levels of accuracy across models, with the CNN model outperforming others, achieving the lowest error rate of 24.36 mm and the highest correlation of 0.986. In contrast, the ANN model yielded the highest error rate at 30.76 mm and the lowest correlation of 0.968, highlighting the relative effectiveness of CNN in this application. Given the high accuracy (92.6%) of the imageprocessing method and CNN model in detecting water surface edges and determining water levels, this system has substantial potential to enhance water resource management and control efficiency.
deeplearning (DL) has revolutionized the field of artificial intelligence by providing sophisticated models across a diverse range of applications, from image and speech recognition to natural language processing and...
详细信息
deeplearning (DL) has revolutionized the field of artificial intelligence by providing sophisticated models across a diverse range of applications, from image and speech recognition to natural language processing and autonomous driving. However, deeplearning models are typically black-box models where the reason for predictions is unknown. Consequently, the reliability of the model becomes questionable in many circumstances. Explainable AI (XAI) plays an important role in improving the transparency and interpretability of the model thereby making it more reliable for real-time deployment. To investigate the reliability and truthfulness of DL models, this research develops image classification models using transfer learning mechanism and validates the results using XAI technique. Thus, the contribution of this research is twofold, we employ three pre-trained models VGG16, MobileNetV2 and ResNet50 using multiple transfer learning techniques for a fruit classification task consisting of 131 classes. Next, we inspect the reliability of models, based on these pre-trained networks, by utilizing Local Interpretable Model-Agnostic Explanations, the LIME, a popular XAI technique that generates explanations for the predictions. Experimental results reveal that transfer learning provides optimized results of around 98% accuracy. The classification of the models is validated on different instances using LIME and it was observed that each model predictions are interpretable and understandable as they are based on pertinent image features that are relevant to particular classes. We believe that this research gives an insight for determining how an interpretation can be drawn from a complex AI model such that its accountability and trustworthiness can be increased.
PurposeVascular distribution is important information for diagnosing diseases and supporting surgery. Photoacoustic imaging is a technology that can image blood vessels noninvasively and with high resolution. In photo...
详细信息
PurposeVascular distribution is important information for diagnosing diseases and supporting surgery. Photoacoustic imaging is a technology that can image blood vessels noninvasively and with high resolution. In photoacoustic imaging, a hemispherical array sensor is especially suitable for measuring blood vessels running in various directions. However, as a hemispherical array sensor, a sparse array sensor is often used due to technical and cost issues, which causes artifacts in photoacoustic images. Therefore, in this study, we reduce these artifacts using deeplearning technology to generate signals of virtual dense array *** 2D virtual array sensor signals using a 3D convolutional neural network (CNN) requires huge computational costs and is impractical. Therefore, we installed virtual sensors between the real sensors along the spiral pattern in three different directions and used a 2D CNN to generate signals of the virtual sensors in each direction. Then we reconstructed a photoacoustic image using the signals from both the real sensors and the virtual *** evaluated the proposed method using simulation data and human palm measurement data. We found that these artifacts were significantly reduced in the images reconstructed using the proposed method, while the artifacts were strong in the images obtained only from the real sensor *** the proposed method, we were able to significantly reduce artifacts, and as a result, it became possible to recognize deep blood vessels. In addition, the processingtime of the proposed method was sufficiently applicable to clinical measurement.
image segmentation plays a crucial role in the roadwork operations of autonomous line-painting machines. However, the limited resources of mobile platforms in intelligent line-painting applications pose a dual challen...
详细信息
image segmentation plays a crucial role in the roadwork operations of autonomous line-painting machines. However, the limited resources of mobile platforms in intelligent line-painting applications pose a dual challenge of ensuring both accuracy and real-time performance in road segmentation. To address this issue, this study introduces a lightweight yet efficient image segmentation model, termed the SLTM Network. Central to this network is the lightweight SLTM module, which significantly reduces the model's parameter count and lowers the computational overhead of the decoder. To enhance the interplay of information at different spatial resolutions, the network incorporates an SE attention-enhanced upsampling module (SAUM) and employs a Spatial Attention Sequence (SAS) unit to improve global environment perception at a low computational cost. Comprehensive experimental evaluations on the Cityscapes dataset demonstrate that the SLTM Network excels in balancing speed and accuracy, achieving an mIoU of 70.5% with only 4.07M parameters and an impressive inference speed of 267.1 FPS. On the embedded device Jetson Xavier NX, it achieves an inference speed of 34.2 FPS. Compared to existing lightweight image segmentation models, the SLTM Network exhibits significant advantages in both processing speed and accuracy, making it particularly suitable for real-time autonomous line-painting machine applications.
暂无评论