Intelligent Visitor Management System (IVMS) is crucial for enhancing security and operational efficiency in smart factories and intelligent office buildings. Leveraging AIoT-driven image analysis will facilitate real...
详细信息
Intelligent Visitor Management System (IVMS) is crucial for enhancing security and operational efficiency in smart factories and intelligent office buildings. Leveraging AIoT-driven image analysis will facilitate realtime visitor authentication and access control. However, the growing volume of interactions and the limited processing power of local terminals complicate the delivery of timely and accurate image analysis. To address these challenges, we propose an edge-terminal collaborative AIoT framework for real-time visitor management. The framework solves the limitations of traditional approaches, where local terminals are unable to handle the computational load and edge solutions experience high latency due to transmission delays. Specifically, it integrates three key components to improve system performance: a local analysis module for initial processing, an image communication module for efficient data transmission, and an edge analysis module for advanced processing. Moreover, the framework jointly optimizes image task offloading, wireless channel allocation, and image compression, all formulated as an optimization problem to ensure fast and accurate analysis. Additionally, a novel multi-level deep Reinforcement learning (DRL) method is further designed to dynamically refine the selection of compression and offloading strategies. By learning in real-time, the DRL model adapts to network variations, addressing the scalability and adaptability limitations of existing methods. Simulation results show that our proposed edge-terminal collaborative AIoT framework significantly outperforms both edge-only and terminal-only methods in terms of latency and accuracy.
The fineness modulus(FM) represents the level of particle size of manufactured sand. real-time feedback of FM of manufactured sand is important for industrial sand production, but extracting the particle profile from ...
详细信息
The fineness modulus(FM) represents the level of particle size of manufactured sand. real-time feedback of FM of manufactured sand is important for industrial sand production, but extracting the particle profile from densely stacked images is a great challenge. In this study, a deeplearning and regression analysis -based online measurement method for FM of manufactured sand is proposed. Firstly, the real fineness modulus of the sand produced by the sand -making machine in realtime was obtained by the vibration -screening method(VSM). Then, the particle size fraction of larger particles (0.6-4.75 mm) was obtained based on machine vision combined with a convolutional neural network and imageprocessing. Secondly, a multiple linear regression model was developed for the percentage of particle size and FM. Finally, the percentage of particle size was substituted into the regression model as the independent variable to achieve a fast prediction of the unknown FM. The experimental results show that the maximum repeatability errors for FM of different manufactured sands are 0.09 and 0.13 respectively, and the maximum absolute errors of the FM prediction results are 0.18 and 0.17 respectively. The calculation efficiency and error level of this research method can meet the online testing at sand making sites.
Chicken meat plays an important role in the healthy diets of many people and has a large global trade volume. In the chicken meat sector, in some production processes, traditional methods are used. Traditional chicken...
详细信息
Chicken meat plays an important role in the healthy diets of many people and has a large global trade volume. In the chicken meat sector, in some production processes, traditional methods are used. Traditional chicken part sorting methods are often manual and time-consuming, especially during the packaging process. This study aimed to identify and classify the chicken parts for their input during the packaging process with the highest possible accuracy and speed. For this purpose, deep-learning-based object detection models were used. An image dataset was developed for the classification models by collecting the image data of different chicken parts, such as legs, breasts, shanks, wings, and drumsticks. The models were trained by the You Only Look Once version 8 (YOLOv8) algorithm variants and the real-time Detection Transformer (RT-DETR) algorithm variants. Then, they were evaluated and compared based on precision, recall, F1-Score, mean average precision (mAP), and Mean Inference time per frame (MITF) metrics. Based on the obtained results, the YOLOv8s model outperformed the other models developed with other YOLOv8 versions and the RT-DETR algorithm versions by obtaining values of 0.9969, 0.9950, and 0.9807 for the F1-score, mAP@0.5, and mAP@0.5:0.95, respectively. It has been proven suitable for real-time applications with an MITF value of 10.3 ms/image.
In industrial applications, surface defect segmentation is a critical task. However, facing challenges such as diverse defect scales, low contrast between defects and background, high interclass similarity and real-ti...
详细信息
In industrial applications, surface defect segmentation is a critical task. However, facing challenges such as diverse defect scales, low contrast between defects and background, high interclass similarity and real-time detection in defect inspection, we propose an efficient lightweight network, named DMC-Net, for real-time surface defect segmentation. The structural optimization of DMC-Net includes the following components: (1) depthwise separable convolution attention module, a lightweight and efficient feature extraction module for extracting multi-scale defect features. (2) Multi-scale feature enhancement module, providing long-range information capture and local information focusing to enhance defect localization capability. (3) Channel shuffle group convolution, enhancing feature interaction and information propagation while reducing the parameter quantity. Based on the experimental results, DMC-Net achieved an mIoU of 73.74% on the NEU-SEG dataset, while achieving an FPS of 211.7. This indicates that we have successfully reduced the complexity and computational cost of the model while improving performance, providing a feasible solution for industrial applications. The relevant code can be obtained at https://***/Michaelzyb/***.
The rapid development of artificial intelligence (AI) and breakthroughs in Internet of Things (IoT) technologies have driven the innovation of advanced autonomous driving systems (ADSs). image classification deep lear...
详细信息
The rapid development of artificial intelligence (AI) and breakthroughs in Internet of Things (IoT) technologies have driven the innovation of advanced autonomous driving systems (ADSs). image classification deeplearning (DL) algorithms immensely contribute to the decision-making process in ADSs, showcasing their capabilities in handling complex real-world driving scenarios, surpassing human driving intelligence. However, these algorithms are vulnerable to adversarial attacks, which aim to fool them in real-time decision- making and compromise the reliability of the autonomous driving functions. This systematic review offers a comprehensive overview of the most recent literature on adversarial attacks and countermeasures on image classification DL models in ADSs. The review highlights the current challenges in applying successful countermeasures to mitigating these vulnerabilities. We also introduce taxonomies for categorizing adversarial attacks and countermeasures and provide recommendations and guidelines to help researchers design and evaluate countermeasures. We suggest interesting future research directions to improve the robustness of image classification DL models against adversarial attacks in autonomous driving scenarios.
In this paper, we proposed a template matching technique using deeplearning to match pairs of wide fields of view and narrow field of view infrared images. The deeplearning network has a similar structure with the A...
详细信息
ISBN:
(纸本)9781510673878;9781510673861
In this paper, we proposed a template matching technique using deeplearning to match pairs of wide fields of view and narrow field of view infrared images. The deeplearning network has a similar structure with the Atrous Spatial Pyramid Pooling (ASPP) module and both wide and narrow fields of view images are input to the same network, so the network weights are shared. Our experiments used the Galaxy S20 (Qualcomm Snapdragon 865) platform and show that the trained network has higher matching accuracy than other template matching techniques and is fast enough to be used in realtime.
Motion compensation in radiation therapy is a challenging scenario that requires estimating and forecasting motion of tissue structures to deliver the target dose. Ultrasound offers direct imaging of tissue in real-ti...
详细信息
Motion compensation in radiation therapy is a challenging scenario that requires estimating and forecasting motion of tissue structures to deliver the target dose. Ultrasound offers direct imaging of tissue in real-time and is considered for image guidance in radiation therapy. Recently, fast volumetric ultrasound has gained traction, but motion analysis with such high-dimensional data remains difficult. While deeplearning could bring many advantages, such as fast data processing and high performance, it remains unclear how to process sequences of hundreds of image volumes efficiently and effectively. We present a 4D deeplearning approach for real-time motion estimation and forecasting using long-term 4D ultrasound data. Using motion traces acquired during radiation therapy combined with various tissue types, our results demonstrate that long-term motion estimation can be performed markerless with a tracking error of 0.35 +/- 0.2 mm and with an inference time of less than 5 ms. Also, we demonstrate forecasting directly from the image data up to 900 ms into the future. Overall, our findings highlight that 4D deeplearning is a promising approach for motion analysis during radiotherapy.
During the converter process, it is crucial to automatically identify and record ladle numbers to track steel product quality and enhance automation levels. However, the steelmaking environment presents several challe...
详细信息
During the converter process, it is crucial to automatically identify and record ladle numbers to track steel product quality and enhance automation levels. However, the steelmaking environment presents several challenges, including intricate ladle scheduling, varying lighting conditions, severe background interference, and significant disparities between manually spray-printed ladle number characteristics and publicly available datasets. The combination of these problems makes it challenging to perform accurate and real-time ladle number identification. In response, this article suggests an automatic ladle number recognition approach based on deeplearning and imageprocessing. First, a double-region object detection model based on YOLOv5 is employed to capture keyframe images of the ladle to be identified from the video stream. Then, a method that can enable the acquisition of an accurate region of ladle numbers in sophisticated industrial settings is proposed to address the distortion of numerical features caused by lighting variations and background interference in industrial environments. Last, leveraging the proprietary dataset found and a ladle number recognition model integrating CNN and multiframe image fusion is designed, developing multithreading design and image queue management to ensure real-time and accurate ladle number recognition. In this study, the video data of a steel plant is used for testing. Through testing 176 steelmaking production cycles, all ladle numbers are accurately identified prior to finishing charging molten iron, indicating the high accuracy and real-time capability of the recognition system.
Road safety can be creatively increased by utilizing systems for reporting and detecting accidents use the YOLO algorithm. Yolo, which stands for "You Only Look Once," is a sophisticated object recognition s...
详细信息
ISBN:
(数字)9798350391565
ISBN:
(纸本)9798350391572
Road safety can be creatively increased by utilizing systems for reporting and detecting accidents use the YOLO algorithm. Yolo, which stands for "You Only Look Once," is a sophisticated object recognition system that can identify and pinpoint objects in live video streams. By identifying accidents and notifying emergency services, the system lowers the reaction times and increases the possibility of saving lives by utilizing the YOLO algorithm. The object detection and alarm systems are the two primary parts of the proposed system. The YOLO method is employed by the object recognition module to look for accidents in live video broadcasts. A module of accident photos was used to teach the system to accurately identify incidents. When an accident is discovered, an alert system is activated. The location of the accident and a brief account of what transpired are communicated to the emergency services by the alert system. This information is communicated to emergency services through a wireless communication network, which expedites response times and increases the likelihood of saving lives. Positive results were obtained by testing the system using the images and an accident module. It was shown that the warning system could react in a couple of seconds and that the YOLO algorithm could identify accidents having a precision of around 94. Highways, busy intersections, and other high-risk areas may have systems in place to increase traffic safety and lower the number of accidents. Developing crash detection and warning systems using deeplearning using the YOLO technique is one way to potentially increase road safety. With the use of technology, problems may be precisely identified in real-time video feeds, alerting emergency services and potentially increasing survival rates and reaction times.
With the increasing complexity of modern football tactics, how to intelligently and accurately analyze tactical changes in real-time during matches has become an important research direction. Traditional manual tactic...
详细信息
With the increasing complexity of modern football tactics, how to intelligently and accurately analyze tactical changes in real-time during matches has become an important research direction. Traditional manual tactical analysis methods are inefficient and susceptible to subjective bias. Therefore, using computer vision and deeplearning technologies for tactical image recognition and analysis in football matches has gradually become a research hotspot. Convolutional Neural Networks (CNNs), as a powerful imageprocessing tool, have been widely applied in video analysis and player detection. However, multi-target motion prediction and tracking management in dynamic football match scenes still face significant challenges. Existing research mainly focuses on static image analysis or simple player tracking, but the high-frequency image updates, player interactions, and occlusion issues in football matches complicate multi-target tracking. While some deeplearning-based methods for multi-target detection and tracking have made progress, challenges remain, such as handling high-density player targets and improving motion trajectory prediction accuracy. To address these shortcomings, this study proposes two core techniques based on CNNs: first, multi-target motion prediction, which accurately forecasts players' future positions based on historical motion data;second, multi-target tracking management, which uses deeplearning to track and manage each player's movement trajectory in real-time. Through these two techniques, this research aims to improve the realtime and accuracy of tactical analysis in football matches, providing coaches and analysts with more scientific and efficient tactical decision-making support.
暂无评论