Polarization is capable of probing microstructures and has unique sensitivity to fibrous anisotropic structure. Polarimetric imaging has demonstrated promising potential in diverse applications ranging from biomedicin...
详细信息
ISBN:
(纸本)9781510647985;9781510647978
Polarization is capable of probing microstructures and has unique sensitivity to fibrous anisotropic structure. Polarimetric imaging has demonstrated promising potential in diverse applications ranging from biomedicine, material science, and atmospheric remote sensing. The polarization properties of samples can be comprehensively described by a Mueller matrix (MM). However, the relationship between individual MM elements and properties of the sample is often not clear. There have been consistent efforts to derive polarization parameters from MM based on certain assumptions for better description of the samples, e.g., MM polar decomposition (MMPD), MM transformation (MMT) and MM differential decomposition. Usually, the MM imaging requires sequential measurements with different polarization states of incident light and the imaging process is time consuming. In addition, for movable samples, we cannot guarantee the consistency during the imaging. This may cause precision issues since the images cannot be well-registered. In this work, we built a statistical translation model to generate polarization parameters from a single Stokes vector which can be obtained by one-shot imaging. This will improve the imaging efficiency, simplify the optical system and avoid introducing errors by the image registration. In the model design, we adopted the generative adversarial network (GAN) where the generator is based on a U-net architecture. We demonstrated the effectiveness of our approach on liver tissue, blood smear and porous anodic alumina (PAA) film, and quantitatively evaluated the results by similarity assessment methods. The model can generate a parameter image within 0.1 second on a desktop computer, which shows the potential to achieve real-time performance.
Scene classification is a hot issue in the field of SAR image interpretation. Many SAR image interpretation tasks can be promoted with the development of highly credible scene classification methods. But the fussy ste...
详细信息
In the current deeplearning research landscape, imageprocessing technology faces significant challenges in lightweight defect detection despite its broad application in object detection. This study introduces the Li...
详细信息
ISBN:
(数字)9798350385991
ISBN:
(纸本)9798350386004
In the current deeplearning research landscape, imageprocessing technology faces significant challenges in lightweight defect detection despite its broad application in object detection. This study introduces the Lightweight real-time DEtection TRansformer (L-RTDETR) model, achieving a balance between high accuracy and model lightness. The model employs a Large Selective Kernel Network (LSKNet) based lightweight backbone, reducing network size and parameters. It also integrates Deformable Attention (DAttention) based Intra-scale Feature Interaction to enhance detection speed with a sparse attention mechanism. By combining L1, Generalized Intersection over Union (GIoU), and Normalized Wasserstein Distance (NWD) loss functions with weighted metrics, the model’s adaptability for object detection tasks improves. Experimental results on the Northeastern University Defect Detection dataset show that L-RTDETR achieves a $35.8 \%$ reduction in model size and increases detection speed to 61.6 fps, outperforming the conventional RT-DETR algorithm. It also shows a $2.5 \%$ improvement in mean Average Precision at $0.5(m A P \text{@} 0.5)$ and a $1.7 \%$ increase in mAP@0.5-0.95. These results underscore the model’s efficiency in lightweight defect detection.
This study presents an advanced gesture recognition and user interface (UI) interaction system developed using deeplearning technologies, emphasizing its transformative impact on UI design and functionality. By emplo...
详细信息
ISBN:
(数字)9798350355413
ISBN:
(纸本)9798350355420
This study presents an advanced gesture recognition and user interface (UI) interaction system developed using deeplearning technologies, emphasizing its transformative impact on UI design and functionality. By employing optimized convolutional neural networks (CNNs), the system achieves high-precision gesture recognition, significantly enhancing user interactions with digital interfaces. Initial steps involve preprocessing collected gesture images to conform to CNN input standards, followed by employing sophisticated feature extraction and classification methodologies. We address class imbalance effectively using Focal Loss as the loss function, ensuring robust model performance across varied gesture types. The experimental results showcase notable improvements in model metrics, with the AUC and Recall increasing progressively as we evolve from simpler models like VGG16 to more complex ones such as DenseNet. Our enhanced model demonstrates a significant advancement with an AUC of 0.83 and a Recall of 0.85, surpassing standard benchmarks. More critically, this system's capacity to support real-time and efficient gesture recognition paves the way for a new era in UI design—where intuitive, natural user gestures can seamlessly integrate into everyday technology use, significantly reducing the learning curve and enhancing user satisfaction. The broad implications of this development are profound, extending beyond mere technical performance to fundamentally reshape how users interact with technology. Such advancements hold considerable promise for the enhancement of smart life experiences, highlighting the pivotal role of gesture-based interactions in the next generation of UI development.
Aiming at addressing the problems of low visibility and poor contrast, this paper proposes a new dark channel prior dehazing. According to the characteristics of the light source, an image is divided into light and no...
Aiming at addressing the problems of low visibility and poor contrast, this paper proposes a new dark channel prior dehazing. According to the characteristics of the light source, an image is divided into light and non-light source areas. Mixed precision operation is used to subsample the dark channel image and deeplearning network and GPU-accelerated method are used to improve the algorithm speed to solve the real-time problem. Experimental results show that compared with similar algorithms, the new algorithm is more balanced in image quality indicators and underwater image indicators, which better working requirements of underwater vehicles. In terms of real-time performance, the new algorithm is superior to similar algorithms. When processingimages a $950\times 550$ pixel, resolving new with an average frame rate of 29.4, which runs 2.46 times faster than dark channel prior, which lays a foundation for underwater robots to carry out underwater operations more efficiently.
A significant amount of research effort is put into studying machine learning (ML) and deeplearning (DL) technologies. real-world ML applications help companies to improve products and automate tasks such as classifi...
详细信息
ISBN:
(数字)9783030672928
ISBN:
(纸本)9783030672911;9783030672928
A significant amount of research effort is put into studying machine learning (ML) and deeplearning (DL) technologies. real-world ML applications help companies to improve products and automate tasks such as classification, image recognition and automation. However, a traditional "fixed" approach where the system is frozen before deployment leads to a sub-optimal system performance. Systems autonomously experimenting with and improving their own behavior and performance could improve business outcomes but we need to know how this could actually work in practice. While there is some research on autonomously improving systems, the focus on the concepts and theoretical algorithms. However, less research is focused on empirical industry validation of the proposed theory. Empirical validations are usually done through simulations or by using synthetic or manually alteration of datasets. The contribution of this paper is twofold. First, we conduct a systematic literature review in which we focus on papers describing industrial deployments of autonomously improving systems and their real-world applications. Secondly, we identify open research questions and derive a model that classifies the level of autonomy based on our findings in the literature review.
Monitoring and maintenance of water resources projects is essential to ensure project safety and environmental protection. Traditional monitoring methods often rely on manual inspections and sensor data, but these met...
详细信息
ISBN:
(数字)9798350369212
ISBN:
(纸本)9798350380002
Monitoring and maintenance of water resources projects is essential to ensure project safety and environmental protection. Traditional monitoring methods often rely on manual inspections and sensor data, but these methods suffer from high cost, low efficiency and limited monitoring range. With the rapid development of computer vision technology, its potential for application in water conservancy engineering has gradually emerged. The purpose of this paper is to study the water conservancy project monitoring and intelligent identification methods based on computer vision technology. We first outline the basic concepts of computer vision and its application scenarios in water conservancy engineering, including structural health monitoring, environmental monitoring and disaster management. Then, the specific applications of imageprocessing techniques, machine learning and deeplearning in intelligent recognition are described in detail, especially how these techniques can be utilized to achieve an efficient and accurate real-time monitoring system. Through several case studies, this paper demonstrates the effects and advantages of the application of computer vision technology in actual water conservancy projects. Finally, it summarizes the main findings of the current study and looks forward to the future technological development and application prospects, pointing out the direction of further research. The findings of this paper show that computer vision-based monitoring and intelligent identification methods not only improve the accuracy and efficiency of monitoring, but also provide new technical means for the management and maintenance of water conservancy projects.
With the development of new technologies such as big data, cloud computing, and the Internet of Things, network attack technology is constantly evolving and upgrading, and network attack detection technology is forced...
详细信息
With the development of new technologies such as big data, cloud computing, and the Internet of Things, network attack technology is constantly evolving and upgrading, and network attack detection technology is forced to undergo corresponding iterative evolution. Three main problems are associated with these technologies: the automatic representation of heterogeneous and complex network traffic data, the uneven network attack samples, and the contradiction between the accuracy of the anomaly detection model and the continuous evolution of attacks. Researchers have proposed several network attack detection techniques based on deeplearning to address these problems. This study reviews and analyzes the studies aimed at dealing with such problems, considering multiple factors, such as models, traffic representation and feature extraction, threat detection model training, and model robustness improvement. Finally, the existing problems and challenges associated with the current research are analyzed with respect to data category imbalance, high-dimensional massive data processing, concept distribution drift, real-time interpretability of the detection model, and the security of the model.
At present, in the field of radar emitter classification, theoretical simulation is mostly used to carry out algorithm research. However, there are few schemes to study signal classification in real electromagnetic en...
详细信息
暂无评论