Heart rate (HR) is a vital parameter for assessing human health. Currently, image photoplethysmography (iPPG) has received widespread attention for its advantages in non-contact HR detection applications. However, iPP...
详细信息
ISBN:
(纸本)9798350386660;9798350386677
Heart rate (HR) is a vital parameter for assessing human health. Currently, image photoplethysmography (iPPG) has received widespread attention for its advantages in non-contact HR detection applications. However, iPPG has poor anti-interference ability, and its application scenarios still have significant limitations. In this paper, we focused on the office environment as a typical application scenario and conducted research on human HR monitoring under different light and noise interference conditions. Experimental data was collected from facial videos of the subjects captured by normal camera. Eulerian video magnification (EVM) algorithm was used to enhance the subtle differences in pixel intensity of facial skin. The extraction of blood volume pulse (BVP) signal was realized based on the combination of improved complete ensemble empirical mode decomposition with adaptive noise (ICEEMDAN) and second-order blind identification (SOBI) in this study. Further, the residual high-frequency noise and harmonic components in the BVP source signal were filtered out using the single value decomposition (SVD) to improve the estimation accuracy of HR. The experimental results showed that the method is highly accurate in monitoring HR of subjects under different light and noise interference environments. Also, the robustness of the method is excellent, and the further expansion of the research has a great application prospect.
Railway plays a leading role in the field of transportation in China and shoulders the important mission of driving the development of national economy. In view of the changeable environment of railway track, the proc...
详细信息
ISBN:
(纸本)9798350350227;9798350350210
Railway plays a leading role in the field of transportation in China and shoulders the important mission of driving the development of national economy. In view of the changeable environment of railway track, the processing algorithm with good real-time performance, strong robustness and high accuracy in the process of foreign body detection is the key to achieve rapid detection. In order to facilitate researchers to compare and analyze the effects of relevant algorithms intuitively and quickly, the GUI visual interactive interface was used to design the simulation platform for track foreign object imageprocessing, conceive the design process, build the overall design of the platform, and divide the static foreign object detection and moving target detection modules according to the realization functions. In the static foreign body detection module, the image of track foreign body is input to determine whether there is track foreign body and give early warning. Meanwhile, the processing effect of different algorithms can be visually compared through the operation results. In the moving target detection module, the moving target in the input video is marked and tracked. The test results of the simulation platform show that the processing platform is simple and easy to operate, and can effectively assist researchers to deepen the understanding and application of orbital foreign object imageprocessing.
To tackle two prevalent challenges in video anomaly detection, misclassification of abnormal frames as normal and the prolonged runtime of existing methods, this article proposes a time-efficient anomaly detection met...
详细信息
To tackle two prevalent challenges in video anomaly detection, misclassification of abnormal frames as normal and the prolonged runtime of existing methods, this article proposes a time-efficient anomaly detection method utilizing twice-reconstruction and parallel computing. The proposed method integrates three key innovations: a twice-reconstruction model to amplify the reconstruction error for abnormal frames, a memory module to better capture and utilize normal distribution patterns, and a parallel computing strategy to significantly accelerate anomaly detection. By reconstructing frames twice, our method enhances the model's anomaly detection capabilities. Meanwhile, a memory module is incorporated to retain normal distribution patterns more effectively, reducing the likelihood of misclassifying abnormal frames. Furthermore, parallel computing is employed to minimize runtime and boost detection efficiency. Unlike existing methods, TPR-VAD achieves both high detection accuracy and time efficiency, making it well-suited for real-world applications such as intelligent surveillance systems. Experimental results on the UCSD Ped2, CUHK Avenue, and ShanghaiTech datasets show that our method achieves superior detection accuracy, reaching 99.01%, 91.21%, and 82.77%, respectively, while significantly reducing runtime.
The use of real-time remote communication has seen significant growth in the last few years. The need for providing the feeling of togetherness in real-time remote communication, combined with new developments in volu...
详细信息
ISBN:
(纸本)9798350393767;9798350393774
The use of real-time remote communication has seen significant growth in the last few years. The need for providing the feeling of togetherness in real-time remote communication, combined with new developments in volumetric video, are expected to lead to immersive and holographic remote communication in the near future. These services will require high bandwidth, low latency and significant processing both at the sender and receiver side, and often within the network itself too. In order to fulfill these requirements, we argue that a holistic cross-layer optimization approach, that takes input from and provides optimization actions to all layers involved in the delivery of these services is necessary. In this paper we provide insights in the design and implementation of a cross-layer system orchestrator for adaptation of real-time remote communication. Based on inputs from the network and application layers, it uses a machine learning (ML) model to maximize the objective video quality metric by finding the best system configuration and taking adaptation actions in both layers. The model performance shows that it learned how to offset any system dynamics coming from the environment with the correct configuration settings.
The effectiveness of autonomous vehicles relies on clear visual input, which rain can significantly obstruct. Rain streaks degrade the quality of captured images and videos, affecting both user perception and the func...
详细信息
ISBN:
(纸本)9781510679344;9781510679351
The effectiveness of autonomous vehicles relies on clear visual input, which rain can significantly obstruct. Rain streaks degrade the quality of captured images and videos, affecting both user perception and the functionality of outdoor vision systems, such as those in autonomous vehicles. This visual degradation impacts the vehicle's ability to interpret its environment, increasing the risk of driving in rainy conditions. Researchers have responded to this challenge by developing various rain removal algorithms, ranging from single-image to video-based approaches, each with its own strengths and weaknesses. This research aims to develop two novel, efficient single-image rain removal algorithms that strike a balance between high performance and quick execution. The proposed algorithms will address the need for an effective de-raining technique suitable for real-time use in autonomous vehicles. By improving visibility in rainy conditions, this innovation will enhance the performance and safety of autonomous vehicles, contributing to advancements in the field. A survey involving ten imageprocessing experts and professionals, who evaluated the results of both algorithms based on perceived quality and improvement, revealed that Algorithm 1 received a higher average rating (0.58) compared to Algorithm 2 (0.43). Although Algorithm 1 is slightly preferred based on average participant ratings, Algorithm 2's superior edge preservation and image sharpness make it more favorable for applications demanding high accuracy and detailed image retention. Overall, the project meets the demand for real-time rain removal in autonomous vehicles and provides valuable insights into the effectiveness of Algorithm 1 in de-raining images compared to Algorithm
With the rise of the Internet of Things (IoT) and edge computing technologies, traditional cloud-dependent convolutional neural network (CNN) imageprocessing methods are facing the challenges of latency and bandwidth...
详细信息
We demonstrate a fully submerged underwater LiDAR transceiver system based on single-photon detection technologies. The LiDAR imaging system used a silicon single-photon avalanche diode (SPAD) detector array fabricate...
详细信息
We demonstrate a fully submerged underwater LiDAR transceiver system based on single-photon detection technologies. The LiDAR imaging system used a silicon single-photon avalanche diode (SPAD) detector array fabricated in complementary metal-oxide semiconductor (CMOS) technology to measure photon time-of-flight using picosecond resolution time-correlated single-photon counting. The SPAD detector array was directly interfaced to a Graphics processing Unit (GPU) for real-timeimage reconstruction capability. Experiments were performed with the transceiver system and target objects immersed in a water tank at a depth of 1.8 meters, with the targets placed at a stand-off distance of approximately 3 meters. The transceiver used a picosecond pulsed laser source with a central wavelength of 532 nm, operating at a repetition rate of 20 MHz and average optical power of up to 52 mW, dependent on scattering conditions. Three-dimensional imaging was demonstrated by implementing a joint surface detection and distance estimation algorithm for real-timeprocessing and visualization, which achieved images of stationary targets with up to 7.5 attenuation lengths between the transceiver and the target. The average processingtime per frame was approximately 33 ms, allowing real-time three-dimensional video demonstrations of moving targets at ten frames per second at up to 5.5 attenuation lengths between transceiver and target.
Novel view synthesis is frequently employed in video streaming, temporal upsampling, or virtual reality. We propose a new representation, potentially visible layered image (PVLI), that uses a combination of a potentia...
详细信息
Novel view synthesis is frequently employed in video streaming, temporal upsampling, or virtual reality. We propose a new representation, potentially visible layered image (PVLI), that uses a combination of a potentially visible set of the scene geometry and layered color images. PVLI encodes the depth implicitly and enables cheap run-time reconstruction. Furthermore, PVLI can also be used to reconstruct pixel and layer connectivities, which is crucial for filtering and post-processing of the rendered images. We use PVLIs to achieve local and server-based real-time ray tracing. In the first case, PVLIs are used as a basis for temporal and spatial upsampling of ray-traced illumination. In the second case, PVLIs are compressed, streamed over the network, and then used by a thin client to perform temporal and spatial upsampling and to hide latency. To shade the view, we use path tracing, accounting for effects such as soft shadows, global illumination, and physically based refraction. Our method supports dynamic lighting, and up to a limited extent, it also handles view-dependent surface interactions.
images and videos captured in poor illumination conditions are degraded by low brightness, reduced contrast, color distortion, and noise, rendering them barely discernable for human perception and ultimately negativel...
详细信息
ISBN:
(纸本)9781510673854;9781510673847
images and videos captured in poor illumination conditions are degraded by low brightness, reduced contrast, color distortion, and noise, rendering them barely discernable for human perception and ultimately negatively impacting computer vision system performance. These challenges are exasperated when processingvideo surveillance camera footage, using this unprocessed video data as-is for real-time computer vision tasks across varying environmental conditions within Intelligent Transportation Systems (ITS), such as vehicle detection, tracking, and timely incident detection. The inadequate performance of these algorithms in real-world deployments incurs significant operational costs. Low-light image enhancement (LLIE) aims to improve the quality of images captured in these unideal conditions. Groundbreaking advancements in LLIE have been recorded employing deep-learning techniques to address these challenges, however, the plethora of models and approaches is varied and disparate. This paper presents an exhaustive survey to explore a methodical taxonomy of state-of-the-art deep learning-based LLIE algorithms and their impact when used in tandem with other computer vision algorithms, particularly detection algorithms. To thoroughly evaluate these LLIE models, a subset of the BDD100K dataset, a diverse real-world driving dataset is used for suitable image quality assessment and evaluation metrics. This study aims to provide a detailed understanding of the dynamics between low-light image enhancement and ITS performance, offering insights into both the technological advancements in LLIE and their practical implications in real-world conditions. The project Github repository can be accessed here.
Tensile testing (aka tension testing) is a widely employed mechanical testing technique for analyzing materials' properties and behavior under applied stress. Tensile testing plays a pivotal role in helping engine...
详细信息
Tensile testing (aka tension testing) is a widely employed mechanical testing technique for analyzing materials' properties and behavior under applied stress. Tensile testing plays a pivotal role in helping engineers to make informed decision about material selection and usage. Despite its importance, there is a limited numbers of studies that explored the potential of AI techniques for realtime monitoring and material behavior prediction in tensile testing. To this end, this work presents a deep learning model designed to predict the material's condition throughout tensile testing and provide an early warning prior to fracture. By leveraging a comprehensive dataset of tension test video samples, the proposed model utilizes both convolution and recurrent neural networks to extract pertinent spatial and temporal visual features, thereby predicting the frames at which material deformation and fracture occur. The evaluation results of our research showed that the proposed model achieved a predictive ability with an F1-score of 97%, on average. The implications of our research are significant for industries and researchers in the field of materials science and engineering. By accurately predicting material status, our model enables automounts, realtime analysis of material behavior during tensile testing, leading to better time and cost efficiency in various applications.
暂无评论