The concept of iris segmentation was created to increase the accuracy of iris recognition. Past recognition methods used the entire eye image directly for recognition classification, which led to poor recognition resu...
The concept of iris segmentation was created to increase the accuracy of iris recognition. Past recognition methods used the entire eye image directly for recognition classification, which led to poor recognition results. For the sake of sovling this problem, this paper proposes the ResU-Net (RU-Net) model, which can guide the network to learn more features that distinguish between iris and non-iris pixels. First, based on the U-Net, the backbone network model is changed to ResNet50 in this paper. This has the advantage of reducing the number of parameters and network complexity, and improving the learning capability. For the sake of solving the problem of sample imbalance between iris region and background region, this paper introduces the Focal Loss loss function. focal loss can effectively deal with the case of sample category imbalance and make the network focus more on the pixels that are difficult to classify. In this paper, the proposed RU-Net model is experimentally evaluated on the CASIA-Iris-Thousand dataset. The experimental results demonstrate that RU-Net achieves significant improvements on NIR iris images, reaching 96.22% MIoU and 98.19% MPA. This indicates that the RU-Net method outperforms other representative iris segmentation methods and has better segmentation capability.
We present a novel rationale-centric framework with human-in-the-loop - Rationales-centric Double-robustness Learning (RDL) - to boost model out-of-distribution performance in few-shot learning scenarios. By using sta...
详细信息
Online videos are a potent tool for educators to disseminate knowledge widely to diverse student audiences. However, collecting student feedback remains a significant challenge for lecturers, particularly in the absen...
详细信息
Online videos are a potent tool for educators to disseminate knowledge widely to diverse student audiences. However, collecting student feedback remains a significant challenge for lecturers, particularly in the absence of feedback. Understanding students’ subjective comprehension levels during online video lectures with sensor technology is yet to be thoroughly researched. This study uses eye-tracking technology to predict self-reported comprehension levels during video lectures. We recruited 20 participants from Germany and Japan who were invited to watch 50-minute lecture videos in three domains. The participants self-annotate the time segment in each lecture video where they dropout using open-source LabelStudio and answer the survey. We applied Long-Short-Term Memory (LSTM) to the preprocessed dataset and achieved an F1 Score of 0.886 for predicting binary self-annotated comprehension levels. We also introduce EyeUnderstand, the web-based application for visualizing the results of the comprehension estimation. We recruited 28 participants for the user study. As a result, 89.3% of the students and 92.9% of the lecturers confirmed that our application is practical.
3D content creation has long been a complex and time-consuming process, often requiring specialized skills and resources. While recent advancements have allowed for text-guided 3D object and scene generation, they sti...
详细信息
This work explores backdoor attack, which is an emerging security threat against deep neural networks (DNNs). The adversary aims to inject a backdoor into the model by manipulating a portion of training samples, such ...
详细信息
With the emergence of 5th generation mobile communication technology, the demand for Virtual Reality (VR) applications is on the rise worldwide. As one of the technologies related to visual content in VR, the quality ...
详细信息
Zero-shot learning (ZSL) aims to classify objects that are not observed or seen during training. It relies on class semantic description to transfer knowledge from the seen classes to the unseen classes. Existing meth...
详细信息
A large amount of data is collected during geological hazard monitoring, which is extremely valuable for further data mining, hazard monitoring and decision analysis. However, in the process of data collection and tra...
A large amount of data is collected during geological hazard monitoring, which is extremely valuable for further data mining, hazard monitoring and decision analysis. However, in the process of data collection and transmission may be affected by interference and other factors, resulting in the generation of abnormal data. To extract higher value information from these basic data, it is necessary to improve the quality of source data first, so it is necessary to detect the abnormal data. The monitoring data of the geological disaster are time-series data, which have the characteristics of time-series correlation. The nearest neighbor difference jump anomaly detection algorithm is suitable for monitoring anomalous data of geological disaster system, but there are shortcomings in selecting floating values and correlation perception. To address these problems, the nearest neighbor difference jumping algorithm is improved and a algorithm is proposed to detect anomalous data of geological disasters, i.e., the data series are segmented by using sliding windows, the selection of floating values is improved, the calculation range of difference values is expanded to enhance the correlation of data, and the concept of change speed is incorporated to better perceive the trend of change before and after the data, and finally the anomaly is determined by using the anomaly probability and correlation. After the experimental comparison and analysis, the accuracy and recall rate of the proposed method on the detection of geological disaster data are improved and meet the expected results.
In coming up with solutions to real-world problems, humans implicitly adhere to constraints that are too numerous and complex to be specified completely. However, reinforcement learning (RL) agents need these constrai...
详细信息
暂无评论