Learner engagement is a significant factor determining the success of implementing an intelligent educational network. Currently the use of Massive Open Online Courses has increased because of the flexibility offered ...
详细信息
Learner engagement is a significant factor determining the success of implementing an intelligent educational network. Currently the use of Massive Open Online Courses has increased because of the flexibility offered by such online learning systems. The COVID period has encouraged practitioners to continue to engage in new ways of online and hybrid teaching. However, monitoring student engagement and keeping the right level of interaction in an online classroom is challenging for teachers. In this paper we propose an engagement recognition model by combining the image traits obtained from a camera, such as facial emotions, gaze tracking with head pose estimation and eye blinking rate. In the first step, a face recognition model was implemented. The next stage involved training the facial emotion recognition model using deeplearning convolutional neural network with the datasets FER 2013. The classified emotions were assigned weights corresponding to the academic affective states. Subsequently, by using the Dlib's face detector and shape predicting algorithm, the gaze direction with head pose estimation, eyes blinking rate and status of the eye (closed or open) were identified. Combining all these modalities obtained from the image traits, we propose an engagement recognition system. The experimental results of the proposed system were validated by the quiz score obtained at the end of each session. This model can be used for realtime video processing of the student's affective state. The teacher can obtain a detailed analytics of engagement statics on a spreadsheet at the end of the session thus facilitating the necessary follow-up actions.
One of the most severe kinds of tumors in people is lung cancer. Identifying lung cancer and its types requires costly and time-consuming procedure research. Furthermore, lung nodules are difficult to identify because...
详细信息
One of the most severe kinds of tumors in people is lung cancer. Identifying lung cancer and its types requires costly and time-consuming procedure research. Furthermore, lung nodules are difficult to identify because of their diversity and visual similarity to neighboring locations. Conventional machine learning methods either treat these components separately or rely on human integration, which can be time-consuming and could fail to fully capture the intricate relationships between these features. deeplearning methods' layered structures enable them to automatically incorporate many features and learn meaningful descriptions. Thus, this framework proposes an efficient deeplearning technique for classifying pulmonary nodules using Computerized Tomography (CT) images. Initially, several pre-processing methods are taken into account to prepare the data. Then, TNet based deeplearning algorithm segments the lung nodule, and the CenterNet-based method extracts the texture and intensity attributes from the segmented image. Following that, the proposed NASNet-based classifier categorizes the nodules as cancerous or not, using the attributes that have been collected. Finally, the presented method will be assessed by the metrics like Dice similarity coefficient (DSC), Sensitivity, Positive predictive Value (PPV), f1-score, precision, recall, and accuracy on the LUNA-16 and Lung image Database Consortium (LIDC-IDRI) datasets, and the outcomes are contrasted with other existing approaches.
Tiny deeplearning Models offer many advantages in various applications. From the perspective of statistical machine learning theory the contributions of this paper is to complement the research advances and results o...
详细信息
Tiny deeplearning Models offer many advantages in various applications. From the perspective of statistical machine learning theory the contributions of this paper is to complement the research advances and results obtained so far in real-time 3D object recognition. We propose a Tiny deeplearning Model named Complementary Spatial Transformer Network (CSTN) for real-time 3D object recognition. It turns out that CSTN's working, and analysis are much simplified in a target space setting. We make algorithmic enhancements to perform CSTN computations faster and keep the learning part of CSTN in minimal size. Finally, we provide the experimental verifications of the results obtained in publicly available point cloud data sets ModelNet40 and ShapeNetCore with our model performing 1.65-2 times better in DPS (Detections/s) rate on GPU hardware for 3D object recognition, when compared to state-of-the-art networks. Complementary Spatial Transformer Network architecture requires only 10-35% of trainable parameters, when compared to state-of-the-art networks, making the network easier to deploy in edge AI devices.
This work aims to simplify the characterization process of coded-apertures for computational imaging (CI) at microwave frequencies. A major benefit of the presented technique is the minimization of the processingtime...
详细信息
This work aims to simplify the characterization process of coded-apertures for computational imaging (CI) at microwave frequencies. A major benefit of the presented technique is the minimization of the processingtime needed to calculate the system sensing matrix for microwave CI-based compressive sensing applications. To achieve this, a deeplearning-based approach which is capable of generating the sensing matrix using features learned directly from the coded-aperture distribution is proposed. To avoid the vanishing gradient problem, the proposed deeplearning network contains skip connections. Using a dataset of 1,000 testing samples, the average normalized mean-squared-error (NMSE) calculated between the sensing matrix generated by the conventional method and that predicted by the proposed network is 0.0036. Moreover, the average mean-squared-error (MSE) calculated between the images reconstructed using the conventional and the predicted sensing matrix is 0.00297. In addition to providing high-fidelity estimations with minimized error, we demonstrate that using the trained network, the prediction of the sensing matrix can be achieved in 0.212 s, corresponding to a 65% reduction in the computation time needed to calculate the sensing matrix. This has significant outcomes in achieving real-time operation of CI-based microwave imaging systems.
Objective With the popularity of high-resolution devices such as high-definition, ultra-high-definition televisions, and smartphones, the demand for high-resolution images is also increasing, which puts forward higher...
详细信息
Objective With the popularity of high-resolution devices such as high-definition, ultra-high-definition televisions, and smartphones, the demand for high-resolution images is also increasing, which puts forward higher requirements for high-resolution imageprocessing and entity recognition *** This article introduced the research progress and application of high-resolution imageprocessing and entity recognition algorithms from the perspective of artificial intelligence (AI). First, the important role of AI in high-resolution imageprocessing and entity recognition was introduced, and then the applications of deeplearning-based algorithms in high-resolution image grayscale equalization, denoising, and deblurring were introduced. Subsequently, the application of AI-based object detection and image segmentation algorithms in entity recognition was explored, and the superiority of AI-based high-resolution imageprocessing and entity recognition algorithms was verified through training and testing. The accuracy of the model was verified through testing experiments. Finally, a summary and outlook were made on high-resolution imageprocessing and entity recognition algorithms based on *** After experimental testing, it was found that high-resolution imageprocessing and entity recognition based on AI had higher efficiency, and the overall image recognition ability was improved by 29.6% compared to traditional image recognition models. The recognition speed and accuracy were also *** High-resolution imageprocessing and element recognition algorithms based on AI enabled observers to see the detailed information in the image more clearly, thus improving the efficiency and accuracy of image analysis. Through continuous improvement of algorithm performance, real-time application, and expansion of cross-disciplinary applications, people can look forward to the development of more advanced and powerful imageprocessing and entity recognition
To enhance the appeal of residential real estate listings and captivate online customers, clean and visually convincing indoor scenes are highly desirable. In this research, we introduce an innovative image inpainting...
详细信息
To enhance the appeal of residential real estate listings and captivate online customers, clean and visually convincing indoor scenes are highly desirable. In this research, we introduce an innovative image inpainting model designed to seamlessly replace undesirable elements within images of indoor residential spaces with realistic and coherent alternatives. While Generative Adversarial Networks (GANs) have demonstrated remarkable potential for removing unwanted objects, they can be resource-intensive and face difficulties in consistently producing high-quality outcomes, particularly when unwanted objects are scattered throughout the images. To empower small- and medium-sized businesses with a competitive edge, we present a novel GAN model that is resource-efficient and requires minimal training time using arbitrary mask generation and a novel half-perceptual loss function. Our GAN model achieves compelling results in removing unwanted elements from indoor scenes, demonstrating the capability to train within a single day using a single GPU, all while minimizing the need for extensive post-processing.
Application of Convolutional neural network in spectrum of Medical image analysis are providing benchmark outputs which converges the interest of many researchers to explore it in depth. Latest preprocessing technique...
详细信息
Application of Convolutional neural network in spectrum of Medical image analysis are providing benchmark outputs which converges the interest of many researchers to explore it in depth. Latest preprocessing technique real ESRGAN (Enhanced super resolution generative adversarial network) and GFPGAN (Generative facial prior GAN) are proving their efficacy in providing high resolution dataset. Objective: Optimizer plays a vital role in upgrading the functioning of CNN model. Different optimizers like Gradient descent, Stochastic Gradient descent, Adagrad, Adadelta and Adam etc. are used for classification and segmentation of Medical image but they suffer from slow processing due to their large memory requirement. Stochastic Gradient descent suffers from high variance and is computationally expensive. Dead neuron problem also proves to detrimental to the performance of most of the optimizers. A new optimization technique Gradient Centralization is providing the unparalleled result in terms of generalization and execution time. Method: Our paper explores the next factor which is the employment of new optimization technique, Gradient centralization (GC) to our integrated framework (Model with advanced preprocessing technique). Result and conclusion: Integrated Framework of real ESRGAN and GFPGAN with Gradient centralization provides an optimal solution for deeplearning models in terms of Execution time and Loss factor improvement.
Underwater images often exhibit color deviation, reduced contrast, distortion, and other issues due to light refraction, scattering, and absorption. Therefore, restoring detailed information in underwater images and o...
详细信息
Underwater images often exhibit color deviation, reduced contrast, distortion, and other issues due to light refraction, scattering, and absorption. Therefore, restoring detailed information in underwater images and obtaining high-quality results are primary objectives in underwater image enhancement tasks. Recently, deeplearning-based methods have shown promising results, but handling details in low-light underwater imageprocessing remains challenging. In this paper, we propose an attention-based color consistency underwater image enhancement network. The method consists of three components: illumination detail network, balance stretch module, and prediction learning module. The illumination detail network is responsible for generating the texture structure and detail information of the image. We introduce a novel color restoration module to better match color and content feature information, maintaining color consistency. The balance stretch module compensates using pixel mean and maximum values, adaptively adjusting color distribution. Finally, the prediction learning module facilitates context feature interaction to obtain a reliable and effective underwater enhancement model. Experiments conducted on three real underwater datasets demonstrate that our approach produces more natural enhanced images, performing well compared to state-of-the-art methods.
A human motion monitoring method based on thermal radiation image system and target detection technology is developed. The heat distribution of human body in motion is captured by thermal image, and the real-time reco...
详细信息
A human motion monitoring method based on thermal radiation image system and target detection technology is developed. The heat distribution of human body in motion is captured by thermal image, and the real-time recognition and analysis of human motion is realized by combining imageprocessing and target detection algorithm. A complete set of thermal radiation optical image monitoring system is designed in this study. A highsensitivity thermal imaging camera is used to capture the thermal radiation images of the human body in the process of motion, and these images are then transmitted to the data acquisition unit for preliminary data collation and storage. The imageprocessing module preprocesses the acquired thermal images, and the preprocessed images are fed into the object detection algorithm, which is based on the deeplearning framework and can recognize and classify different movements of the human body. The thermal radiation image monitoring system can accurately capture the thermal image of the human body in different motion states, and identify the movement type of the athlete in realtime through the target detection algorithm. The system has a strong ability to capture the details of actions, and can identify the beginning, progress and end stages of actions. Compared with traditional monitoring methods, the thermal radiation light image monitoring system has obvious advantages in terms of data accuracy and real-time performance. This method can not only provide high precision movement recognition, but also has the advantages of non-contact and real-time monitoring, which greatly improves the efficiency and accuracy of sports training monitoring.
Automation in horticulture with computer vision and deeplearning revolutionizes the industry by enabling precise and efficient harvesting and disease detection. This technology enhances productivity and quality contr...
详细信息
ISBN:
(纸本)9798331541859;9798331541842
Automation in horticulture with computer vision and deeplearning revolutionizes the industry by enabling precise and efficient harvesting and disease detection. This technology enhances productivity and quality control, benefiting farmers and food manufacturers alike. The present work offers a comparative analysis of advanced deeplearning detection models-YOLO V8, RT-DETR with ResNet and with EfficientNet backbones-focused on identifying diseased (canker) and healthy (fresh) oranges. A comprehensive evaluation of these three models was conducted to assess their performance. YOLO V8 emerged with superior results, showcasing its robustness and effectiveness for real-time detection. Amongst the 3models YOLO V8 attained the best accuracy value of 99.6%. The other performance metrics underscore the model's high accuracy and efficiency in identifying and classifying objects with minimal errors, affirming its suitability for real-time horticultural diagnostics. The comparative analysis highlights YOLO V8's superiority over RT-DETR models with ResNet and EfficientNet backbones, positioning it as a highly reliable choice for detecting healthy and diseased oranges. This work advances the field by proving the superior performance of YOLO V8 for this specific application and validating the performance with the other sophisticated deeplearning architectures.
暂无评论