The goal of functional error correction is to preserve neural network performance when stored network weights are corrupted by noise. To achieve this goal, a selective protection (SP) scheme was proposed to optimally ...
详细信息
One of the most consequential public health issues in the world and a major factor in women's mortality is breast cancer. Early detection and diagnosis can significantly improve the likelihood of survival. Therefo...
详细信息
One of the most consequential public health issues in the world and a major factor in women's mortality is breast cancer. Early detection and diagnosis can significantly improve the likelihood of survival. Therefore, this study suggests a deep end-to-end heterogeneous ensemble approach by using deep convolutional neural networks models for breast histological images classification tested on the BreakHis dataset. The proposed approach showed a significant increase of performances compared to their base learners. Thus, seven deep learning architectures (vGG16, vGG19, ResNet50, Inception v3, Inception ResNet v2, Xception, and MobileNet v2) were trained using fivefold cross-validation. Thereafter, deep end-to-end heterogeneous ensembles of two up to seven models were constructed based on three selection criteria's (by accuracy, by diversity, and by both accuracy and diversity) and combined with two voting methods: majority voting by tacking the mode of the distribution of the predicted labels, and weighted voting by taking the average of predicted probabilities. Results showed the effectiveness of deep end-to-end ensemble learning techniques for histopathological breast cancer images classification since the ensembles designed using weighted voting with the selection by accuracy strategy method exceeded the ones designed using the selection by diversity or by accuracy and diversity strategies. The accuracy values of the proposed approach have shown a significant amelioration compared to the least performing base learner used as a baseline ResNet 50 with an accuracy increased from 78.14%, 78.57%, 82.80 and 79.43% to 93.8%, 93.4%, 93.3%, and 91.8% through the BreakHis dataset's four magnification factors: 40X, 100X, 200X, and 400X respectively.
Defect detection is a crucial quality control process in the manufacturing industry, aimed at identifying and classifying imperfections or anomalies in products before they reach customers. Traditional manual inspecti...
详细信息
The emergent ecosystems of intelligent edge devices in diverse Internet-of-Things (IoT) applications, from automatic surveillance to precision agriculture, increasingly rely on recording and processing a variety of im...
详细信息
The emergent ecosystems of intelligent edge devices in diverse Internet-of-Things (IoT) applications, from automatic surveillance to precision agriculture, increasingly rely on recording and processing a variety of image data. Due to resource constraints, e.g., energy and communication bandwidth requirements, these applications require compressing the recorded images before transmission. For these applications, image compression commonly requires: 1) maintaining features for coarse-grain pattern recognition instead of the high-level details for human perception due to machine-to-machine communications;2) high compression ratio that leads to improved energy and transmission efficiency;and 3) large dynamic range of compression and an easy tradeoff between compression factor and quality of reconstruction to accommodate a wide diversity of IoT applications as well as their time-varying energy/performance needs. To address these requirements, we propose, MAGIC, a novel machine learning (ML)-guided image compression framework that judiciously sacrifices the visual quality to achieve much higher compression when compared to traditional techniques, while maintaining accuracy for coarse-grained vision tasks. The central idea is to capture application-specific domain knowledge and efficiently utilize it in achieving high compression. We demonstrate that the MAGIC framework is configurable across a wide range of compression/quality and is capable of compressing beyond the standard quality factor limits of both JPEG 2000 and WebP. We perform experiments on representative IoT applications using two vision data sets and show 42.65x compression at similar accuracy with respect to the source. We highlight low variance in compression rate across images using our technique as compared to JPEG 2000 and WebP.
image inpainting is the process of reconstructing missing or damaged regions in an image and is an important task in computer visionapplications for restoration and enhancement. However, repair algorithms are often s...
详细信息
ISBN:
(纸本)9798350365740;9798350365757
image inpainting is the process of reconstructing missing or damaged regions in an image and is an important task in computer visionapplications for restoration and enhancement. However, repair algorithms are often sensitive to noise and yield suboptimal results. To address this challenge, a new integrated two-stage framework is introduced to improve the performance of image inpainting. In the first stage, an effective Noise2void denoising is applied to learn meaningful representations of image patches and effectively denoise the input image. The proposed N2v model considers the structural links between pixels and retains contextual information at the same time, while suppressing noise. In the 2nd stage, an advanced enhanced DeepFill inpainting model employing deep neural networks is applied. Experimental results showed that the method proposed will outperform traditional repair methods. The denoising step tunes the accuracy of reconstructing missing areas, and greatly improves the quality of inpainting. Applied on huge benchmark datasets, the performance is evaluated and demonstrated that N2v integrated with DeepFill outperforms individual inpainting techniques. Furthermore, we carry out an ablation study to evaluate the contribution of each constituent part of our proposed framework. This outcome underscores the complementary nature of the denoising and repair stages and points to the need for noise control before repairs. In general, our technique provides a strong and effective approach to image restoration tasks and allows for improving inpainting methods under real-world conditions.
Object recognition, an essential technique in computer vision, enables machines to identify and understand real-time objects and environments based on input images. The main aim of this technology is to accurately rec...
详细信息
UAv-based intelligent data acquisition for 3D reconstruction and monitoring of infrastructure has experienced an increasing surge of interest due to recent advancements in imageprocessing and deep learning-based tech...
详细信息
UAv-based intelligent data acquisition for 3D reconstruction and monitoring of infrastructure has experienced an increasing surge of interest due to recent advancements in imageprocessing and deep learning-based techniques. view planning is an essential part of this task that dictates the information capture strategy and heavily impacts the quality of the 3D model generated from the captured data. Recent methods have used prior knowledge or partial reconstruction of the target to accomplish view planning for active reconstruction;the former approach poses a challenge for complex or newly identified targets while the latter is computationally expensive. In this work, we present Bag-of-views (Bov), a fully appearance-based model used to assign utility to the captured views for both offline dataset refinement and online next-best-view (NBv) planning applications targeting the task of 3D reconstruction. With this contribution, we also developed the view Planning Toolbox (vPT), a lightweight package for training and testing machine learning-based view planning frameworks, custom view dataset generation of arbitrary 3D scenes, and 3D reconstruction. Through experiments which pair a Bov-based reinforcement learning model with vPT, we demonstrate the efficacy of our model in reducing the number of required views for high-quality reconstructions in dataset refinement and NBv planning.
In this demonstration paper, we present "e2evideo" a versatile Python package composed of domain-independent modules. These modules can be seamlessly customised to suit specialised tasks by modifying specifi...
详细信息
This paper proposes machinevision-based live traffic monitoring aided with various imageprocessing methods. The application is designed to monitor live traffic on a road or private property premises with features li...
详细信息
To address the problems such as the difficulty of traffic sign detection and recognition under low illumination, a new low illumination traffic sign detection and recognition algorithm is proposed. The algorithm first...
详细信息
To address the problems such as the difficulty of traffic sign detection and recognition under low illumination, a new low illumination traffic sign detection and recognition algorithm is proposed. The algorithm firstly uses an illumination judgement algorithm to filter out low-illumination images, then uses a New Illumination Enhancement algorithm to adjust the brightness and contrast of the low-illumination images, and finally uses mask RCNN (mask region-based convolutional neural network, mask RCNN) to detect and recognize traffic signs. The New Illumination Enhancement Algorithm is based on Illumination-Reflection model, firstly converting the image RGB space into HSv space, applying guided filtering to the v channel to obtain the illumination component, using the illumination component to extract the reflection component, and adjusting the reflection component by linear pull-up. Next, the distribution characteristics of the illumination component are used to adjust the 2D gamma function and obtain the optimized illumination component. Subsequently, the illumination component is used to obtain the detail component. Finally, a hybrid spatial enhancement method is used to obtain the enhanced v-channel and reconstruct the image. The experimental results show that the New Illumination Enhancement algorithm can effectively improve image brightness and sharpness in both low illumination traffic scenes, ensure that the image is not distorted, retain image information and enhance the prominence of traffic signs in traffic scenes. In the ZCTSDB-lightness test set, the combined algorithm of new light image enhancement and Mask RCNN improved object detection mAP(bb) and instance segmentation mAP(seg) by 2.810% and 1.176%, respectively, compared to Mask RCNN. In the ZCTSDB test set, the performance of the new low illumination traffic sign detection and recognition algorithm outperformed all other algorithms.
暂无评论