We propose a new method to visualize the centerline of a single tree canopy based on a depth camera. Firstly, the depth camera captures the image of the target tree to obtain the 3D point cloud data, which is filtered...
详细信息
ISBN:
(纸本)9783031500718;9783031500725
We propose a new method to visualize the centerline of a single tree canopy based on a depth camera. Firstly, the depth camera captures the image of the target tree to obtain the 3D point cloud data, which is filtered and denoised. Then, we used the Poisson surface reconstruction method to reconstruct the 3D spatial surface of the point cloud data to restore the real scene accurately. In addition, we used Random Sampling Consensus (RANSAC) and Least Square Circle (LSC) in MATLAB software to fit circles to the 3D point cloud. We proposed a new spatial straight-line fitting method to visualize the centerline of the tree crown. The method has the advantage of no error in the spatial scattering Z-coordinate, and the fitted straight line is perpendicular to the xoy plane. The new method produces a smaller root mean square error (RMSE) than the traditional spatial straight line fitting method. This method can be effectively applied to practical applications such as tree crown pruning, providing accurate positional information for the positioning of tools during the pruning process. Ultimately, the pruning time can be shortened, and the accuracy of the pruning process can be improved.
Diffusion models have shown remarkable success in visual synthesis, but have also raised concerns about potential abuse for malicious purposes. In this paper, we seek to build a detector for telling apart real images ...
详细信息
ISBN:
(纸本)9798350307184
Diffusion models have shown remarkable success in visual synthesis, but have also raised concerns about potential abuse for malicious purposes. In this paper, we seek to build a detector for telling apart real images from diffusiongenerated images. We find that existing detectors struggle to detect images generated by diffusion models, even if we include generated images from a specific diffusion model in their training data. To address this issue, we propose a novel image representation called DIffusion reconstruction Error (DIRE), which measures the error between an input image and its reconstruction counterpart by a pre-trained diffusion model. We observe that diffusion-generated images can be approximately reconstructed by a diffusion model while real images cannot. It provides a hint that DIRE can serve as a bridge to distinguish generated and real images. DIRE provides an effective way to detect images generated by most diffusion models, and it is general for detecting generated images from unseen diffusion models and robust to various perturbations. Furthermore, we establish a comprehensive diffusion-generated benchmark including images generated by various diffusion models to evaluate the performance of diffusion-generated image detectors. Extensive experiments on our collected benchmark demonstrate that DIRE exhibits superiority over previous generated-image detectors.
Synthetic data is considered to be a promising solution for data privacy and scarcity. Some studies have shown that synthetic data generated from a simple GAN-based model enables privacy-preserving data sharing and da...
详细信息
ISBN:
(纸本)9781510660335;9781510660342
Synthetic data is considered to be a promising solution for data privacy and scarcity. Some studies have shown that synthetic data generated from a simple GAN-based model enables privacy-preserving data sharing and data augmentation also in the medical imaging field. However, there are some limitations in applying this approach to real world situations: 1) Since generative models needs large amount of data to be trained, it is hard to be applied for small data situation. 2) Even after successfully training generative models, it is hard to guarantee which class the synthesized data corresponds to, especially for non-conditional generative models, so it needs to be re-labeled. Here, we propose GAN based ROI conditioned synthesis of medical image for data augmentation. We used StyleGAN2 to learn the distribution of CXR and Bayesian imagereconstruction for ROI-conditioned synthesis from the distribution. In the 4-class classification of CXRs showing normal, pneumonia, pleural effusion, and pneumothorax, using synthetic data for data sharing showed comparable performance to centralized learning, slightly better in terms of AUROC. Also, using synthetic data for augmentation, the accuracy and AUROC showed up to 6.5% and 8.9% increases, respectively.
Positron emission tomography (PET) is a quantitative imaging modality widely used in oncology, neurology, and pharmacology. The data acquired by a PET scanner correspond to projections of the concentration activity, a...
详细信息
ISBN:
(数字)9781665496209
ISBN:
(纸本)9781665496209
Positron emission tomography (PET) is a quantitative imaging modality widely used in oncology, neurology, and pharmacology. The data acquired by a PET scanner correspond to projections of the concentration activity, assumed to follow a Poisson distribution. The reconstruction of images from tomographic projections corrupted by Poisson noise is a challenging ill-posed large-scale inverse problem. Several available solvers use the majorization-minimization (MM) principle, though relying on various construction strategies with a lack of unifying framework. This work fills the gap by introducing the concept of Bregman majorization. This leads to a unified view of MM-based methods for imagereconstruction in the presence of Poisson noise. from this general approach, we exhibit three algorithmic solutions and compare their computational efficiency on a problem of dynamic PET imagereconstruction, either using GPU or CPU processing.
In view of the indisputable power of the inpainting technique for filling out visual information in images, this paper presents an approach for reconstructing occluded images from a single sample using the inpainting ...
详细信息
Creation of 3D models from a single RGB image is challenging problem in image processing these days, as the technology is in its early development stage. However, the demands for 3D technology and 3D reconstruction ha...
详细信息
Various supervised learning-based medical imagereconstruction methods have been developed with the goal of improving image quality (IQ). These methods typically use loss functions that minimize pixel-level difference...
详细信息
ISBN:
(纸本)9781510685963;9781510685970
Various supervised learning-based medical imagereconstruction methods have been developed with the goal of improving image quality (IQ). These methods typically use loss functions that minimize pixel-level differences between the reconstructed and high-quality target images. While they may seemingly perform well based on traditional image quality metrics such as mean squared error, they do not consistently improve objective IQ measures based on diagnostic task performance. This work introduces a task-informed learned imagereconstruction method. To establish the method, a measure of signal detection performance is incorporated in a hybrid loss function that is used for training. The proposed method is inspired by null space learning, and a task-informed data-consistent (DC) U-Net is utilized to estimate a null space component of the object that enhances task performance, while ensuring that the measurable component is stably reconstructed using a regularized pseudo-inverse operator. The impact of changing the specified task or observer at inference time to be different from that employed for model training, a phenomenon we refer to as "task-shift" or "observer-shift", respectively, was also investigated.
image forgery, as a classic form of academic mis-conduct, has garnered increasing interest from researchers in the field of research integrity. Concurrently, automated detection and localization methods for image forg...
详细信息
Recently, inversion methods have been exploring the incorporation of additional high-rate information from pretrained generators (such as weights or intermediate features) to improve the refinement of inversion and ed...
详细信息
ISBN:
(纸本)9798350318920;9798350318937
Recently, inversion methods have been exploring the incorporation of additional high-rate information from pretrained generators (such as weights or intermediate features) to improve the refinement of inversion and editing results from embedded latent codes. While such techniques have shown reasonable improvements in reconstruction, they often lead to a decrease in editing capability, especially when dealing with complex images that contain occlusions, detailed backgrounds, and artifacts. To address this problem, we propose a novel refinement mechanism called Domain-Specific Hybrid Refinement (DHR), which draws on the advantages and disadvantages of two mainstream refinement techniques. We find that the weight modulation can gain favorable editing results but is vulnerable to these complex image areas and feature modulation is efficient at reconstructing. Hence, we divide the image into two domains and process them with these two methods separately. We first propose a Domain-Specific Segmentation module to automatically segment images into in-domain and out-of-domain parts according to their invertibility and editability without additional data annotation, where our hybrid refinement process aims to maintain the editing capability for in-domain areas and improve fidelity for both of them. We achieve this through Hybrid Modulation Refinement, which respectively refines these two domains by weight modulation and feature modulation. Our proposed method is compatible with all latent code embedding methods. Extension experiments demonstrate that our approach achieves state-of-the-art in real image inversion and editing. Code is available at https://***/caopulan/Domain-Specific_Hybrid_Refinement_Inversion.
The proceedings contain 79 papers. The special focus in this conference is on Medical image Computing. The topics include: Keynote: Autonomous Surgery from a Surgeon’s Perspective;image Registration for a Dynamic Bre...
ISBN:
(纸本)9783658474218
The proceedings contain 79 papers. The special focus in this conference is on Medical image Computing. The topics include: Keynote: Autonomous Surgery from a Surgeon’s Perspective;image Registration for a Dynamic Breathing Model;surrogate-based Respiratory Motion Estimation using Physics-enhanced Implicit Neural Representations;comparison of Framewise Video Classification in Laryngoscopies;real-time Fiberscopic image Improvement for Automated Lesion Detection in the Urinary Bladder;robust Statistical Shape Modelling with Implicit Neural Representations;iRBSM: A Deep Implicit 3D Breast Shape Model;Diffusion Models for Conditional Brain Tumor MRI Generation with Tumor-induced Deformations;LLM-driven Baselines for Medical image Segmentation: A Systematic Analysis;efficient Deep Learning-based Forward Solvers for Brain Tumor Growth Models;is Self-supervision Enough?: Benchmarking Foundation Models Against End-to-end Training for Mitotic Figure Classification;look, No Convs! Permutation- and Rotation-invariance for MetaFormers;evaluating the Fidelity of Explanations for Convolutional Neural Networks in Alzheimer’s Disease Detection;real-time Landmark Guidance for Radial Head Localization in Ultrasound Imaging;ultrasound-based 3D reconstruction of Residual Limbs using Electromagnetic Tracking;autocalibration for 3D Ultrasound reconstruction in Infant Hip Dysplasia Screening;Weakly Supervised Segmentation of Hyper-reflective Foci with Compact Convolutional Transformers and SAM 2;Bridging Gaps in Retinal Imaging: Fusing OCT and SLO Information with Implicit Neural Representations for Improved Interpolation and Segmentation;Histologic dataset of Normal and Atypical Mitotic Figures on Human Breast Cancer (AMi-Br);preservation of image Content in Stain-to-stain Translation for Digital Pathology;Automation Bias in AI-assisted Medical Decision-making under Time Pressure in Computational Pathology;Anatomy-aware data Augmentation for Multi-organ Segmentation in CT: AnatoMix.
暂无评论