We introduce FabricDiffusion, a method for transferring fabric textures from a single clothing image to 3D garments of arbitrary shapes. Existing approaches typically synthesize textures on the garment surface through...
详细信息
ISBN:
(纸本)9798400711312
We introduce FabricDiffusion, a method for transferring fabric textures from a single clothing image to 3D garments of arbitrary shapes. Existing approaches typically synthesize textures on the garment surface through 2D-to-3D texture mapping or depth-aware inpainting via generative models. Unfortunately, these methods often struggle to capture and preserve texture details, particularly due to challenging occlusions, distortions, or poses in the input image. Inspired by the observation that in the fashion industry, most garments are constructed by stitching sewing patterns with flat, repeatable textures, we cast the task of clothing texture transfer as extracting distortion-free, tileable texture materials that are subsequently mapped onto the UV space of the garment. Building upon this insight, we train a denoising diffusion model with a large-scale synthetic dataset to rectify distortions in the input texture image. This process yields a flat texture map that enables a tight coupling with existing Physically-Based Rendering (PBR) material generation pipelines, allowing for realistic relighting of the garment under various lighting conditions. We show that FabricDiffusion can transfer various features from a single clothing image including texture patterns, material properties, and detailed prints and logos. Extensive experiments demonstrate that our model significantly outperforms state-to-the-art methods on both synthetic data and real-world, in-the-wild clothing images while generalizing to unseen textures and garment shapes.
Featured Application This research presents a novel imagereconstruction method for Boron Neutron Capture Therapy (BNCT)-Single Photon Emission Computed Tomography (SPECT) using Bayesian estimation with limited view-a...
详细信息
Featured Application This research presents a novel imagereconstruction method for Boron Neutron Capture Therapy (BNCT)-Single Photon Emission Computed Tomography (SPECT) using Bayesian estimation with limited view-angle projection data. The method aims to address the inherent challenges in BNCT-SPECT imaging where conventional algorithms struggle with incompletedatafrom restricted projection angles. By improving the image accuracy under such conditions, this technique has potential applications in clinical settings, where accurate tumor localization and dose distribution are critical for the success of BNCT treatments. This method could lead to enhanced treatment planning and monitoring, ultimately improving patient *** Boron Neutron Capture Therapy (BNCT) is an emerging radiation treatment for cancer, and its challenges are being explored. Systems capable of capturing real-time observations of this treatment's effectiveness, particularly BNCT-SPECT methods that measure gamma rays emitted instantaneously from outside the body during nuclear reactions and that reconstruct images using Single Photon Emission Computed Tomography (SPECT) techniques, remain unavailable. BNCT-SPECT development is hindered by two main factors, the first being the projection angle. Unlike conventional SPECT, the projection angle range which is achievable by rotating a detector array cannot exceed approximately 90 degrees. Consequently, Fourier-based imagereconstruction methods, requiring projections from at least 180 degrees, do not apply to BNCT-SPECT. The second limitation is the measurement time. Given these challenges, we developed a new sequential approximation imagereconstruction method using Bayesian estimation, which is effective under the stringent BNCT-SPECT conditions. We also compared the proposed method with the existing Maximum Likelihood-Expectation Maximization (ML-EM) imagereconstruction method. Numerical experiments were conducted by obtaining BNCT-SPE
With the escalating concern about global warming, the environmental impact of electronic devices must be scrutinized. Life Cycle Assessments (LCA) reveal that Integrated Circuits (ICs) are the primary contributors to ...
详细信息
ISBN:
(纸本)9798350380392;9798350380385
With the escalating concern about global warming, the environmental impact of electronic devices must be scrutinized. Life Cycle Assessments (LCA) reveal that Integrated Circuits (ICs) are the primary contributors to greenhouse gas emissions in these devices. However, performing an inventory to determine the ICs impact is a complex task due to missing data and the existing studies on ICs have been neglecting CMOS image Sensors (CIS). Despite the surge in CIS usage, particularly in smartphones, there is a lack of comprehensive models to assess their environmental impact. This paper proposes a multi-level set of models that leverage available information while considering the specificities of CIS. The most comprehensive model incorporates factors such as the total silicon area, geographical location (influencing the energy mix), and the technology node. To accommodate scenarios with incompletedata, subsequent models are designed to effectively utilize averaged parameters. The proposed models are applied to sensors manufactured by STMicroelectronics and Sony, and the results are compared with existing LCA results from Fairphone. Our approach provides a more comprehensive understanding of the environmental impact of CIS, contributing to the broader goal of reducing the carbon footprint of electronic devices. Our results suggest that the carbon impact of a Fairphone 4 image sensor is likely higher than previously estimated, with a significant gap between our findings and the expected value.
作者:
Chang, XuGao, DonglaiHarbin Inst Technol
Minist Ind & Informat Technol Key Lab Smart Prevent & Mitigat Civil Engn Disaste Harbin 150090 Peoples R China Harbin Inst Technol
Minist Educ Key Lab Struct Dynam Behav & Control Harbin 150090 Peoples R China
The present study on the recognition of coherent structures in flow fields was conducted using three typical data-driven modal decomposition methods: proper orthogonal decomposition (POD), dynamic mode decomposition (...
详细信息
The present study on the recognition of coherent structures in flow fields was conducted using three typical data-driven modal decomposition methods: proper orthogonal decomposition (POD), dynamic mode decomposition (DMD), and Fourier mode decomposition (FMD). Two real circular cylinder wake flows (forced and unforced), obtained from two-dimensional particle image velocimetry (2D PIV) measurements, were analyzed to extract the coherent structures. It was found that the POD method could be used to extract the large-scale structures from the fluctuating velocity in a wake flow, the DMD method showed potential for dynamical mode frequency identification and linear reconstruction of the flow field, and the FMD method provided a significant computational efficiency advantage when the dominant frequency of the flow field was known. The limitations of the three methods were also identified: The POD method was incomplete in the spatial-temporal decomposition and each mode mixed multiple frequencies leading to unclear physics, the DMD method is based on the linear assumption and thus the highly nonlinear part of the flow field was unsuitable, and the FMD method is based on global power spectrum analysis while being overwhelmed by an unknown high-frequency flow field.
Hyperspectral image unmixing estimates a collection of constituent materials (called endmembers) and their corresponding proportions (called abundances), which is a critical preprocessing step in many remote sensing a...
详细信息
Few-shot image classification aims to provide accurate predictions for novelty by learning from a limited number of samples. Classical few-shot image classification methods usually use data augmentation and self-super...
详细信息
ISBN:
(纸本)9798350300673
Few-shot image classification aims to provide accurate predictions for novelty by learning from a limited number of samples. Classical few-shot image classification methods usually use data augmentation and self-supervision to compensate for the lack of training sample, and introduce migration learning and meta-learning to pre-train the model or accelerate the model optimization, which improves the classification performance of the model. However, with a small amount of labeled sample data, these methods cannot meet the requirements of the model's ability to characterize sample features, resulting in a model that is highly susceptible to overfitting problems. In this paper, we propose a Dual Feature reconstruction Network (DFRN) for few-shot image classification. The network constructs the double feature vector by two modules, in which the first-level feature module generates an attention mask based on the image to make the feature vector characterize more of the target region, and the secondary feature module interferes with the feature vector to improve its generalization performance. In addition, the network also enhances the classification performance of the model by considering the contextual information of the support classes through an auxiliary loss function. Through extensive experiments, the network proposed in this paper achieves excellent performance on Flowers, CUB and Cars datasets and outperforms other reference fine-grained image classification methods such as FRN.
Non-blind deconvolution aims to restore a sharp imagefrom its blurred counterpart given an obtained kernel. Existing deep neural architectures are often built based on large datasets of sharp ground truth images and ...
详细信息
ISBN:
(纸本)9798350307443
Non-blind deconvolution aims to restore a sharp imagefrom its blurred counterpart given an obtained kernel. Existing deep neural architectures are often built based on large datasets of sharp ground truth images and trained with supervision. Sharp, high quality ground truth images, however, are not always available, especially for biomedical applications. This severely hampers the applicability of current approaches in practice. In this paper, we propose a novel non-blind deconvolution method that leverages the power of deep learning and classic iterative deconvolution algorithms. Our approach combines a pre-trained network to extract deep features from the input image with iterative Richardson-Lucy deconvolution steps. Subsequently, a zero-shot optimisation process is employed to integrate the deconvolved features, resulting in a high-quality reconstructed image. By performing the preliminary reconstruction with the classic iterative deconvolution method, we can effectively utilise a smaller network to produce the final image, thus accelerating the reconstruction whilst reducing the demand for valuable computational resources. Our method demonstrates significant improvements in various real-world applications non-blind deconvolution tasks.
In this paper we study the applications of deep-learning to the problem of imagereconstruction in Compton scatter tomography, a field where deep-learning techniques are still unexplored. Particularly, we focus on a n...
详细信息
Although existing fMRI-to-imagereconstruction methods could predict high-quality images, they do not explicitly consider the semantic gap between training and testing data, resulting in reconstruction with unstable a...
详细信息
Deep learning-based algorithms for single MR image (MRI) super-resolution have shown great potential in enhancing the resolution of low-quality images. However, many of these methods rely on supervised training with p...
详细信息
ISBN:
(纸本)9783031439063;9783031439070
Deep learning-based algorithms for single MR image (MRI) super-resolution have shown great potential in enhancing the resolution of low-quality images. However, many of these methods rely on supervised training with paired low-resolution (LR) and high-resolution (HR) MR images, which can be difficult to obtain in clinical settings. This is because acquiring HR MR images in clinical settings requires a significant amount of time. In contrast, HR CT images are acquired in clinical routine. In this paper, we propose a CT-guided, unsupervised MRI super-resolution reconstruction method based on joint cross-modality image translation and super-resolution reconstruction, eliminating the requirement of high-resolution MRI for training. The proposed approach is validated on two datasets respectively acquired from two different clinical sites. Well-established metrics including Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index Metrics (SSIM), and Learned Perceptual image Patch Similarity (LPIPS) are used to assess the performance of the proposed method. Our method achieved an average PSNR of 32.23, an average SSIM of 0.90 and an average LPIPS of 0.14 when evaluated on data of the first site. An average PSNR of 30.58, an average SSIM of 0.88, and an average LPIPS of 0.10 were achieved by our method when evaluated on data of the second site.
暂无评论