In cardiovascular magnetic resonance (CMR), typical acquisitions often involve a limited number of short and long axis slices. However, reconstructing the 3D chambers is crucial for accurately quantifying heart geomet...
详细信息
ISBN:
(纸本)9783031524479;9783031524486
In cardiovascular magnetic resonance (CMR), typical acquisitions often involve a limited number of short and long axis slices. However, reconstructing the 3D chambers is crucial for accurately quantifying heart geometry and assessing cardiac function. neuralimplicit Representations (NIR) learn implicitfunctions for anatomical shapes from sparse measurements by leveraging a learned continuous shape prior, without the need for high-resolution ground truth data. In this study, we utilized coronary computed tomography (CCTA) images to simulate CMR sparse label maps of two types: standard (10mm spaced short axis and 2 long axis slices) and 3-slice (single short and 2 long axis slices). Whole heart NIR reconstructions were compared to a Label Completion U-Net (LC-U-Net) network trained on the dense segmentations. The findings indicate that the LC-U-Net is not robust when tested with fewer slices than those used during training. In contrast, the NIR consistently achieved Dice scores above 0.9 for the left ventricle, left ventricle myocardium, and right ventricle labels, irrespective of changes in the training or test set. Predictions from standard views achieved average Dice scores across all labels of 0.84 +/- 0.03 and 0.88 +/- 0.03, when training on 3-slice and standard data respectively. In conclusion, this study presents promising results for 3D shape reconstruction invariant to slice position and orientation without requiring full resolution training data, offering a robust and accurate method for cardiac chamber reconstruction in CMR.
We present an approach for the reconstruction of textured 3D meshes of human heads from one or few views. Since such few-shot reconstruction is underconstrained, it requires prior knowledge which is hard to impose on ...
详细信息
We present an approach for the reconstruction of textured 3D meshes of human heads from one or few views. Since such few-shot reconstruction is underconstrained, it requires prior knowledge which is hard to impose on traditional 3D reconstruction algorithms. In this work, we rely on the recently introduced 3D representation- neural implicit functions- which, being based on neural networks, allows to naturally learn priors about human heads from data, and is directly convertible to textured mesh. Namely, we extend NeuS, a state-of-the-art neuralimplicit function formulation, to represent multiple objects of a class (human heads in our case) simultaneously. The underlying neural net architecture is designed to learn the commonalities among these objects and to generalize to unseen ones. Our model is trained on just a hundred smartphone videos and does not require any scanned 3D data. Afterwards, the model can fit novel heads in the few-shot or one-shot modes with good results.
In this paper, we propose a novel point cloud representation method based on neural implicit functions - spatial fields. This method utilizes neural implicit functions to transform three-dimensional coordinate points ...
详细信息
ISBN:
(数字)9789819756001
ISBN:
(纸本)9789819755998;9789819756001
In this paper, we propose a novel point cloud representation method based on neural implicit functions - spatial fields. This method utilizes neural implicit functions to transform three-dimensional coordinate points into local spatial fields, converting the original "discrete-discrete" point cloud representation into a "discrete-continuous" geometric representation, thereby to obtain continuous point cloud representations and richer geometric detail expression. In this method, each three-dimensional coordinate point in the original sparse point set is transformed into a local spatial field embedding multi-layer neighborhood information by means of implicitfunctions. Eventually, multiple such local spatial fields are aggregated into a continuous high-resolution spatial field to approximate the object surface as closely as possible. At last, arbitrary-scale sampling can be conducted in the high-resolution spatial field for point cloud densification needs at arbitrary resolutions in downstream applications such as 3D medical image reconstruction and autonomous driving. This paper provides an example to illustrate how to utilize the results of the proposed solution for 3D model reconstruction.
We propose united implicitfunctions (UNIF), a part-based method for clothed human reconstruction and animation with raw scans and skeletons as the input. Previous part-based methods for human reconstruction rely on g...
详细信息
ISBN:
(数字)9783031200625
ISBN:
(纸本)9783031200618;9783031200625
We propose united implicitfunctions (UNIF), a part-based method for clothed human reconstruction and animation with raw scans and skeletons as the input. Previous part-based methods for human reconstruction rely on ground-truth part labels from SMPL and thus are limited to minimal-clothed humans. In contrast, our method learns to separate parts from body motions instead of part supervision, thus can be extended to clothed humans and other articulated objects. Our Partition-from-Motion is achieved by a bone-centered initialization, a bone limit loss, and a section normal loss that ensure stable part division even when the training poses are limited. We also present a minimal perimeter loss for SDF to suppress extra surfaces and part overlapping. Another core of our method is an adjacent part seaming algorithm that produces non-rigid deformations to maintain the connection between parts which significantly relieves the part-based artifacts. Under this algorithm, we further propose "Competing Parts", a method that defines blending weights by the relative position of a point to bones instead of the absolute position, avoiding the generalization problem of neural implicit functions with inverse LBS (linear blend skinning). We demonstrate the effectiveness of our method by clothed human body reconstruction and animation on the CAPE and the ClothSeq datasets. Our code is available at https://***/ShenhanQian/***.
In this work, we propose a novel image reconstruction framework that directly learns a neuralimplicit representation in k-space for ECG-triggered non-Cartesian Cardiac Magnetic Resonance Imaging (CMR). While existing...
详细信息
ISBN:
(纸本)9783031340475;9783031340482
In this work, we propose a novel image reconstruction framework that directly learns a neuralimplicit representation in k-space for ECG-triggered non-Cartesian Cardiac Magnetic Resonance Imaging (CMR). While existing methods bin acquired data from neighboring time points to reconstruct one phase of the cardiac motion, our framework allows for a continuous, binning-free, and subject-specific k-space representation. We assign a unique coordinate that consists of time, coil index, and frequency domain location to each sampled k-space point. We then learn the subject-specific mapping from these unique coordinates to k-space intensities using a multi-layer perceptron with frequency domain regularization. During inference, we obtain a complete k-space for Cartesian coordinates and an arbitrary temporal resolution. A simple inverse Fourier transform recovers the image, eliminating the need for density compensation and costly non-uniform Fourier transforms for non-Cartesian data. This novel imaging framework was tested on 42 radially sampled datasets from 6 subjects. The proposed method outperforms other techniques qualitatively and quantitatively using data from four and one heartbeat(s) and 30 cardiac phases. Our results for one heartbeat reconstruction of 50 cardiac phases show improved artifact removal and spatio-temporal resolution, leveraging the potential for real-time CMR. (Code available: https://***/wenqihuang/NIK MRI).
Recent research works have focused on generating human models and garments from their 2D images. However, state-of-the-art researches focus either on only a single layer of the garment on a human model or on generatin...
详细信息
ISBN:
(纸本)9783031263187;9783031263194
Recent research works have focused on generating human models and garments from their 2D images. However, state-of-the-art researches focus either on only a single layer of the garment on a human model or on generating multiple garment layers without any guarantee of the intersection-free geometric relationship between them. In reality, people wear multiple layers of garments in their daily life, where an inner layer of garment could be partially covered by an outer one. In this paper, we try to address this multi-layer modeling problem and propose the Layered-Garment Net (LGN) that is capable of generating intersection-free multiple layers of garments defined by implicit function fields over the body surface, given the person's near front-view image. With a special design of garment indication fields (GIF), we can enforce an implicit covering relationship between the signed distance fields (SDF) of different layers to avoid self-intersections among different garment surfaces and the human body. Experiments demonstrate the strength of our proposed LGN framework in generating multi-layer garments as compared to state-of-the-art methods. To the best of our knowledge, LGN is the first research work to generate intersection-free multiple layers of garments on the human body from a single image.
Combining human body models with differentiable rendering has recently enabled animatable avatars of clothed humans from sparse sets of multi-view RGB videos. While state-of-the-art approaches achieve a realistic appe...
详细信息
ISBN:
(数字)9783031198243
ISBN:
(纸本)9783031198236;9783031198243
Combining human body models with differentiable rendering has recently enabled animatable avatars of clothed humans from sparse sets of multi-view RGB videos. While state-of-the-art approaches achieve a realistic appearance with neural radiance fields (NeRF), the inferred geometry often lacks detail due to missing geometric constraints. Further, animating avatars in out-of-distribution poses is not yet possible because the mapping from observation space to canonical space does not generalize faithfully to unseen poses. In this work, we address these shortcomings and propose a model to create animatable clothed human avatars with detailed geometry that generalize well to out-of-distribution poses. To achieve detailed geometry, we combine an articulated implicit surface representation with volume rendering. For generalization, we propose a novel joint root-finding algorithm for simultaneous ray-surface intersection search and correspondence search. Our algorithm enables efficient point sampling and accurate point canonicalization while generalizing well to unseen poses. We demonstrate that our proposed pipeline can generate clothed avatars with high-quality pose-dependent geometry and appearance from a sparse set of multi-view RGB videos. Our method achieves state-of-the-art performance on geometry and appearance reconstruction while creating animatable avatars that generalize well to out-of-distribution poses beyond the small number of training poses.
暂无评论