Human face anthropometric measurements are used in forensics, orthodontics, face modelling and many other domains, wherein distance between set of facial landmarks play an important role to make inferences. 3D facial ...
详细信息
ISBN:
(纸本)9781467385640
Human face anthropometric measurements are used in forensics, orthodontics, face modelling and many other domains, wherein distance between set of facial landmarks play an important role to make inferences. 3D facial data captured using specialized acquisition methods can be used to reduce the time and tedious task involved in order to compute these measurements. the proposed method is developed to compute fifteen canonical linear measurements between facial landmarks using Kinect camera. Results obtained from this system are compared withthe traditional method of measurement using digital Vernier caliper. the experimental results indicate that measurements using RGB-D data obtained from Kinect are good enough for a quick preliminary assessment of the subject as compared to traditional method.
As vision algorithms mature with increasing inspiration from the learning community, statistically independent pseudo random number generation (PRNG) becomes increasingly important. At the same time, execution time de...
详细信息
ISBN:
(纸本)9781424442195
As vision algorithms mature with increasing inspiration from the learning community, statistically independent pseudo random number generation (PRNG) becomes increasingly important. At the same time, execution time demands have seen algorithms being implemented on evolving parallel hardware such as GPUs. the Mersenne Twister (MT) [7] has proven to be the current state of the art for generating high quality random numbers, and the Nvidia provided software for parallel MT is in widespread use. While execution time is important, development time is also critical. As processor cardinality changes, a foundation for generating simulations that will vary only in execution time and not in the actual result is useful;otherwise the development time will be impacted. In this paper we present an implementation of the Lagged Fibonacci Generator (LFG) considered to be of quality equal [7] to MT on the GPU. Unlike MT LFG has this important processor-cardinality agnostic capability that is - as the number of processing resources changes, the overall sequence of random numbers remains the same. this feature notwithstanding, our basic implementation is roughly as fast as the parallel MT;an in-memory version is actually 25% faster in execution time. Both parallel MT as well as parallel LFG show enormous speedup over their sequential counterparts. Finally, a prototype particle filter tracking application shows that our method works not just in parallel computing theory, but also in practice for vision applications, providing a decrease of 60% in execution time.
In this paper, an embedded entropy based image registration scheme has been proposed. Here, Tsallis and Renyi's entropy have been embedded to form a new entropic measure. this parametrized entropy has been used to...
详细信息
ISBN:
(纸本)9781479915880
In this paper, an embedded entropy based image registration scheme has been proposed. Here, Tsallis and Renyi's entropy have been embedded to form a new entropic measure. this parametrized entropy has been used to determine the weighted mutual information (MI) for the CT and MR brain images. the embedded mutual information has been maximized to obtain registration. this notion of embedded mutual information has also been validated in feature space registration. the mutual information with respect to the registration parameter has been found to be a nonlinear curve. It has been found that the feature space registration resulted in higher value mutual information and hence registration process could be smoother. We have used Simulated Annealing algorithm to determine the maximum of this embedded mutual information and hence register the images.
the recent era of digitization is expected to digitized many old important documents which are degraded due to various reasons. Degraded document image binarization has many challenges like intensity variation, backgr...
详细信息
ISBN:
(纸本)9781467385640
the recent era of digitization is expected to digitized many old important documents which are degraded due to various reasons. Degraded document image binarization has many challenges like intensity variation, background contrast variation, bleed through, text size variation and so on. Many approaches are available for document image binarization, but none can handle all types of degradation at once. We proposed an approach which consists of three stages such as preprocessing, Text-Area detection and post-processing. Preprocessing enhances the contrast of the image. Next stage involves identifying Text-Area. Postprocessing technique takes care of false positives and false negative based on intensity values of preprocessed and gray image. the Performance is evaluated based on various quantitative measures and is compared withthe method regarded best so far. the algorithm is also expected to be independent of the script, hence is tested on Gujarati degraded document images.
In this paper, a multi-view stereo image watermarking scheme is proposed to resist the RST (rotation, scaling and translation) attack. To make the scheme resilient to RST, the coefficients of Singular Value Decomposit...
详细信息
ISBN:
(纸本)9781467385640
In this paper, a multi-view stereo image watermarking scheme is proposed to resist the RST (rotation, scaling and translation) attack. To make the scheme resilient to RST, the coefficients of Singular Value Decomposition (SVD) from both left and right views have been used for insertion of the watermark bits. 2D-DWT (Discrete wavelet transform) is used as a preprocessing step to get more correlated SVD coefficients of the left and right view such that the visual degradation due to embedding can be reduced. In this work, a blind embedding scheme is proposed by altering the selected SVD coefficients to improve the robustness of the embedding scheme. A comprehensive set of experiments have been performed to justify the robustness of the proposed scheme against RST attack. Moreover, this scheme can be used to detect the view swapping attack using DIBR technique.
Constructing a high-resolution (HR) image from low-resolution (LR) image(s) has been a very active research topic recently with focus shifting from multi-frames to learning based single-frame super-resolution (SR). Mu...
详细信息
ISBN:
(纸本)9781424442195
Constructing a high-resolution (HR) image from low-resolution (LR) image(s) has been a very active research topic recently with focus shifting from multi-frames to learning based single-frame super-resolution (SR). Multi-frame SR algorithms attempt the exact reconstruction of reality, but are limited to small magnification factors. Learning based SR algorithms learn the correspondences between. LR and HR patches. Accurate replacements or revealing the exact underlying information is not guaranteed in many scenarios. In this paper we propose an alternate solution. We propose to capture images at right zoom such that it has just sufficient amount of information so that further resolution enhancements can be easily achieved using an v off the shelf single-frame SR algorithm. this is true under the assumption that such a zoom factor is not very high, which is true for most man-made structures. the low-resolution. image is divided into small patches and ideal resolution is predicted for every patch. the contextual information is incorporated using a Markov Random Field based prior. Training data is generated from high-quality images and can use any single-frame SR algorithm. Several constraints are proposed to minimize the extent of zoom-in. We validate the proposed approach on synthetic data and real world images to show the robustness.
In this paper, an uncompressed domain video watermarking scheme resilient to temporal adaptation is proposed for scalable video coding. In the proposed scheme, each temporal layer has been separately embedded with a d...
详细信息
ISBN:
(纸本)9781467385640
In this paper, an uncompressed domain video watermarking scheme resilient to temporal adaptation is proposed for scalable video coding. In the proposed scheme, each temporal layer has been separately embedded with a different watermark which is generated by DCT domain decomposition of a single watermark image. A zigzag sequence of block wise DCT coefficients of the watermark image is partitioned into non-overlapping sets and each set is embedded separately into different temporal layers. the base layer is embedded withthe first set of DCT coefficient (which includes DC coefficient of each block) and successive layers are embedded with successive nonoverlapping coefficient sets. the coefficients of each set is chosen in such a fashion that uniform energy distribution across all temporal layers can be maintained. Experimental results show that the proposed scheme is robust against temporal scalability and robustness of the watermark increases withthe addition of successive enhancement layers.
this paper addresses the problem of reconstruction of specular surfaces using a combination of Dynamic Programming and Markov Random Fields formulation. Unlike traditional methods that require the exact position of en...
详细信息
ISBN:
(纸本)9781479915880
this paper addresses the problem of reconstruction of specular surfaces using a combination of Dynamic Programming and Markov Random Fields formulation. Unlike traditional methods that require the exact position of environment points to be known, our method requires only the relative position of the environment points to be known for computing approximate normals and infer shape from them. We present an approach which estimates the depth from dynamic programming routine and MRF stereo matching and use MRF optimization to fuse the results to get the robust estimate of shape. We used smooth color gradient image as our environment texture so that shape can be recovered using just a single shot. We evaluate our method using synthetic experiments on 3D models like Stanford bunny and show the real experiment results on golden statue and silver coated statue.
In this paper, a fractional order total variation (TV) model is presented for estimating the optical flow in the image sequences. the proposed fractional order model is introduced by generalizing a variational flow mo...
详细信息
ISBN:
(纸本)9781467385640
In this paper, a fractional order total variation (TV) model is presented for estimating the optical flow in the image sequences. the proposed fractional order model is introduced by generalizing a variational flow model formed with a quadratic and a total variation terms. However, it is difficult to solve this generalized model due to the non-differentiability of the total variation regularization term. the Grunwald-Letnikov derivative is used to discretize the fractional order derivative. the resulting formulation is solved by using an efficient numerical algorithm. the experimental results verify that the proposed model yields a dense flow and preserves discontinuities in the flow field. Moreover, It also provides a significant robustness against outliers.
Recent years have witnessed an exponential growth in the use of digital images due to development of high quality digital cameras and multimedia technology. Easy availability of image editing software has made digital...
详细信息
ISBN:
(纸本)9781479915880
Recent years have witnessed an exponential growth in the use of digital images due to development of high quality digital cameras and multimedia technology. Easy availability of image editing software has made digital imageprocessing very popular. Ready to use software are available on internet which can be easily used to manipulate the images. In such an environment, the integrity of the image can not be taken for granted. Malicious tampering has serious implication for legal documents, copyright issues and forensic cases. Researchers have come forward with large number of methods to detect image tampering. the proposed method is based on hash generation technique using singular value decomposition. Design of an efficient hash vector as proposed will help in detection and localization of image tampering. the proposed method shows that it is robust against content preserving manipulation but extremely sensitive to even very minute structural tampering.
暂无评论