Using the edge detection techniques we propose a new enhancement scheme for noisy digital images. This uses inhomogeneous anisotropic diffusion scheme via the edge indicator provided by well known edge detection metho...
详细信息
ISBN:
(纸本)9781424442195
Using the edge detection techniques we propose a new enhancement scheme for noisy digital images. This uses inhomogeneous anisotropic diffusion scheme via the edge indicator provided by well known edge detection methods. Addition of a fidelity term facilitates the proposed scheme to remove the noise while preserving edges. This method is general in the sense that it can be incorporated into any of the nonlinear anisotropic diffusion methods. Numerical results show the promise of this hybrid technique on real and noisy images.
Usage of statistical classifiers, namely AdaBoost and its modifications, in object detection and pattern recognition is a contemporary and popular trend The computatiponal performance of these classifiers largely depe...
详细信息
ISBN:
(纸本)9781424442195
Usage of statistical classifiers, namely AdaBoost and its modifications, in object detection and pattern recognition is a contemporary and popular trend The computatiponal performance of these classifiers largely depends on low level image features they are using: both from the point of view of the amount of information the feature provides and the executional time of its evaluation. Local Rank Difference is an image feature that is alternative to commonly used Haar features. It is suitable for implementation in programmable (FPGA) or specialized (ASIC) hardware as well as graphics hardware (GPU). Additionally, as shown in this paper, it performs very well on common CPU's. The paper discusses the LRD features and their properties, describes an experimental implementation of LRD using the multimedia instruction set of current general-purpose processors, presents its empirical performance measures compared to alternative approaches, and suggests several notes on practical usage of LRD and proposes directions for future work.
Radiation therapy is a standard treatment for a large variety of cancers. Radiation treatment planning involves delineating tumor, areas-at-risk, as well as normal anatomical structures. Automated techniques for detec...
详细信息
ISBN:
(纸本)9781467385640
Radiation therapy is a standard treatment for a large variety of cancers. Radiation treatment planning involves delineating tumor, areas-at-risk, as well as normal anatomical structures. Automated techniques for detection and segmentation of tumor is a challenging task because of the absence of any standard model to identify anatomical structures. As radiation therapy is delivered in multiple daily doses (called fractions) treatment fields need to be positioned so that the daily uncertainties in the target position are covered adequately. This is done by adding a margin called the Planning Target Volume (PTV) margin around the clinical target volume. The PTV margin is determined from an understanding of the daily uncertainties. In this paper, we intend to aid doctors in determining the uncertainties of daily pelvis positioning for radiotherapy by estimating the degree of change in the pelvic bone position. We propose a technique for automatically detecting regions of pelvic bone anatomy from CBCT images. We track the change in the bone positions in pelvic regions on a daily basis. A small change suggests that the uncertainty involved in the position of tumorous cell is less, indicating that the planning target margin may be reduced substantially.
One of the major challenges in no-reference (NR) image quality assessment (IQA) is the ability to generalize to diverse quality assessment applications. Recently, multi-modal vision-language models are found to be ver...
详细信息
ISBN:
(纸本)9798400710759
One of the major challenges in no-reference (NR) image quality assessment (IQA) is the ability to generalize to diverse quality assessment applications. Recently, multi-modal vision-language models are found to be very promising in this direction. They are beginning to form a part of several state of the art NR IQA methods. On the other hand, multi-modal large language models (LLMs) are increasingly being studied for various computervision applications including IQA. In this work, we perform a thorough study of the ability of multi-modal LLMs for NR IQA by training some of its components and testing for its generalizability. In particular, we keep the LLM frozen and learn parameters corresponding to the querying transformer, LLM prompt and some layers that process the embedding output by the LLM. We observe that some of these components offer a generalization performance far superior to any existing NR IQA algorithm.
Comprehensive database that contains all possible variations of handwriting is crucial for training and recognition. The primary challenge for an optical character recognizer (OCR) is that a number of interclass chara...
详细信息
ISBN:
(纸本)9781467385640
Comprehensive database that contains all possible variations of handwriting is crucial for training and recognition. The primary challenge for an optical character recognizer (OCR) is that a number of interclass characters bear structural resemblance whereas images within a class render much dissimilarity. Acquisition of such a large database that ensures robust training of the recognizer is a painstaking task. Therefore, recent research interests have been to create, from a few samples of handwriting, a comprehensive synthetic database which not only ensures naturalness, but provides much needed pattern variability. In this paper, we propose a new approach of synthetic handwritten numeral generation for Odia language using interclass deformation. We experimentally evaluate the generated databases using the state-of-the-art recognition systems. The recognition results are compared on two benchmark databases (ISI Kolkata and IIT Bhubaneswar Odia numeral) as well as two newly created synthetic databases. The Odia numeral database sizes are increased by 20-fold each using our proposed approach. The introduction of nonlinear pattern variance because of interclass deformation is proved to pose better challenge to conventional recognizers. We also experimented on a mixture of original and synthetic database for training the OCR to achieve robustness and higher accuracy.
Cognitive load is defined as the mental workload imparted on brain while doing a task. The amount of cognitive load experienced depends on individual's ability of perception, assimilation and response to a task. R...
详细信息
ISBN:
(纸本)9781467385640
Cognitive load is defined as the mental workload imparted on brain while doing a task. The amount of cognitive load experienced depends on individual's ability of perception, assimilation and response to a task. Real-time measurement of the level of cognitive load using low cost Electroencephalogram (EEG) signal enables understanding of personal cognitive skills. In this paper, we propose a methodology of selecting a reference task whose bio-markers closely match with a given task while probing different cognitive abilities. The benefit of this approach is to have a limited set of training models for the reference tasks related to various cognitive categories and use the same for a variety of unknown tasks. Experiment is performed for two levels of cognitive load with three different tasks namely Stroop color task, logical reasoning task and usage of on-screen keyboards. The training models of the reference tasks, selected by cluster analysis of low and high cognitive levels are used to evaluate an unknown task. Experimental results indicate that the Stroop is a better reference for On-Screen keyboard test compared to the Logical reasoning test. Support vector machine (SVM) and principal component analysis (PCA) followed by SVM (PCASVM) are used as the classifiers for the testing.
The different tissues namely gray matter (GM) white matter (WM), and cerebrospinal fluid (CSF) are spread over the entire brain. It is difficult to demarcate them individually when a brain image is considered. The bou...
详细信息
ISBN:
(纸本)9781479915880
The different tissues namely gray matter (GM) white matter (WM), and cerebrospinal fluid (CSF) are spread over the entire brain. It is difficult to demarcate them individually when a brain image is considered. The boundaries are not well defined. Modified fuzzy C means (MFCM) and level sets segmentation based methodology is proposed in this paper for automated brain MRI image segmentation into WM, GM and CSF. The initial segmentation is done by MFCM approach and the results thus obtained are input to the level set methodology. We have tested the methodology on 100 different brain MRI images. The results are compared by using individual MFCM and level set segmentation methods. We took the opinion of 10 expert radiologists to corroborate our results. The results are validated by radiologists as 'Accurate', 'Satisfactory', 'Adequate' and 'Not acceptable'. The results obtained using only level set are 'not acceptable'. Most of the results obtained using MFCM are 'Adequate'. The results obtained using combined method are 'Satisfactory'. Hence, the results obtained using combined MFCM and level sets based segmentation are considered better than using individual MFCM and level set segmentation methods. The manual intervention is avoided in the combined approach. The time required to segment using combined approach is also less compared to level set method. The segmentation using proposed methodology is helpful for radiologists in hospitals for brain MRI image analysis.
This paper presents a fundamentally new algebraic approach to the analysis and synthesis of discrete orthogonal basis functions. It provides the theoretical background to unify Fourier Gabor and discrete orthogonal po...
详细信息
ISBN:
(纸本)9781424442195
This paper presents a fundamentally new algebraic approach to the analysis and synthesis of discrete orthogonal basis functions. It provides the theoretical background to unify Fourier Gabor and discrete orthogonal polynomial moments. For the first time, a set of objective tests are proposed to measure the quality of basis functions. It consists of two main sections: the theoretical background on the generation and orthogonalization of basis functions together with a new solution for the computation of spectra from incomplete data, as well as the implementation of interpolation for all orthogonal basis functions;a new approach to discrete orthogonal polynomials, proving that there is one and only one unitary discrete polynomial basis. Furthermore, the concept of anisotropic moments is introduced and applied to 2D seismic data, which is an imageprocessing problem. The new polynomial basis is numerically better conditioned than the discrete cosine transform. This opens the door to new image compression algorithms, offering a higher compression ratio than the well known JPEG method, for the same numerical effort.
This paper presents a new color document image binarization that is suitable for palm leaf manuscripts and color document images. The proposed method consists of two steps: a pre-processing procedure using low-pass Wi...
详细信息
ISBN:
(纸本)9781424442195
This paper presents a new color document image binarization that is suitable for palm leaf manuscripts and color document images. The proposed method consists of two steps: a pre-processing procedure using low-pass Wiener filter, and contrast adaptive binarization for segmentation of text from the background. Firstly, in the pre-processing stage, low-pass Wiener filter is used to eliminate noisy areas, smoothing of background texture as well as contrast enhancement between background and text areas. Finally, binarization is performed by using contrast adaptive binarization method in order to extract useful text information from low quality document images. The techniques are tested on a set of palm leaf manuscript images/color document images. The performance of the algorithm is demonstrated on by palm leaf manuscripts/color documents distorted with show-through effects, uneven background color and localized spot.
The performance of an Infrared imaging system is often measured by its Detection, Recognition and Identification (DRI) ranges. In thermal cameras, acquired images will be of high dynamic range (typically 15 bits) and ...
详细信息
ISBN:
(纸本)9781467385640
The performance of an Infrared imaging system is often measured by its Detection, Recognition and Identification (DRI) ranges. In thermal cameras, acquired images will be of high dynamic range (typically 15 bits) and lack adequate contrast. An Indium Antimonide (InSb) detector operating in 1-5 micron wavelength has responsivity of the order of 30m V /degrees K resulting in very low differential pixel values even after non-uniformity correction. Mapping the grey levels of high dynamic range thermal data (typically 15 bits) to invariably less dynamic range displays (typically 8 bits), without loss of information, is a challenging task. Hence, image enhancement technique that preserves all the information content and displays with proper contrast is inevitable. Conventional enhancement techniques like AGC, Histogram Equalization and their many variants have the tendency of generating excess noise and assigning more display dynamic range to the dominating temperature data while suppressing the non-dominating temperature data. In this paper, a new enhancement technique of modified Adaptive Histogram Equalization (AHE) is proposed with its real time implementation on FPGA for high dynamic range thermal data. Algorithm aims to retain information content of high dynamic range thermal data, with improved contrast, low noise while compressing the dynamic range to suit the display resolution.
暂无评论