In this paper, a new approach is proposed for the detection of JPEG anti-forensic operations. It is based on the fact that when a JPEG anti-forensic operation is applied, the values of DCT coefficients are changed. th...
详细信息
Recognition of unconstrained handwritten texts is always a difficult problem, particularly if the style of handwriting is a mixed cursive one. Among various indian scripts, only Bangla has this additional difficulty o...
详细信息
ISBN:
(纸本)9781450347532
Recognition of unconstrained handwritten texts is always a difficult problem, particularly if the style of handwriting is a mixed cursive one. Among various indian scripts, only Bangla has this additional difficulty of tackling mixed cursiveness of its handwriting style in the pipeline of a method towards its automatic recognition. A few other common recognition difficulties of handwriting in an indian script include the large size of its alphabet and the extremely cursive nature of the shapes of its alphabetic characters. these are among the reasons of achieving only limited success in the study of unconstrained handwritten Bangla text recognition. Artificial Neural Network (ANN) models have often been used for solving difficult real-life pattern recognition problems. Recurrent Neural Network models (RNN) have been studied in the literature for modeling sequence data. In this study, we consider Long Short Term Memory (LSTM) network model, a useful member of this family. In fact, Bidirectional Long Short-Term Memory (BLSTM) neural networks is a special kind of RNN and have recently attracted special attention in solving sequence labelling problems. In this article, we present a BLSTM architecture based approach for unconstrained online handwritten Bangla text recognition.
Malaria is a deadly infectious disease affecting red blood cells in humans due to the protozoan of type Plasmodium. In 2015, there is an estimated death toll of 438, 000 patients out of the total 214 million malaria c...
详细信息
ISBN:
(纸本)9781450347532
Malaria is a deadly infectious disease affecting red blood cells in humans due to the protozoan of type Plasmodium. In 2015, there is an estimated death toll of 438, 000 patients out of the total 214 million malaria cases reported worldwide. thus, building an accurate automatic system for detecting the malarial cases is beneficial and has huge medical value. this paper addresses the detection of Plasmodium Falciparum infected RBCs from Leishman's stained microscope slide images. Unlike the traditional way of examining a single focused image to detect the parasite, we make use of a focus stack of images collected using a bright field microscope. Rather than the conventional way of extracting the specific features we opt for using Convolutional Neural Network that can directly operate on images bypassing the need for hand-engineered features. We work withimage patches at the suspected parasite location there by avoiding the need for cell segmentation. We experiment, report and compare the detection rate received when only a single focused image is used and when operated on the focus stack of images. Altogether the proposed novel approach results in highly accurate malaria detection.
Modern 3D imaging technologies often generate large scale volume datasets that may be represented as 3-way tensors. these volume datasets are usually compressed for compact storage, and interactive visual analysis of ...
详细信息
ISBN:
(纸本)9781450347532
Modern 3D imaging technologies often generate large scale volume datasets that may be represented as 3-way tensors. these volume datasets are usually compressed for compact storage, and interactive visual analysis of the data warrants efficient decompression techniques at real time. Using well known tensor decomposition techniques like CP or Tucker decomposition the volume data can be represented by a few basis vectors, the number of such vectors, called the rank of the tensor, determining the visual quality. However, in such methods, the basis vectors used between successive ranks are completely different, thereby requiring a complete recomputation of basis vectors whenever the visual quality needs to be altered. In this work, a new progressive decomposition technique is introduced for scalar volumes wherein new basis vectors are added to the already existing lower rank basis vectors. Large scale datasets are usually divided into bricks of smaller size and each such brick is represented in a compressed form. the bases used for the different bricks are data dependent and are completely different from one another. the decomposition method introduced here uses the same basis vectors for all the bricks at all hierarchical levels of detail. the basis vectors are data independent thereby minimizing storage and allowing fast data reconstruction.
Automated segmentation of brain structure in magnetic resonance imaging (MRI) scans is an important first step in diagnosis of many neurological diseases. In this paper, we focus on segmentation of the constituent sub...
详细信息
ISBN:
(纸本)9781450347532
Automated segmentation of brain structure in magnetic resonance imaging (MRI) scans is an important first step in diagnosis of many neurological diseases. In this paper, we focus on segmentation of the constituent sub-structures of basal ganglia (BG) region of the brain that are responsible for controlling movement and routine learning. Low contrast voxels and undefined boundaries across sub-regions of BG pose a challenge for automated segmentation. We pose the segmentation as a voxel classification problem and propose a Deep Neural Network (DNN) based classifier for BG segmentation. the DNN is able to learn distinct regional features for voxel-wise classification of BG area into four sub-regions, namely, Caudate, Putamen, Pallidum, and Accumbens. We use a public dataset with a collection of 83 T-1 weighted uniform dimension structural MRI scans of healthy and diseased (Bipolar with and without Psychosis, Schizophrenia) subjects. In order to build a robust classifier, the proposed classifier has been trained on a mixed collection of healthy and diseased MRs. We report an accuracy of above 94% (as calculated using the dice coefficient) for all the four classes of healthy and diseased dataset.
Haze during the bad weather, degrades the visibility of the scene drastically. Degradation of scene visibility varies with respect to the transmission coefficient/map (T-c) of the scene. Estimation of accurate T-c is ...
详细信息
ISBN:
(纸本)9781450366151
Haze during the bad weather, degrades the visibility of the scene drastically. Degradation of scene visibility varies with respect to the transmission coefficient/map (T-c) of the scene. Estimation of accurate T-c is key step to reconstruct the haze free scene. Previously, local as well as global priors were proposed to estimate the T-c. We, on the other hand, propose integration of local and global approaches to learn both point level and object level T-c. the proposed local encoder decoder network (LEDNet) estimates the scene transmission map in two stages. During first stage, network estimates the point level T-c using parallel convolutional filters and spatial invariance filtering. the second stage comprises of a two level encoder-decoder architecture which anticipates the object level T-c. We also propose, local air-light estimation (LAE) algorithm, which is able to obtain the air-light component of the outdoor scene. Combination of LEDNet and LAE improves the accuracy of haze model to recover the scene radiance. Structural similarity index, mean square error and peak signal to noise ratio are used to evaluate the performance of the proposed approach for single image haze removal. Experiments on benchmark datasets show that LEDNet outperforms the existing state-of-the-art methods for single image haze removal.
the contour tree represents the topology of level sets of a scalar function. Nodes of the tree correspond to critical level sets and arcs of the tree represent a collection of topologically equivalent level sets conne...
详细信息
ISBN:
(纸本)9781450366151
the contour tree represents the topology of level sets of a scalar function. Nodes of the tree correspond to critical level sets and arcs of the tree represent a collection of topologically equivalent level sets connecting two critical level sets. the augmented contour tree contains degree-2 nodes on the arcs that represent regular level sets. the degree-2 nodes correspond to regular points of the scalar function and other critical points that do not affect the number of level set components. the augmented contour tree is significantly larger in size and requires more effort to compute when compared to the contour tree. Applications of the contour tree to data exploration and visualization require the augmented contour tree. Current approaches propose algorithms to compute the contour tree and the augmented contour tree from scratch. Precomputing and storing the large augmented contour tree will not be necessary if the contour tree can be augmented on-demand. this paper poses the problem of computing the augmented contour tree given a contour tree as input. Computational experiments demonstrate that the on-demand augmentation can be computed fast while resulting in good memory savings.
Recent trends in image segmentation algorithms have shown various large scale networks with impressive performance for natural scene images. However most of the networks come with costly overheads such has large memor...
详细信息
ISBN:
(纸本)9781450366151
Recent trends in image segmentation algorithms have shown various large scale networks with impressive performance for natural scene images. However most of the networks come with costly overheads such has large memory requirements or dependence on huge number of parallel processing units. In most cases costly graphicsprocessing units or GPUs are used to boost computational capability. However for creating products in the real world we need to consider speed, performance as well as cost of deployment. We propose a novel "spark" module which is a combination of the "fire" module of SqueezeNet and depth-wise separable convolutions. Along withthis modified SqueezeNet as an encoder we also propose the use of depth-wise separable transposed convolution for a decoder. the resultant encoder-decoder network has approximately 49 times lesser number of parameters than SegNet and almost 223 times lesser number of parameters than fully convolutional networks(FCN). Even in a CPU the network completes a forward pass for a single sample in approximately 0.39 seconds which is almost 5.1 times faster as compared to SegNet and almost 8.7 times faster compared to FCN.
Curb detection is a critical component of driver assistance and autonomous driving systems. In this paper, we present a discriminative approach to the problem of curb detection under diverse road conditions. We define...
详细信息
ISBN:
(纸本)9781450347532
Curb detection is a critical component of driver assistance and autonomous driving systems. In this paper, we present a discriminative approach to the problem of curb detection under diverse road conditions. We define curbs as the intersection of drivable and non-drivable area which are classified using dense Conditional random fields(CRF). In our method, we fuse output of a neural network used for pixelwise semantic segmentation with depth and color information from stereo cameras. CRF fuses the output of a deep model and height information available in stereo data and provides improved segmentation. Further we introduce temporal smoothness using a weighted average of Segnet output and output from a probabilistic voxel grid as our unary potential. Finally, we show improvements over the current state of the art neural networks. Our proposed method shows accurate results over large range of variations in curb curvature and appearance, without the need of retraining the model for the specific dataset.
the two-volume set LNCS 1842/1843 constitutes the refereed proceedings of the 6th European conference on computervision, ECCV 2000, held in Dublin, Ireland in June/July 2000. the 116 revised full papers presented wer...
详细信息
ISBN:
(数字)9783540450535
ISBN:
(纸本)9783540676867
the two-volume set LNCS 1842/1843 constitutes the refereed proceedings of the 6th European conference on computervision, ECCV 2000, held in Dublin, Ireland in June/July 2000. the 116 revised full papers presented were carefully selected from a total of 266 submissions. the two volumes offer topical sections on recognitions and modelling; stereoscopic vision; texture and shading; shape; structure from motion; image features; active, real-time, and robot vision; segmentation and grouping; vision systems engineering and evaluation; calibration; medical image understanding; and visual motion.
暂无评论