Stereo computation is one of the vision problems where the presence of outliers cannot be neglected. Most standard algorithms make unrealistic assumptions about noise distributions, which leads to erroneous results th...
详细信息
Stereo computation is one of the vision problems where the presence of outliers cannot be neglected. Most standard algorithms make unrealistic assumptions about noise distributions, which leads to erroneous results that cannot be corrected in subsequent postprocessing stages. In this paper we present a modification of the standard area-based correlation approach so that it can tolerate a significant number of outliers. The approach exhibits a robust behavior not only in the presence of mismatches but also in the case of depth discontinuities. The confidence measure of the correlation and the number of outliers provide two complementary sources of information which, when implemented in a multiresolution framework, result in a robust and efficient method. We present the results of this approach on a number of synthetic and real images.
The basic limitations of the current appearance-based matching methods using eigenimages are non-robust estimation of coefficients and inability to cope with problems related to occlusions and segmentation. In this pa...
详细信息
The basic limitations of the current appearance-based matching methods using eigenimages are non-robust estimation of coefficients and inability to cope with problems related to occlusions and segmentation. In this paper we present a new approach which successfully solves these problems. The major novelty of our approach lies in the way how the coefficients of the eigenimages are determined. Instead of computing the coefficients by a projection of the data onto the eigenimages. we extract them by a hvpothesize-and-test paradigm using subsets of image points. Competing hypotheses arc then subject to a selection procedure based on the Minimum Description Length principle. The approach enables us not only lo reject outliers and to deal with occlusions but also to simultaneously use multiple classes of eigenimages.
In ATM networks cell loss causes data to be dropped in the channel. When digital video is transmitted over these networks one must be able to reconstruct the missing data so that the impact of these errors is minimize...
详细信息
In ATM networks cell loss causes data to be dropped in the channel. When digital video is transmitted over these networks one must be able to reconstruct the missing data so that the impact of these errors is minimized. In this paper we describe a Bayesian approach to conceal these errors. Assuming that the digital video has been encoded using the MPEG1 or MPEG2 compression scheme, each frame is modeled as a Markov random field. A maximum a posteriori estimate of the missing macroblocks and motion vectors is described based on the model.
This paper proposes a method of 3-dimensional measurement of a planar surface by using two fixed light sources and a TV camera. A set of two images is recorded by the camera switching on each light alternately. The pe...
详细信息
This paper proposes a method of 3-dimensional measurement of a planar surface by using two fixed light sources and a TV camera. A set of two images is recorded by the camera switching on each light alternately. The peak of the luminance distribution in each image is detected, and the orientation and distance of the planar surface are calculated. The position of the peak of a luminance distribution can be estimated accurately by using an imageprocessing. The light sources can be conventional apparatus with no particular structure. The method is simple and suitable for a vision system on an indoor mobile robot.
作者:
Oda, MAkamatsu, SFukamachi, HATR Human Information Processing Laboratories
Kyoto Japan 619-02 Shigeru Akamatsu received his B.E.
M.E. and Dr. of Eng. degrees in Mathematical Enginwring and Instrumentation Physics in 1975 1977 and 1994 respectively from the University of Tokyo Japan. In 1977 he joined the Electrical Communications Laboratories Nippon Telegraph and Telephone Public Corporation Yokosuka Japan. From 1977 through 1985 he was engaged in the research and development of optical character recognition systems for handwritten Chmese characters. During the academic year 1985-86 he was a Visiting Researcher at the University of California Irvine California U.S.A. Until 1992 he was a Senior Research Engineer Supervisor at W Human Interface Laboratories and conducted research on human image recognition with a special interest in face recognition. Since 1992 he has been with ATR Human Information Processing Research Laboratories Kyoto Japan where he is currently Head of Department 2. His research interests include computational and cognitive studies on high-level vision with a special interest in facial information processing by man and computer. He is a member of the Institute of Electronics Information and Communication Engineers of Japan the Information Processing Society of JapanACMand the IEEE Computer Society.Hidm Fukamachi received his B.E. degree in Image Technology from the Tokyo Institute of Polytechnics in 1990. He joined NTT Software Corp. in 1990
where he was engaged in the development of image processing software. Since 1993 he has been with ATR Human Information Processing Research Laboratories.
Some facial images are not necessarily clear images in the human brain. However, it can be easily judged whether a face matches the image in our mind;this is true even when the drawing or the expression of the target ...
详细信息
Some facial images are not necessarily clear images in the human brain. However, it can be easily judged whether a face matches the image in our mind;this is true even when the drawing or the expression of the target image is difficult to comprehend. This is because even if the image cannot be imagined like a picture, it does exist. The authors have proposed a system with which a retriever can retrieve an ambiguous target image from an image database. The system's retrieval efficiency and ease-of-retrieval were clarified by experiments using line drawn facial images. In this paper, the system is extended to photographs of faces. The most difficult problem in treating photographs is determining the retrieval key. One consideration is to add the subjective impressions or features of physical figures as keywords for each image data. Regardless of the method, however, significant manpower is necessary. The K-L expansion technique is used here, which has been verified as effective for face identification, and expect to need only a small amount of manpower. This paper investigates whether the technique can also be useful in ambiguous image retrieval.
This paper presents a texture segmentation algorithm based on a hierarchical wavelet decomposition. Using Daubechies' four-tap filter, an original image is decomposed into three detail images and one approximate i...
详细信息
This paper presents a texture segmentation algorithm based on a hierarchical wavelet decomposition. Using Daubechies' four-tap filter, an original image is decomposed into three detail images and one approximate image. The decomposition can be recursively applied to the approximate image to generate a lower resolution of the pyramid. The segmentation starts at the lowest resolution using the K-means clustering scheme and textural features obtained from various sub-bands. The result of segmentation is propagated through the pyramid to a higher resolution with continuously improving the segmentation. The lower resolution levels help to build the contour of the segmented texture, while higher levels refine the process, and correct possible errors.
We describe a multiresolution approach to edge detection using a sequential search algorithm. The use of a multiresolution image pyramid allows the integration of global edge information contained in lower resolutions...
详细信息
We describe a multiresolution approach to edge detection using a sequential search algorithm. The use of a multiresolution image pyramid allows the integration of global edge information contained in lower resolutions to guide the sequential search at higher resolutions. As a consequence, the dependence on a priori knowledge of the image edges is greatly reduced. Estimating the sequential search parameters from lower resolution images provides for a more accurate and less costly search of edge paths in the image.
Video parsing is a fundamental operation used in many digital video applications such as digital libraries and video servers. The accuracy and execution speed of the parsing algorithm is critical if large amounts of v...
详细信息
Video parsing is a fundamental operation used in many digital video applications such as digital libraries and video servers. The accuracy and execution speed of the parsing algorithm is critical if large amounts of video data are to be processed, particularly in real-time. We present a new algorithm to reconstruct DC coefficient images of a DCT and motion compensation compressed video sequence, e.g. MPEG. The histograms of the DC coefficient images can be used to detect scene changes.
In this paper we describe two error-recovery approaches for MPEG encoded video over ATM networks. The first approach aims at reconstructing each lost pixel by spatial interpolation from the nearest undamaged pixels. T...
详细信息
In this paper we describe two error-recovery approaches for MPEG encoded video over ATM networks. The first approach aims at reconstructing each lost pixel by spatial interpolation from the nearest undamaged pixels. The second approach recovers lost macroblocks by minimizing intersample variations within each block and across its boundaries. Moreover, a new technique for packing ATM cells with compressed data is also proposed.
The problem inherent with any digital image (or digital video) system is the large amount of bandwidth required for transmission or storage. This has driven the research area of image compression to develop more compl...
详细信息
暂无评论