The first step of the coding technique proposed in the MPEG standard is motion compensation. It reduces the residual error energy using a fraction of the total bit rate to transmit motion information. Motion compensat...
详细信息
ISBN:
(纸本)0819420425
The first step of the coding technique proposed in the MPEG standard is motion compensation. It reduces the residual error energy using a fraction of the total bit rate to transmit motion information. Motion compensation is performed using a block matching approach though the algorithm to compute motion vectors is not given in the MPEG standard. Usually, an exhaustive search around the macroblock position is used. This solution (proposed in the test model) gives the lowest error but has the highest complexity. In this work we propose an algorithm that reduces the complexity of the block matching procedure while achieving comparable performance with the exhaustive search. The proposed solution is particularly attractive for the spatially scalable version of the coder when both a full resolution and a spatially downsampled sequence are transmitted. The algorithm uses a multiresolution motion compensation scheme. Exhaustive search block matching is performed in the downsampled sequence and the vector field computed is used as an estimate of the motion vectors for the full resolution sequence. Thus, only a refinement needs to be computed. This allows a consistent reduction of the computation time with respect to exhaustive search at the full resolution level, while the residual error energy increases only slightly.
This paper investigates motion estimation and compensation in object-oriented analysis-synthesis coding. Object- oriented coding employs a mapping parameter technique for estimating motion information in each object. ...
详细信息
ISBN:
(纸本)0819423564
This paper investigates motion estimation and compensation in object-oriented analysis-synthesis coding. Object- oriented coding employs a mapping parameter technique for estimating motion information in each object. The mapping parameter technique using gradient operators requires high computational complexity. The main objective of this paper is to propose a hybrid mapping parameter estimation method using the hierarchical structure in object-oriented coding. The hierarchical structure employed constructs a low- resolution image. Then six mapping parameters for each object are estimated from the low-resolution image and these parameter values are verified based on the displaced frame difference (DFD). If the verification test succeeds, the parameters and object boundaries are coded. Otherwise, eight mapping parameters are estimated in a low-resolution image and the verification test is again applied to an image reconstructed by estimated parameters. If it succeeds, the parameters and object boundaries are coded, otherwise, the regions are coded by second-order polynomial approximation. Theoretical analysis and computer simulation show that the peak signal to nose ratio (PSNR) of the image reconstructed by the proposed method lies between those of images reconstructed by the conventional 6- and 8-parameter estimation methods with reduction of the computation time by a factor of about four.
CCD image sensors are capable of delivering at least 600% of linear video to the signal processing. This headroom is needed to reproduce large scene contrasts in a convincing way on displays not capable to handle thes...
详细信息
The rate control algorithm plays an important role for improving and stabilizing the playback quality for video coded with the MPEG standard. Several optimal control techniques have been proposed to aim at the best po...
详细信息
ISBN:
(纸本)0819420425
The rate control algorithm plays an important role for improving and stabilizing the playback quality for video coded with the MPEG standard. Several optimal control techniques have been proposed to aim at the best possible quality for a given channel rate and buffer size. Some of these approaches are complex in that they require the rate and distortion characteristics of the input data to be measured. This motivates us to pursue a method for approximating the rate and distortion functions to reduce the computations. Previous work has been based to a large extent on modeling the distortion as negative exponential functions of the rate. This type of model ignores many factors in a real MPEG encoding process and is not general enough for all video sources. In this paper, we use piece-wise polynomials to approximate the frame-level rate and distortion. The frame dependency between the predictive frame and its reference frames is also considered in our model. Compared to other models, our method is relatively more complex but gives more accurate results. We observe low average relative model errors, which indicates that the model is accurate for most of the quantization settings. We use the model within our gradient-based rate control algorithm and show how using the model one can closely approximate the solution obtained using the actual data. Finally, we apply a simplified version of the model to a new fast algorithm derived from the MPEG Test Model 5, and demonstrate that both the quality (in terms of PSNR) and stability of the quality can be improved.
This paper presents a videocompression format for HDTV video to be used in studio applications. digital HDTV signals need to be compressed in order to be routed or stored within a studio. The bitrates targeted suppor...
详细信息
The diversity in the nature of TV images is increasing due to the mixing of graphics data with natural scenery. In this paper we study the lossless coding of such graphics data with the objective to come to recommenda...
详细信息
ISBN:
(纸本)0819423564
The diversity in the nature of TV images is increasing due to the mixing of graphics data with natural scenery. In this paper we study the lossless coding of such graphics data with the objective to come to recommendations for videocompression of such images. For this purpose, a comparison has been made between contour coding, template coding and run-length coding of graphical imagery. It is concluded that template coding is suitable for TXT-like images but it is not practical for more generic graphics images. For menus and similar imagery, contour and run-length coding provide interesting options.
The proceedings contains 91 papers from the 1996 IEEE 39th Midwest Symposium on Circuits & Systems. Topics discussed include: fuzzy logic and control;VLSI circuits;application specific integrated circuits (ASICs);...
详细信息
The proceedings contains 91 papers from the 1996 IEEE 39th Midwest Symposium on Circuits & Systems. Topics discussed include: fuzzy logic and control;VLSI circuits;application specific integrated circuits (ASICs);transconductance amplifiers;color image processing;vector quantization;image segmentation;object and character recognition;digital filters;Kalman and adaptive filtering techniques;video coding and compression;digital signal processing;CMOS devices;dynamic programming;adaptive algorithms;neural networks;inference engines;spread spectrum communication systems;multiplexing;and robotics.
Currently, a new and fascinating area of multimedia applications is emerging. This area embodies the traditionally distinct application areas of telecommunications, broadcasting, computers, and consumer electronics an...
Currently, a new and fascinating area of multimedia applications is emerging. This area embodies the traditionally distinct application areas of telecommunications, broadcasting, computers, and consumer electronics and offers consumers a wide variety of new and en- hanced applications. The push behind these applications is the convergence of a number of technologies (digital signal processing, transmission, storage, and integrated circuit de- sign) in combination with a number of standardization activities (JPEG, MPEG, DVB, and DAVIC). The success of these new multimedia applications will depend on the degree of interoperability that will exists between multimedia systems. With interoperability, users can enjoy a large number of applications with a limited number of devices. The degree of interoperability can be increased by compressing and/or representing the video components of multimedia applications in a scalable way. Scalable means here that the video signal can be reproduced at various quality levels by decoding different portions of the compressed data. Such a feature is useful when, for example, a multimedia device has to visualize a high resolution video signal on a low resolution display, or when the delivery of the video signal is error-prone. In the latter case, the scalably compressed video signals can be made to degrade in a graceful way when delivery errors occur. video signals are very suitable for scalable compression. This is partly due to the large intelligibility of video signals, which makes them recognisable at a wide variety of signal qualities. For video signals, four basic types of scalability can be defined: namely, amplitude, frequency, spatial, and temporal scalability. Amplitude and frequency scalability are both related to the SNR quality of the signal, while spatial and temporal scalability are related to the spatial resolution and temporal resolution of the video signal, respectively. To implement scalability, multimedia systems may us
In this paper we propose an optimal quad-tree (QT)-based motion estimator for videocompression. It is optimal in the sense that for a given bit budget for encoding the displacement vector field (DVF) and the QT segme...
详细信息
ISBN:
(纸本)0819423564
In this paper we propose an optimal quad-tree (QT)-based motion estimator for videocompression. It is optimal in the sense that for a given bit budget for encoding the displacement vector field (DVF) and the QT segmentation, the scheme finds a DVF and a QT segmentation which minimizes the energy of the resulting displaced frame difference (DFD). We find the optimal QT decomposition and the optimal DVF jointly using the Lagrangian multiplier method and a multilevel dynamic program. The resulting DVF is spatially inhomogeneous since large blocks are used in areas with simple motion and small blocks in areas with complex motion. We present results with the proposed QT-based motion estimator which show that for the same DFD energy the proposed estimator uses about 30% fewer bits than the commonly used block matching algorithm.
This paper presents a very low bit-rate coding algorithm based on image split in order to represent it through an adaptive multigrid supported by a binary tree structure. Independently of its tree representation, the ...
详细信息
ISBN:
(纸本)0819420425
This paper presents a very low bit-rate coding algorithm based on image split in order to represent it through an adaptive multigrid supported by a binary tree structure. Independently of its tree representation, the picture is segmented via a watershed procedure and several criteria are combined to automatically extract interesting areas of the image. This object information is not transmitted but used to reduce picture complexity, and therefore the bit-rate, while keeping a good subjective quality. This is achieved by a merge procedure which homogenizes values of the tree subblocks belonging to a same non-interesting object. This treatment affects both intra- and inter-images. For intra-images, the resulting tree structure is entropy coded while its leaves are encoded through a DPCM procedure followed by a multi-Huffman coder. For inter-images, a motion field is adaptated by an adaptative block matching algorithm which is a kind of BMA for which blocksize is chosen in order to reach a sufficient level of confidence. Residues, essential to correct motion compensation artifacts, are sent through local intra-trees or, if the bit-rate allows it, through DCT blocks, allowing to reach an arbitrary level of quality. During the reconstruction step, an object oriented approach combined with the use of overlapping functions allows to reduce block artifacts while keeping sharp edges.
暂无评论