Providing high bit rate real-time video services has been a major driving factor in the advancement of high speed networking technology such as ATM based BISDN. In this paper, we describe MPEG2Tool, an X-window-based ...
详细信息
ISBN:
(纸本)0819420425
Providing high bit rate real-time video services has been a major driving factor in the advancement of high speed networking technology such as ATM based BISDN. In this paper, we describe MPEG2Tool, an X-window-based software implementation of the MPEG-2 videocompression algorithm with many additional useful functions. The ultimate goal of designing this toolkit was to facilitate the study of MPEG video transmission over ATM-based networks. The toolkit consists of four major modules, which appear as four push-buttons in the main Motif menu: (1) encoding, (2) statistical analysis, (3) transmission simulation, and (4) decoding.
Embedded captions in TV programs such as news broadcasts, documentaries and coverage of sports events provide important information on the underlying events. In digitalvideo libraries, such captions represent a highl...
详细信息
ISBN:
(纸本)0819420425
Embedded captions in TV programs such as news broadcasts, documentaries and coverage of sports events provide important information on the underlying events. In digitalvideo libraries, such captions represent a highly condensed form of key information on the contents of the video. In this paper we propose a scheme to automatically detect the presence of captions embedded in video frames. The proposed method operates on reduced image sequences which are efficiently reconstructed from compressed MPEG video and thus does not require full frame decompression. The detection, extraction and analysis of embedded captions help to capture the highlights of visual contents in video documents for better organization of video, to present succinctly the important messages embedded in the images, and to facilitate browsing, searching and retrieval of relevant clips.
Block based motion estimation is an efficient interframe predictor, making it an important component in video coding schemes. A significant portion of a video codec's computational budget however, is allocated to ...
详细信息
ISBN:
(纸本)0819420425
Block based motion estimation is an efficient interframe predictor, making it an important component in video coding schemes. A significant portion of a video codec's computational budget however, is allocated to the task of computing motion vectors. For low bit-rate video coding applications such as teleconferencing, motion vector information occupies a substantial percentage of the available channel bandwidth. In this paper we present a method that accelerates motion vector computation by using spatio-temporal prediction to bias the search (in a statistical sense) towards the most probable direction of the motion using object trajectories from previously computed frames. Furthermore, since the motion vectors are linearly predicted, they can be coded efficiently. Linear predictive motion vector coding compares favorably to other motion estimation methods and can be incorporated within existing videocompression standards.
Screening of color-separated continuous-tone photographic images produces large high-resolution black-and-white images (up to 5000 dpi). Storing such images on disk or transmitting them to a remote imagesetter is an e...
详细信息
ISBN:
(纸本)0819423564
Screening of color-separated continuous-tone photographic images produces large high-resolution black-and-white images (up to 5000 dpi). Storing such images on disk or transmitting them to a remote imagesetter is an expensive and time-consuming task, which makes lossless compression desirable. Since a screened photographic image may be viewed as a rotated rectangular grid of large half-tone dots, each of them being made up of an amount of microdots, we suspect that compression results obtained on the CCITT test images might not apply to high-resolution screened images and that the default parameters of many existing compressionalgorithms may not be optimal. In this paper we compare the performance of lossless one-dimensional general-purpose byte-oriented statistical and dictionary-based coders as well as lossless coders designed for compression of two- dimensional bilevel images on high-resolution screened images. The general-purpose coders are: GZIP (LZ77 by GNU), TIFF LZW and STAT (an optimized PPM compressor by Bellard). The non-adaptive two-dimensional black-and-white coders are: TIFF Group 3 and TIFF Group 4 (former published fax- standards by CCITT). The adaptive two-dimensional coders are: BILEVEL coding (by Witten et al.) and JBIG (latest fax- standard). First we compared the methods without tuning their parameters. We found that both in compression ratio (CR) and speed, JBIG (CR 7.3) was best, followed by STAT (CR 6.3) and BILEVEL coding (CR 6.0). Some results are remarkable: STAT works very well, despite its one- dimensional approach; JBIG beats BILEVEL coding on high- resolution images though BILEVEL coding is better on the CCITT images, and finally, TIFF Group 4 (CR 3.2) and TIFF Group 3 (2.7) can't compete with any of these three methods. Next, we fine-tuned the parameters for JBIG and BILEVEL coding, and this resulted in an increased compression ratio of 8.0 and 6.7 respectively.
The forthcoming introduction of helical scan digital data tape recorders with high access bandwidth and large capacity will facilitate the recording and retrieval of a wide variety of multimedia information from diffe...
详细信息
ISBN:
(纸本)0819420425
The forthcoming introduction of helical scan digital data tape recorders with high access bandwidth and large capacity will facilitate the recording and retrieval of a wide variety of multimedia information from different sources, such as computer data and digital audio and video. For the compression of digital audio and video, the MPEG standard has internationally been accepted. Although helical scan tape recorders can store and playback MPEG compressed signals transparently they are not well suited for carrying out special playback modes, in particular fast forward and fast reverse. Only random portions of a original MPEG bitstream are recovered on fast playback. Unfortunately these shreds of information cannot be interpreted by a standard MPEG decoder, due to loss of synchronization and missing reference pictures. In the EC-sponsored RACE project DART (digital Data Recorder Terminal) the possibilities for recording and fast playback of MPEG video on a helical scan recorder have been investigated. In the approach we present in this paper, we assume that not transcoding is carried out on the incoming bitstream at recording time, nor that any additional information is recorded. To use the shreds of information for the reconstruction of interpretable pictures, a bitstream validator has been developed to achieve conformance to the MPEG-2 syntax during fast playback. The concept has been validated by realizing hardware demonstrators that connect to a prototype helical scan digital data tape recorder.
In this paper, we present an approach to characterize video sequences using information theoretic measures. This characterization is then used to efficiently represent a volume of video. In a typical video sequence, s...
详细信息
ISBN:
(纸本)0819420425
In this paper, we present an approach to characterize video sequences using information theoretic measures. This characterization is then used to efficiently represent a volume of video. In a typical video sequence, sometimes texture reveals structure, in other cases motion does it. In addition, the temporal and spatial extents are variables. The attempt of this work is to build this structure by looking at a given region over a multiplicity of frames and scales using entropy measures. We then present a hierarchically structured class of coders that efficiently represent this volume of video. The structure built in the analysis stage is used to control and select amongst this class of coders.
A variety of image compressionalgorithms exists for applications where reconstruction errors are tolerated. When lossless coding is mandatory, compression ratios greater than 2 or 3 are hard to obtain. DPCM technique...
详细信息
ISBN:
(纸本)0819423564
A variety of image compressionalgorithms exists for applications where reconstruction errors are tolerated. When lossless coding is mandatory, compression ratios greater than 2 or 3 are hard to obtain. DPCM techniques can be implemented in a hierarchical way, thus producing high- quality intermediate versions (tokens) of the input images at increasing spatial resolutions. Data retrieval and transmission can be achieved in a progressive fashion, either by stopping the process at the requested resolution level, or by recognizing that the image being retrieved is no longer of interest. However, progressiveness is usually realized with a certain performance penalty with respect to the reference DPCM (i.e., 4-pel optimum causal AR prediction). A generalized recursive interpolation (GRINT) algorithm is proposed and shown to be the most effective progressive technique for compression of still images. The main advantage of the novel scheme with respect to the standard hierarchical interpolation (HINT) is that interpolation is performed in a separable fashion from all error-free values, thereby reducing the variance of interpolation errors. Moreover, the introduction of a parametric half-band interpolation filter produces further benefits and allows generalized interpolation. An adaptive strategy consists of measuring image correlation both along rows and along columns and interpolating first along the direction of minimum correlation. The statistics of the different subband-like sets of interpolation errors are modeled as generalized Gaussian PDFs, and individual codebooks are fitted for variable length coding. The estimate of the shape factor of the PDF is based on a novel criterion matching the entropy of the theoretical and actual distributions. Performances are evaluated by comparing GRINT with HINT, and a variety of other multiresolution techniques. Optimum 4-pel causal DPCM and lossless JPEG are also considered for completeness of comparisons, although they are not p
Over the last few years three technologies have reached the stage of maturation where then can become synergistic. These are wideband, high speed networking high quality videocompression (MPEG-I and II), and high cap...
详细信息
Over the last few years three technologies have reached the stage of maturation where then can become synergistic. These are wideband, high speed networking high quality videocompression (MPEG-I and II), and high capacity affordable digital storage media. This paper addresses the interaction of these three technologies. In particular, it examines the problem of taking a compressed video data stream that is stored on a server, and transmitting it over an ATM channel which has a capacity smaller than that required by the data stream. The conventional approach to this problem would be to transcode by decoding the video data, and then re-encoding so as to meet the channel constraints. Currently this is not a cost effective solution since, while MPEG decoders are relatively inexpensive, encoders are not. Our approach to this problem is to partially decompress the video bitstream. Then, perform the transcoding in the quantized data domain. Finally, a valid bitstream is reassembled and transmitted. This approach has the advantage providing nearly identical quality as the traditional transcoding approach, at a fraction of the hardware cost.
Software implementations of MPEG decompression provide flexibility at low cost but suffer performance problems, including poor cache behavior. For MPEG video, decompressing the video in the implied order does not take...
详细信息
ISBN:
(纸本)0819420425
Software implementations of MPEG decompression provide flexibility at low cost but suffer performance problems, including poor cache behavior. For MPEG video, decompressing the video in the implied order does not take advantage of coherence generated by dependent macroblocks and, therefore, undermines the effectiveness of processor caching. In this paper, we investigate the caching performance gain which is available to algorithms that use different traversal algorithms to decompress these MPEG streams. We have found that the total cache miss rate can be reduced considerably at the expense of a small increase in instructions. To show the potential gains available, we have implemented the different traversal algorithms using the standard Berkeley MPEG player. Without optimizing the MPEG decompression code itself, we are able to obtain better cache performance for the traversal orders examined. In one case, faster decompression rates are achieved by making better use of processor caching, even though additional overhead is introduced to implement the different traversal algorithm. With better instruction-level support in future architectures, low cache miss rates will be crucial for the overall performance of software MPEG video decompression.
We make a case that taking the number of bits to code each motion vector into account when estimating motion for videocompression results in significantly better performance at low bit rates, using simulation studies...
详细信息
ISBN:
(纸本)0819420425
We make a case that taking the number of bits to code each motion vector into account when estimating motion for videocompression results in significantly better performance at low bit rates, using simulation studies on established benchmark videos. First, by modifying a `vanilla' implementation of the H.261 standard, we show that choosing motion vectors explicitly to minimize rate (in a greedy manner), subject to implicit constraints on distortion, yields better rate-distortion tradeoffs than minimizing notions of prediction error. Locally minimizing a linear combination of rate and distortion results in further improvements. Using a heuristic function of the prediction error and the motion vector code-length results in compression performance comparable to the more computationally intensive coders while requiring a practically small amount of computation. We also show that making coding control decisions to minimize rate yields further improvements.
暂无评论