We have developed a wide-area-distributed storage system for multimedia databases that minimizes the possibility of simultaneous failure of multiple disks in the event of a major disaster. It features a RAID system wh...
详细信息
We have developed a wide-area-distributed storage system for multimedia databases that minimizes the possibility of simultaneous failure of multiple disks in the event of a major disaster. It features a RAID system whose member disks are spatially distributed over a wide area. Each node has a device that includes the controller of the RAID and the controller of the member disks controlled by other nodes. The devices in the node are connected to a computer using fiber optic cables and communicate using Fibre-Channel technology. Any computer at a node can utilize multiple devices connected by optical fibers as a single 'virtual disk'. The advantage of this system structure is that devices and fiber optic cables are shared by the computers. In this report, we first describe our proposed system and a prototype we used for testing. Then we discuss its performance, i.e., how the read and write throughputs are affected by data-access delay, the RAID level, and queuing.
In this paper we present a framework for content based query and retrieval of information from large videodatabases. This framework enables content based retrieval of video sequences by characterizing the sequences u...
详细信息
In this paper we present a framework for content based query and retrieval of information from large videodatabases. This framework enables content based retrieval of video sequences by characterizing the sequences using motion, texture and colorimetry cues. This characterization is biologically inspired and results in a compact parameter space where every segment of video is represented by an 8 dimensional vector. Searching and retrieval is done in real-time with high accuracy in this parameter space. The present version of the videoBook has 165 video sequences, each 15 seconds long at 30 frames a second representing storage of 65 Giga Bytes. The videoBook is able to search and retrieve video sequences with 92% accuracy in real-time. Experiments thus demonstrate that the characterization is capable of extracting higher level structure from raw pixel values.
various methods of automatic shot boundary detection have been proposed and claimed to perform reliably. Although the detection of edits is fundamental to any kind of video analysis since it segments a video into its ...
详细信息
various methods of automatic shot boundary detection have been proposed and claimed to perform reliably. Although the detection of edits is fundamental to any kind of video analysis since it segments a video into its basic components, the shots, only few comparative investigations on early shot boundary detection algorithms have been published. These investigations mainly concentrate on measuring the edit detection performance, however, do not consider the algorithms' ability to classify the types and to locate the boundaries of the edits correctly. This paper extends these comparative investigations. More recent algorithms designed explicitly to detect specific complex editing operations such as fades and dissolves are taken into account, and their ability to classify the types and locate the boundaries of such edits are examined. The algorithms' performance is measured in terms of hit rate, number of false hits, and miss rate for hard cuts, fades, and dissolves over a large and diverse set of video sequences. The experiments show that while hard cuts and fades can be detected reliably, dissolves are still an open research issue. The false hit rate for dissolves is usually unacceptably high, ranging from 50% up to over 400%. Moreover, all algorithms seem to fail under roughly the same conditions.
In this paper we present most recent evolution of JACOB, a system we developed for image and video content-based storage and retrieval. The system is based on two separate archives: a ''Features DB'' a...
详细信息
ISBN:
(纸本)0819423181
In this paper we present most recent evolution of JACOB, a system we developed for image and video content-based storage and retrieval. The system is based on two separate archives: a ''Features DB'' and a ''Raw-data DB''. When a user puts a query, a search is done in the ''Features DB'';the selected items are taken from the ''Raw-data DB'' and shown to the user. Two kinds of sessions are allowed: ''database population'' and ''database querying''. During a ''database population'' session the user inserts new data into the archive. The input data can consist of digital images or videos. videos are split into shots and for each shot one or more representative frames (r-frames) are automatically extracted. Shots and r-frames are then characterized, either in automatic or semi-automatic way, and stored in the archives. Automatic features' extraction consists of computing some low-level global features. Semi-automatic features' extraction is done by using annotation tools that perform operations that aren't currently possible with fully automatic methods. To this aim semi-automatic motion based segmentation and labeling tools have been developed. During a ''database querying'' session, queries direct or by example are allowed. Queries (direct or by example) may be iterated and variously combined to satisfy the query in the smallest number of steps. Multifeature querying is based on statistical analysis of the feature space.
The amount of digitized video in video archives is becoming so huge that easier access and content browsing tools are desperately needed. Also, video is no longer one big piece of data but a collection of useful small...
详细信息
ISBN:
(纸本)0819431273
The amount of digitized video in video archives is becoming so huge that easier access and content browsing tools are desperately needed. Also, video is no longer one big piece of data but a collection of useful smaller building blocks that can be accessed and used independently from the original context of presentation. In this paper, we demonstrate a content model for audio-video sequences with the purpose of enabling the automatic generation of video summaries. The model is based on descriptors which indicate various properties and relations of audio and video segments. In practice, these descriptors could either be generated automatically by analysis methods, or produced manually (or computer-assisted) by the content provider himself. We analyze the requirements and characteristics of the different data segments with respect to the problem of summarization, and we define our model as a set of constraints which allow to produce good quality summaries.
videodatabases are very demanding systems as far as mass storage requirements and computational resources necessary to perform common database operations, such as browsing and retrieval, are required. These operation...
详细信息
ISBN:
(纸本)0780362985
videodatabases are very demanding systems as far as mass storage requirements and computational resources necessary to perform common database operations, such as browsing and retrieval, are required. These operations can be simplified both in terms of computational complexity and of processing time by performing them on an ensemble of frames, called key frames, representative of the content units (shots) in which a video can be segmented. In this contribution an adaptive key frames extraction method based on a wavelet based multiresolution analysis in a perceptually uniform color space is presented. Experimental results that show the effectiveness of the proposed technique to select key frames summarizing the video's content, are finally provided.
In this research paper, we propose novel features based on information theory for imageretrieval. We propose the novel concept of "probabilistic filtering". We propose a hybrid approach for imageretrieval ...
详细信息
In this research paper, we propose novel features based on information theory for imageretrieval. We propose the novel concept of "probabilistic filtering". We propose a hybrid approach for imageretrieval that combines annotation approach with content based imageretrieval approach. Also rough set theory is proposed as a tool for audio/video object retrieval from multi-media databases. (C) 2015 The Authors. Published by Elsevier B.v.
We propose an image clustering algorithm which uses fuzzy graph theory. First, we define a fuzzy graph and the concept of connectivity for a fuzzy graph. Then, based on our definition of connectivity we propose an alg...
详细信息
ISBN:
(纸本)0819435902
We propose an image clustering algorithm which uses fuzzy graph theory. First, we define a fuzzy graph and the concept of connectivity for a fuzzy graph. Then, based on our definition of connectivity we propose an algorithm which finds connected subgraphs of the original fuzzy graph. Each connected subgraph can be considered as a cluster. As an application of our algorithm, we consider a database of images. We calculate a similarity measure between any pairs of images in the database and generate the corresponding fuzzy graph. Then, we find the subgraphs of the resulting fuzzy graph using our algorithm. Each subgraph corresponds to a cluster. We apply our image clustering algorithm to the key frames of news programs to find the anchorperson clusters. Simulation results show that our algorithm is successful to find most of anchorperson frames from the database.
Illumination invariance is of paramount importance to annotate video sequences stored in large videodatabases consistently. Yet, popular texture analysis methods such as multichannel filtering techniques do not yield ...
详细信息
Illumination invariance is of paramount importance to annotate video sequences stored in large videodatabases consistently. Yet, popular texture analysis methods such as multichannel filtering techniques do not yield illumination-invariant texture representations. In this paper, we assess the effectiveness of three illumination normalisation schemes for texture representations derived from Gabor filter outputs. The schemes aim at overcoming intensity scaling effects due to changes in ilumination conditions. A theoretical analysis and experimental results enable us to select one scheme as the most promising one. In this scheme, a normalising factor is derived at each pixel by combining the energy responses of different filters at that pixel. The scheme overcomes illumination variations well, while still preserving discriminatory textural information. Further statistical analysis may shed light on other interesting properties or limitations of the scheme.
In the past decades the advancement in the area of database management systems shifts towards multimedia. Multimedia information is very expressive, self explanatory, narrative, etc. Now a day the development of digit...
详细信息
暂无评论