The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
This paper proposes an algorithm for automatic detection of 3D video shots with different perceptual features. The proposed algorithm is able to identify distinct three-dimensional visual scenes by detecting 3D video shot boundaries based on clustering of depth-temporal features. A combination of texture variation along the temporal dimension and depth variance is used by K-means clustering to find...
We present a novel interface for large-scale video archives that uses content-based filtering of search results. The interface has been used for the Interactive Known-Item Search (Interactive KIS) task of TRECVID 2012 and achieved good search performance. We found that for KIS tasks content-based filtering as used in our interface is convenient and able to successfully narrow down interactive search...
In this paper we propose an in-depth evaluation of the performance of video descriptors to multimodal video genre categorization. We discuss the perspective of designing appropriate late fusion techniques that would enable to attain very high categorization accuracy, close to the one achieved with user-based text information. Evaluation is carried out in the context of the 2012 Video Genre Tagging...
This article investigates a novel use of the well-known audio words representations to detect specific audio events, namely gunshots and explosions, in order to get more robustness towards soundtrack variability in Hollywood movies. An audio stream is processed as a sequence of stationary segments. Each segment is described by one or several audio words obtained by applying product quantization to...
In this paper, we present a complete system for audio indexing. This system is based state-of-the-art methods of Speech-Music-Noise segmentation and Monophonic/Polyphonic estimation. After those methods we propose an original system of superposed sources detection. This approach is based on the analysis of the evolution of the predominant frequencies. In order to validate the whole system we used...
This paper addresses the problem of video summarization through an automatic selection of a single representative keyframe. The proposed solution is based on the mutual reinforcement paradigm, where a keyframe is selected thanks to its highest and most frequent similarity to the rest of considered frames. Two variations of the algorithm are explored: a first one where only frames within the same video...
Clip art databases can be composed by raster images or by vector drawings. There are technologies for searching and retrieving clip arts for both image formats but research has been done separately, focusing on either format, without taking benefits of both research fields as a whole. This paper describes a study where the benefits of combining information extracted from vector and raster images to...
This paper describes a new approach for multimedia documents indexing and addresses the problem of automatically detecting a large number of visual concepts. Though using a multi-label approaches are used in some works, concepts detectors are often trained independently. We propose a model that takes into account the detection of not only a target concept but also other ones and regroups in terms...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.