Multimedia Information Retrieval
||A major contributor to this article appears to have a close connection with its subject. (July 2014)|
||This article possibly contains original research. (July 2014)|
Multimedia Information Retrieval (MMIR or MIR) is a research discipline of computer science that aims at extracting semantic information from multimedia data sources.[not in citation given] Data sources include directly perceivable media such as audio, image and video, indirectly perceivable sources such as text, biosignals as well as not perceivable sources such as bioinformation, stock prices, etc. The methodology of MMIR can be organized in three groups:
- Methods for the summarization of media content (feature extraction). The result of feature extraction is a description.
- Methods for the filtering of media descriptions (for example, elimination of redundancy)
- Methods for the categorization of media descriptions into classes.
Feature Extraction Methods
Feature extraction is motivated by the sheer size of multimedia objects as well as their redundancy and, possibly, noisiness.[not in citation given] Generally, two possible goals can be achieved by feature extraction:
- Summarization of media content. Methods for summarization include in the audio domain, for example, Mel Frequency Cepstral Coefficients, Zero Crossings Rate, Short-Time Energy. In the visual domain, color histograms such as the MPEG-7 Scalable Color Descriptor can be used for summarization.
- Detection of patterns by auto-correlation and/or cross-correlation. Patterns are recurring media chunks that can either be detected by comparing chunks over the media dimensions (time, space, etc.) or comparing media chunks to templates (e.g. face templates, phrases). Typical methods include Linear Predictive Coding in the audio/biosignal domain, texture description in the visual domain and n-grams in text information retrieval.
Merging and Filtering Methods
Multimedia Information Retrieval implies that multiple channels are employed for the understanding of media content. Each of this channels is described by media-specific feature transformations. The resulting descriptions have to be merged to one description per media object. Merging can be performed by simple concatenation if the descriptions are of fixed size. Variable-sized descriptions - as they frequently occur in motion description - have to be normalized to a fixed length first.
Frequently used methods for description filtering include factor analysis (e.g. by PCA), singular value decomposition (e.g. as latent semantic indexing in text retrieval) and the extraction and testing of statistical moments. Advanced concepts such as the Kalman filter are used for merging of descriptions.
Generally, all forms of machine learning can be employed for the categorization of multimedia descriptions[not in citation given] though some methods are more frequently used in one area than another. For example, Hidden Markov models are state-of-the-art in speech recognition, while Dynamic Time Warping - a semantically related method - is state-of-the-art in gene sequence alignment. The list of applicable classifiers includes the following:
- Metric approaches (Cluster Analysis, Vector Space Model, Minkowski Distances, Dynamic Alignment)
- Nearest Neighbor methods (K-nearest neighbors algorithm, K-Means, Self-Organizing Map)
- Risk Minimization (Support Vector Regression, Support Vector Machine, Linear Discriminant Analysis)
- Density-based Methods (Bayes Nets, Markov Processes, Mixture Models)
- Neural Networks (Perceptron, Associative Memories, Spiking Nets)
- Heuristics (Decision Trees, Random Forests, etc.)
The quality of MMIR Systems depends heavily on the quality of the training data. Discriminative descriptions can be extracted from media sources in various forms. Machine learning provides categorization methods for all types of data. However, the classifier can only be as good as the given training data. On the other hand, it requires considerable effort to provide class labels for large databases. The future success of MMIR will depend on the provision of such data. The annual TRECVID competition is currently one of the most relevant sources of high-quality ground truth.
MMIR provides an overview over methods employed in the areas of information retrieval. Methods of one area are adapted and employed on other types of media. Multimedia content is merged before the classification is performed. MMIR methods are, therefore, usually reused from other areas such as:
- Bioinformation Analysis
- Biosignal Processing
- Content-based Image and Video Retrieval
- Face Recognition
- Audio and Music Classification
- Speech Recognition
- Technical Chart Analysis
- Text Information Retrieval
The Journal of Multimedia Information Retrieval documents the development of MMIR as a research discipline that is independent of these areas. See also  for a complete overview over this research discipline.
- H Eidenberger. " Fundamental Media Understanding ", atpress, 2011, p. 1.
- H Eidenberger. " Fundamental Media Understanding ", atpress, 2011, p. 2.
- A Del Bimbo. " Visual Information Retrieval ", Morgan Kaufmann, 1999.
- HG Kim , N Moreau, T Sikora. " MPEG-7 Audio and Beyond", Wiley, 2005.
- MS Lew (Ed.). " Principles of Visual Information Retrieval ", Springer, 2001.
- H Eidenberger. " Fundamental Media Understanding ", atpress, 2011,p. 125.
- JC Nordbotten. "Multimedia Information Retrieval Systems". Retrieved 14 October 2011.
- H Eidenberger. " Frontiers of Media Understanding ", atpress, 2012.
- H Eidenberger. " Professional Media Understanding ", atpress, 2012.
- "Journal of Multimedia Information Retrieval", Springer, 2011, Retrieved 21 October 2011.
- H Eidenberger. " Handbook of Multimedia Information Retrieval ", atpress, 2012.