@inbook{bc41b65a23924a8d9d78fd674a7d655a,
title = "Visual Alphabets: Video classification by end users",
abstract = "The work presented here introduces a real-time automatic scene classifier within content-based video retrieval. In our envisioned approach end users like documentalists, not image processing experts, build classifiers interactively, by simply indicating positive examples of a scene. Classification consists of a two-stage procedure. First, small image fragments called patches are classified. Second, frequency vectors of these patch classifications are fed into a second classifier for global scene classification (e.g., city, portraits, or countryside). The first stage classifiers can be seen as a set of highly specialized, learned feature detectors, as an alternative to letting an image processing expert determine features a priori. The end user or domain expert thus builds a visual alphabet that can be used to describe the image in features that are relevant for the task at hand.We present results for experiments on a variety of patch and image classes. The scene classifier approach has been successfully applied to other domains of video content analysis, such as content-based video retrieval in television archives, automated sewer inspection, and porn filtering.",
keywords = "IR-58737, METIS-243090, Image Processing, Classification, content, scenes, Video Retrieval, Real Time, HMI-MR: MULTIMEDIA RETRIEVAL, EWI-20854, patches",
author = "Menno Isra{\"e}l and {van den Broek}, Egon and {van der Putten}, Peter and {den Uyl}, {Marten J.}",
year = "2007",
doi = "10.1007/978-1-84628-799-2_10",
language = "Undefined",
isbn = "978-1-84628-436-6",
series = "Chapter 10",
publisher = "Springer",
pages = "185--206",
editor = "Petrushin, {Valery A.} and Latifur Khan",
booktitle = "Multimedia Data mining and Knowledge Discovery",
address = "Germany",
}