Title
Describing when and where in vision
Abstract
Different from the what and where pathways in the organization of the visual system, we address representations that describe dynamic visual events in a unified way. Representations are an essential tool for any kind of process that operates on data, as they provide a language to describe, store and retrieve that data. They define the possible properties and aspects that are stored, and govern the levels of abstraction at which the respective properties are described. In the case of visual computing (computer vision, image processing), a representation is used to describe information obtained from visual input (e.g. an image or image sequence and the objects it may contain) as well as related prior knowledge (experience). The ultimate goal, to make applications of visual computing be part of our daily life, requires that vision systems operate reliably, nearly anytime and anywhere. Therefore, the research community aims to solve increasingly more complex scenarios. Vision both in humans and computers is a dynamic process, thus variations (change) always appear in the spatial and the temporal dimensions. Nowadays significant research efforts are undertaken to represent variable shape and appearance, however, joint representation and processing of spatial and temporal domains is not a well-investigated topic yet. Visual computing tasks are mostly solved by a two-stage approach of frame-based processing and subsequent temporal processing. Unfortunately, this approach reaches its limits in scenes with high complexity or difficult tasks e.g. action recognition. Therefore, we focus our research on representations which jointly describe information in space and time and allow to process data of space-time volumes (several consecutive frames). In this keynote we relate our own experience and motivations, to the current state of the art of representations of shape, of appearance, of structure, and of motion. Challenges for such representations are in applications like multiple object tracking, tracking non-rigid objects and human action recognition.
Year
DOI
Venue
2011
10.1007/978-3-642-25085-9_2
CIARP
Keywords
Field
DocType
subsequent temporal processing,visual input,visual computing,computer vision,dynamic process,image processing,visual computing task,visual system,frame-based processing,dynamic visual event
Visual computing,Computer vision,Abstraction,Human visual system model,Computer science,Spacetime,Action recognition,Image processing,Video tracking,Artificial intelligence,Image sequence
Conference
Volume
ISSN
Citations 
7042
0302-9743
0
PageRank 
References 
Authors
0.34
0
3
Name
Order
Citations
PageRank
Walter G. Kropatsch1896152.91
Adrian Ion222221.11
Nicole M. Artner3789.04