HOME
  COMMITTEE
  CALL FOR PAPERS
  IMPORTANT DATES
  AWARDS
  TECHNICAL PROGRAM
  SPONSORSHIP
  THEME SESSION
  PLENARY TALK
  HP Competition
Plenary Talk
::: Wavelets and Differential Operators: from Fractals to Marr’s Primal Sketch :::

Overview: Invariance is an attractive principle for specifying image processing algorithms. In this presentation, we promote affine invariance—more precisely, invariance with respect to translation, scaling and rotation. As starting point, we identify the corresponding class of invariant 2D operators: these are combinations of the (fractional) Laplacian and the complex gradient (or Wirtinger operator). We then specify some corresponding differential equation and show that the solution in the real-valued case is either a fractional Brownian field (Mandelbrot and Van Ness, 1968) or a polyharmonic spline (Duchon, 1976), depending on the nature of the system input (driving term): stochastic (white noise) or deterministic (stream of Dirac impulses). The affine invariance of the operator has two important consequences: (1) the statistical self-similarity of the fractional Brownian field, and (2) the fact that the polyharmonic splines specify a multiresolution analysis of $L_2(\R^2)$ and lend themselves to the construction of wavelet bases. The other fundamental implication is that the corresponding wavelets behave like multi-scale versions of the operator from which they are derived; this makes them ideally suited for the analysis of multidimensional signals with fractal characteristics (whitening property of the fractional Laplacian).
The complex extension of the approach yields a new complex wavelet basis that replicates the behavior of the Laplace-gradient operator and is therefore adapted to edge detection. We introduce the Marr wavelet pyramid which corresponds to a slightly redundant version of this transform with a Gaussian-like smoothing kernel that has been optimized for better steerability. We demonstrate that this multiresolution representation is well suited for a variety of image-processing tasks. In particular, we use it to derive a primal wavelet sketch—a compact description of the image by a multiscale, subsampled edge map—and provide a corresponding iterative reconstruction algorithm.

Dr. Michael Unser
Ecole Polytechnique Fédérale de Lausanne (EPFL)
Switzerland

Brief Biography: Michael Unser is Professor and Director of EPFL's Biomedical Imaging Group, Lausanne, Switzerland. His main research area is biomedical image processing. He has a strong interest in sampling theories, multiresolution algorithms, wavelets, and the use of splines for image processing. He has published over 150 journal papers on those topics, and is one of ISI’s Highly Cited authors in Engineering (http://isihighlycited.com).
From 1985 to 1997, he was with the Biomedical Engineering and Instrumentation Program, National Institutes of Health, Bethesda USA, conducting research on bioimaging and heading the Image Processing Group.
Dr. Unser is a fellow of the IEEE and the recipient of three Best Paper Awards from the IEEE Signal Processing Society. He is a member of the Swiss Academy of Engineering Sciences.

::: Light Field Photography and Microscopy :::

Overview: Light field photography is a technique for recording light intensity as a function of position and direction in a 3D scene. Unlike conventional photographs, light fields permit manipulation of viewpoint and focus after the imagery has been recorded. At Stanford we have built a number of devices for capturing light fields, including (1) an array of 128 synchronized video cameras, (2) a handheld camera in which a microlens array has been inserted between the main lens and sensor plane, and (3) a microscope in which a similar microlens array has been inserted at the intermediate image plane.
The third device permits us to capture light fields of microscopic biological (or industrial) objects in a single snapshot. Although diffraction limits the product of spatial and angular resolution in these light fields, we can nevertheless produce useful perspective flyarounds and 3D focal stacks from them. Since microscopes are inherently orthographic devices, perspective flyarounds represent a new way to look at microscopic specimens. Focal stacks are not new, but manual techniques for capturing them are time-consuming and hence not applicable to moving or light-sensitive specimens. Applying 3D deconvolution to these focal stacks, we can produce a set of cross sections, which can be visualized using volume rendering. Ours is the first technology (of which we are aware) that can produce volumetric models from a single photograph.
In this talk, I will describe a prototype light field microscope and show perspective views, focal stacks, and reconstructed volumes for a variety of biological specimens. I will also survey some promising directions for this technology. For example, by introducing a second microlens array and a video projector, we can control the light field arriving at a specimen as well as the light field leaving it. Potential applications of this idea include microscope scatterometry - measuring reflectance as a function of incident and reflected angle, and "designer illumination" - illuminating one part of a microscopic object while avoiding illuminating another.

Dr. Marc Levoy
Stanford University, California

Brief Biography: Marc Levoy is a Professor of Computer Science and (jointly) Electrical Engineering at Stanford University. He received a Bachelor's and Master's in Architecture from Cornell University in 1976 and 1978, and a PhD in Computer Science from the University of North Carolina at Chapel Hill in 1989. In the 1970's Levoy worked on computer animation, developing an early computer-assisted cartoon animation system. This system was used by Hanna-Barbera Productions from 1983 until 1996 to produce The Flintstones, Scooby Doo, and other shows. In the 1980's Levoy worked on volume rendering, a family of techniques for displaying sampled three-dimensional functions, for example computed tomography (CT) or magnetic resonance (MR) data. In the 1990's he worked on technology and algorithms for digitizing three-dimensional objects. This led to the Digital Michelangelo Project, in which he and a team of researchers spent a year in Italy digitizing the statues of Michelangelo using laser scanners. His current interests include light field sensing and display, computational photography, and applications of computer graphics in microscopy and biology. Awards: Charles Goodwin Sands Medal for best undergraduate thesis (1976), National Science Foundation Presidential Young Investigator (1991), ACM SIGGRAPH Computer Graphics Achievement Award (1996), ACM Fellow (2007). Recent professional service: Papers Chair of SIGGRAPH 2007.

::: Scene Segmentation with Latent Topic Markov Field Models - and - Classification and Dimensionality Reduction using Convex Class Models :::

Overview: The talk will be in two parts. In the first part I will present work with Jakob Verbeek on semantic-level scene segmentation by combining spatial coherence models such as Markov and Conditional Random Fields with latent topic based local image content models such as Probabilisitic Latent Semantic Analysis over bag-of-words representations. In the second part I will present some work with Hakan Cevikalp, Frederic Jurie and Robi Polikar on using simple convex approximations to high-dimensional classes for multi-class classification and discriminant dimensionality reduction.

Dr. Bill Triggs
Centre National de Recherche Scientifique (CNRS)
France

Brief Biography: Professor Triggs is an ex mathematical physicist turned computer vision and pattern recognition researcher employed by France's main national research organisation, the Centre National de Recherche Scientifique (CNRS). He works in (and is deputy director of) the Laboratoire Jean Kuntzmann (LJK) in Grenoble in the heart of the French Alps. LJK is a newly created applied mathematics and computer science research laboratory associated with the CNRS and three Grenoble universities (UJF, INPG, UPMF) and with links to INRIA's Rhône-Alpes research unit.
Nowadays he works mainly on applying statistical pattern recognition and machine learning to visual object recognition and scene understanding, but he is also known for his work on vision geometry (matching constraints, scene reconstruction, auto-calibration, etc) and on human body tracking. He has been associated with several robotics and computer vision research teams in Grenoble : SHARP (now e-Motion) from 1993 to 1995; MOVI (now PERCEPTION) from 1995 to 2003; and LEAR from 2003 to 2007. Before that he worked on robotic planning and mobile robotics in Oxford University's Robotics Research Group.

::: Scene and Pose Associated Camera Motion Synthesis :::

Overview: Synthesis and control of camera motion is one of the essential components in animation. In this talk we present our work that demonstrates authoring of camera motion in two scenarios: a) planning walkthroughs of a scene, where camera responds to the scene contents, and b) creating 3D character animation with moving camera, where camera accounts for the pose of the character specified.
In the first scenario, the emphasis is given on designing a walkthrough of a scene with the knowledge of interesting and uninteresting parts of the scene. The approach developed offers a walkthrough so that the camera looks at the interesting objects for the most part of the tour. Further the speed can also be controlled from the quality of the scene being seen. A graph theoretic optimization scheme is adopted that computes optimal gaze and speed of the camera at sampled points on a specified path.
For the second scenario, we introduce a framework of view-dependent animation that embodies the association of the character pose and the camera (or the view). Typically, an animator while making the storyboard specifies the character pose with desired camera for the key views. A space called "view space" is generated using the key cameras and associated character poses. A desired animation holding the camera pose association can be generated by tracing a path in the view space.
This work is done in collaboration with Subhashis Banerjee, Subhajit Sanyal and Parag Chaudhury.

Dr. Prem K Kalra
Indian Institute of Technology, Delhi

Brief Biography: Prem Kalra is a professor in the Department of Computer Science and Engineering at the Indian Institute of Technology (IIT), Delhi. He is with IIT Delhi since Dec 1997. Before joining IIT Delhi, he was at University of Geneva (Switzerland). He did BSc Mechanical Engineering from (Dayalbagh) Agra University, India, MTech in Industrial Engineering from IIT Delhi, MS in Computer Science from University of New Brunswick, Canada and PhD in Computer Science from Swiss Federal Institute of Technology- Lausanne (EPFL), Switzerland. His research interests include computer vision based modeling and rendering, 3D visualization and animation, and image/video super-resolution. He has published more than 50 papers in the reputed international journals and conferences.

© All rights reserved, IEEE Kharagpur Section, IIT Kharagpur