Demo Program

Demo Chair: Matthew Ma (Email: mattma@ieee.org)
Demo Logistics: Rebecca Chau (Email: ceb@ctshk.com)

The ICPR 2006 will hold demo sessions in conjunction with conference meetings to encourage discussions and information exchanges among attendees.

Demo Schedule

Date:August 22 (Tuesday) and August 23 (Wednesday)
Time:10am - 12:30pm and 1:30pm – 5pm
Location:Room 409

List of Registered Demos

Abstracts of Registered Demos

3D Geometry Reconstruction and Statistical Shape Modeling

Authors:H. Lamecker, T. Wenckebach, H.-C. Hege
Affiliation:Zuse Institute Berlin, Germany
Contact:Hans Lamecker, lamecker@zib.de

Abstract:

Generating geometric models of anatomical objects from 2D or 3D image data is a prerequisite for many tasks in computer-aided medical diagnostics and planning. Each imaging modality (typically CT, MRT, US, X-ray, etc.) exhibits its own specific properties to be considered in the feature extraction or segmentation process. Manual methods are time-consuming, prone to errors and hardly reproducible. Incorporating a-priori knowledge about the object and the data to be segmented seems to be a feasible way to automate this task.

In this demo, we will present a software, based on the 3D visualization and volume modeling system Amira, which allows to generate and utilize statistical models of arbitrary 3D shapes. Amira already offers a variety of image filtering methods, interactive segmentation tools and a surface reconstruction algorithm for creating surfaces as the basis for the training of a statistical model. The focus of this presentation will be on how to establish correspondence across different shapes, register different shapes to a common reference coordinate system and perform statistical analysis.

Statistical shape models can be applied to a range of different applications: segmentation of 3D image data (especially low-contrast image data), 3D reconstruction from 2D data (X-ray images), reconstruction of pathological, incomplete or sparse surface data. Different anatomical models and their application will be presented in this demo.

^ Back to top

Fast Linear Feature Detection using Multiple Directional Non-Maximum Suppression

Authors:Changming Sun and Pascal Vallotton
Affiliation:CSIRO Mathematical & Information Sciences NSW 1670, Australia
Contact:Changming Sun changming.sun@csiro.au

Abstract:

The capacity to detect linear features is central to image analysis, computer vision, and pattern recognition; and it has practical applications in areas such as neurite outgrowth detection, retinal vessel extraction, skin hair removal for malonoma detection, plant root analysis, and road detection. Linear features detection often represents the starting point for segmentation and image interpretation. Here, we present a new algorithm for linear feature detection using multiple directional non-maximum suppression. Given its low computational complexity, the algorithm is very fast. We show on several examples that it performs remarkably well in terms of sensitivity and continuity of detected features.

^ Back to top

Robust Head Tracking and Illumination Robust Facial Expression Recognition System

Authors:Wooju Ryu, Daehwan Kim, Hyung-Soo Lee, Jaewon Sung, Daijin Kim
Affiliation:Department of Computer Engineering, POSTECH, KOREA
Contact:Hyung-Soo Lee, sooz@postech.ac.kr

Abstract:

Active Appearance Model (AAM) is a well-known model that can represent a non-rigid object effectively. However, the AAM often fails to fit the input image under the severely changing illumination condition, because it uses the fixed appearance basis vectors that are usually obtained in the training phase. To overcome this disadvantage, we propose an adaptive AAM that updates the appearance basis vectors with the input image using incremental principal component analysis (PCA). We also propose a layered generalized discriminant analysis (GDA) classifier which combines shape and appearance information to improve the recognition performance of person independent facial expressions. This system runs at 6~8 frames per second on a Pentium 3.2GHz PC and the average expression recognition performance is 90.1%. We also propose another demo which represents a robust head tracking using 3D cylindrical model. Eye detection and active appearance model are used for initializing the 3D cylinder model. Iterative re-weighted least squares technique has been used for fitting algorithm and dynamic template technique approach can be robustly tracking about gradual lighting changes. Our system runs in real-time about 10 fps and can robustly track the slowly changing head motion. Possible tracking ranges are -90 degrees to 90 degrees horizontally and -60 degrees to60 degrees vertically.

^ Back to top

Gesture Recognition using Temporal Templates

Authors:Wooju Ryu, Daehwan Kim, Hyung-Soo Lee, Jaewon Sung, Daijin Kim
Affiliation:Department of Computer Engineering, POSTECH, KOREA
Contact:Hyung-Soo Lee, sooz@postech.ac.kr

Abstract:

We present a gesture recognition system using a stereo camera for the intelligent robot in the typical indoor environment. The gesture recognition system represents a specific motion by MEI (Motion Energy Image) and MHI (Motion History Image) view-specific temporal templates, where MEI is a binary representation of which motion has occurred in an image sequence and MHI is a scalar-valued image whose intensity is a function of regency of motion. The gesture recognition system classifies each motion using GDA ensemble via two steps: spotting the meaningful gesture section in the image sequence and recognizing the meaningful section into one of 10 specific gestures. The gesture recognition system runs at 6~8 frames per second on a 3.2GHz PC and shows an 80% accuracy of segmenting image sequence to the meaningful gesture section and a 90% accuracy of the gesture recognition.

^ Back to top

Real-Time Document Image Retrieval Using Web Cameras

Authors:Tomohiro Nakai, Koichi Kise and Masakazu Iwamura
Affiliation:Graduate School of Engineering, Osaka Prefecture University, Japan
Contact:Tomohiro Nakai, nakai@m.cs.osakafu-u.ac.jp

Abstract:

Camera-based document image retrieval is a task of searching document images from the database based on query images captured using digital cameras. We have already proposed a method called "locally likely arrangement hashing (LLAH)" which enables us a fast (about 100 msec.) and accurate (more than 94%) retrieval from a large database (including 10,000 document images). In this presentaion, we propose a real-time document image retrieval based on LLAH. The proposed method repeats retrieval and display the result using a web camera. The system currently achieves 7 fps for the retrieval from the database of 20,000 images. As an application of the real-time retrieval, we also propose a method of augmented reality for documents. This method is to superimpose relevant information onto the camera-captured image naturally by using the parameters of perspective transformation calculated as a subsidiary result of retrieval. We consider that the results indicate a new possibility of document images as media of displaying information.

^ Back to top

Stereo Vision, Spherical Video, and High Frame-rate cameras

Authors:
Affiliation:Point Grey Research, Inc., Vancouver, BC, Canada
Contact:Don Murray, donm@ptgrey.com

Abstract:

Point Grey Reseach will be demonstrating a variety of advanced cameras suitable for computer vision reseach. These include stereo vision, spherical/panoramic video and high frame-rate cameras.

Bumblebee2 Stereo Vision camera - this is a real-time stereo camera that runs at speeds up to 48 frames/second. The camera uses high quality Sony CCD sensors and is available in color or B&W models. It transmits images to the host computer over a 1394 FireWire interface. It is factory calibrated and stays in calibration. The Triclops C/C++ stereo SDK is provided with the camera. Real-time rectified and disparity images will be demonstrated, as well as live 3D point clouds.

Ladybug2 Spherical Video system - this is a video system that captures 75% of a full sphere of video data at speeds up to 30 frames/second. The camera is a cluster of six sensors tightly packaged to minimize the effects of parallax. This number of sensors allows the system to deliver 4.7 megapixels of image data with every frame. The Ladybug software development kit allows geometric processing from individual sensor images as well as a suite of routines for rendering panoramic images in various formats. The Ladybug2 camera will be demonstrated running live, as well as pre-recorded outdoor images. Free DVDs with sample data will be available.

Dragonfly Express high-framerate VGA camera - this camera produces images at 200 frames/second with VGA resolution over off-the-shelf 1394B interface technology. This camera is well-suited for high-speed tracking and provides impressive performance for a modest price. 200 frames/second operation will be demonstrated.

Point Grey Research makes a range of high quality 1394 cameras suitable for vision research. Resolutions range from 640x480 to 1600x1200. PGR cameras auto-synchronize with other cameras on the 1394 bus, making them ideal for multi-camera applications. The cameras are feature-packed and completely configurable via computer over a single interface. They are Linux compatible. A variety of cameras will be on-hand for demonstration to interested parties.

Please visit http://www.ptgrey.com for further information on Point Grey Research cameras.

^ Back to top