team members ming-chun chang lungisa matshoba steven preston supervisors dr james gain dr patrick...

25
Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

Upload: edwina-horton

Post on 21-Jan-2016

219 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

Team Members

Ming-Chun Chang

Lungisa Matshoba

Steven Preston

Supervisors

Dr James Gain

Dr Patrick Marais

Page 2: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

PROJECT OBJECTIVES

• Create application where 3D objects can be manipulated using hand gestures.

• Interface must be simple and intuitively easy to use.

• Translation, rotation and selection of objects must be possible.

Page 3: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

• 3D Objects to be manipulated will be molecules.

• Hand gestures will be input using a web camera.

• Set of hand gestures will be specified where each gesture relates to a specific task.

Page 4: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

• Project decomposed into three phases:

2D Image Processing (M. Chang)

Data Analysis Phase (S. Preston)

Front-end Visualisation (L. Matshoba)

Page 5: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

Skeletonization

NeuralNetwork

PrincipleComponent

Analysis

Visualisation and Testing

Sequence of 2D Images

Image of Basic Skeleton Model

Appropriate Hand Gesture Extraction

Motion Capture of User’s Hand

Appropriate Object Translation

Key

Input / Output

Process

Test Results

System Architecture

Page 6: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

•AIMS

–Feature extraction of sequence of hand images.

–Elimination of noise.

–Adequate performance in real-time

Page 7: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

•INPUT

–Logitech Webcam is used as the capturing device.

–Sequence of hand images captured by the webcam.

–Capable of capturing 30 frames per second.

Page 8: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

•IMPLEMENTATION

–Segmentation of image to isolate the hand.

–Image smoothing and filtering to eliminate noise.

–Threshold of image to isolate the desired features.

Page 9: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

•OUTPUT

–Sets of features extracted from the sequence of hand images.

–Basic representation of the hand structure.

Page 10: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

•CHALLENGES

–Efficient algorithm implementation capable of real-time processing.

–Clearing of background noise.

–Precise and accurate identification of hand features.

Page 11: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

•Success Factor

–Processing of 24 frames of hand images per second.

–95% accuracy of feature extraction.

–Elimination of noise.

Page 12: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

• AIMS– To analyse data provided by image processing phase.– Determine what hand gesture the user has carried out.

• 2 training methods will be used:– Neural Network– Principle Component Analysis (PCA)

Page 13: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

• A pattern classification problem.– Common application of training techniques such as neural networks and PCA.

• Many similar examples that suggest it is feasible:

– Face recognition using neural networks.

Page 14: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

• Input is data extracted from 2D images.

• Logitech Webcam captures at most 30 frames per second.

• Input will consist of representation of 24 frames.

Page 15: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

• PCA and neural network will require a training data set.

• Hundreds of inputs will be required.

• Not likely to pose a problem as data collection requires no expense and little resources.

Page 16: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

• Output will be provided for the front-end visualisation phase.

Simple output:

• One variable indicating the gesture that has been performed.

• Possible speed variable as well.

Page 17: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

PROBLEM

• Input data is captured from a single still camera – thus input data is in 2D form.

• But the users performs gestures in 3D world.

• Tilting and rotating of the hand could make it difficult to detect the correct gesture.

Page 18: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

SOLUTION

• Need appropriate set of gestures.

• Need a well designed neural network.

OTHER PROBLEMS?

• Speed and efficiency not a concern.

Page 19: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

• Whether neural network implementation recognises at least 95% hand gestures correctly.

• Whether PCA implementation recognises at least 95% hand gestures correctly.

• Whether hand gestures recognised agree with at least 95% of those recognised by polhemus tracker.

Page 20: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

AIMS- To produce a usable application for the gesture recognition interface

- Test the usability of the interface for a real world application

- Create a testing system to compare 2D and 3D gesture driven interface

- To test the usability of the gesture system.

Page 21: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

■ The front end visualization will deal with two main kinds of input.

■ Input from the 2D hand gestures as extracted by the Data Analysis Phase

■ Input from the 3D hand gestures – assumed to be more accurate.

■ A metric will be generated to measure the gesture recognition capabilities of the 2D hand gesture extraction.

Page 22: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

■ visual feedback of the system

■ Accuracy metric measuring difference between 2D and 3D gesture recognition

Page 23: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

■ Interface for the viewing of 3D molecule structures.

■ Different molecule level of detail views offered – selected regions more detailed

■ Visible section rotation

■ Seamless changes between ‘Ribbon’ and ‘Ball & Stick’ representations

Page 24: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

• Whether the system can run in real time.

• Accuracy of data extracted from 2D images.

• 95% hand gestures are recognised correctly.

• Whether motion capture and learning technique implementations agree on 95% of gestures.

Page 25: Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais

• Not entirely new concept.

• At the very least build an application where basic transformations can be done.

• Compare effectiveness of using learning techniques approach against motion-capture approach.