Multi-Modal Human-Machine Communication for Instructing Robot Grasping Tasks

Computer Science – Human-Computer Interaction

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

7 pages, 8 figures

Scientific paper

A major challenge for the realization of intelligent robots is to supply them with cognitive abilities in order to allow ordinary users to program them easily and intuitively. One way of such programming is teaching work tasks by interactive demonstration. To make this effective and convenient for the user, the machine must be capable to establish a common focus of attention and be able to use and integrate spoken instructions, visual perceptions, and non-verbal clues like gestural commands. We report progress in building a hybrid architecture that combines statistical methods, neural networks, and finite state machines into an integrated system for instructing grasping tasks by man-machine interaction. The system combines the GRAVIS-robot for visual attention and gestural instruction with an intelligent interface for speech recognition and linguistic interpretation, and an modality fusion module to allow multi-modal task-oriented man-machine communication with respect to dextrous robot manipulation of objects.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

Multi-Modal Human-Machine Communication for Instructing Robot Grasping Tasks does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with Multi-Modal Human-Machine Communication for Instructing Robot Grasping Tasks, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Multi-Modal Human-Machine Communication for Instructing Robot Grasping Tasks will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-72643

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.