3W for HRI

From Robot Intelligence
Jump to: navigation, search

Project Outline

  • Research Period: 2012.6 ~ 2017.5
  • Funded by the Ministry of Trade, Industry and Energy (Grant No: 10041629)

Introduction and Research Targets

1. This project is for the purpose of the implementation of technologies for identification (WHO), behavior (WHAT) and location (WHERE) of human based sensor network fusion program.
2. In this project we develop a robot-assisted management system for promptly coping with abnormal events in classroom environments.

- Reliably detect the occurrence of human-caused emergency situations via audio-visual perception modules
- Make an urgent SMS transmission to let someone know that an emergency event occurs, and relay spot information to them
- Perform an immediate reaction for the happening by using robot navigation and interaction technologies on behlaf of the remote user
Simonpic overview.png

Developing Core Technologies


  • WHERE: Human detection and localization
  • WHO: Face recognition and ID tracking
  • WHAT: Recognition of individual and group behavior
  • 3W data association on the perception sensor network


  • Automatic message transmission for human-caused emergencies
  • Analysis of student attitude
  • Remote monitoring via Web technologies and stream server


  • Human-friendly robot behavior
  • Gaze control and robot navigation
  • Human following with recovery mechanism


SimonPiC Remote handling for emergency on 3w data.png

(SimonPiC) 3W final.png

(SimonPiC) Audio-Visual Integration.png

(SimonPiC) WHERE WHO Fusion.jpg

(SimonPiC3) Human Leg Detection.jpg

(SimonPiC2) 3W FUSION Simulation.jpg