eISSN:2278-5299

International Journal of Latest Research in Science and Technology

DOI:10.29111/ijlrst   ISRA Impact Factor:3.35

A News Letter Sign UP!
AN ENHANCED MULTIMODAL SOUND LOCALIZATION WITH HUMANLIKE AUDITORY SYSTEM FOR INTELLIGENT SERVICE ROBOTS

Research Paper Open Access

International Journal of Latest Research in Science and Technology Vol.2 Issue 6, pp 26-31,Year 2013

AN ENHANCED MULTIMODAL SOUND LOCALIZATION WITH HUMANLIKE AUDITORY SYSTEM FOR INTELLIGENT SERVICE ROBOTS

Keun-Chang Kwak

Correspondence should be addressed to :

Received : 25 December 2013; Accepted : 28 December 2013 ; Published : 31 December 2013

Share
Download 125
View 177
Article No. 10224
Abstract

This paper is concerned with an enhanced multimodal sound localization with humanlike auditory system for a network-based intelligent service robot, which exploits strong information technology infrastructure. The objective of this paper is to integrate several audiovisual-based Human-Robot Interaction (HRI) components that can naturally interact between human and robot through audiovisual information obtained from robot camera and microphones in the noisy environments or the presence of multiple persons. The proposed approach is comprised of two main stages. The first stage performs speech recognition, sound localization, and speaker recognition to know whether the user calls the robot or not as well as the direction and identification of the caller respectively, when someone calls robot’s name. In the second stage, an intelligent robot moves forward to the specific caller based on multiple face detection/recognition with the aid of the information identified by speaker recognition among multiple persons. The robot platform used in this study is WEVER, which is a network-based intelligent service robot developed in Electronics and Telecommunication Research Institute. This robot refers to an Ubiquitous Robotic Companion (URC) that provides necessary services anytime and anywhere. The effectiveness of the proposed approach is compared with other multimodal methods and sound localization itself.

Key Words   
Multimodal Sound Localization, Human-Robot Interaction, Ubiquitous Robotic Companion, Spe
Copyright
References
  1. G. Ha, J. C. Sohn, Y. J. Cho, and H. Yoon, “Towards ubiquitous robotic companion: design and implementation of ubiquitous robotic service framework”, ETRI Journal, vol. 27, no. 6, pp. 666-676, 2005.
  2. S. Choi, M. Kim, and H. D. Kim, “Probabilistic speaker localization in noisy environments by audio-visual integration”, Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 4704-4709, Beijing, Oct., 2006.
  3. Hara, F. Asano, Y. Kawai, F. Kanehiro, and K. Yamamoto, “Robust speech interface based on audio and video information fusion for humanoid HRP-2”, Proceedings of the 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp.2404-2410, Sendai, Sep., 2004.
  4. Huang, T. Supaongprapa, I. Terakura, F. Wang, N. Ohnishi, and N. Sugie, “A model-based sound localization system and its application to robot navigation”, Robotics and Autonomous Systems, vol. 27, pp. 199-209, 1999.
  5. H. Johnson and D. E. Dudgeon, Array signal processing: concepts and techniques, Prentice-Hall, Englewood Cliffs, 1993.
  6. O. Schmidt, “Multiple emitter location and signal parameter estimation”, IEEE Trans. on Antennas and Propagation, vol.34, no.3, pp. 276-280, 1986.
  7. Stoica and K.C. Sharman, “Maximum likelihood method for direction-of-arrival estimation”, IEEE Trans. on Acoustics and Speech Signal Processing, vol. 38, no.7, pp.1131-1143, 1990.
  8. M. Valin, F. Michaud, B. Hadjou, and J. Rouat, “Localization of simultaneous moving sound sources for mobile robot using a frequency-domain steered beamformer approach”, Proceeding of the 2004 IEEE International Conference on Robotics and Automation, pp. 1033-1038, 2004.
  9. Nakadi, H. G. Okuno, and H. Kitano, “Real-time sound source localization and separation for robot audition”, In Processing IEEE International Conference on Spoken Language Processing, pp. 193-196, 2002.
  10. Choi, J. Lee, S. Jeong, K. C. Kwak, S. Y. Chi, M. Hahn, “Multimodal sound source localization for intelligent service robot”, The 3rd International Conference on Ubiquitous Robots and Ambient Intelligence (URAI06), 2006.
  11. S. Lee, “Spontaneous dialogue recognition with large out-of-vocabularies”, IEEE Asia Pacific Conference on circuits and systems, pp. 247-251, 1996.
  12. Kim, M. Ji, H. Kim, K. C. Kwak, and S. Y. Chi, “Text-independent speaker recognition for ubiquitous robot companion”, The 3rd International Conference on Ubiquitous Robots and Ambient Intelligence (URAI06), 2006.
  13. H. Kim, H. S. Yoon, S. Y. Chi, Y. J. Cho, “Face identification for human robot interaction: intelligent security system for multi-user working environment on PC”, The 15th IEEE International Symposium on Robot and Human Interactive Communication (ROMAN06), pp. 617-622, 2006
  14. H. Kim, J. Y. Lee, H. S. Yoon, H. J. Kim, Y. J. Cho, and E. Y. Cha, “A vision-based user authentication system in robot environments by using semi-biometrics and tracking”, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2005), pp. 246-251, 2005.
  15. A. Reynolds and R. C. Rose, “Robust text-independent speaker identification using Gaussian mixture speaker models,” IEEE Trans. on Speech and Audio Processing, vol. 3, no. 1, pp. 72–83. 1995.
  16. A. Reynolds, T. F. Quatieri, and R. B. Dunn, “Speaker verification using adapted Gaussian mixture models,” Digital Signal Processing, vol. 10, pp. 19-41, 2000.
  17. C. Kwak, K. D. Ban, K. S Bae, H. J. Kim, S. Y. Chi, and Y. J. Cho, “Speech-based Human-Robot Interaction Components for URC intelligent service robots”, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2006), Video session, 2006.
To cite this article

Keun-Chang Kwak , " An Enhanced Multimodal Sound Localization With Humanlike Auditory System For Intelligent Service Robots ", International Journal of Latest Research in Science and Technology . Vol. 2, Issue 6, pp 26-31 , 2013


Responsive image

MNK Publication was founded in 2012 to upholder revolutionary ideas that would advance the research and practice of business and management. Today, we comply with to advance fresh thinking in latest scientific fields where we think we can make a real difference and growth now also including medical and social care, education,management and engineering.

Responsive image

We offers several opportunities for partnership and tie-up with individual, corporate and organizational level. We are working on the open access platform. Editors, authors, readers, librarians and conference organizer can work together. We are giving open opportunities to all. Our team is always willing to work and collaborate to promote open access publication.

Responsive image

Our Journals provide one of the strongest International open access platform for research communities. Our conference proceeding services provide conference organizers a privileged platform for publishing extended conference papers as journal publications. It is deliberated to disseminate scientific research and to establish long term International collaborations and partnerships with academic communities and conference organizers.