Sound Source Localization and Separation for Emotional Robot

감성로봇을 위한 음원의 위치측정 및 분리

  • 김경환 (한국과학기술원 기계공학과 대학원) ;
  • 김연훈 (한국과학기술원 기계공학과 대학원) ;
  • 곽윤근 (한국과학기술원 기계공학과)
  • Published : 2003.05.01

Abstract

These days, the researches related with the emotional robots are actively investigated and in progress. And human language, expression, action etc. are merged in the emotional robot to understand the human emotion. However, there are so many sound sources and background noise around the robot, that the robots should be able to separate the mixture of these sound sources into the original sound sources, moreover to understand the meaning of voice of a specific person. Also they should be able to turn or move to the direction of a specific person to observe his expression or action effectively. Until now, the researches on the localization and separation of sound sources have been so theoretical and computative that real-time processing is hardly possible. In this reason for the practical emotional robot, fast computation should be realized by using simple principle. In this paper the methods for detecting the direction of sound sources by using the phase difference between peaks on spectrums, and the separating the sound sources by using fundamental frequency and its overtones of human voice, are proposed. Also by using these methods, it is shown that the effective and real-time localization and separation of sound sources in living room are possible.

Keywords

References

  1. Lee, K. H., 'Human stability and sensible technology,' J. of KSPE, Vol. 18, No. 2, pp. 9-13, 2001
  2. Bregman, A. S., 'Auditory scene analysis: the perceptual organization of sound,' MIT Press, Cambridge, MA, 1990
  3. Bregman, A. S., 'Auditory scene analysis: hearing in complex environment' in McAdams and Bigand(eds.) 'Thinking in sounds,' Oxford University press, New York, 1993
  4. Nakatani, T. and Okuno, H. G., 'Harmonic sound stream segregation using localization and its application to speech stream segregation,' Speech Communication 27, pp. 209-222, 1999 https://doi.org/10.1016/S0167-6393(98)00079-X
  5. Huang, J., et al., 'Echo avoidance in a computational model of the precedence effect,' Speech Communication 27, pp. 223-233, 1999 https://doi.org/10.1016/S0167-6393(98)00075-2
  6. Unoki, M. and Akagi, M., 'A method of signal extraction from noisy signal based on auditory scene analysis,' Speech Communication 27, pp. 261-279, 1999 https://doi.org/10.1016/S0167-6393(98)00077-6
  7. Nakadai, K. et al., 'Humanoid active audition system,' IEEE Humanoids, 2000