logo Idiap Research Institute        
 [BibTeX] [Marc21]
Speech/Non-Speech Detection in Meetings from Automatically Extracted Low Resolution Visual Features
Type of publication: Idiap-RR
Citation: Hung_Idiap-RR-20-2009
Number: Idiap-RR-20-2009
Year: 2009
Month: 7
Institution: Idiap
Note: submitted to icmi-mlmi
Abstract: In this paper we address the problem of estimating who is speaking from automatically extracted low resolution visual cues from group meetings. Traditionally, the task of speech/non-speech detection or speaker diarization tries to find who speaks and when from audio features only. Recent work has addressed the problem audio-visually but often with less emphasis on the visual component. Due to the high probability of losing the audio stream during video conferences, this work proposes methods for estimating speech using just low resolution visual cues. We carry out experiments to compare how context through the observation of group behaviour and task-oriented activities can help improve estimates of speaking status. We test on 105 minutes of natural meeting data with unconstrained conversations.
Keywords:
Projects Idiap
IM2
Authors Hung, Hayley
Ba, Silèye O.
Added by: [ADM]
Total mark: 0
Attachments
  • Hung_Idiap-RR-20-2009.pdf (MD5: 16799bb2e79691af71fce12862d18708)
Notes