I am a speech technology researcher at Microsoft. My research has been evolving with the goal of creating computers that can recognize human speech in complex acoustic scenes. My current research focuses on multi-talker speech recognition, far-field conversational speech recognition, speaker diarization, ad-hoc microphone arrays, and meeting transcription. I helped the development of some capabilities of Microsoft Azure Speech Services, including Speech Devices SDK and Conversation Transcription.
Prior to joining Microsoft in 2016, I worked at NTT Communication Science Laboratories, where I co-invented the weighted prediction error (WPE) method, a dereverberation algorithm that is widely used in both the research community and the industry, with my colleagues. I also conducted research at the University of Cambridge as a Visiting Scholar in 2013 and worked for Doshisha University as a Part-Time Lecturer in 2015.
I have been a member of the Speech and Language Processing Technical Committee (SLTC) of the IEEE Signal Processing Society since 2018. I served as an organizing committee member for the REVERB Challenge and the accompanying workshop in 2013.
My publication list on this website may not be up to date. See my Google Scholar profile (linked below) instead.