MENU

Nobutaka Ono Laboratory

Professor Nobutaka Ono

Professor Nobutaka Ono

Microphone Array / Source Separation / Acoustic Scene Recognition / Time-Frequency Analysis / Acoustic Signal Processing

Sound is an important media containing a variety of information, and in everyday life, we communicate with speech, enjoy music, and perceive various situations around us through sound. Aiming to realize the advanced acoustic information processing similar to or exceeding that of humans, Ono Laboratory is working on signal processing and information processing for sounds such as speech and music. Some specific research topics are as follows.


Blind Source Separation

In the real environment, various sounds can be heard as a mixture. But for example, speech recognition does not work well if noise or multiple speech are overlapped. Blind source separation is a technique to separate such mixed sounds into individual sound sources without prior information. The conventional methods need many computations, but Ono lab developed the world's fastest algorithm called auxiliary function based independent vector analysis (AuxIVA) in 2011, and implemented it as an iPhone application to prove its high speed.

In recent years, aiming at hearing aid applications, we are working on developing low-latency algorithms and real-time implementations.

Fast blind source separation on iPhone

Asynchronous Distributed Microphone Array

A system that acquires and processes spatial information of sound using multiple microphones is called a microphone array. In a microphone array, a minute time difference between multiple microphone signals (for example, about 0.1 ms) is an important cue for sound source localization and sound source separation, so it was necessary for multiple microphones to record the sound synchronously. On the other hand, there are many devices with recording functions around us (PCs, smartphones, etc). We are studying new signal processing techniques for array signal processing on such non-synchronized recording devices. For example, we aim an application such that, if meeting participants record the conversation with their smartphones and upload them to the cloud after the meeting, the signals are automatically synchronized, the sound sources are separated, each speech is recognized, and the meeting minutes are sent by e-mail.

Acoustic Sensing Based on Sound-Light Conversion and Video Camera

We are starting to study new acoustic sensing based on special sensor nodes that illuminate LEDs according to the sound intensity, and a video camera. In this framework, the video camera works as a super multi-channel acoustic device. Sound spectrum information cannot be obtained, but instead, a wide range of spatial information can be acquired at once. Because security cameras, surveillance cameras, etc. can be used, this sensing will be widely applicable for sound environment recognition and abnormality detection. We have already developed an inexpensive and small device with a battery that can work alone.

Back to previous page