/ˈaː.ri/

Aditya

Arie

Nugraha

Research Scientist @ RIKEN-AIP (Japan) — Dr. in Informatics from University of Lorraine (France)

prof_pic.jpg

Kyoto University Artificial Intelligence Research Unit,

Dr. Ichikawa Commemorative Laboratory, Room #202,

Yoshida-honmachi, Sakyo-ku, Kyoto, 606-8501 JAPAN

I am a research scientist in the Sound Scene Understanding Team, Center for Advanced Intelligence Project (AIP), RIKEN and a visiting researcher in the Speech and Audio Processing Group, Kyoto University.

I received a Doctorate in Informatics from University of Lorraine, France for a doctoral research on multichannel audio source separation based on deep neural networks conducted at Inria Nancy – Grand-Est, France under the supervision of Dr. Antoine Liutkus and Dr. Emmanuel Vincent. The doctoral thesis covers the applications of our separation methods to various tasks, including speech enhancement, singing voice separation, and musical instrument separation.

My current research interests include audio source separation, audio-visual scene understanding, and machine learning.

news

Oct 24, 2023 Our paper “Time-Domain Audio Source Separation Based on Gaussian Processes with Deep Kernel Learning” was presented at IEEE WASPAA 2023.
Aug 20, 2023 Our team provided a tutorial entitled “Foundations, Extensions and Applications of Statistical Multichannel Speech Separation Models” at Interspeech 2023.
Jun 8, 2023 Our paper “Exploiting Sparse Recovery Algorithms for Semi-Supervised Training of Deep Neural Networks for Direction-of-Arrival Estimation” was presented at IEEE ICASSP 2023.
Oct 26, 2022 Our paper “Direction-Aware Adaptive Online Neural Speech Enhancement with an Augmented Reality Headset in Real Noisy Conversational Environments” was presented at IEEE/RSJ IROS 2022.
Sep 21, 2022 Our paper “Direction-Aware Joint Adaptation of Neural Speech Enhancement and Recognition in Real Multiparty Conversational Environments” was presented at Interspeech 2022.
Sep 7, 2022 We presented two papers at IWAENC 2022: ① “DNN-free Low-Latency Adaptive Speech Enhancement Based on Frame-Online Beamforming Powered by Block-Online FastMNMF” and ② “Joint Localization and Synchronization of Distributed Camera-Attached Microphone Arrays for Indoor Scene Analysis”.
May 7, 2022 Our Sound Scene Understanding Team presented two papers at IEEE ICASSP 2022: ① “Flow-Based Fast Multichannel Nonnegative Matrix Factorization for Blind Source Separation” and ② “Neural Full-Rank Spatial Covariance Analysis for Blind Source Separation”.
Apr 28, 2022 Our article “Generalized Fast Multichannel Nonnegative Matrix Factorization Based on Gaussian Scale Mixtures for Blind Source Separation” has been accepted for publication in IEEE/ACM Transactions on Audio, Speech, and Language Processing. It is now available on IEEE Xplore.
Apr 1, 2022 I’m happy to share that I’m starting a new position as Research Scientist (研究員) at RIKEN!
Jan 22, 2022 Our paper “Flow-Based Fast Multichannel Nonnegative Matrix Factorization for Blind Source Separation” has been accepted to IEEE ICASSP 2022.

selected publications

  1. WASPAA
    Time-Domain Audio Source Separation Based on Gaussian Processes with Deep Kernel Learning
    Aditya Arie Nugraha, Diego Di Carlo, Yoshiaki Bando, Mathieu Fontaine, and Kazuyoshi Yoshii
    In Proceedings of IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), 2023
  2. IROS
    Direction-Aware Adaptive Online Neural Speech Enhancement with an Augmented Reality Headset in Real Noisy Conversational Environments
    Kouhei Sekiguchi*, Aditya Arie Nugraha*, Yicheng Du, Yoshiaki Bando, Mathieu Fontaine, and Kazuyoshi Yoshii
    In Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022
  3. IWAENC
    DNN-Free Low-Latency Adaptive Speech Enhancement Based on Frame-Online Beamforming Powered by Block-Online FastMNMF
    In Proceedings of International Workshop on Acoustic Signal Enhancement (IWAENC), 2022
  4. ICASSP
    Flow-Based Fast Multichannel Nonnegative Matrix Factorization for Blind Source Separation
    In Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022
  5. SPL
    Neural Full-Rank Spatial Covariance Analysis for Blind Source Separation
    IEEE Signal Processing Letters, 2021
  6. TASLP

    15th IEEE Signal Processing Society (SPS) Japan Student Journal Paper Award

    Fast Multichannel Nonnegative Matrix Factorization with Directivity-Aware Jointly-Diagonalizable Spatial Covariance Matrices for Blind Source Separation
    IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2020
  7. SPL
    Flow-Based Independent Vector Analysis for Blind Source Separation
    IEEE Signal Processing Letters, 2020
  8. CSL

    ISCA Award for the Best Review Paper published in Computer Speech and Language (2016-2020)

    An analysis of environment, microphone and data simulation mismatches in robust speech recognition
    Computer Speech & Language, 2017
  9. TASLP

    6th IEEE Signal Processing Society (SPS) Japan Young Author Best Paper Award

    Multichannel audio source separation with deep neural networks
    Aditya Arie Nugraha, Antoine Liutkus, and Emmanuel Vincent
    IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2016