Publications

Simultaneous Music Separation and Generation Using Multi-Track Latent Diffusion Models

Read original: arXiv:2409.12346 – Published 20/09/2024 by Tornike Karchkhadze, Mohammad Rasool Izadi, Shlomo Dubnov. Abstract: Diffusion models have recently shown strong potential in both music generation and music source separation tasks. Although in early stages, a trend is emerging towards integrating these tasks into a single framework, as both involve generating musically aligned parts and can […]

Media Music Residences Workshops

Valérie Philippin Résidence REACH / Somax2

En 2023, Mikhail Malt a invité Valérie Philippin pour une résidence de recherche création à l’Ircam pendant la saison 2023-24, pour développer Somax2 autour de la (re)composition de la parole et créer une forme artistique mêlant les improvisations de l’intelligence artificielle et des deux musiciens. Une première étape à l’Ircam a permis de développer le logiciel […]

Publications

Creativity and Visual Communication from Machine to Musician: Sharing a Score through a Robotic Camera

Read original. Article published by Ross Greer, Laura Fleig, Shlomo Dubnov. Abstract: This paper explores the integration of visual communication and musical interaction by implementing a robotic camera within a « Guided Harmony » musical game. We aim to examine co-creative behaviors between human musicians and robotic systems. Our research explores existing methodologies like improvisational game pieces […]

Publications

Multi-Track MusicLDM: Towards Versatile Music Generation with Latent Diffusion Model

Read original. Published by Tornike Karchkhadze, Mohammad Rasool Izadi, Ke Chen, Gerard Assayag, Shlomo Dubnov. Abstract: Diffusion models have shown promising results in cross-modal generation tasks involving audio and music, such as text-to-sound and text-to-music generation. These text-controlled music generation models typically focus on generating music by capturing global musical attributes like genre and mood. […]

Publications

Improving Generalization of Speech Separation in Real-World Scenarios: Strategies in Simulation, Optimization, and Evaluation

View original. Published by Ke Chen, Jiaqi Su, Taylor Berg-Kirkpatrick, Shlomo Dubnov, Zeyu Jin. Abstract: Achieving robust speech separation for overlapping speakers in various acoustic environments with noise and reverberation remains an open challenge. Although existing datasets are available to train separators for specific scenarios, they do not effectively generalize across diverse real-world scenarios. In […]

Residences Workshops

Artistic Residence Valérie Philippin & Mikhail Malt

Restitution de la résidence de recherche artistique, de Valérie Philippin et Mikhail Malt, dans le cadre du projet REACH et présentation de Somax2 full video at https://medias.ircam.fr/x77ccfb_presentation-et-module-applicatif-suite-a In 2023, Mikhail Malt invited Valérie Philippin to a creative research residency at Ircam during the 2023-24 season, to develop Somax2 around the (re)composition of speech and create […]

Publications

Retrieval Guided Music Captioning via Multimodal Prefxes

Read full publication. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24)Special Track on AI, the Arts and Creativity. Published by Nikita Srivatsan, Ke Chen, Shlomo Dubnov and Taylor Berg-Kirkpatrick. Abstract: In this paper we put forward a new approach to music captioning, the task of automatically generating natural language descriptions for songs. […]

Concerts Conferences Events

Improtech Paris – Tokyo is over.

Improtech is a musical festival and an interdisciplinary workshop bringing together research and creation in musician – machine intelligent musical interaction. After praised editions in New York, Philadelphia, Athens and Uzeste, Improtech Paris-Tokyo has landed down in Tokyo, on july 29, 2024, organised by REACH. This edition of Improtech in Tokyo has welcomed a number of brillant personalities during a week long […]

Publications

Musicldm: Enhancing novelty in text-to-music generation using beat-synchronous mixup strategies

Read full publication. Published by Ke Chen, Yusong Wu, Haohe Liu, Marianna Nezhurina, Taylor Berg-Kirkpatrick, Shlomo Dubnov. Abstract: Diffusion models have shown promising results in cross-modal generation tasks, including text-to-image and text-to-audio generation. However, generating music, as a special type of audio, presents unique challenges due to limited availability of music data and sensitive issues […]