Rehearsals for our upcoming participation, 8/11 : 7 pm – NowNet Arts Conference Session Time – Temematic – NY/Paris – Concert Demonstration: Co-improvisation in the virtual space: corporeality, memory, and artificial intelligence, Cássia Carrascoza Bonfim, University of São Paulo, Mikhail Malt, Ircam-STMS, France, Somax2 Reach cocreative software
Auteur/autrice : Gerard Assayag
ARTE Documentary on the Future of Music
Elaine Chew and Gerard Assayag are part of director Anna Neuhaus’ ARTE documentary on The Future of Music focusing on what AI means for music, featuring jazz pianist Michael Wollny and classical pianist Kit Armstrong. This segment was filmed in Paris at IRCAM, the Institut de Recherche et Coordination Acoustique/Musique. Elaine is joined by Gérard Assayag (ERC […]
Simultaneous Music Separation and Generation Using Multi-Track Latent Diffusion Models
Read original: arXiv:2409.12346 – Published 20/09/2024 by Tornike Karchkhadze, Mohammad Rasool Izadi, Shlomo Dubnov. Abstract: Diffusion models have recently shown strong potential in both music generation and music source separation tasks. Although in early stages, a trend is emerging towards integrating these tasks into a single framework, as both involve generating musically aligned parts and can […]
Valérie Philippin Résidence REACH / Somax2
En 2023, Mikhail Malt a invité Valérie Philippin pour une résidence de recherche création à l’Ircam pendant la saison 2023-24, pour développer Somax2 autour de la (re)composition de la parole et créer une forme artistique mêlant les improvisations de l’intelligence artificielle et des deux musiciens. Une première étape à l’Ircam a permis de développer le logiciel […]
Creativity and Visual Communication from Machine to Musician: Sharing a Score through a Robotic Camera
Read original. Article published by Ross Greer, Laura Fleig, Shlomo Dubnov. Abstract: This paper explores the integration of visual communication and musical interaction by implementing a robotic camera within a « Guided Harmony » musical game. We aim to examine co-creative behaviors between human musicians and robotic systems. Our research explores existing methodologies like improvisational game pieces […]
Multi-Track MusicLDM: Towards Versatile Music Generation with Latent Diffusion Model
Read original. Published by Tornike Karchkhadze, Mohammad Rasool Izadi, Ke Chen, Gerard Assayag, Shlomo Dubnov. Abstract: Diffusion models have shown promising results in cross-modal generation tasks involving audio and music, such as text-to-sound and text-to-music generation. These text-controlled music generation models typically focus on generating music by capturing global musical attributes like genre and mood. […]
Improving Generalization of Speech Separation in Real-World Scenarios: Strategies in Simulation, Optimization, and Evaluation
View original. Published by Ke Chen, Jiaqi Su, Taylor Berg-Kirkpatrick, Shlomo Dubnov, Zeyu Jin. Abstract: Achieving robust speech separation for overlapping speakers in various acoustic environments with noise and reverberation remains an open challenge. Although existing datasets are available to train separators for specific scenarios, they do not effectively generalize across diverse real-world scenarios. In […]
Artistic Residence Valérie Philippin & Mikhail Malt
Restitution de la résidence de recherche artistique, de Valérie Philippin et Mikhail Malt, dans le cadre du projet REACH et présentation de Somax2 full video at https://medias.ircam.fr/x77ccfb_presentation-et-module-applicatif-suite-a In 2023, Mikhail Malt invited Valérie Philippin to a creative research residency at Ircam during the 2023-24 season, to develop Somax2 around the (re)composition of speech and create […]
Retrieval Guided Music Captioning via Multimodal Prefxes
Read full publication. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24)Special Track on AI, the Arts and Creativity. Published by Nikita Srivatsan, Ke Chen, Shlomo Dubnov and Taylor Berg-Kirkpatrick. Abstract: In this paper we put forward a new approach to music captioning, the task of automatically generating natural language descriptions for songs. […]
Improtech Paris – Tokyo is over.
Improtech is a musical festival and an interdisciplinary workshop bringing together research and creation in musician – machine intelligent musical interaction. After praised editions in New York, Philadelphia, Athens and Uzeste, Improtech Paris-Tokyo has landed down in Tokyo, on july 29, 2024, organised by REACH. This edition of Improtech in Tokyo has welcomed a number of brillant personalities during a week long […]