Read original: arXiv:2409.12346 – Published 20/09/2024 by Tornike Karchkhadze, Mohammad Rasool Izadi, Shlomo Dubnov. Abstract: Diffusion models have recently shown strong potential in both music generation and music source separation tasks. Although in early stages, a trend is emerging towards integrating these tasks into a single framework, as both involve generating musically aligned parts and can […]
Publications
Creativity and Visual Communication from Machine to Musician: Sharing a Score through a Robotic Camera
Read original. Article published by Ross Greer, Laura Fleig, Shlomo Dubnov. Abstract: This paper explores the integration of visual communication and musical interaction by implementing a robotic camera within a « Guided Harmony » musical game. We aim to examine co-creative behaviors between human musicians and robotic systems. Our research explores existing methodologies like improvisational game pieces […]
Multi-Track MusicLDM: Towards Versatile Music Generation with Latent Diffusion Model
Read original. Published by Tornike Karchkhadze, Mohammad Rasool Izadi, Ke Chen, Gerard Assayag, Shlomo Dubnov. Abstract: Diffusion models have shown promising results in cross-modal generation tasks involving audio and music, such as text-to-sound and text-to-music generation. These text-controlled music generation models typically focus on generating music by capturing global musical attributes like genre and mood. […]
Improving Generalization of Speech Separation in Real-World Scenarios: Strategies in Simulation, Optimization, and Evaluation
View original. Published by Ke Chen, Jiaqi Su, Taylor Berg-Kirkpatrick, Shlomo Dubnov, Zeyu Jin. Abstract: Achieving robust speech separation for overlapping speakers in various acoustic environments with noise and reverberation remains an open challenge. Although existing datasets are available to train separators for specific scenarios, they do not effectively generalize across diverse real-world scenarios. In […]
Retrieval Guided Music Captioning via Multimodal Prefxes
Read full publication. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24)Special Track on AI, the Arts and Creativity. Published by Nikita Srivatsan, Ke Chen, Shlomo Dubnov and Taylor Berg-Kirkpatrick. Abstract: In this paper we put forward a new approach to music captioning, the task of automatically generating natural language descriptions for songs. […]
Musicldm: Enhancing novelty in text-to-music generation using beat-synchronous mixup strategies
Read full publication. Published by Ke Chen, Yusong Wu, Haohe Liu, Marianna Nezhurina, Taylor Berg-Kirkpatrick, Shlomo Dubnov. Abstract: Diffusion models have shown promising results in cross-modal generation tasks, including text-to-image and text-to-audio generation. However, generating music, as a special type of audio, presents unique challenges due to limited availability of music data and sensitive issues […]
Microphone-based Data Augmentation for Automatic Recognition of Instrumental Playing Techniques
Paper written by Nicolas Brochec*, Tsubasa Tanaka*, and Will Howie*† has been presented at the ICMC2024 (International Computer Music Conference 2024). * Tokyo University of the Arts† Japan Society for the Promotion of Science Full publication Abstract: Within existing research on the automatic classification of musical instrument playing techniques, few available datasets include enough playing […]
Maths & Musique #3 : Musique, combinatoire des mots et improvisation par ordinateur, par Marc Chemillier
Lire la publication originale. Par Marc Chemillier, Directeur d’études de l’EHESS La musique entretient depuis toujours des liens étroits avec les mathématiques. Les pythagoriciens à partir du Ve siècle avant J.-C. commencent à s’intéresser aux rapports entre la musique et les nombres. Beaucoup plus proche de nous, Joseph Fourier (1768-1830) jette les bases de l’analyse harmonique […]
Maths & Musique #1 : Les maths dans la musique, la musique des maths : réflexions autour de la « dynamique mathémusicale », par Moreno Andreatta
Lire l’article original. Par Moreno Andreatta, Directeur de recherche CNRS. Peut-on partir des problèmes posés par la musique pour faire avancer la recherche en mathématique et, via les maths, stimuler la créativité musicale ? Voilà une question qui aurait sans doute fait sourire les scientifiques et musiciennes ou musiciens professionnels il y a juste une trentaine […]
Maths & Musique : une série d’articles mêlant science et art
Lire l’article original. 1, 2, 3, 4 : comment adopter le bon tempo ? Pourquoi le son est-il meilleur dans une salle de concert que dans notre salon ? Pourquoi les styles musicaux ont des battements par minute (bpm) différents ? En quoi le rythme musical, ses fréquences et ses harmonies, sont des grandeurs mathématiques […]