Conferences Publications Research

Introducing EG-IPT and ipt~: a novel electric guitar dataset and a new Max/MSP object for real-time classification of instrumental playing techniques

Article by Marco Fiorini*, Nicolas Brochec*, Joakim Borg and Riccardo Pasini (* equal contribution) has been accepted for the New Interfaces for Musical Expression (NIME) Conference in Canberra, Australia.

Read the full paper

Video demo

Abstract: This paper presents two key contributions to the real-time classi- fication of Instrumental Playing Techniques (IPTs) in the context of NIME and human-machine interactive systems: the EG-IPT dataset and the ipt∼ Max/MSP object. The EG-IPT dataset, specif- ically designed for electric guitar, encompasses a broad range of IPTs captured across six distinct audio sources (five microphones and one direct input) and three pickup configurations. This di- versity in recording conditions provides a robust foundation for training accurate models. We evaluate the dataset by employing a Convolutional Neural Network-based classifier (CNN), achieving state-of-the-art performance across a wide array of IPT classes, thereby validating the dataset’s efficacy. The ipt∼ object is a new Max/MSP external enabling real-time classification of IPTs via pre-trained CNN models. While in this paper it’s demonstrated with the EG-IPT dataset, the ipt∼ object is adaptable to models trained on various instruments. By integrating EG-IPT and ipt∼, we introduce a novel, end-to-end workflow that spans from data collection, model training to real-time classification and human- computer interaction. This workflow exemplifies the entangle- ment of diverse components (data acquisition, machine learning, real-time processing, and interactive control) within a unified system, advancing the potential for dynamic, real-time music performance and human-computer interaction in the context of NIME.

Laisser un commentaire

Votre adresse e-mail ne sera pas publiée. Les champs obligatoires sont indiqués avec *