{"id":979,"date":"2024-02-06T16:56:34","date_gmt":"2024-02-06T15:56:34","guid":{"rendered":"https:\/\/reach.ircam.fr\/?p=979"},"modified":"2024-02-29T18:12:01","modified_gmt":"2024-02-29T17:12:01","slug":"cocreative-interaction-somax2-and-the-reach-project","status":"publish","type":"post","link":"https:\/\/reach.ircam.fr\/index.php\/2024\/02\/06\/cocreative-interaction-somax2-and-the-reach-project\/","title":{"rendered":"Cocreative Interaction: Somax2 and the REACH Project"},"content":{"rendered":"\t\t<div data-elementor-type=\"wp-post\" data-elementor-id=\"979\" class=\"elementor elementor-979\">\n\t\t\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-1b9e083 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"1b9e083\" data-element_type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-74fa216\" data-id=\"74fa216\" data-element_type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-659fee4 elementor-widget elementor-widget-text-editor\" data-id=\"659fee4\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p><a href=\"https:\/\/direct.mit.edu\/comj\/article-abstract\/doi\/10.1162\/comj_a_00662\/119103\/Cocreative-Interaction-Somax2-and-the-REACH?redirectedFrom=fulltext\">Full publication<\/a><\/p><div>By G\u00e9rard Assayag, Laurent Bonnasse-Gahot, Joakim Borg<\/div><div>\u00a0<\/div><div class=\"al-author-name\"><strong>Abstract<\/strong>: Somax2 is an artificial intelligence (AI)-based multiagent system for human\u2013machine co-improvisation that generates stylistically coherent streams while continuously listening and adapting to musicians or other agents. The model on which it is based can be used with little configuration to interact with humans in full autonomy, but it also allows fine real-time control of its generative processes and interaction strategies, closer in this case to a \u201csmart\u201d digital instrument. An offspring of the Omax system, conceived at the Institut de Recherche et Coordination Acoustique\/Musique, the Somax2 environment is part of the European Research Council Raising Cocreativity in Cyber\u2013Human Musicianship (REACH) project, which studies distributed creativity as a general template for symbiotic interaction between humans and digital systems. It fosters mixed musical reality involving cocreative AI agents. The REACH project puts forward the idea that cocreativity in cyber\u2013human systems results from the emergence of complex joint behavior, produced by interaction and featuring cross-learning mechanisms. Somax2 is a first step toward this ideal, and already shows life-size achievements. This article describes Somax2 extensively, from its theoretical model to its system architecture, through its listening and learning strategies, representation spaces, and interaction policies.<\/div>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<\/div>\n\t\t","protected":false},"excerpt":{"rendered":"<p>Full publication By G\u00e9rard Assayag, Laurent Bonnasse-Gahot, Joakim Borg\u00a0Abstract: Somax2 is an artificial intelligence (AI)-based multiagent system for human\u2013machine co-improvisation that generates stylistically coherent streams while continuously listening and adapting to musicians or other agents. The model on which it is based can be used with little configuration to interact with humans in full autonomy, [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":1022,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"footnotes":""},"categories":[46,73],"tags":[],"class_list":["post-979","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-publications-research","category-software"],"aioseo_notices":[],"blog_post_layout_featured_media_urls":{"thumbnail":["https:\/\/reach.ircam.fr\/wp-content\/uploads\/2024\/02\/Capture_decran_le_2023-03-30_a_11.52.37_217kqSo-218x285-1-150x150.png",150,150,true],"full":["https:\/\/reach.ircam.fr\/wp-content\/uploads\/2024\/02\/Capture_decran_le_2023-03-30_a_11.52.37_217kqSo-218x285-1.png",218,285,false]},"categories_names":{"46":{"name":"Publications","link":"https:\/\/reach.ircam.fr\/index.php\/category\/research\/publications-research\/"},"73":{"name":"Software","link":"https:\/\/reach.ircam.fr\/index.php\/category\/research\/software\/"}},"tags_names":[],"comments_number":"0","wpmagazine_modules_lite_featured_media_urls":{"thumbnail":["https:\/\/reach.ircam.fr\/wp-content\/uploads\/2024\/02\/Capture_decran_le_2023-03-30_a_11.52.37_217kqSo-218x285-1-150x150.png",150,150,true],"cvmm-medium":["https:\/\/reach.ircam.fr\/wp-content\/uploads\/2024\/02\/Capture_decran_le_2023-03-30_a_11.52.37_217kqSo-218x285-1.png",218,285,false],"cvmm-medium-plus":["https:\/\/reach.ircam.fr\/wp-content\/uploads\/2024\/02\/Capture_decran_le_2023-03-30_a_11.52.37_217kqSo-218x285-1-218x207.png",218,207,true],"cvmm-portrait":["https:\/\/reach.ircam.fr\/wp-content\/uploads\/2024\/02\/Capture_decran_le_2023-03-30_a_11.52.37_217kqSo-218x285-1.png",218,285,false],"cvmm-medium-square":["https:\/\/reach.ircam.fr\/wp-content\/uploads\/2024\/02\/Capture_decran_le_2023-03-30_a_11.52.37_217kqSo-218x285-1.png",218,285,false],"cvmm-large":["https:\/\/reach.ircam.fr\/wp-content\/uploads\/2024\/02\/Capture_decran_le_2023-03-30_a_11.52.37_217kqSo-218x285-1.png",218,285,false],"cvmm-small":["https:\/\/reach.ircam.fr\/wp-content\/uploads\/2024\/02\/Capture_decran_le_2023-03-30_a_11.52.37_217kqSo-218x285-1-130x95.png",130,95,true],"full":["https:\/\/reach.ircam.fr\/wp-content\/uploads\/2024\/02\/Capture_decran_le_2023-03-30_a_11.52.37_217kqSo-218x285-1.png",218,285,false]},"_links":{"self":[{"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/posts\/979","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/comments?post=979"}],"version-history":[{"count":7,"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/posts\/979\/revisions"}],"predecessor-version":[{"id":1025,"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/posts\/979\/revisions\/1025"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/media\/1022"}],"wp:attachment":[{"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/media?parent=979"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/categories?post=979"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/reach.ircam.fr\/index.php\/wp-json\/wp\/v2\/tags?post=979"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}