{"id":14,"date":"2025-01-01T00:00:19","date_gmt":"2025-01-01T00:00:19","guid":{"rendered":"https:\/\/wp.lancs.ac.uk\/botornot\/?page_id=14"},"modified":"2026-02-21T23:08:36","modified_gmt":"2026-02-21T23:08:36","slug":"music-editions","status":"publish","type":"page","link":"https:\/\/wp.lancs.ac.uk\/botornot\/home\/music-editions\/","title":{"rendered":"Music editions"},"content":{"rendered":"<p>The <a href=\"https:\/\/lancasteruni.eu.qualtrics.com\/jfe\/form\/SV_78vpS0sTfeOovK6\"><em>Music Edition<\/em><\/a> (v1) of <strong>Bot or Not?<\/strong> (est. 2025) forms part of an ESRC-supported programme of research led by <a href=\"http:\/\/www.lancaster.ac.uk\/linguistics\/about\/people\/claire-hardaker\">Prof Claire Hardaker<\/a>, <a href=\"http:\/\/www.lancaster.ac.uk\/linguistics\/about\/people\/georgina-brown\">Dr Georgina Brown<\/a>, and <a href=\"http:\/\/www.lancaster.ac.uk\/linguistics\/about\/people\/hope-mcvean\">Hope McVean<\/a> at Lancaster University. It examines a deceptively simple question:<\/p>\n<blockquote><p><strong>how accurately can listeners distinguish between AI-generated and human singing voices, and what cues inform those judgements?<\/strong><\/p><\/blockquote>\n<div id=\"attachment_64\" style=\"width: 260px\" class=\"wp-caption alignright\"><a href=\"http:\/\/wp.lancs.ac.uk\/botornot\/files\/2026\/02\/qr_bonme_v1.png\"><img loading=\"lazy\" decoding=\"async\" aria-describedby=\"caption-attachment-64\" class=\"wp-image-64 size-full\" src=\"http:\/\/wp.lancs.ac.uk\/botornot\/files\/2026\/02\/qr_bonme_v1.png\" alt=\"BoNME QR code\" width=\"250\" height=\"250\" srcset=\"https:\/\/wp.lancs.ac.uk\/botornot\/files\/2026\/02\/qr_bonme_v1.png 250w, https:\/\/wp.lancs.ac.uk\/botornot\/files\/2026\/02\/qr_bonme_v1-150x150.png 150w\" sizes=\"auto, (max-width: 250px) 100vw, 250px\" \/><\/a><p id=\"caption-attachment-64\" class=\"wp-caption-text\">Want to share this quiz?<br \/>Feel free to use this QR code<\/p><\/div>\n<p>While synthetic <em>speech<\/em> has now reached extremely high levels of perceptual plausibility, singing presents a more complex test case. It combines pitch stability, vibrato, breath control, phrasing, timbral nuance, stylistic convention, and emotional projection &#8211; all of which may function as perceptual signals of \u201chumanness\u201d. The project investigates whether these signals remain diagnostic as generative models improve, or if they are also gradually being removed from our AI-detection toolkit.<\/p>\n<p>The wider relevance is non-trivial. If AI-generated vocals become reliably indistinguishable from human performance, the implications extend across:<\/p>\n<ul>\n<li>performer identity and attribution<\/li>\n<li>copyright and neighbouring rights<\/li>\n<li>fraud and impersonation risks<\/li>\n<li>creative labour markets<\/li>\n<li>export and regulatory frameworks<\/li>\n<\/ul>\n<p>Public controversies &#8211; such as the <a href=\"https:\/\/www.bbc.co.uk\/news\/articles\/cp8mjnn7eqno\">The Velvet Sundown<\/a> case &#8211; illustrate how quickly questions of authorship, authenticity, and disclosure become socially and legally charged.<\/p>\n<h2>Design overview<\/h2>\n<p>The Music Edition draws from a curated larger bank of both human-created and AI-generated songs. Each participant is randomly presented with <strong>nine samples<\/strong>, which may comprise any combination (all human, all AI, or mixed). For each sample, participants:<\/p>\n<ol>\n<li>Classify the voice as <strong>human<\/strong> or <strong>AI-generated<\/strong><\/li>\n<li>Provide <strong>confidence ratings<\/strong> (pre- and post-task)<\/li>\n<li>Offer <strong>qualitative explanations<\/strong> of the cues or reasoning underlying their decisions (submitted prior to score feedback)<\/li>\n<\/ol>\n<p>This structure enables analysis at multiple levels:<\/p>\n<ul>\n<li><strong>Overall detection accuracy<\/strong><\/li>\n<li><strong>Calibration of confidence vs. performance<\/strong><\/li>\n<li><strong>Metacognitive shift<\/strong> (change in confidence across exposure)<\/li>\n<li><strong>Cue salience<\/strong> (via thematic and corpus-linguistic analysis of explanations)<\/li>\n<li><strong>Error patterns<\/strong> across voice types, genres, or production styles<\/li>\n<\/ul>\n<p>Importantly, participants are instructed to focus on the voice only. Instrumentation may be synthetic in both human- and AI-voiced tracks, and human performances may legitimately include processing effects (e.g., pitch correction). The experimental question is not whether production sounds modern, but whether vocal origin is perceptually identifiable.<\/p>\n<h3>Current performance snapshot<\/h3>\n<p>How well do people do at this particular quiz? Our latest summary statistics (Feb 2026) are as follows:<\/p>\n<table style=\"border-collapse: collapse;width: 100%;height: 48px\">\n<tbody>\n<tr style=\"height: 24px\">\n<td style=\"width: 16.6667%;height: 24px\"><strong>Responses<\/strong><\/td>\n<td style=\"width: 16.6667%;height: 24px\"><strong>Lowest score<\/strong><\/td>\n<td style=\"width: 16.6667%;height: 24px\"><strong>Highest score<\/strong><\/td>\n<td style=\"width: 16.6667%;height: 24px\"><strong>Mean<\/strong><\/td>\n<td style=\"width: 16.6667%;height: 24px\"><strong>SD<\/strong><\/td>\n<td style=\"width: 16.6667%;height: 24px\"><strong>Variance<\/strong><\/td>\n<\/tr>\n<tr style=\"height: 24px\">\n<td style=\"width: 16.6667%;height: 24px\">374<\/td>\n<td style=\"width: 16.6667%;height: 24px\">0<\/td>\n<td style=\"width: 16.6667%;height: 24px\">9<\/td>\n<td style=\"width: 16.6667%;height: 24px\">4.90<\/td>\n<td style=\"width: 16.6667%;height: 24px\">1.48<\/td>\n<td style=\"width: 16.6667%;height: 24px\">2.20<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"size-full wp-image-47 aligncenter\" src=\"http:\/\/wp.lancs.ac.uk\/botornot\/files\/2026\/02\/75b9138b-2db8-43a8-9975-88e8a553104b.jpg\" alt=\"Feb 2026: BoNME results\" width=\"720\" height=\"296\" srcset=\"https:\/\/wp.lancs.ac.uk\/botornot\/files\/2026\/02\/75b9138b-2db8-43a8-9975-88e8a553104b.jpg 720w, https:\/\/wp.lancs.ac.uk\/botornot\/files\/2026\/02\/75b9138b-2db8-43a8-9975-88e8a553104b-300x123.jpg 300w, https:\/\/wp.lancs.ac.uk\/botornot\/files\/2026\/02\/75b9138b-2db8-43a8-9975-88e8a553104b-676x278.jpg 676w\" sizes=\"auto, (max-width: 720px) 100vw, 720px\" \/><\/p>\n<h2>Data and collaboration<\/h2>\n<p>The dataset comprises perceptual judgements, confidence metrics, and free-text reasoning data linked to known ground truth labels. The project is designed to support:<\/p>\n<ul>\n<li>perceptual and cognitive research on synthetic media<\/li>\n<li>linguistic analysis of AI vocal production<\/li>\n<li>human-AI discrimination modelling<\/li>\n<li>regulatory and policy-oriented research<\/li>\n<li>risk assessment in an array of contexts<\/li>\n<\/ul>\n<p>At its core, the Music Edition asks a simple question with increasingly complex consequences: <strong>when we hear singing, what makes us believe there is a person behind it, and when are we wrong?<\/strong> It also takes us into deeply philosophical questions around the nature of being and identity: must an artist really exist for the associated art to be perceived as legitimate?<\/p>\n<p>We believe that the answers to these questions are no longer merely aesthetic. They are becoming both existential and evidential.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>The Music Edition (v1) of Bot or Not? (est. 2025) forms part of an ESRC-supported programme of research led by Prof Claire Hardaker, Dr Georgina Brown, and Hope McVean at Lancaster University. It examines a deceptively simple question: how accurately can listeners distinguish between AI-generated and human singing voices, and what cues inform those judgements? [&hellip;]<\/p>\n","protected":false},"author":77,"featured_media":45,"parent":2,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-14","page","type-page","status-publish","has-post-thumbnail","hentry"],"_links":{"self":[{"href":"https:\/\/wp.lancs.ac.uk\/botornot\/wp-json\/wp\/v2\/pages\/14","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/wp.lancs.ac.uk\/botornot\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/wp.lancs.ac.uk\/botornot\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/wp.lancs.ac.uk\/botornot\/wp-json\/wp\/v2\/users\/77"}],"replies":[{"embeddable":true,"href":"https:\/\/wp.lancs.ac.uk\/botornot\/wp-json\/wp\/v2\/comments?post=14"}],"version-history":[{"count":10,"href":"https:\/\/wp.lancs.ac.uk\/botornot\/wp-json\/wp\/v2\/pages\/14\/revisions"}],"predecessor-version":[{"id":111,"href":"https:\/\/wp.lancs.ac.uk\/botornot\/wp-json\/wp\/v2\/pages\/14\/revisions\/111"}],"up":[{"embeddable":true,"href":"https:\/\/wp.lancs.ac.uk\/botornot\/wp-json\/wp\/v2\/pages\/2"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/wp.lancs.ac.uk\/botornot\/wp-json\/wp\/v2\/media\/45"}],"wp:attachment":[{"href":"https:\/\/wp.lancs.ac.uk\/botornot\/wp-json\/wp\/v2\/media?parent=14"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}