Although the results from Tuesday’sAdaptistration Classical Music Challenge are unlikely to be the final word on the matter, it appears that it’s still pretty easy to tell the difference between an orchestra of human beings and one of computer generated sounds…
The challenge consisted of participants listening to two different recordings of Edvard Greig’s Holberg Suite, Mvt. #1. One version was created using the Vienna Symphonic Library and the other one with a human orchestra. Afterward, participants identified which recording was which or if they couldn’t tell the difference.
27 readers had the time to compare each recording and 25/27 (93%) were able to accurately identify each sample, 1/27 (4%) was incorrect, and one 1/27 (4%) could not tell the difference, as the chart below illustrates.
(click to enlarge)
I think this was a worthwhile test because the Guardian Unlimited technology correspondent, David Smith, claimed that new Vienna Symphonic Library software package is capable of mimicking “human musicians in the performance of greats such as Bach, Beethoven and Mozart so convincingly that a casual listener to Classic FM would be unable to tell the difference”.
Apparently, the casual listener that also reads this blog would beg to differ.
If you missed the challenge and didn’t vote, no worries, you can still feel free to compare the two recordings:
Drew,
I listened to both clips and could tell immediately which one was the real thing…but, I do have a background in classical music. I played the clips for my wife and a friend of ours (both who have zero formal music training) and they both had no trouble telling the difference. They picked it out after listening for about 6 seconds.
So – I think the message is that although your reading audience is predisposed to be musically inclined, and thus will likely be a bit better than average at picking out the “real” performance, even those who have no formal training are clearly capable of picking out the human vs. android orchestra.
An additional observation: if we were allowed to listen to only a very small sample of the sound clip (like 1/4 of a second), I think the numbers would be drastically different. The actual tone doesn’t sound all that different; it’s the subtelties, the way the orchestra swells and recedes, the accenting of specific notes that provide context for the meter, and so forth. The real orchestra “plays” the rests – the android orchestra just suspends animation.
This experiment reminds me of some research I uncovered for my Master’s thesis. I did work on timbre theory, approaching it from a neurocognitive perspective, and read a lot about various timbral subtleties that allow the brain to differentiate sounds (mainly to identify them).
Incidentally, the human brain is capable of identifying discrepancies in synchony of attacks up to 15 milliseconds! In other words, a difference of 15 milliseconds between the entrance of one player and another can be heard. I think that particular aspect plays a significant role in distinguishing the two ensembles in your listening samples.
Everyone should check out this story from the Guardian: http://arts.guardian.co.uk/news/story/0,,1854371,00.html
My fears about commercial applications of this have already become reality.
Drew,
I listened to both clips and could tell immediately which one was the real thing…but, I do have a background in classical music. I played the clips for my wife and a friend of ours (both who have zero formal music training) and they both had no trouble telling the difference. They picked it out after listening for about 6 seconds.
So – I think the message is that although your reading audience is predisposed to be musically inclined, and thus will likely be a bit better than average at picking out the “real” performance, even those who have no formal training are clearly capable of picking out the human vs. android orchestra.
An additional observation: if we were allowed to listen to only a very small sample of the sound clip (like 1/4 of a second), I think the numbers would be drastically different. The actual tone doesn’t sound all that different; it’s the subtelties, the way the orchestra swells and recedes, the accenting of specific notes that provide context for the meter, and so forth. The real orchestra “plays” the rests – the android orchestra just suspends animation.
My two cents.
Jeff
Listening to 1/4 of a second of sound. I like that idea, it made me smile as the thought of musician fracking a note while playing Cage’s 4’33″…
This experiment reminds me of some research I uncovered for my Master’s thesis. I did work on timbre theory, approaching it from a neurocognitive perspective, and read a lot about various timbral subtleties that allow the brain to differentiate sounds (mainly to identify them).
Incidentally, the human brain is capable of identifying discrepancies in synchony of attacks up to 15 milliseconds! In other words, a difference of 15 milliseconds between the entrance of one player and another can be heard. I think that particular aspect plays a significant role in distinguishing the two ensembles in your listening samples.
Everyone should check out this story from the Guardian:
http://arts.guardian.co.uk/news/story/0,,1854371,00.html
My fears about commercial applications of this have already become reality.