Let me see if I can break this down in a way that helps you understand why we don't need to invent a better double blind methodology, even if any of us were qualified to design one...
Whether or not the results are statistically significant is not determined by the score. It is determined by the testing methodology, the number of trials, and the margin of error.
If the methodology is sound, and enough trials are run to exceed the margin of error, the audiophile's score of 10 correct out of 20 concludes, statistically, that he could not reliably hear the difference. Period. Not I, or anyone with any sense, would argue that ABX testing is conclusive regardless of the quality of the execution. I've run quite a few casual, blind A/B listening test on myself, at home. I understand that I've only made a point for myself when I've done so. I can't speak for anyone else, but that kind of ABX testing is not what I'm talking about here. I'm talking about the kind Meyer and Moran did. The kind Stereophile did in comparing amps years ago, the kind Harman is doing all the time in their labs.
Many have. It is well-understood, in scientific, not audiophile circles, that subtle differences are most reliably detected through quick switching. All hearing tests are based on this understanding. Would you endeavor to re-design your opthamologists vision test if you didn't like the results? Would you insist on a whole new, self-designed testing methodology because you were seeing an H instead of an A on that fourth line down? Or would you, perhaps, accept that the collected wisdom of the people who have long designed and excecuted vision tests might be more rational than your desire to have 20/20 vision?
Eliminate "cost" from the sentence above and you have embraced the collective scientific wisdom.
I don't think anything is perfect and I have no interest in calling you names, Gregg, but we're going to need an audiologist, a research scientist and a statistician to come up with a better methodology. This one is a tested, tweaked and widely accepted method of testing the perception of sensory differences including and beyond audio. The only place I'm aware of that it is broadly considered to be invalid is on audiophile forums. Even wine snobs seem to accept the results more gracefully.
Tim
Okay then we have a situation where an audiophile claims to hear something that is not confirmed by conventional measurements. Said audiophile accepts an ABX challenge. Hs results are statistically insignificant. He say score 10 out of 20.
Whether or not the results are statistically significant is not determined by the score. It is determined by the testing methodology, the number of trials, and the margin of error.
We can come to several conclusions. He could not reliably hear the difference. Or there is something wrong with the way he approached the test. The test is fundamentally flawed. Or I have suggested the he did not want thear a difference.
If the methodology is sound, and enough trials are run to exceed the margin of error, the audiophile's score of 10 correct out of 20 concludes, statistically, that he could not reliably hear the difference. Period. Not I, or anyone with any sense, would argue that ABX testing is conclusive regardless of the quality of the execution. I've run quite a few casual, blind A/B listening test on myself, at home. I understand that I've only made a point for myself when I've done so. I can't speak for anyone else, but that kind of ABX testing is not what I'm talking about here. I'm talking about the kind Meyer and Moran did. The kind Stereophile did in comparing amps years ago, the kind Harman is doing all the time in their labs.
Many have suggested the problem is auditory memory.
Many have. It is well-understood, in scientific, not audiophile circles, that subtle differences are most reliably detected through quick switching. All hearing tests are based on this understanding. Would you endeavor to re-design your opthamologists vision test if you didn't like the results? Would you insist on a whole new, self-designed testing methodology because you were seeing an H instead of an A on that fourth line down? Or would you, perhaps, accept that the collected wisdom of the people who have long designed and excecuted vision tests might be more rational than your desire to have 20/20 vision?
Although it is often stated you could listen as long as you like, rapid switching is a much more cost effective method.
Eliminate "cost" from the sentence above and you have embraced the collective scientific wisdom.
Surely even proponents of ABX do not think it is perfect. Perhaps we can come up with a better double blind methodology with getting offended and calling each other names.
I don't think anything is perfect and I have no interest in calling you names, Gregg, but we're going to need an audiologist, a research scientist and a statistician to come up with a better methodology. This one is a tested, tweaked and widely accepted method of testing the perception of sensory differences including and beyond audio. The only place I'm aware of that it is broadly considered to be invalid is on audiophile forums. Even wine snobs seem to accept the results more gracefully.
Tim