I agree with that but, if the test is blind, why should it matter?
I think we used some double negatives. "Being back to square one" and "being a wash is synonmous to me. Failing to prove the world is round is not proof that it's flat.Not at all. The test failed to prove a difference. It did not prove there is none. So, we are back to square one unless the anticipated statistical analysis informs us otherwise.
Right. That is why one needs a sufficiently large number of different subjects.If one does not want to hear a diffeerence, how do we account for it ? E.G. If i have already decided brand X taste the same as Cocoa Cola, how can a blind test eliminate that bias? In a tube vs solid state the "gross " distortions of tubes should be ready appearent? The results are almost always consistent with guessing.
Right. That is why one needs a sufficiently large number of different subjects.
Indeed. That is one reason I was disappointed with the tests. Only 3 participants. Even if they repeated it ad nauseam, the results would not completely satisfy.The problem is also that the ABXers assume that 100% of the people will hear the difference. Say that's not so and only 30% of the listeners can hear a difference between DUT#1 and DUT#2. Then the number of needed participants to show a significant difference would swell to far beyond the ability to test. That's why, and I'm sure you're well aware of, that drug trials are inter-institutional so as to be able to accumulate patients and adequate stats.
Rich, I sympathise with you: the people who are gung-ho about ABX tests never seem to get it, that the differences that are discernable are subtle enough, and complex enough in the causes for the differences, that the actual procedure for doing the ABX can often swamp or just dilute the differences enough so the testing yields poor resultsJeff seems to be taking great glee in the fact that we had difficulty in recognizing which machine was which in our sessions. But, again, I believe that if the tests were conducted differently and with wires running directly into the amps, etc. that there would have been more noticeable differences in the sound. I definitely do not like ABX testing, question their validity and may decide not to participate in this kind of testing again in the future. It puts undue pressure on the listener and I also believe that some subtleties may well be lost in this testing method. Additionally, I do not like the usage of a switching box and believe that it may well "level the playing field" to the point that it removes much of the advantage of a superior product. In addition, it may take substantially more time than only about 1 minute to recognize the performance of 1 product over another.
Summary of this test: differences could be heard sighted but not blind. I've participated in blind tests (not ABX), failed miserably and realized I had foolishly spent about $6000 on equipment that I could not hear improvement in.
I am familiar with all of the arguments against all forms of blind testing. BUT FOR ME, if it is THAT hard to easily hear differences that cost real money, I spend my money elsewhere.
YMMV!
A bit silly. A simple, though expensive way: buy two unmodded units, confirm they are close enough in quality as is to make them equivalent. Have one modified, with no external indication of a change. Then slot one at random into your system until the setup stabilises, over a number of days if necessary; assess it. Then do this a number of times: should be good enough to get a clear result, or non-result ...I maintain that there is NO TEST that could be devised and performed whose results would be accepted by Rich - and apparently some others here - if the modded unit was not a clear winner. There would always be something else at fault because the benefit of the mods is unquestioned by him. He KNOWS that they have improved his system. If testing doesn't bear that belief out, then the test was flawed.
Jeff
I in particular was under extreme stress and was unable to distinguish between the units in the X portion of the session.
Rich
This particular mod shop puts labels and seals on the modified product.Have one modified, with no external indication of a change.
Honestly, Rich, the only pressure on you was what you placed on yourself for fear that you might not be able to tell the difference in a fair test, i.e the blind portion.Actually that is not the case. The differences seem to have been heard both "sighted" (as I say I could not see them around the corner on the floor) as well as in the AB portion of the testing (where we had a tendency to pick one of the two units for the sound preference) and there may have been a preference noted without statistically high enough levels. I in particular was under extreme stress and was unable to distinguish between the units in the X portion of the session.
In science new data causes a theory to be questioned. In religion - and faith-based audio - new data causes the data to be questioned.
I maintain that there is NO TEST that could be devised and performed whose results would be accepted by Rich - and apparently some others here - if the modded unit was not a clear winner. There would always be something else at fault because the benefit of the mods is unquestioned by him. He KNOWS that they have improved his system. If testing doesn't bear that belief out, then the test was flawed.
Jeff
That's because of your parochial engineering viewpoint, ignoring the vast literature on the biology of hearing, testing and responses and statistics. Ever hear of the inverted U in relationship to perception? If not, then you should take a look at it. Ever read up on biological adaptation and our bodies reaction to stressors? Ever read the literature on how our hearing varies so widely from person to person as to make interaural testing methods worthless? Most of all, have you studied how long and short term memory works? If you had, you would find essentially all that ABX testing does is show that our short term memory has a limited capacity. Big whoopti do. That's been known for years.
Taking DBT testing methodology used say in the pharma industry, and applying it willy nilly to all situations doesn't work. And that holds true for many scientific tests and methodologies. This subject has been discussed ad nauseum on here many times.
I am curious about something Myles. If there were gross differences, would you expect them to be heard in the type of test that was run?That's because of your parochial engineering viewpoint, ignoring the vast literature on the biology of hearing, testing and responses and statistics.
I find it interesting that all you need to do is to cast doubt on the test without any real data that invalidates it - or data that invalidates A/B/X testing for audio gear in general. So your bar is very low, while the one for us testors is impossibly high.That's because of your parochial engineering viewpoint, ignoring the vast literature on the biology of hearing, testing and responses and statistics. Ever hear of the inverted U in relationship to perception? If not, then you should take a look at it. Ever read up on biological adaptation and our bodies reaction to stressors? Ever read the literature on how our hearing varies so widely from person to person as to make interaural testing methods worthless? Most of all, have you studied how long and short term memory works? If you had, you would find essentially all that ABX testing does is show that our short term memory has a limited capacity. Big whoopti do. That's been known for years.
Taking DBT testing methodology used say in the pharma industry, and applying it willy nilly to all situations doesn't work. And that holds true for many scientific tests and methodologies. This subject has been discussed ad nauseum on here many times.