Lots of things are hidden under masks, such as all of the techno-filth that is hidden under the mask of sighted evaluations. Stereophile, anybody?
Are you referring to Convention Paper 9174 by Meridian? Luckily they tested filter and wordlength/dither aspects separately. I hope they found a few well trained listeners. Looks interesting.Looking at the program for October's AES Convention in Los Angeles, it looks as if there will be a paper repeating the essence of the Meyer-Moran tests but conforming to the requirements listed in ITU BS.1116-2.
Are you referring to Convention Paper 9174 by Meridian? Luckily they tested filter and wordlength/dither aspects separately. I hope they found a few well trained listeners. Looks interesting.
Tim - the tests are to establish the limits of what is audibly perceivable. Do you suggest listening tests to establish audibility limits should only use music to establish these limits?
What you do with the knowledge of the known limits is up to you but establishing them first is important. You may consider them irrelevant to music listening? You may personally not be able to hear above 14KHz so you may consider that a limited bandwidth is suitable for you. ArnyK's has established he has hearing damage so his selection criteria for audio reproduction may well be even more limited than yours.
Spot on, Tim.We now have a bit of data that says Amir and a few others have been able to hear some undefined difference when they knew precisely what to listen for. Is that useful? Yes. It contradicts previous data indicating that no audible difference could be heard between hi-res and RB files. Does it have any bearing on the experience of listening to music? Maybe. But that has not only not been established, we haven't even added to the data set for that.
Science is hard.
Tim
(...)
Science is hard.
Tim
Tim, that's not science you are citing - that's your bias packaged to look like science - the fundamental point that you are blind to in this thread. Did you read the title of the paper "The Audibility of Typical Digital Audio Filters in a High-Fidelity Playback System"? Did you read BS1116 or at least scan it? See the bit about trained listeners?
Did JJ test a broad range of listeners when testing the audibility of codecs. No! They were trained listeners who knew what to listen for & used specific audio to best reveal audible characteristics that were theorised might exist. Does it have any bearing on listening to music by th egeneral public? Yes!
Can long term listening tests with music be done at a later stage to establish the relevance of the results to the general public - sure but it's not the first test that is done when trying to establish "The Audibility of Typical Digital Audio Filters in a High-Fidelity Playback System" - is it?
I hope they found a broad range of listeners. What can be heard when you know exactly what to listen for and where is not necessarily relevant to what can be heard when you're listening to music. This is the biggest flaw in the test protocols we're currently discussing. No matter how well-controlled they are, they are telling us very little about our actual listening experience.
I meant what I said, John, no more, no less. Is establishing "the known limits" of what is audible important? I don't know, but we certainly haven't done that, or even attempted to do that, here. We have established what Amir (and a few others), who has said that his hearing drops off rapidly above 12khz, can hear. Given that there are, no doubt, some adolescent girls out there who can hear 18, maybe even 20khz, we are not establishing any known limits. We now have a bit of data that says Amir and a few others have been able to hear some undefined difference when they knew precisely what to listen for. Is that useful? Yes. It contradicts previous data indicating that no audible difference could be heard between hi-res and RB files. Does it have any bearing on the experience of listening to music? Maybe. But that has not only not been established, we haven't even added to the data set for that. And that's what I meant, John, no more, no less.
Science is hard.
Tim
Yes, Tim, what you said was in direct reply to Kees who quoted the paper - the full post:Let's review, John. Here's what I said:
I hope they found a broad range of listeners. What can be heard when you know exactly what to listen for and where is not necessarily relevant to what can be heard when you're listening to music. This is the biggest flaw in the test protocols we're currently discussing. No matter how well-controlled they are, they are telling us very little about our actual listening experience.Originally Posted by Kees de Visser View Post
Are you referring to Convention Paper 9174 by Meridian? Luckily they tested filter and wordlength/dither aspects separately. I hope they found a few well trained listeners. Looks interesting.
Tim
Really? So how would testing with "a broad range of listeners" & using music as the test signal be an advantage or of use in this particular test?My hoping for tests involving a broad range of listeners and tests that are relevant to real listening does not mean I don't understand what these tests intended to do.
You are mistakenly conflating two testing concepts - one to establish the limits of audibility & a completely different one to establish the importance of the limits to "normal listening experience". The problem Tim is that you can't define this "normal listening experience" - there is a wide spectrum of conditions, focus, attentiveness, etc. over which normal listening takes place. You also seem to miss the bit in BS1116 where there is a necessity for statistical analysis to confirm or otherwise the perception of small differences. Statistics are easily made meaningless by not controlling the variables in the test which is one of the main points in this thread - mix enough untrained listeners in with trained listeners & you will dilute the test results to a null result!!I scanned the paper and BS1116, and I get it. I get that they are testing for the audibility of small differences. I get that the differences are so small that the tests demand listeners trained to hear those specific differences, and listening protocols aimed at finding them. This is exactly why I think these tests are irrelevant to the normal listening experience, and would like to see tests run under real listening conditions, with music lovers as well as trained listeners. Got it?
Tim,
We can easily agree on this. But fortunately high-end has not being waiting for the scientific confirmation of existing stereo audio knowledge to implement excellent SOTA audio products that we can currently enjoy, while a few enjoy believing that it is all bias and illusion, as it was not peer reviewed.
If differences were obvious there would not be any need for testing in the first place.
Statistics are easily made meaningless by not controlling the variables in the test which is one of the main points in this thread - mix enough untrained listeners in with trained listeners & you will dilute the test results to a null result!!
I think there might be some gap between obvious and audible to trained listeners listening to carefully chosen examples. I think that gap is where most of enjoy our music and it might be worth exploring.
I'm not, as John says, misunderstanding what these tests are meant to do, Jack. I'm simply wishing for tests of the audibility of hi res to real listeners listening to real music; tests more relevant to our listening experience. While I understand that John can argue endlessly with almost anything, I'm not sure how I can clarify my simple comment any more.
Tim
(...) Really? So how would testing with "a broad range of listeners" & using music as the test signal be an advantage or of use in this particular test? You are mistakenly conflating two testing concepts - one to establish the limits of audibility & a completely different one to establish the importance of the limits to "normal listening experience". The problem Tim is that you can't define this "normal listening experience" - there is a wide spectrum of conditions, focus, attentiveness, etc. over which normal listening takes place. You also seem to miss the bit in BS1116 where there is a necessity for statistical analysis to confirm or otherwise the perception of small differences. Statistics are easily made meaningless by not controlling the variables in the test which is one of the main points in this thread - mix enough untrained listeners in with trained listeners & you will dilute the test results to a null result!!
Good summary John. I have posted many times that the key word is statistics - without a clear understanding of statistical analysis it is not possible to debate these issues. IMHO it is why we are in a circular debate.
In audio quality evaluation, ITU-R BS.1534-1, commonly known as MUSHRA, is widely used for the subjective assessment of intermediate audio quality. Studies have identified limitations of the MUSHRA methodology [1][2], which can influence the robustness to biases and errors introduced during the testing process. Therefore ITU-R BS.1534 was revised to reduce the potential for introduction of systematic errors and biases in the resulting data. These modifications improve the validity and the reliability of data collected with the MUSHRA method. The main changes affect the post screening of listeners, the inclusion of a mandatory mid-range anchor, the number and length of test items as well as statistical analysis. In this paper the changes and reasons for modification are given.
Convention Paper 9172
Knowledge and levels John, control them, as any competently organised blind-test does, and the listeners are solely evaluating sound. Perfectly valid and meaningful.
As for your trained listener red-herring. Need I point out again that none of us who differentiated Arny's and Ethan's files were, Vital, Adamdea, you, me, etc, bar Amir, apparently.
![]() | Steve Williams Site Founder | Site Owner | Administrator | ![]() | Ron Resnick Site Owner | Administrator | ![]() | Julian (The Fixer) Website Build | Marketing Managersing |