Popular Post The Computer Audiophile Posted June 23, 2020 Author Popular Post Share Posted June 23, 2020 On 6/18/2020 at 3:16 PM, plissken said: Correct, it will channel hunt and hop. But you should be able to see DFS alerts in the controller. I was thinking about this over the weekend. I know, that's what I do and it's ridiculous. At least I was sipping the tea you sent me while doing it :~) I can't wait for us audiophiles to start comparing the sound quality of WiFi frequency bands and channels. If we can't make fun of ourselves, we are in trouble :~) P.S. I'm a card carrying, knuckle dragging audiophile and include myself as the target of all jokes I make. plissken and Superdad 2 Founder of Audiophile Style | My Audio Systems Link to comment
bluesman Posted June 25, 2020 Share Posted June 25, 2020 On 6/22/2020 at 11:19 AM, cat6man said: I personally define objective proof in audio as requiring measurements that confirm a hypothesis. I do not include blind or double blind testing as 'objective' (your mileage may vary) so let's continue on my version of a possibly 'objective' answer. any engineers or scientists here? As a tenured professor at a university medical center, I and my colleagues rely on DBT as a valid research tool. Well designed, properly powered double blinded clinical trials have been the basis for many major scientific achievements that have saved countless lives. If the p value on well chosen, appropriately tests of the delta between a placebo or control cohort and the active study cohort is 0.02, there was a 98% probability that the observed effect occurred because of the intervention and a 2% probability it happened purely by random chance. I don’t understand how you can dismiss this as not being objective. The problem with most amateur DBT is how it’s done, not the principle behind it. pkane2001 1 Link to comment
The Computer Audiophile Posted June 25, 2020 Author Share Posted June 25, 2020 10 minutes ago, bluesman said: The problem with most amateur DBT is how it’s done, not the principle behind it. 100% true. fas42 1 Founder of Audiophile Style | My Audio Systems Link to comment
kennyb123 Posted June 25, 2020 Share Posted June 25, 2020 5 hours ago, bluesman said: The problem with most amateur DBT is how it’s done, not the principle behind it. DBT proponents seem to see this as feature and not a bug. Their desired outcome is guaranteed and made bulletproof (so they think) because properly-conducted DBT has been demonstrated to be a valid research tool. Valid questions about how the test was conducted are ignored. That's why I've said that those who press for amateur DBTs are actually the ones behaving like snake oil salesmen. sandyk 1 Digital: Sonore opticalModule > Uptone EtherRegen > Shunyata Sigma Ethernet > Antipodes K30 > Shunyata Omega USB > Gustard X26pro DAC < Mutec REF10 SE120 Amp & Speakers: Spectral DMA-150mk2 > Aerial 10T Foundation: Stillpoints Ultra, Shunyata Denali v1 and Typhon x1 power conditioners, Shunyata Delta v2 and QSA Lanedri Gamma Revelation and Infinity power cords, QSA Lanedri Gamma Revelation XLR interconnect, Shunyata Sigma Ethernet, MIT Matrix HD 60 speaker cables, GIK bass traps, ASC Isothermal tube traps, Stillpoints Aperture panels, Quadraspire SVT rack, PGGB 256 Link to comment
plissken Posted June 25, 2020 Share Posted June 25, 2020 2 hours ago, kennyb123 said: Valid questions about how the test was conducted Links please... Link to comment
Popular Post bluesman Posted June 26, 2020 Popular Post Share Posted June 26, 2020 3 hours ago, kennyb123 said: DBT proponents seem to see this as feature and not a bug. Their desired outcome is guaranteed and made bulletproof (so they think) because properly-conducted DBT has been demonstrated to be a valid research tool. Valid questions about how the test was conducted are ignored. That's why I've said that those who press for amateur DBTs are actually the ones behaving like snake oil salesmen. Maybe they just don't know enough about it to do it right (or that it's not a valid methodology for answering the question being asked). You're talking about a fairly sophisticated methodology that requires specialized knowledge. When you encounter a specious application of it, you'd be of much more help identifying the flaws in it than simply dismissing it. Name calling reduces you to the level of those you're criticizing. As they say at the airport, if you see something.......say something. But you're only part of the solution if what you say is meaningful and constructive. Otherwise, you're part of the problem. pkane2001 and plissken 1 1 Link to comment
kennyb123 Posted June 26, 2020 Share Posted June 26, 2020 27 minutes ago, bluesman said: Maybe they just don't know enough about it to do it right (or that it's not a valid methodology for answering the question being asked). You're talking about a fairly sophisticated methodology that requires specialized knowledge. When you encounter a specious application of it, you'd be of much more help identifying the flaws in it than simply dismissing it. Name calling reduces you to the level of those you're criticizing. As they say at the airport, if you see something.......say something. But you're only part of the solution if what you say is meaningful and constructive. Otherwise, you're part of the problem. False accusations reduce one to an even lower level. Are you accusing me of not pointing out the flaws? I agree that name calling is bad but that’s different than employing a metaphor to illustrate the hypocrisy in holding certain positions. The snake oil salesman pitching pseudo-scientific arguments in favor of a product that he knows isn’t going to produce positive results is very much akin to putting forth a blind test methodology one knows won’t produce positive results. Digital: Sonore opticalModule > Uptone EtherRegen > Shunyata Sigma Ethernet > Antipodes K30 > Shunyata Omega USB > Gustard X26pro DAC < Mutec REF10 SE120 Amp & Speakers: Spectral DMA-150mk2 > Aerial 10T Foundation: Stillpoints Ultra, Shunyata Denali v1 and Typhon x1 power conditioners, Shunyata Delta v2 and QSA Lanedri Gamma Revelation and Infinity power cords, QSA Lanedri Gamma Revelation XLR interconnect, Shunyata Sigma Ethernet, MIT Matrix HD 60 speaker cables, GIK bass traps, ASC Isothermal tube traps, Stillpoints Aperture panels, Quadraspire SVT rack, PGGB 256 Link to comment
Popular Post cat6man Posted June 26, 2020 Popular Post Share Posted June 26, 2020 11 hours ago, bluesman said: As a tenured professor at a university medical center, I and my colleagues rely on DBT as a valid research tool. Well designed, properly powered double blinded clinical trials have been the basis for many major scientific achievements that have saved countless lives. If the p value on well chosen, appropriately tests of the delta between a placebo or control cohort and the active study cohort is 0.02, there was a 98% probability that the observed effect occurred because of the intervention and a 2% probability it happened purely by random chance. I don’t understand how you can dismiss this as not being objective. The problem with most amateur DBT is how it’s done, not the principle behind it. of course DBT is a valid and immensely useful scientific method, and i agree that most amateur DBT are relatively useless. however in most(?) cases at the medical center, i'd guess that you have some (what i'll call) objective output measure such as blood pressure, survival rate, visual acuity, heart ejection fraction........i.e. things you can measure. compare that with 'the soundstage is wider' or 'it sounds more real' or 'i hear more breath on the vocals' i never meant to imply that DBT is not valid, but that with objective/measurable criteria (versus subjective opinion, even if in large controlled numbers) it is much clearer what is going on. i assume DBT on psychological research methodology gets a bit messier? barrows and motberg 2 Link to comment
Audiophile Neuroscience Posted June 26, 2020 Share Posted June 26, 2020 2 hours ago, cat6man said: of course DBT is a valid and immensely useful scientific method, and i agree that most amateur DBT are relatively useless. however in most(?) cases at the medical center, i'd guess that you have some (what i'll call) objective output measure such as blood pressure, survival rate, visual acuity, heart ejection fraction........i.e. things you can measure. Yes, I agree and it is a point that seems often overlooked IMO. The best outcome measures pass what I have referred to previously as objective 'test of test' parameters. Also, those tests are ideally out of the subjects direct control or influence or even knowledge. Ideally subjects don't have a clue what is being tested like a blood measure they have never even heard of, let alone something that tests their skill or potentially their reputation. Things however cannot be always ideal. Without test of test parameters however, I won't go into detail here but, one is basically using an uncalibrated tool. Audio perceptual testing is interesting in this context, adding a layer of increased difficulty as you have to develop methodology that in essence uses a perceptual test to test perception. There is no gold standard to which we can 'calibrate' against such as using a blood test serving as a surrogate marker of a particular cancer but comparing results against gold standard biopsy results. Even with well done medical research confounders often lead to less than conclusive results and this often has nothing to do with blinding. Just look at the number of peer reviewed publications on the role of statins in heart disease, or the place of coronary CT calcium scores, what you might expect should be fairly straight forward is anything but. Eliminating bias is not the issue , it is the "how" you do it. Methodology is key. Blinding does not automatically make a methodology valid and indeed you cannot guarantee it does not actually increase false negatives through other means such as interdependent variables . In order to know this you would need to know the false negative rates and this poses difficulty without a gold standard reference for calibration. Sound Minds Mind Sound Link to comment
Audiophile Neuroscience Posted June 26, 2020 Share Posted June 26, 2020 11 hours ago, kennyb123 said: DBT proponents seem to see this as feature and not a bug. Their desired outcome is guaranteed and made bulletproof (so they think) because properly-conducted DBT has been demonstrated to be a valid research tool. Valid questions about how the test was conducted are ignored. That's why I've said that those who press for amateur DBTs are actually the ones behaving like snake oil salesmen. I agree it seems to be assumed by some as fairly simple and infallible procedure, or at least highly validated, and can even be used by amateurs as a weapon to push a particular view. It is just a test tool and as alluded to above, like all tools, one needs to know how good the tool is at telling you what you think it is telling you - in the context and setting of the whole test methodology .........I think we are heading OT..... motberg 1 Sound Minds Mind Sound Link to comment
pkane2001 Posted June 26, 2020 Share Posted June 26, 2020 3 hours ago, Audiophile Neuroscience said: I agree it seems to be assumed by some as fairly simple and infallible procedure, or at least highly validated, and can even be used by amateurs as a weapon to push a particular view. It is just a test tool and as alluded to above, like all tools, one needs to know how good the tool is at telling you what you think it is telling you - in the context and setting of the whole test methodology .........I think we are heading OT..... All things being equal, DBTs are better than sighted testing, unless you are testing for confirmation bias or a placebo effect. Whether DBTs are infallible or not is not a question, at least in my mind. Of course they can be done wrong, and the results may not be accurate. It's important to understand the limitations and follow a proper test design. But, it's an audio malpractice, IMHO, to deny that DBTs can, with proper controls, produce a much more accurate, repeatable and reproducible result than sighted testing. And such controls are not voodoo or a mystery. They are studied, reported and well documented. A sighted test cannot be done right: there's no way to eliminate or separate the effect of biases related to other senses and preconceived ideas and knowledge. A DBT can be done wrong, but when properly planned and conducted, it is a scientific instrument that can be measured and calibrated. -Paul DeltaWave, DISTORT, Earful, PKHarmonic, new: Multitone Analyzer Link to comment
bluesman Posted June 26, 2020 Share Posted June 26, 2020 8 hours ago, cat6man said: However in most(?) cases at the medical center, i'd guess that you have some (what i'll call) objective output measure such as blood pressure, survival rate, visual acuity, heart ejection fraction........i.e. things you can measure. compare that with 'the soundstage is wider' or 'it sounds more real' or 'i hear more breath on the vocals' The critical factor in choosing a test is whether it can answer the question being asked. As you point out, the question asked by audiophiles is often unanswerable and therefore untestable as asked. It’s not possible to determine if a given cable improves SQ with blinded testing - for a start, one audiophile’s smooth is another’s muddy. But it’s possible to determine if there’s a consistent difference between two that subjects can identify with 95+% consistency in enough well done blinded listening trials to be statistically significant. Correctly identifying one alternative from another with 95% certainty in well designed and well conducted DBT is objective. And this can be useful info. The “best” questions for any study are objectively measurable, as you point out. But many healthcare decisions are made on the basis of parameters that can be as vague and nebulous as “how real it sounds”, e.g. sense of well-being, intensity of pain, Quality of Life Years, patient satisfaction, and likelihood of recommending. Picking and using good tests to get valid, repeatable results is dependent on what question is asked, how it’s asked, and in what form an answer is sought. And measurement systems are less reliable than we believe. Even “simple” blood pressure measurement is not simple, e.g. 3 consecutive readings 5 minutes apart can vary widely. Other factors ignored by those who oversimplify DBT include consistency among multiple raters and consistency of the same rater over time. If a subject correctly identifies A or B in 95% of presentations in one trial session but averages 53% over 5 identical sessions, that 95% session is not representative - it’s random chance. Controls are also needed. Present the same alternative to raters multiple times and see if they think they perceive differences that aren’t there. Link to comment
bluesman Posted June 26, 2020 Share Posted June 26, 2020 32 minutes ago, bluesman said: Present the same alternative to raters multiple times and see if they think they perceive differences that aren’t there. To bring this back OT, well done DBT could help determine if network platforms and components make discernible difference in SQ. Link to comment
The Computer Audiophile Posted June 26, 2020 Author Share Posted June 26, 2020 14 minutes ago, bluesman said: To bring this back OT, well done DBT could help determine if network platforms and components make discernible difference in SQ. I suppose we would also have to include in that statement, well done DBT could help determine if network platforms and components make discernible difference in SQ, to those tested. There will always be people who, either real or not, believe they are outliers or different from the test group and the results don't ring true for them. Founder of Audiophile Style | My Audio Systems Link to comment
bluesman Posted June 26, 2020 Share Posted June 26, 2020 22 minutes ago, The Computer Audiophile said: There will always be people who, either real or not, believe they are outliers or different from the test group and the results don't ring true for them. And there really are a few clinging to each end of the bell shaped curve. But most of us are like most of us. Only in Lake Wobegon are the children all above average. The Computer Audiophile 1 Link to comment
The Computer Audiophile Posted June 26, 2020 Author Share Posted June 26, 2020 4 minutes ago, bluesman said: And there really are a few clinging to each end of the bell shaped curve. But most of us are like most of us. Only in Lake Wobegon are the children all above average. Agree. Plus, I live close to Lake Wobegon :~) Founder of Audiophile Style | My Audio Systems Link to comment
bluesman Posted June 26, 2020 Share Posted June 26, 2020 5 minutes ago, The Computer Audiophile said: Agree. Plus, I live close to Lake Wobegon :~) You’re also not that far from Frostbite Falls. You may even have met some of our audiophile colleagues who studied statistics at Wassamatta U. The Computer Audiophile 1 Link to comment
Popular Post ray-dude Posted June 26, 2020 Popular Post Share Posted June 26, 2020 A detour, then back to OT I've done a lot of user experience testing over the years (a LOT). I've also been heavily involved in more formal scientific testing (drug clinical trials, etc). The former seems more germane to our humble hobby, but that's neither here nor there. What I found with user experience testing is that the best insights and outcomes didn't come from asking people which they preferred, but from watching them think about it, then designing for that insight. User experience testing that focused on direct measurement ("Do you prefer the menu on the left or the right?") would more often than not lead us astray because we were misinterpreting what we were measuring. For example, most of us are of an age where we remember the Pepsi challenge. People would try blinded Pepsi vs blinded Coke, and people would prefer Pepsi. Pepsi is better than Coke, right? Heck, let's make it double blind by having the person handing out the samples in white papercups not know which is which. That will make us even more scientific. If you prefer Coke, you're wrong! The problem is that that test was measure first use preference, not long term use preference. Pepsi had (has) a stronger sugar slam, but it lacks depth and subtle over time. Based on first use, Pepsi wins more often than their market share. However, purchasing preference and people trying Pepsi but going back to Coke indicates that based on long term use, Coke wins over Pepsi for most people. This paralleled my experience with user experience testing, where most testing results were interesting but irrelevant because we were measuring first use preference. You really had to observe long term behavior to infer proper preference, and even then, more often than not, the value was getting an insight that then fed into the next design iteration. Back to audio. Same pattern gives us dynamic range compression. First listen sounds better when things are "louder". Over time, it is an ick of an experience for a lot of us. Loudness wars mastering is Pepsi, high dynamic range masterings are Coke. Some people prefer Pepsi, I prefer Coke. I guarantee that if you were to do extended DBT testing of low DR masterings vs high DR masterings and ask what people prefer, a huge majority would prefer the low DR mastering based on first listen (yes you could carefully volume match, etc, but I'm offering this as an example, not as an experimental design). I bring this up on this thread, because SFPs are falling into that same pattern for me. This is a class of component where I'm finding I have to ignore my initial experience and preference, and gauge my mood after extended listening. Do I want to listen to music longer? Am I able to focus on work more or less when listening to music with a particular SFP pair? Like my user experience example above, once I identify a long term preference, I then look for what short term identifyable characteristic is a "tell" for what that long term experience will be, then look for components that have more or less of that tell. I'm starting to find that "tell", but I still have to lean into extended listening Audiophile Neuroscience, motberg and kennyb123 3 ATT Fiber -> EdgeRouter X SFP -> Taiko Audio Extreme -> Vinnie Rossi L2i-SE w/ Level 2 DAC -> Voxativ 9.87 speakers w/ 4D drivers Link to comment
kennyb123 Posted June 26, 2020 Share Posted June 26, 2020 2 hours ago, ray-dude said: I bring this up on this thread, because SFPs are falling into that same pattern for me. This is a class of component where I'm finding I have to ignore my initial experience and preference, and gauge my mood after extended listening. Do I want to listen to music longer? Am I able to focus on work more or less when listening to music with a particular SFP pair? Like my user experience example above, once I identify a long term preference, I then look for what short term identifyable characteristic is a "tell" for what that long term experience will be, then look for components that have more or less of that tell. I'm starting to find that "tell", but I still have to lean into extended listening. Very well said. For me, spending more time listening to a SFPs has told me more. Listening to music with varying levels of recording quality reveals what might be missed with a quick shootout. Also the Pepsi and Coke example is very apt as something that might have sounded great after a shootout could very well drive you mad after extended listening. Digital: Sonore opticalModule > Uptone EtherRegen > Shunyata Sigma Ethernet > Antipodes K30 > Shunyata Omega USB > Gustard X26pro DAC < Mutec REF10 SE120 Amp & Speakers: Spectral DMA-150mk2 > Aerial 10T Foundation: Stillpoints Ultra, Shunyata Denali v1 and Typhon x1 power conditioners, Shunyata Delta v2 and QSA Lanedri Gamma Revelation and Infinity power cords, QSA Lanedri Gamma Revelation XLR interconnect, Shunyata Sigma Ethernet, MIT Matrix HD 60 speaker cables, GIK bass traps, ASC Isothermal tube traps, Stillpoints Aperture panels, Quadraspire SVT rack, PGGB 256 Link to comment
Popular Post cat6man Posted June 26, 2020 Popular Post Share Posted June 26, 2020 4 hours ago, bluesman said: You’re also not that far from Frostbite Falls. You may even have met some of our audiophile colleagues who studied statistics at Wassamatta U. I'm a big fan of WU! (and still have my coffee mug) This discussion reminds me of situations that happened repeatedly at work over the years. We'd measure something in the lab and some arrogant PhD would say it was impossible because he had determined that such and such was optimum, blah blah blah. I always insisted that the theoreticians in my department/lab get their hands dirty and verified that their algorithms behaved as expected when implemented. In my personal experience, I obtained some of my most interesting insights (and patents) from the lab (i.e. real world) experiments not lining up with the oversimplified theories.........like "Why the hell is that happening?".............answer: the experiment is trying to tell you something dummy! Back to audio: It really bugs the heck out of me that putting a better power supply on a digital music server can have such a profound impact on my musical enjoyment................I love the enjoyment part but am frustrated that I don't understand the mechanism (actually worse, I don't have a clue!) by which this happens. I have no specific theory or hypothesis to fit my experience but I think a framework is in sight. My engineering intuition (or hubris?) suggests that it is related to behavior on the tail of distributions (a phenomenon I've seen repeatedly in my professional work). A possible framework, though unproven, might look as follows: (YMMV, ignore the rest if you detest speculation) Whatever mechanism is at work is likely not measurable in an average sense but is likely, IMHO, to be a low probability event. Let's hypothesize 1% of the time this "something happens". For red book, this would be ~440 times/sec. I can certainly imagine that this might be discernible if it affected 440 samples/sec but would yet be unmeasurable on average due to being masked by the other 43,000+ samples. Occasional long latencies between samples? Clock jitter spikes? Short infrequent power supply glitches? I've no idea but something is going on. [start lecture] Case in point/analogy: In wireless communications, using 2 spatially separated antennas give improved immunity to fading and is called space diversity reception, the idea being that if the fading is independent on each antenna the probability that they both are bad (i.e. faded) is greatly reduced. That's the theory.....and we saw exactly that with a fading simulator in the lab, but not so when testing around a real cell site with space diversity antennas. WTF? It turned out that the two antennas have almost independent fading but there was a residual 1-2% correlation in the fading statistics. On average that would make no difference to the measurements, however the wireless receiver behaved differently. Now a 1% correlated signal could be simply modeled as 99% of the time uncorrelated and 1% of the time completely correlated (i.e. both fading at the same time). The radio receiver was happy as a clam 99% of the time but very unhappy the other 1% of the time, and whenever that 1% of the time occurred the receiver would have to change its operating point and required SNR and the entire receiver loop would have to re-converge. When we programed the fading simulator to have 1% correlation between the two receiver inputs, we obtained the exact same results as in the outside world. The 1% correlation did not make any change to the average fading statistics but it had a huge impact on the receiver due to dynamic effects. [end lecture] So let me put down a friendly challenge. In the interest of being 'objective' and not 'subjective' in this sub-forum, how about a sub-sub-group interested in getting to the bottom of this technically and not just acting like my old 'arrogant PhD' colleagues who already knew all the answers (but had oversimplified the situation and therefore hadn't formulated the problem accurately)? This will be my last shot at trying to see if anyone else is interested in approaching this in a similar manner to my current way of thinking. I'm not interested in debating philosophy of testing and will not reply to such. Anyone want to try to figure out what is going on here? Audiophile Neuroscience, Superdad, motberg and 1 other 4 Link to comment
ray-dude Posted June 26, 2020 Share Posted June 26, 2020 1 hour ago, cat6man said: So let me put down a friendly challenge. In the interest of being 'objective' and not 'subjective' in this sub-forum, how about a sub-sub-group interested in getting to the bottom of this technically and not just acting like my old 'arrogant PhD' colleagues who already knew all the answers (but had oversimplified the situation and therefore hadn't formulated the problem accurately)? This is the fun stuff!! If PhD's are allowed in, count me in! (FYI, I put my hypothesis out there in part 1 of my Extreme review...reference voltage, ground plane, and reference timing are the father/son/holy ghost of digital audio, I think, and everything always seems to come back to those fundamentals) ATT Fiber -> EdgeRouter X SFP -> Taiko Audio Extreme -> Vinnie Rossi L2i-SE w/ Level 2 DAC -> Voxativ 9.87 speakers w/ 4D drivers Link to comment
plissken Posted June 26, 2020 Share Posted June 26, 2020 5 hours ago, ray-dude said: Like my user experience example above, once I identify a long term preference, I then look for what short term identifyable characteristic is a "tell" for what that long term experience will be, then look for components that have more or less of that tell. I'm starting to find that "tell", but I still have to lean into extended listening Would this include a setup with two SFP+ modules were someone other than the listener could turn off a tranceiver? That is you could have it in rack for an extended period of time? Say 3-6 months? Link to comment
Audiophile Neuroscience Posted June 27, 2020 Share Posted June 27, 2020 12 hours ago, bluesman said: The critical factor in choosing a test is whether it can answer the question being asked. As you point out, the question asked by audiophiles is often unanswerable and therefore untestable as asked. It’s not possible to determine if a given cable improves SQ with blinded testing - for a start, one audiophile’s smooth is another’s muddy. But it’s possible to determine if there’s a consistent difference between two that subjects can identify with 95+% consistency in enough well done blinded listening trials to be statistically significant. Correctly identifying one alternative from another with 95% certainty in well designed and well conducted DBT is objective. And this can be useful info. The “best” questions for any study are objectively measurable, as you point out. But many healthcare decisions are made on the basis of parameters that can be as vague and nebulous as “how real it sounds”, e.g. sense of well-being, intensity of pain, Quality of Life Years, patient satisfaction, and likelihood of recommending. Picking and using good tests to get valid, repeatable results is dependent on what question is asked, how it’s asked, and in what form an answer is sought. And measurement systems are less reliable than we believe. Even “simple” blood pressure measurement is not simple, e.g. 3 consecutive readings 5 minutes apart can vary widely. Other factors ignored by those who oversimplify DBT include consistency among multiple raters and consistency of the same rater over time. If a subject correctly identifies A or B in 95% of presentations in one trial session but averages 53% over 5 identical sessions, that 95% session is not representative - it’s random chance. Controls are also needed. Present the same alternative to raters multiple times and see if they think they perceive differences that aren’t there. I mostly agree but a couple of comments. In the absence of objective direct outcome measures, correctly identifying a perceptual difference to a statistically significant degree is indeed useful and objective information. As you point out p=.05 is the 'usual' figure nominated but some would like a bit better than this depending on the situation. @manisandher in the Red Pill/ Blue Pill scored p=0.1, a 99% consistency ie probability that his perception was not a product of random chance or guesswork. But, without direct objective measures the problems of inductive reasoning are amplified - there is no inductive law to prove a hypothesis. Hence, that thread was very long and hotly debated - basically the result was challenged, by some, based on supposed methodological issues. So while "well designed and well conducted DBT" is prerequisite, achieving same is not always, and I would suggest often not, without challenge. A bit like audio, its 'all in the implementation'. You touched on intra-observer false positives. These are important but so are intra-observer false negatives. IMO false negatives in general are really more the issue in the context of blind listening tests. It has to do with the whole "blind test it or it didn't happen" claim. If a perception disappears on a blind listening test, is it due to the perception never existing (beyond biased imagination) or is it that the test has failed in some way ie it is a false negative. It's a rhetorical question as seemingly most have preconceived unshakable positions. Whatever the case, if you don't know the false positives or false negative rates in the first place (together with other test of test parameters) of the test procedure you are working in the dark. The blind leading the blinded. It is analogous to using an uncalibrated tool. At the very least one needs to accept certain assumptions of validity about the test procedure. I think there is more light shed on the situation if the subject passes the test with flying colors, but in either case it comes down to acceptance of the methodology. Not directed at you but asking the question how do we know what we know is not just philosophical musing but strikes at scientific method. Whether we are assessing SFP modules or considering the credibility of reviewers on this very forum site, there will always be claims that "if it wasn't blind tested it didn't happen". In the absence of proof one way or another I remain open but skeptical about reported observations. sandyk 1 Sound Minds Mind Sound Link to comment
fas42 Posted June 27, 2020 Share Posted June 27, 2020 7 hours ago, cat6man said: So let me put down a friendly challenge. In the interest of being 'objective' and not 'subjective' in this sub-forum, how about a sub-sub-group interested in getting to the bottom of this technically and not just acting like my old 'arrogant PhD' colleagues who already knew all the answers (but had oversimplified the situation and therefore hadn't formulated the problem accurately)? This will be my last shot at trying to see if anyone else is interested in approaching this in a similar manner to my current way of thinking. I'm not interested in debating philosophy of testing and will not reply to such. Anyone want to try to figure out what is going on here? Go for it ... 🙂. The overall answer for "what is going on" is that electrical noise from a variety of sources internal and external to the rig impacts the analogue areas of the replay chain; just enough to be audible - this was true 3 decades ago, and is just as true, still, right now. Doesn't matter that the music player is "right over there, way, way from the sensitive stuff!!" ... nasty stuff gets around with the greatest of ease, and your challenge, should you choose to accept it 😝, is to track down and nail every last one of these pathways for the SQ to be degraded by interference mechanisms. The precise, technical explanation for what's going on in a particular setup would be handy to know - but ultimately far less important than knocking the relevant interference pathways on the head ... 😉. Link to comment
bluesman Posted June 27, 2020 Share Posted June 27, 2020 39 minutes ago, Audiophile Neuroscience said: As you point out p=.05 is the 'usual' figure nominated but some would like a bit better than this depending on the situation. @manisandher in the Red Pill/ Blue Pill scored p=0.1, a 99% consistency ie probability that his perception was not a product of random chance or guesswork. A p value of 0.1 means a 90% probability, not 99%. Link to comment
Recommended Posts
Create an account or sign in to comment
You need to be a member in order to leave a comment
Create an account
Sign up for a new account in our community. It's easy!
Register a new accountSign in
Already have an account? Sign in here.
Sign In Now