Home
AudioAsylum Trader
Propeller Head Plaza

Technical and scientific discussion of amps, cables and other topics.

For Sale Ads

FAQ / News / Events

 

Use this form to submit comments directly to the Asylum moderators for this forum. We're particularly interested in truly outstanding posts that might be added to our FAQs.

You may also use this form to provide feedback or to call attention to messages that may be in violation of our content rules.

You must login to use this feature.

Inmate Login


Login to access features only available to registered Asylum Inmates.
    By default, logging in will set a session cookie that disappears when you close your browser. Clicking on the 'Remember my Moniker & Password' below will cause a permanent 'Login Cookie' to be set.

Moniker/Username:

The Name that you picked or by default, your email.
Forgot Moniker?

 
 

Examples "Rapper", "Bob W", "joe@aol.com".

Password:    

Forgot Password?

 Remember my Moniker & Password ( What's this?)

If you don't have an Asylum Account, you can create one by clicking Here.

Our privacy policy can be reviewed by clicking Here.

Inmate Comments

From:  
Your Email:  
Subject:  

Message Comments

   

Original Message

RE: Searching for truth?

Posted by Dave_K on November 10, 2014 at 15:15:48:

Rick,

I don't necessarily agree with the author that if you don't know how to improve the detectability (SNR) then the hypothesis probably not scientific.

As a counter-example, where would particle physics be if we had rejected the standard model back in the 1960s on the basis of warning sign #3 - that many of its predictions were essentially unverifiable with known technology and involved nearly undetectable weakly interacting particles? Or if we had rejected it on the basis of warning sign #7? Instead, we invested heroic amounts of money and effort trying to confirm the theory's predictions, and it turned out to be right.

I agree that a good first step to tackling a problem is "observing the problem while futzing with as many variables as possible to ferret out those that affect it." That's how I usually start troubleshooting. However, I don't think it's nearly as effective at optimizing. When troubleshooting, you are starting from having observed a problem and it's usually a pretty easy observation to repeat and confirm. But when you experiment to find improvements, I find the observations to be a lot noisier and the variables are not always independent.

And then you have to deal with patternicity - human nature seeking explanations for patterns that are essentially random, or at least caused by some uncontrolled variable that we weren't interested in. This seems to be a common problem in machine learning and other fields involving statistical classification, where the classifier adapts to features evident in random data which are not caused by the underlying function/mechanism of interest. There is a similar tendency in human perception, which evolved to evade predators and hunt prey. Our minds are well adapted to identify patterns in sensory data and then learn to recognize them. It seems like we're hard wired to to over-classify; we tend to see and interpret patterns even in random data, and don't easily recognize randomness unless we're looking at a sufficiently large data set where the distribution is obvious. I think this is part of the reason why there's a lot of tail chasing in audiophiledom. When you approach the optimum then it becomes harder to control variables and harder to separate an effect from noise, and you can get off-track chasing meaningless patterns in noise.

The only way I know of to avoid tail chasing is to make a lot of observations over a long period before making any conclusions.

Dave