Athon, I’m trying to keep up with all the ideas! First, you say alpha is the chance of being wrong, given one has concluded H. Thus, we have alpha = 1-Pr(H given the data).
Firstly, my apologies for oversimplifying. I guess sometimes I slip back into 'teaching year 10 probability' mode where the details are occasionally dropped in order to get kids to understand the concept of 'probable' versus 'possible'.
Alpha is simply the chance that your null hypothesis is correct. The lower it is, the less likely it is the situation which explains the phenomena you are observing, and the greater my confidence can be in my hypothesis being a correct justification for the observation.
Alpha would then be a bayesian probability.
I must admit, I do side more with Bayesian philosophy than the frequentists, but that's another discussion. How is alpha = 1 - P strictly Bayesian? Perhaps there's something about that philosophy I'm missing, or I've missed in your argument.
You have demonstrated my point that, even with substantial training (as it seems you have had), people routinely misinterpret the standard results. If experts can’t get it right, what hope is there for the rest of us? And, as I asked in my first post, why do we continue to report alpha? When we call it “inferential,” we are just giving people enough rope to hang themselves with.
I'd hardly qualify as an expert. I've used statistics professionally and have taught it at low high school level, but must admit many of the finer points still lose me. Which is why I'm not arguing this vehemently, but rather because I honestly wonder if I've lost something in your explanations.
You still haven't explained why alpha is arbitrary (or at least, where I can find your explanation).
On your “black hair” experiment: By equating “how significance this number is” with “the chances my probability would be wrong in a larger population,” you are again erroneously calling alpha a bayesian probability. Alpha is pretty tricky, isn’t it? But my point is that these errors are inescapable. As is your erroneous reference to “risk.” Measurement of risk necessitates bayesian reasoning.
Bayesian reasoning takes into account the chance that variables outside of our scope are at work influencing the probability. It's a real world application, even though it is a little vague, as it accepts that we cannot see all of the variables in an experiment. Alpha is still useful, even though it is not applied as strictly as frequency might dictate.
“confidence in the validity of any knowledge we gather:” In statistics, do we truly “gather knowledge?” Rather, we gather data. What does that have to do with “increasing knowledge.” And now that you continue to speak about levels of confidence and increasing our knowledge, I'm unclear about what you meant earlier by saying "statistics is only a tool."
Ok, I wondering when we would get into 'definitions' territory.
Data is raw observable information we interpret from our surroundings. Knowledge is the interpretation of that data in reference to the context it's taken from. Statistics is a way of interpreting the relevance of data in context with its environment. Therefore, we can only construct knowledge personally, although data is something objective that exists outside of our observation of it. Statistics is a tool for classifying and attributing values to information we gather from the environment.
Athon