p is the probability of getting an outcome when the null hypothesis is true, not the probability THAT the null hypothesis is true.
*laugh* Correct. Figures there'd be someone around who wants to be precise.
In any case,
p is effectively the uncertainty factor, i.e. the chance that you haven't proven what you thought you had, which is all I was using it for in my argument.
I thought the study was supposed to test whether prayer helps improve the health of cancer patients, not whether prayer affects random number generators.
The study is supposed to test whether prayer creates an effect in this study. It is set up especially so that if it creates a HRQOL effect on people prayed for, that can get detected... but who am I to be picky?
You don't have to claim that you personally and wholeheartedly believe prayer works, and that you believe it works in a particular way.
I don't; I'm an agnostic and have never been a theist. I simply want to test it.
But you have to decide what sort of prayer you want to test for---how it would work if it did work---because otherwise you won't be able to decide how to test for it.
Per above, I'm testing for additive prayer on seriously ill patients. The choice of cancer patients is largely arbitrary; I'd be open to switching it to any group that has rapid (w/in 1 year) changes of health condition (so as to create a bell curve that would make group differences sensitive), is appealing to pray for, etc.
This is a different question from the previous one, and it has, in general, a different answer, even though both questions could be phrased, imprecisely, as, "What is the probability of a false positive?". In both, we're interested in the probability of the combination: prayer doesn't work and the study result is positive. However, in one, we already know that prayer doesn't work but we don't know whether the study result will be positive, while in the other, we already know that the study result was positive but we don't know whether prayer works.
What's the probability it's raining, if it's cloudy? What's the probability it's cloudy, if it's raining? Not the same thing.
Per above, this is certainly true. However, I think we are getting outside the range of valid objections to my protocol, and into simply general objections to all research, particularly speculative research. As it is not related to any particular flaw of *my* methodology, I'd rather not get into it.
Not only must you know what the probability of a positive result was on the assumption that prayer doesn't work (i.e., p), you also must know what the probability of a positive result was on the assumption that prayer does work.
And that's not possible to know. (Though theists may claim otherwise.)
Certainly it's not possible to even discuss what that
p(positive
| it works [in manner X]) without discussing X. And I will not get into any discussion about X, as I consider it a waste of time given the dearth of valid non-contradictory evidence to determine X.
You have not explained what you hope to accomplish with this study and how it will be achieved. The general indication is that you wish to provide results that will push the research forward among serious researchers. That is, a positive result would be considered valid and worthy of further consideration by those who are currently unconvinced. There has already been a lot of research on this topic. What you need to explain is exactly what the methodologic concerns were from previous studies and how your study will overcome these concerns. And how your results can be received in a way that they can be taken seriously. Normally, studies are published in peer-reviewed journals in order to make sure that there has been at least a minimum degree of oversight to assess validity. If you can not accomplish that, it seems likely that the very people you wish to reach will not accept the results as valid. You do not mention any connection to an academic institution or ethical approval from an independent review board. Normally, both of those are required for consideration for publication.
While I thank you for your concern, that is not something I am interested in discussing, as it falls within the "what will you do with it if you win" category.
By false hope, I don't mean the presupposition that the prayer itself will offer benefit, but the presupposition that participation in this study can advance this area of research - i.e. that participation in this study can be meaningful. How can it be meaningful if the chances of it finding a real effect are miniscule and if "positive" findings will probably be ignored by serious researchers?
That is only one aspect of it.
You could just as well claim that serious researchers would
never accept a positive finding, and that therefore
any research is completely fruitless. I happen to disagree. However, per above, I do not want to discuss this further, as it is not related to a specific critique of (and preferably, improvement to) my methodology.
You are collecting a lot of data and it is vaguely defined.
It is not vaguely defined, except the comments section, which I have said that I do not intend to use for the purposes of conclusion-relevant analysis.
Once you sit down to analyze it, you will probably be able to find dozens of ways in which to compare the two groups (the "data mining" you refer to). Since you have set your p to <0.05, by chance you will find several outcome variables that are different between the two groups.
Certainly. Which is why I set the analysis before the data is collected, per standard rigorous protocol. No sharpshooter fallacy here.
These ideas are discussed in greater detail in this
paper.
Thank you for the reference. I'll have to read it later, since I'm a bit busy at the moment.
Positive results will likely be subject to extra scrutiny for validity (without additional support from other research). Outcome measures whose validity has already been established (e.g. a visual analog scale for pain) should be used.
I tentatively intend to use the well-established SF36v2 HRQOL as a measure for round 1. What I use for round 2 will be decided after round 1 is complete.
It would be preferable to have a neutral third party analyze the results. The outcome measures should be coded (converted into a form suitable for analysis) blind.
I intend to collect all data by internet application, and only use paper for verification / signature collection. So the neutral third party is a computer program, with the relevant parts of it being open sourced.
You refer to "data mining" as selection bias. Selection bias refers more to how you select the population from which your samples will be drawn (in this case, people aware of your site who have cancer, have an interest in prayer and healing, and make the effort to participate) which leads to issues of generalizability and confounding. Although, to be fair, I find that often selection bias as a term gets used as a catch-all for different kinds of bias - sample biases in particular - so it's use may not really be constrained to a particular type of bias.
Sorry for the lax use of terms. What I mostly was referring to is formally known as the Texas sharpshooter's fallacy.
It is not "impossible to prove a negative". It is no more or less possible to prove a negative than it is a positive. However, I see this bit of "wisdom" repeated frequently, including on this forum, so that's probably a whole separate discussion.
Indeed it is.
Argument from ignorance is generally valid only in some very limited circumstances, where you have proven the ability of the test to detect the thing tested for, and are claiming that the (new) negative results are therefore evidence that the thing tested for does not exist in the
place it was newly tested for.
See my "Charlie the Treasure Hunter" analogy; should come up on a forum search.
Somehow I got the impression from my quick reading that you had already decided upon an N of 25, which is what my comment about being underpowered was based upon. I see that that number is still undetermined.
Correct. I would like 50<n<500 but it's primarily a pragmatic question, of how many qualified participants can be recruited.
I mentioned confounders and this has also been discussed. With so many variables (likely mostly unmeasured) that can affect outcome, the concern is that unequal sorting of the confounders/independent variables could lead to a false positive, and that there is a chance that this unequal sorting could happen in the same direction for each study.
How would it do so [at likelihood > p], given that the sorting into groups is random?
Depending upon the strength of the association between confounder and outcome, the chance of this happening may be greater than the one in a thousand standard from JREF (i.e. the chance of confounding may be greater than the chance that the null should be rejected). Starz' Monte Carlo sim doesn't eliminate this concern as it tested different parameters than the ones that we are concerned about.
If you believe that the sim is invalid, please propose an alternate sim so that we can test your hypothesis.
Again, thanks for the reference; will read later.
I don't think that this one affects my methodology, however.