Hi Jelo. Thanks for replying.
I would be using Amazon mTurk. That means that as long as you have the money you can get respondents. You can get about 500 responses in about an hour, if you pay a bit more than minimum effective rate (like $0.50).
My problem is not getting participants. It is not even them finishing - I usually incentivise completion and I get about 90% completion rates. My problem is that mTurkers talk amongst themselves and discuss optimum strategies to get maximum bonus. There are forums for that.
So, I do not want it to be clear while the experiment is running what it is about and how to increase your chances of maximizing utility. My argument to the ethics board is exactly that. I will disclose, but after data collection. If needs be, I'll send them an email with the debriefing notes. Once I am done. This is OT.
The maximum number of participants I can reach is about 10 million. They are not forced to participate, but do so willingly, because they get paid to do it. I don't really have the funds to pay 10 million participants, anyway. And am not looking for that many.
As I said before, in the second experiment I am looking into a low probability event. There is an opportunistic sample, of which about 2% are affected. I do not want to pay a hundred people each time in order to get two usable responses.
The problem with > 300 in group sample size is that some journals (like Psychological Science) will reject studies like that -> the sample size is too big and with such big samples all of a sudden everything has a significant effect on everything else and the observed power has long since become 1.0. Big sample sizes would be OK in EFA and CFA and if I wanted to look at interactions in AN(C)OVA, but I am interested in main effects, thus far.
If I get 95,110,100,95,120,130 and prune down, then I have an issue with some journals like, for example, Behavioural Research Methods. I guess I could argue that I wanted to stay within GLM constraints and randomly remove the appropriate number of responses and all of that would work, if I in parallel presented the results of an analysis where respondents were not removed and reached roughly the same results. That, however, would not work in my second experiment, where I would for example have 2500 vs 100 in some of the more lower-low-probability-event groups

. There is no way I can justify removing 2400 responses from a sample of 2500. And certainly I would be asked, why did I pay for all those responses and collected them, if I am chucking them now.
Of course I agree with you that there are always outliers and responses that need to be cut. In the first experiment we agreed (with my co-author) that we wouldn't attempt to control at all, so I am with you a 100% there. And thank you for your comments.
But in my second experiment, that is simply not an option for reasons mentioned above.
Thank you for taking the time. I appreciate it.
David