• Quick note - the problem with Youtube videos not embedding on the forum appears to have been fixed, thanks to ZiprHead. If you do still see problems let me know.

How would you test for a PEAR level effect?

Zeph

Scholar
Joined
Dec 23, 2010
Messages
80
I found these forums when a search for non-vendor info on the Psyleron REG-1 pointed to a 2008 thread about same. Reading that thread prompted a thought experiment (er..) I'd like to pose to the group.

The PEAR research *claims* to have found good evidence that the human mind can affect genuinely random bit streams generated by physical processes (like reverse biased zener diode noise), but not deterministic pseudo-random number generators. However, they claim the effect is in the order of one part in then thousand. So for example, when the subject is trying to generate more 1's than 0's (or vice versa), there would be around 100001 ones to 10000 zeros (or vice versa), which did not happen when the subject was not trying to influence the bitstream. (They made other claims too, but the thought experiment is only about this limited aspect).

Some of the response here was to assume that they had probably cherry picked the data or used naive techniques. My general scepticism about "extraordinary claims" would make me suspect that is more likely than not to be the case (but that's more of a hunch than a fact and I don't have a set-in-concrete position). However it brings to mind an interesting question of "could it be done right and how?".

Obviously to sort out this effect, you would have to do a good number of experiments, to pull such a weak signal out of the noise. Does that mean that it's inherently beyond any possibility of scientific experimentation? I think not, but I'd like some thought from scientists and statisticians here on it.

How would YOU design an experiment to objectively test this as a hypothesis? I expect that one might, for example, record timestamped bitstreams along with timestamped periods in which the subject was instructed to favors 0's, favor 1's, or not try for any influence. This could be done blind in the sense that the experimenter need not themselves know which of the three modes a subject is being asked to take (eg: one/zero/neither periods could be shown on a screen visible only to the subject). One could also switch between true random and pseudo-random bitstreams without knowledge of the experimenter or subject, or even the statistician doing the initial analysis (while recording the switch for later validation).

Note that we don't need to know that the random number generator is "perfect" (the very definition of which is a difficult question). If there were a statistically significant excess of ones when the subject is so instructed, and zeros when the subject is so instructed, using the same hardware, that is in itself an interesting finding. (In fact, one could argue that if the bitstream is influenceable by the human mind, then it's obviously not perfectly random; so the question is more about whether a non-deterministic bitstream is influenceable).

Like many physics experiments, I suspect the results would be a bracketing: either "we find an effect in the range of 1 in X bits +/- error" or, "the effect could be no larger than 1 in Y bits (or we would have seen it)". That is, you could not per se "disprove" the effect, but you COULD constrain its maximum magnitude, not unlike many CERN experiments.

For the statistically well informed: what specific statistical methods would you use to correlate a stream of 1's and 0's with a slower recording of "intention" periods? How many tests would it take? How would you define your p for "showing an effect" and for "constraining the maximum magnitude of an effect not found"?

After answering the simpler question, let's make the thought experiment a bit harder. Suppose the effect is stipulated to potentially work better for some people, or on some days than on others, just as some people play golf better on some days than others - that is, as a normal aspect of varying human performance rather than as a deliberate scam. An acknowledged real ability at the limits of perception or cognition might have such variability (something which comes to mind from conventional science is a recent experiment in which soccer experts predicted the score of a game from limited information and were rated as doing better or worse than random chance under variations in attention and distraction). You would of course have all the data, for good and bad days, and would need to find a way to analyze it fairly without biased cherry picking. Experimenters in other subjective psychological areas have dealt with variable performance, but maybe their primary "signal to noise" ratio was higher than 1 in 10^4.

Would this combination (statistically small effect AND potentially varying performance) make such a hypothesized effect theoretically or practically impossible to investigate under any circumstance, rendering it inherently outside the envelope and authority of science? Or could an experimental regime be designed which would eventually either demonstrate an effect (with increasing confidence in the error bounds), or constrain the upper limits of such an effect (with decreasing limits)?

Zeph
 
I found these forums when a search for non-vendor info on the Psyleron REG-1 pointed to a 2008 thread about same. Reading that thread prompted a thought experiment (er..) I'd like to pose to the group.

The PEAR research *claims* to have found good evidence that the human mind can affect genuinely random bit streams generated by physical processes (like reverse biased zener diode noise), but not deterministic pseudo-random number generators. However, they claim the effect is in the order of one part in then thousand. So for example, when the subject is trying to generate more 1's than 0's (or vice versa), there would be around 100001 ones to 10000 zeros (or vice versa), which did not happen when the subject was not trying to influence the bitstream. (They made other claims too, but the thought experiment is only about this limited aspect).

Some of the response here was to assume that they had probably cherry picked the data or used naive techniques. My general scepticism about "extraordinary claims" would make me suspect that is more likely than not to be the case (but that's more of a hunch than a fact and I don't have a set-in-concrete position). However it brings to mind an interesting question of "could it be done right and how?".

Obviously to sort out this effect, you would have to do a good number of experiments, to pull such a weak signal out of the noise. Does that mean that it's inherently beyond any possibility of scientific experimentation? I think not, but I'd like some thought from scientists and statisticians here on it.

How would YOU design an experiment to objectively test this as a hypothesis? I expect that one might, for example, record timestamped bitstreams along with timestamped periods in which the subject was instructed to favors 0's, favor 1's, or not try for any influence. This could be done blind in the sense that the experimenter need not themselves know which of the three modes a subject is being asked to take (eg: one/zero/neither periods could be shown on a screen visible only to the subject). One could also switch between true random and pseudo-random bitstreams without knowledge of the experimenter or subject, or even the statistician doing the initial analysis (while recording the switch for later validation).

Note that we don't need to know that the random number generator is "perfect" (the very definition of which is a difficult question). If there were a statistically significant excess of ones when the subject is so instructed, and zeros when the subject is so instructed, using the same hardware, that is in itself an interesting finding. (In fact, one could argue that if the bitstream is influenceable by the human mind, then it's obviously not perfectly random; so the question is more about whether a non-deterministic bitstream is influenceable).

Like many physics experiments, I suspect the results would be a bracketing: either "we find an effect in the range of 1 in X bits +/- error" or, "the effect could be no larger than 1 in Y bits (or we would have seen it)". That is, you could not per se "disprove" the effect, but you COULD constrain its maximum magnitude, not unlike many CERN experiments.

For the statistically well informed: what specific statistical methods would you use to correlate a stream of 1's and 0's with a slower recording of "intention" periods? How many tests would it take? How would you define your p for "showing an effect" and for "constraining the maximum magnitude of an effect not found"?

After answering the simpler question, let's make the thought experiment a bit harder. Suppose the effect is stipulated to potentially work better for some people, or on some days than on others, just as some people play golf better on some days than others - that is, as a normal aspect of varying human performance rather than as a deliberate scam. An acknowledged real ability at the limits of perception or cognition might have such variability (something which comes to mind from conventional science is a recent experiment in which soccer experts predicted the score of a game from limited information and were rated as doing better or worse than random chance under variations in attention and distraction). You would of course have all the data, for good and bad days, and would need to find a way to analyze it fairly without biased cherry picking. Experimenters in other subjective psychological areas have dealt with variable performance, but maybe their primary "signal to noise" ratio was higher than 1 in 10^4.

Would this combination (statistically small effect AND potentially varying performance) make such a hypothesized effect theoretically or practically impossible to investigate under any circumstance, rendering it inherently outside the envelope and authority of science? Or could an experimental regime be designed which would eventually either demonstrate an effect (with increasing confidence in the error bounds), or constrain the upper limits of such an effect (with decreasing limits)?

Zeph

Hey Zeph, are you asking if the PEAR claims should be considered meritorious at first blush, or are you asking if these claims can actually be investigated within the "authority of science" as you put it? Because the former is an interesting question, but the latter is a pseudoscientific dodge that is seen all the time.
 
I'm asking a question about what kind of serious scientific experimental design could in theory investigate a hypothetical phenomena such as PEAR posits, if one wanted to start afresh. Let's for the moment assume that perhaps they didn't do it right, so their results are not sufficient to convince skeptics. However, they also didn't disprove it. (I'm asserting that the nature of "disproof" would be more along the lines of a statistically generated upper bound on the magnitude of an effect, a not-uncommon case with science).

I'm puzzled by some kind of unstated assumptions behind your question, and I imagine that it relates to some old argument and some "pseudoscientific dodge" you perceived somebody to be making back then, and thus you imagine I might have the same agrumentative motivation. I do not, please let that go and take it on face value.

One can just read the following paragraph and skip the rest of this post to get the essence:
***
Would it be possible to design an experimental regime which would either (1) detect and measure to the satisfaction of rational sceptics such an effect if it indeed existed, or (2) constrain the maximum magnitude of such an effect which would have been below detection thresholds. If so, how? If not, why?
***

Optional elaboration:

I'd like to have us design some procedures and statistical tests which would yield statistically sound results like:
1) The effect exists, with people able to change about 85 +/- 20 bits per million, or
2) The effect if it existed would have to be below 1 bit per million (or our tests would have detected it). Either with p=.95

Either outcome would be valuable; I'm not trying to prove either case, but to ask how we could definitively produce one or the other of these outcomes with proper scientific process. Or discover that such an outcome seems impossible within the scientific method, and that nobody here could come up with such a make-or-break scientific procedure.

I have given some suggestions about experimental design (three intention states recorded and displayed to the subject but not made visible to the experimenter, randomized switching between non-deterministic random number generator and pseudorandom number generator unknown to both subject and experimenter,
independent statistical analysis of the result, etc). I would add full publication of the data for additional analysis.

I think the blind switch between random and pseudorandom aspect would be very enlightening. Suppose that each week the device randomly chose to use a non-deterministic random noise source, or a pseudorandom number generating algorithm, but neither the experimenter, the subject, or the initial statistician knew which weeks were which. If the statistician show significant effects on weeks using both, we have reason to doubt the analysis (at least according to the effect as asserted by PEAR). On the other hand, if the post-experimental statistician show more effect on weeks using the true random number generator, that would be a (weak) level of triple blind analysis.

The idea here is to design an unbiased scientific and statistical procedure, which (1) should genuinely demonstrate the effect if real and large enough, and (2) should otherwise demonstrate the lack of effect (or upper bound to its magnitude). The point is to leave few "pseudoscientific dodges" for either side to hide behind at the end of the experiment. (Sigh, it can never be zero - one could assert that the stars were in the wrong place. But we could eliminate all reasonable credible objections from either side).

Or to understand why such a clean procedure is not even theoretically possible within the scientific method.

Zeph
 
I'm asking a question about what kind of serious scientific experimental design could in theory investigate a hypothetical phenomena such as PEAR posits, if one wanted to start afresh. Let's for the moment assume that perhaps they didn't do it right, so their results are not sufficient to convince skeptics.

You type a lot. To assume they (PEAR) didn't do it "right" is an assumption that many with credentials far better than mine have already posited. Convincing "skeptics" isn't necessary; what is is evidence that PSI effects can be replicated or demonstrated with results more frequent than chance. Such results would make posts like these superfluous.
 
Resume, you seem to be trying to answer the question "do PEAR's existing results hold up to scrutiny?". That has been discussed at length in other threads, of which you are probably aware if you have more to contribute on that topic.

I'm asking a different question which I haven't yet seen considered much less examined to death: apart from what PEAR did or did not do, would it be possible to do the investigation "right", and if so how? If not, why?

It's easy to say "oh, they must have cherry picked the data" than to put oneself forward to describe how it could have been performed and statistically analysed, so I'm not surprised that some would find this "too much work for too little bloodletting". But others seem to have some knowledge of statistics, and willingness to actually think things out rather than rely on instinct. So for example, one simple question would be:

Suppose experimenters released a sequence of time consecutive value pairs - the first number being the 1 or 0 produced by a random process, the second being 1, 0, or X representing the subject's intention at that time. The former would change as the random generator cycled; the latter would change much more slowly, as the subject's instructed intention changed. What statistical test would be appropriate for determining whether there was a signficant corelation between intent and the generated bits? How specifically would you calculate and quantize the number of excess "hits" over the course of a sequence of such sessions, and the error bars? Or how would you calculate the p = .95 maximum level of such hypothetical effect (eg: in terms of one bit of deviation from norms per N bits of the sequence) which could have escaped detection in the testing series?

I would be quite satisfied to see that the upper bound was set at, say, no more than one in 10^6 bits. That would not "prove" that any such effect didn't exist (it's pretty much impossible to do that in an absolute sense - what if the mind's effect was only about one per quadrillion bits? pretty hard to disprove that!), but it would reduce the maximum size of the effect to irrelevancy and cast appropriate doubt on any effect.

Or I would find an appropriately analyzed result which did show and statistically characterize an effect a valuable result. Either one would be a result.

My goal isn't to validate some pre-existing belief I have about the existence of the PEAR-suggested mental effects, but to demonstrate that the scientific method would be capable of making such a determination to the satisfaction of scientists and sceptics.

Again, this question can stand alone, it's not about PEAR's actual result or methods, but about how it COULD be done right. PEAR's documents are just the inspiration for this thought experiment.

Suppose a group of non-woo people were to purchase one of those REG-1 devices and wanted to conduct truly meaningful tests done correctly. How would the skeptics here suggest the tests be done and analyzed - such that they would accept the results as valid. (Of course the next step would be to seek to reproduce the effects if any were shown - but that's beyond the current question).
 
Hey Zeph, I came across this today and I remembered this thread. Thought you might find it interesting.

THE CAPRICIOUS, ACTIVELY EVASIVE, UNSUSTAINABLE NATURE OF PSI: A SUMMARY AND HYPOTHESES

ABSTRACT

Many parapsychological writers have suggested that psi may be capricious or actively evasive. The evidence for this includes the unpredictable, significant reversal of direction for psi effects, the loss of intended psi effects while unintended secondary or internal effects occur, and the pervasive declines in effect for participants, experimenters, and lines of research. Also, attempts to apply psi typically result in a few very impressive cases among a much larger number of unsuccessful results. The term unsustainable is applicable because psi is sometimes impressive and reliable, but then becomes actively evasive. One of the most testable models for this property is that psi effects occur against a background of supporting and opposing motivation and psi influence due to the extreme polarization of attitudes toward psi in the population. These attitudes may have genetic and gender associated components. Another possible explanation is that the primary function of psi is to induce a sense of mystery and wonder. Other possible functions of psi also need to be investigated. For example, psi could contribute to evolution by briefly influencing random processes to enhance diversity, without specifically guiding evolution or having sustained effects. Some type of higher consciousness may influence or control psi effects.

[...]

Shift From Intended Effects to Unintended Secondary Effects

Another seemingly capricious or defiant psi manifestation is when the overall intended effect becomes nonsignificant, but unintended secondary effects provide evidence for psi. The Princeton Engineering Anomalies Research (PEAR) laboratory provides a recent example. Studies with electronic random event generators (REGs) had small but significant effects for a decade. A recent large-scale replication effort obtained nonsignificant results overall, but Jahn et al. (2000) reported internal structural effects that appeared to indicate psi. The analyses for these effects were based on findings in the previous data, and the effects were reported as significant after adjusting for multiple analyses. However, the effects had different patterns than the earlier results and were not consistent across the three laboratories participating in the project.

Jahn and Dunne (2001) summarized the situation as follows: “At the end of the day, we are confronted with an archive of irregular, irrational, yet indismissable data that testifies, almost impishly, to our enduring lack of comprehension of the basic nature of these phenomena” (p. 300). They noted that these changes in psi manifestations are not consciously intended or desired by those conducting the studies, and suggested that unconscious processes may have a major role in psi effects and the associated inconsistencies.

The evolution of research at the Princeton laboratory is notably similar to the earlier experience at the Duke laboratory. At Duke the initial research was remarkably successful in demonstrating the intended effects. However, a decade later, unintended, internal effects were increasingly being reported as the primary finding. J. B. Rhine (1946a) commented: [M]any of the experiments yielding only chance totals have proved to be fruitful in other respects. Analyses of different character than those initially applied have in some cases revealed hidden relationships that were first overlooked. Some of the most important discoveries concerning PK, like those concerning ESP, have emerged in this way. (p. 73) In fact, J. B. Rhine (1974) argued that these internal effects were some of the best evidence for psi because...

[...]
 
Thanks, Limbo. I'll look more into that when I have a bit of time, it sound interesting.

If, for the purposes of discussion one considers the implications of asserted PEAR type dynamics (mental modification of results without any path of causation explainable by current science), one could conclude that experiments could be very hard to conduct without experimenter bias. That is, if the subject can mentally influence results, the experimenter might also. And a skeptical experimenter who was consciously or unconsciously trying to disprove any effect could hypothetically "override" the subject's effects.

This prospect could easily lead to a morass, as skeptics rightly object to results which cannot be objectively demonstrated to a skeptical or hostile examiner. The scientific method strongly depends on effects being stable and replicatable in the face of such skepticism. A hypothetical effect which due to its nature could be suppressed by the mental attitude of the hostile observer would be rather hard to investigate! At that point one would come to a philosophical divide (not unlike how quantum mechanical investigators have different interpretations of Schroedinger's Cat).

Some people would in effect define "reality" as no more and no less that that which can be investigated by the scientific method: if something cannot be so investigate, then it would by their definiation be automatically false and untrue. Other people would put such a phenomenon as "outside the scope of science", but not necessarily unreal or untrue (nor necessarily real or true!). It's important to note that this is a philosophical choice, not an objective scientific one. The methods of science probably cannot be used to "prove" that every aspect of the universe which is "real" must be able to be investigated by science - that would be an article of unsupported faith.

It's interesting however that the opposite can be proven within the more limited field of abstract mathematics. It is possible to prove, using the tools of mathematics, that there can exist true hypotheses within a sufficiently complex mathematical system, which absolutely CANNOT be proven with the tools of that system - that is, it's provably impossible, not just something we haven't yet come up with.

The fields are different, so this is no more than suggestive rather than "proof" or even "evidence" - but if some true things in complex mathematical systems can never be mathematically proven, it's not an inherently absurd conjecture that some true things in the physical universe might not be provable or disprovable by the scientific method.

I do not offer this as reason to "just believe whatever you want" in a woo woo way. A great MANY things *can* be fruitfully investigated with the scientific method, and that may include PEAR-like effects, with proper care. I'm only suggesting a lack of hubris, not a lowering of standards where they do apply.

I personally have no fear of the concept that some things might exist which are not provable as true or false scientifically. However, I have some trepidations about the social ramifications of irrationality (vs arationality), and the lack of popular discernment of that distinction. Even if Psi is inherently capricious, evasive and unsustainable as the source you cite suggests - that still doesn't in the least mean that every new age gadget or practice should get a free ride. If they make scientifically testable assertions, those should be subjected to proper scientific scrutiny!

Perhaps the source which most affected my concerns about irrationality personally was "The Crucible" by Miller, a fictional account of the Salem witch trials based in part on the actual historical situation. That drove home how unsafe a world would feel where, say, a perfect alibi that somebody was at definitely home at the time of an alleged assault elsewhere is not a defense, because in their worldview the accused's "spirit body" could have simultaneously done the crime, for which their real body would now have to pay. It would be nice to think that, absent the constraints of logical scientific worldview, we could have a world of love and light - but historically it's more likely that the dark demons of our psyche would run more rampant, and we'd all too often be hanging neighbors for mystically causing our cows to run dry. So I'm cautious about expanding the view that some things might be true but hard to verify by science, until we can get people to better accept the authority of science and rationality regarding the many things for which it IS appropriate.

The very idea that some things might be true but unprovable will make me sound woo woo or like a woo woo apologists to the more dogmatic here, while the more mystical may find the great weight I do still give to scientific means too oppressive. So it goes. I'd like make contact with more kindred souls looking for nuanced discernment rather than blanket dogmatic ratification (on either "side"), however.
 
Some of the response here was to assume that they had probably cherry picked the data or used naive techniques. My general scepticism about "extraordinary claims" would make me suspect that is more likely than not to be the case (but that's more of a hunch than a fact and I don't have a set-in-concrete position). However it brings to mind an interesting question of "could it be done right and how?".

I'm probably going to regret this, but I will assume you are asking the question in good faith and looking for an answer from the scientific perspective.

It's actually fairly easy to do. Any competent statistician can set up an experiment for you. Basically, you need to calculate beforehand what the needed sample size is, based on the estimated difference between the two. This site allows you do to the calculations; for the difference you're looking at, you'd need about a billion observations under each condition to in order to have reasonable confidence in the accuracy of your results, but also sufficient power to find what you were looking for.

The trick, though, is that you get ONE shot to look for the effect. You don't get to reanalyze the data, or try to do any sort of re-sampling or repeated measures, or anything like that. Post hoc statistics like this are almost always suspect for this reason; basically, it amounts to running twenty or more experiments until random chance throws up a "significant" finding at the 0.05 level.
 
Thanks, Limbo. I'll look more into that when I have a bit of time, it sound interesting.


No problemo, consider it an X-mas present.

This prospect could easily lead to a morass, as skeptics rightly object to results which cannot be objectively demonstrated to a skeptical or hostile examiner.


Yup. I've given the problem a bit of thought. You are talking about known parapsychological effects. You are talking about the sheep-goat effect and the parapsychological experimenter effect.

A large enough group of talented long-term expert meditators, acting together in psychic unison, might be able to overwhelm the morass. There is a correlation between meditation and psychic strength.
 
Last edited:
Yup. I've given the problem a bit of thought. You are talking about known parapsychological effects. You are talking about the sheep-goat effect and the parapsychological experimenter effect.

Both of which are well-understood by people trained in statistics to be non-existent effects and almost purely an artifact of the kind of oversampling of data I described in the previous post.


A large enough group of talented expert meditators, acting in psychic unison, might be able to overwhelm the morass.

Or Superman might be able to fly against the earth's rotation fast enough to go back in time and fix the random number generator. I mean, if you're going to make up fairy tales, why not go all the way?
 
[/URL].

A large enough group of talented long-term expert meditators, acting together in psychic unison, might be able to overwhelm the morass. There is a correlation between meditation and psychic strength.

Geez, citation?
 
There is a fundamental problem with PEAR's experiments, not the least of which is distinguishing significant data from noise.

Let's assume your premise is that mental activity can influence the output of an electronic device. Your test will consist of (1) a control, where no mental influence is attempted, and (2) a test when mental influence is attempted.

Your goal would be for the results of (2) to be statistically significant compared to (1).

Now the problem here is: How do you know that the baseline isn't being affected by random brain "waves"? Conversely, how do you know that test (2) is being bombarded by brain "waves" at all? What if a lab rat or someone nearby was thinking really hard and affected (1), or if your experimenter was kidding when he said he was trying to mentally alter (2)?

Unless there is some alternate, reliable, repeatable way of detecting and quantifying mental waves, these problems will doom all tests to absolute failure. You cannot design a test to eliminate the unwelcome possibilities.
 
I'm probably going to regret this, but I will assume you are asking the question in good faith and looking for an answer from the scientific perspective.

Wow, you must be pretty jaded or battle hardened from previous encounters to have so much trepidation. My sympathies. But I am indeed asking in good faith.

Thanks for the link. I have to go soon to other things, but I'll follow up later.

It's actually fairly easy to do. Any competent statistician can set up an experiment for you. Basically, you need to calculate beforehand what the needed sample size is, based on the estimated difference between the two. This site allows you do to the calculations; for the difference you're looking at, you'd need about a billion observations under each condition to in order to have reasonable confidence in the accuracy of your results, but also sufficient power to find what you were looking for.
Wow. Does that mean a billion bits in the continuously and rapidly running bitstream, or a billion periods of subject setting an intention (eg: for 1 minute at a time)? The former might be no problem, if the random bitstream can be run fast, but a billion periods-of-intention-setting would be a bit burdensome.

The trick, though, is that you get ONE shot to look for the effect. You don't get to reanalyze the data, or try to do any sort of re-sampling or repeated measures, or anything like that. Post hoc statistics like this are almost always suspect for this reason; basically, it amounts to running twenty or more experiments until random chance throws up a "significant" finding at the 0.05 level.
Yes, I get that. But we need more nuance than "you have one chance", in that it's common for scientific data to be reanalyzed by other scientists, perhaps using other techniques which they assert to be more relevant. I think you are more talking about mining the data by changing time periods, or sample sizes, or other parameters and ignoring any negative results, until you finally find some measure which yields statistical significance. Your point is good tho.

Anyway, do you have a specific statistical test you would suggest?

Suppose (to make this more concrete) one were to generate & record 1000 random bits per second, and ask the subject to set an intention (three states: 1's, 0's or no preference) for periods of 1 minute at a time, recording along with the bitstream what intention was being set while those bits were generated. Using, say, R, what specific statistical tests would one use?

Thanks for any help. I'll continue looking elsewhere as well, but one of the points in this "thought experiment" is: "What experimental regimen and statistical tests would rational skeptics accept (if it quantized an effect), and rational supporters accept as constraining the effect?".

Imagining the experimenter picking the statistical methods independently, and then later the skeptic (or would-be-believer) gets to say retrospectively - "oh, that was the wrong way, your results are meaningless". What a waste. Better of skeptics and believers agreed that a particular statistical approach was appropriate in advance, and neither got to selectively reject it post-experiment.

For reference - digging into climate change science, I see scientists and statisticians on both sides frequently dispute the appropriateness of a given statistical measure, depending on what they want to prove. So I know it's very possible to later say "you wasted your time gathering that kind of data, because I now think it's not a valid way to analyze things".

Zeph
 
Last edited by a moderator:
Hmmm... Well, I actually suggested this to one of the PEAR researchers about 15 years ago, and they couldn't really be arsed. I did some work with George Marsaglia, one of the world's foremost researchers on random numbers, and we did a lot of stuff to test whether they were really random.

My suggestion was that an independent group produce number generators. Some of these would be strictly deterministic pseudo-random number generators, and some would be real random number generators, using something good like radioactive decay instead of thermal noise, which has a lot of problems. The researchers who used these to test their anomalies would have no way of telling which was which. I even suggested making them available over the Internet.
 
epepke,

Your thought (long before me) of using both random and pseudo-random generators, kept blind from the experimenters, agrees with the very tentative protocols I was suggesting above. (I was also suggesting that the random/pseudorandom status of particular tests be kept blind from the initial statistical analysis but later exposed for review and further analysis). I also agree with the idea of publishing the bitstream level results for independent post analysis; if the analysis was done in R, that too could be published for review.

I'm less certain whether the difference between thermal noise and radioactive decay (for example) really matter. Certainly a positive result (mental intention can slightly influence random numbers generated by thermal noise/radioactive decay) would be similarly significant in either case. Either would indicate a phenomenon violating current concepts of causation.

Of course, if one could demonstrate and characterize the effect with one source, it could be interesting to test the other using the same protocols and see if the effect was larger, smaller, or absent.

Using both forms of random generators from the start would be interesting in one way - in the case of negative results, there would be no room left for saying "well sure, there was no effect with thermal noise, but there would have been with radioactive decay". That is, if some believers were asserting that the noise source made a difference, then testing all the asserted "works with" sources would be useful in producing a comprehensive negative result. As I understand it, PEAR asserts that the random number generator source doesn't matter much (but that it doesn't work with pseudorandom number generators). That doesn't mean they are correct.

Likewise I don't see that the experiment depends on extra high quality random numbers. To take an example, if one used a generator which produced 49% ones and 51% zeros, BUT it could be demonstrated that human intentions would modify these ratios by one in 10^4 bits when comparing the same not-smoothly-random streams with and without intentions, it would still be a significant result. It's the unexplained modification of non-deterministic events (or lack thereof) which matters. (I guess that a particularly bad random number generator - like one the kept getting stuck for a while with all 1s or all 0s occassionally - might make it harder to statistically pull the signal out of the noise).

But these are my thoughts. If there is some way in which the purity of the random numbers matters more than I realize, I'd like to understand.
 
epepke,

Your thought (long before me) of using both random and pseudo-random generators, kept blind from the experimenters, agrees with the very tentative protocols I was suggesting above. (I was also suggesting that the random/pseudorandom status of particular tests be kept blind from the initial statistical analysis but later exposed for review and further analysis).

Hmmm. What if multiple random number generators were running and being recorded at the same time, as subjects changed their intentions? For example, a thermal noise generator and a radioactive decay or photon based generator, and a pseudorandom generator, could all be operating and recorded.

This would have to be taken into account in the statistical analysis. Obviously, if you ran 20 simultaneous random number generators, you not be surprised if you get "statistically significant results" from at least one of them at the p=0.95 level! On the other hand, if 10 of the 20 showed p=0.95 level results, that might be even stronger evidence than a single generator at that level. So I believe that multiple simultaneous random bitstream generators could be compensated for by raising the bar appropriately. That's where a real statistician would come in.

The advantage is not having to run N times as many tests. Also, this might test whether different random sources give different results, under the same conditions (same subjects and intentions).
 
Wow. Does that mean a billion bits in the continuously and rapidly running bitstream, or a billion periods of subject setting an intention (eg: for 1 minute at a time)? The former might be no problem, if the random bitstream can be run fast, but a billion periods-of-intention-setting would be a bit burdensome.

A billion bits in the bitstream at under each condition you want to study; a billion bits for the control stream, a billion under the "think 'one' thoughts," and a billion under "think 'zero'" if you want to test that as well.

Basically, the variation in the number of 1's you get in a random stream is about the square root of the number of observations. If you flip a hundred (fair*) coins, you'll get a fifty heads, plus or minus 10 (the square root of 100). If you flip 10,000 coins, you'll get five thousand heads, plus or minus a hundred. (There's a constant multiplier in there that I'm neglecting.)

So you need a large enough n that sqrt(n) is much smaller than one part in 10,000. Since 10,000 squared is a hundred million, you will need about a billion bits. In theory, with a billion bits, if the PEAR hypothesis is right, you will get about 500,050,000 ones instead of the 500,000,000 you expect with the control condition -- in either case, plus or minus about 30,000, so "chances are" you'll see more ones than can be explained by chance alone.

At this point, a simple binomial test will give you the p-values if you need it. Statistically, this is pretty iron-clad (if you get an actual statistician to do the numbers right instead of relying on an Internet posting that's correct to an order of magnitude....), but the execution and more importantly interpretation will still be a problem.

If, as you suggested, you can generate random numbers at 1000/second, it will take about a million seconds, or approximately thirty years, to run the experiment. And that's running 24/7, with no time off for Christmas or August Bank Holiday; on a more normal schedule, that would be something like 120-150 years of lab time. Good luck finding a funding agency that would approve that.

And you'd need at least double this long, because you'd need to get an equally long run to establish a control set; one of the standard alternative explanations for PEAR's findings is that no one can calibrate their equipment to one part in 10,000. You'd need to find a single set of equipment that could be expected to run without changing operating characteristics for three hundred years, and you'd need to be able to confirm that it was, in fact, meeting those expectations (in this case, the calibration would take as long as the entire experiment, for the same reason).

And even if you could run the experiment, it would still probably not be accepted by either side. Limbo's already explained about the "known parapsychological effects" like "the sheep-goat effect and the parapsychological experimenter effect." These effects basically "mean" that any time an incompetent researcher produces a finding that evaporates in the lab of a competent one, that means the competent researcher is doing something wrong; in other words, the PEAR hypothesis is unfalsifiable, because anyone who can't duplicate it must be subconsciously suppressing it.

On the skeptic side,.... there are too many well-documented instances of incompetence amounting to fraud in the data handling of PEAR and PEAR-like groups; I no more trust their findings than I trust the cancer research done by the Tobacco Institute.

Beyond this, there's the simple fact that such an ability, even if confirmed, would be of no practical significance whatsoever; even if the statistical significance were tremendous, the "clinical" significance, or usefulness in explaining real-world phenomena, would be negligible.
 
Hmmm. What if multiple random number generators were running and being recorded at the same time, as subjects changed their intentions?

You'd need to prove that there was no possible effect that could cause correlations among the RNGs. If, for example, cosmic rays generated by the solar wind affected radioactive decay (as they are known to do), then every decay generator on Earth would be affected at the same time. For example, a thermal noise generator and a radioactive decay or photon based generator, and a pseudorandom generator, could all be operating and recorded.

Since I don't think such a proof is possible, I think multiple generators would invalidate the results.
 
Something here does not jell, but I don't have time to follow it up at the moment.

Jahn claims strong statistical significance for the meta analysis of experiments over 12 years.

We can doubt that this is valid on the basis that it is unlikely that they had such rigid, failsafe lab procedures over that time that there could be no bias in the samples committed to the database.

But I have never heard anyone question his math.
 

Back
Top Bottom