[This thread was originally posted in the "Skeptic vs. Proponent" forum.]
This paper was referred to by Jeffrey Mishlove in his recent interview with Roger Nelson. In it, Dean Radin claims to have trained an artificial neural network using micro-PK data, so that the network was able to identify individual participants from their data more accurately than would be expected by chance:
http://deanradin.com/articles/1989%20neu...rk%201.pdf
This is what he did. The data were from PEAR Lab experiments in which participants tried to influence the behaviour of random number generators under three different conditions - high, low and neutral. The result of each trial was the sum of 200 randomly generated bits. 50 trials made a run, and series of 50 runs in each of the conditions made a full experimental session.
The first part of the paper uses one session for each of 32 participants. Each session was divided into two halves. One contained the first 25 runs of each series, and was used to train an artificial neural network to associate data with individual participants. The other half contained the last 25 runs of each series, and was used to test whether the neural network could really recognise the "signatures" left in the data by different participants. The simplest way of looking at this is to consider how many of the participants the network managed to identify correctly from the second half of the session. (Radin also presents results for some more complicated ways of looking at it.) There were 32 participants, and on the null hypothesis the whole process must be equivalent to picking each one at random, so the probability of getting each one right is 1/32. So on average we'd expect just 1 participant to be correctly identified by chance.
The other key feature of the experiment is that the training of the neural network is done by an iterative process starting from a randomised initial state. The neural network is highly non-linear, so if the initial state is changed, this process leads to a different final state. So Radin carried out the whole procedure of training and testing 100 times, starting with 100 different random initial conditions. The end result is one figure for the average number of participants identified correctly (i.e. the number averaged over these 100 repetitions of the process) and another figure for the standard deviation of the number of participants identified correctly.
He tried three different neural networks, and for the first one, the average number of correct identifications was 1.46 and the standard deviation was 1.13. The average is higher than 1, the number we'd expect by chance. But the question is whether that difference is statistically significant.
To try to answer that question, Radin generated a set of control data using a pseudo-random number generator, and applied exactly the same procedure to that. (He also did the same for a "scrambled" version of the experimental data, but the principle is similar.) The result was that the average number of correct identifications was 1.02 and the standard deviation was 0.91. Based on the averages and standard deviations for the PEAR data and the pseudo-random control data, he calculated a t statistic of 3.02 (for 198 degrees of freedom), which corresponds to a highly significant p value of 0.0014.
That seems to answer the question, but is that answer statistically valid?
This paper was referred to by Jeffrey Mishlove in his recent interview with Roger Nelson. In it, Dean Radin claims to have trained an artificial neural network using micro-PK data, so that the network was able to identify individual participants from their data more accurately than would be expected by chance:
http://deanradin.com/articles/1989%20neu...rk%201.pdf
This is what he did. The data were from PEAR Lab experiments in which participants tried to influence the behaviour of random number generators under three different conditions - high, low and neutral. The result of each trial was the sum of 200 randomly generated bits. 50 trials made a run, and series of 50 runs in each of the conditions made a full experimental session.
The first part of the paper uses one session for each of 32 participants. Each session was divided into two halves. One contained the first 25 runs of each series, and was used to train an artificial neural network to associate data with individual participants. The other half contained the last 25 runs of each series, and was used to test whether the neural network could really recognise the "signatures" left in the data by different participants. The simplest way of looking at this is to consider how many of the participants the network managed to identify correctly from the second half of the session. (Radin also presents results for some more complicated ways of looking at it.) There were 32 participants, and on the null hypothesis the whole process must be equivalent to picking each one at random, so the probability of getting each one right is 1/32. So on average we'd expect just 1 participant to be correctly identified by chance.
The other key feature of the experiment is that the training of the neural network is done by an iterative process starting from a randomised initial state. The neural network is highly non-linear, so if the initial state is changed, this process leads to a different final state. So Radin carried out the whole procedure of training and testing 100 times, starting with 100 different random initial conditions. The end result is one figure for the average number of participants identified correctly (i.e. the number averaged over these 100 repetitions of the process) and another figure for the standard deviation of the number of participants identified correctly.
He tried three different neural networks, and for the first one, the average number of correct identifications was 1.46 and the standard deviation was 1.13. The average is higher than 1, the number we'd expect by chance. But the question is whether that difference is statistically significant.
To try to answer that question, Radin generated a set of control data using a pseudo-random number generator, and applied exactly the same procedure to that. (He also did the same for a "scrambled" version of the experimental data, but the principle is similar.) The result was that the average number of correct identifications was 1.02 and the standard deviation was 0.91. Based on the averages and standard deviations for the PEAR data and the pseudo-random control data, he calculated a t statistic of 3.02 (for 198 degrees of freedom), which corresponds to a highly significant p value of 0.0014.
That seems to answer the question, but is that answer statistically valid?