After machines become intelligent, since they will be able to process information a million times faster than biological neurons, that raises a question I’ve been wondering about for years: what will be the approximate magnitude of their IQs, relative to human IQ tests?
I once naively assumed the answer would be just the average human IQ (= 100) times one million, but I should have known better: IQ is not a linear measurement, since it is based on the normal distribution (“bell curve”) from statistics:
While one standard deviation is 15 points, and two SDs are 30 points, and so on, this does not imply that mental ability is linearly related to IQ, such that IQ 50 means half the cognitive ability of IQ 100. In particular, IQ points are not percentage points.
http://en.wikipedia.org/wiki/Intelligence_quotient
The crux of the situation is that one must relate processing speed to test performance. This is quite tricky. Once that is achieved using standard deviations, one can then merely divide by 15 to involve standard deviations to derive an IQ measurement. I finally found an article online called “Setting Time Limits on Tests” (Wim J. van der Linden, 2011, Applied Psychological Measurement, 35(3) 183–199) that indirectly addressed this question. After I saved the PDF file, I could no longer find the article online, so I uploaded the article to here, if you want a copy:
http://www.divshare.com/download/20336325-412
Most tests, especially IQ tests, have time limits, and the above study was able to statistically relate the risk (= pi sub theta in the article) of a test taker not finishing a test before the time limit, dependent on that test taker’s speed parameter (= tau, in the article). Of course, with very brief time spans in which to finish the test, the risk of not finishing the test on time is almost 1, and with enough time, the risk of not finishing the test on time drops to almost 0. Inbetween those two extremes is obviously a curve that drops from 1 to 0, which can be seen in the article’s Figure 2, plotted for five different values of tau.
To twist the article’s results into a form that answers my question, it’s easier to regard the article’s probabilistic curve as a measure of the test score outcome as a percentage, subtracted from 1. With that mental shift, the formulas in the article should allow nearly direct calculation of test performance based on processing speed. In particular, a human’s test performance for a given speed as measured in standard deviations (= tau) would be compared to a machine’s theoretical performance (= tau * 1,000,000 / 15), since the machine is 1,000,000 times as fast as the human, so the machine’s risk of non-completion would reach 0 one million times faster than the human. Of course the answer depends somewhat on which part of the curve is used, but an order of magnitude estimation should be very nearly the same in any case.
I was really hoping to give you all the answer, but the article turned out to be maddeningly frustrating since it did not give a clear-cut formula that spanned the early derivation to the final results, so I was not able to make these calculations in a reasonable period of time. I even went to the trouble to find an online calculator for the lognormal function and then to plot the results using typical parameters they showed, but my curves did not very closely resemble their results. I won’t say they have irreproducible results, but they do have a lot of missing parts that the reader has to fill in at what is likely an impractically long time expenditure. I’ll admit it: I failed. :-(
Therefore if anybody has time, inclination, and sufficient mathematical prowess, I encourage them to answer this question for all of us regarding the order of magnitude of intelligent machine IQ based on my earlier comments. The article demonstrates a lot of heavy knowledge of statistics that I don’t have, like the trick of using cumulants to overcome the problems of using moments…
“In some cases theoretical treatments of problems in terms of cumulants are simpler than those using moments.”
http://en.wikipedia.org/wiki/Cumulant
...and the tricks of simplification that allowed the problem to reduce to using the cumulative distribution function (cdf) from the lognormal distribution, so I was impressed, although the critical parts I am presumably not understanding seem to come from the details of the lognormal curve they are using, such as the exact parameters.
http://en.wikipedia.org/wiki/Log-normal_distribution