I picked a lousy (slow, pre-Fourth) day to post a criticism of “IQ” last week, and badly titled it “Test Your IQ With These Puzzles! (Not So Easy!)“. Something like “A Hidden Bias in IQ Tests” would have been better, perhaps. That, and asking people to have a go at puzzles can be intimidating, or perhaps seem pointless. Anyway, too clever by half, as they say.
(Which reminds me, follow me on my YouTube channel. Do this as quickly as you can.)
Briefly, all IQ test questions have a “bias”, where I use that word in an entirely neutral way. You cannot remove conditions, i.e. “bias”, from questions, just as you cannot remove tacit premises in any logical proposition that always include the definitions and meanings of the words and symbols used. That being necessarily so, the IQ test itself cannot be a perfect measure of intelligence, and must be in part, however small, also a test of bias recall. (This adds to my other criticism of attempts to quantify intelligence.)
This isn’t a devastating blow “against” IQ tests, but it is a good point, and it’s well to keep in mind when, as some do, thinking about differences in group scores.
Even though my post died the death, one answer I did get was a gentleman insinuating my critique was wrong because he said IQ tests are the best measures we have of intelligence. Accepting that IQ tests are the best ways of assessing intelligence, at least for the sake of argument, it does not follow in any way, not even by a scintilla, that therefore there is no such thing as inherent and irremovable bias in questions. This would be like a caveman arguing that since sharp sticks are the best known way to hunt animals, all demonstrations showing they will not penetrate the hides of some animals are invalid.
I grant IQ tests are the best way, to date, of assessing a certain kind of inherent analytical abilities. And I have said innumerable times, and say once more, that IQ tests are useful and helpful, and that the many stereotypes we hold about intelligence are true. But I also insist that “IQ” is not intelligence: intelligence is intelligence, and that if you mean intelligence and not some score on a biased, but useful, test, then say intelligence and do not say “IQ”.
That off my chest, the best criticism I had was from my dad. I gave him the first question on the quiz (What is the next number in this sequence: 1, 2, 3, 4, 5, ___?) After I gave him the answer (which isn’t 6) and its explanation, he asked, politely and calmly, “What good does any of this do?”
One, and a minor one, to critique a specific aspect of IQ test. But mainly to help us notice there is much over-certainty in science.
Quoting the speech I gave a Hillsdale (from an even greater number of similar examples; also at YouTube):
Ioannidis examined forty nine top papers. Here’s what he found: “…7 (16%) were contradicted by subsequent studies, 7 others (16%) had found effects that were stronger than those of subsequent studies, 20 (44%) were replicated, and 11 (24%) remained largely unchallenged.”
…
The British Medical Journal 2017 review of New & Improved cancer drugs found that for only about 35% of new drugs was there an important effect, and that “The magnitude of the benefit on overall survival ranged from 1.0 to 5.8 months.” That’s it. An average of three months.
Richard Horton, editor of The Lancet, in 2015 announced that half of science is wrong. He said: “The case against science is straightforward: much of the scientific literature, perhaps half, may simply be untrue. Afflicted by studies with small sample sizes, tiny effects, invalid exploratory analyses, and flagrant conflicts of interest, together with an obsession for pursuing fashionable trends of dubious importance, science has taken a turn towards darkness.”
I’ve warned us many times about new drugs. These do well in the hands of drug companies. Companies duly report their mandatory statistics, using approved and customary methods of analysis, and then the drugs are made official and released into the wild. At which time their initial successes diminish. Or even disappear. Or become failures.
Leaving aside all questions of cheating, which of course is with us always, and even passing by all questions of analysis methods, of which I am hyper-critical, a reason for this typical diminishment of signal is unacknowledged and unseen biases.
One way to say this is that in these cases we bring cause to data, and data do not bring cause to us. We begin in all these analyses by assuming the new drug (or whatever, as this applies to all science) is causative in the direction we desire. We next make a guess of how strong the cause is, which brings in yet more assumptions. These are both biases, which don’t seem to be there, think many, because of the bizarre way the analyses proceed (which gives us wee Ps and all that).
The example I like to give is that we do not include the color of the physicians’ socks who administered the new drugs in the experiments in the models of the drugs’ efficacies. Because why? Because of our bias—again, using the word in it neutral meaning—against sock color being causative.
Bias is not bad! Nor good. Not inherently. It only becomes good or bad when the bias we assume does or does not match Reality. Consider not only do we leave sock color out of these models, we leave out an infinity of possible other causes! Hence the real point of last week’s post. Most of these omissions will be correct, as most things we can imagine (say, aliens in the Andromeda galaxy shooting G-rays at us) won’t be causative in our experiment.
I don’t only mean scientific experiments, but all cases in life where we are assessing what caused whatever we take an interest in.
If you’re doing an experiment testing a new drug, it is both natural and necessary to suppose it is a potential cause in whatever effect you are measuring. This is a good bias. The problem, the bad bias, comes in not supposing other causes which may be the only ones, or the main ones, operating in your experiment. And that is even natural, too, since many times these other causes won’t be measured well because the experiment is not about these, but about the drug.
And because you are so used to leaving out potential causes—after all, there are an infinity of them—that it’s really nothing to leave out a few more. Then comes the weeping and gnashing of teeth.
Not noticing what’s implicit or tacit in questions, whether in IQ tests or science experiments, is common. Trying to tease out what you don’t see is hard work; unseen good biases can elude even the best minds. Consider the very best experiments (always on things and not people) in science are those that try to control every possible cause, and manipulate only one of these, so that any effect can be ascribed to the one cause that was manipulated. But even these still leave out an infinity of other potential causes, and so it is an assumption, a bias, that you have found the one single correct one.
The main point being that this is inescapable. It is not a weakness that can be eradicated. It is part of our nature, because we do not have the ability, and never will, to see all things.
Subscribe or donate to support this site and its wholly independent host using credit card click here. Or use the paid subscription at Substack. Cash App: \$WilliamMBriggs. For Zelle, use my email: matt@wmbriggs.com, and please include yours so I know who to thank. BUY ME A COFFEE.
Discover more from William M. Briggs
Subscribe to get the latest posts sent to your email.
@WMB – I think you are off on a false trail here. Using “not perfect” as a reason for rejecting IQ as a measurement is a critique that is way overpowered; since it would reject pretty much everything anybody (including yourself) ever said on any subject. Or else you are critiquing the worst examples of misunderstanding IQ.
The history of IQ measurement in the UK is that it was used to try and reduce the effect of family background and amount/ quality of schooling on educational evaluations; when the alternative was subject based exams.
http://iqpersonalitygenius.blogspot.com/2012/08/pioneering-studies-of-iq-and.html
Prior to this, kids were evaluated using exams in subjects such mathematics, Latin, ancient history – and if you hadn’t had the needful years of education and access to books, you couldn’t do the exams. IQ tests were designed to minimize the effects of schooling, and they “discovered” that *some* kids from poor families and rural areas had the ability to benefit from academic education – the IQ scores usefully predicted their performance.
It was about doing better, not about attaining perfection.
But my general point is that I think your approach to discussing the validity of IQ seems like special pleading, and will not go anywhere useful.
bruce,
Absolutely nowhere will you find me “rejecting” IQ. The opposite is the case. I go out of my way, each time, to say it is reasonable and has good uses as a necessarily incomplete measure of a certain part of intelligence.
But “IQ” does not exist. It is not intelligence. Intelligence is intelligence is the only message I want to convey with “IQ”. If I could get people to say “intelligence” when they mean intelligence, and not say “IQ”, I will be very happy.
I believe, though, I am destined to be unhappy about this.
It’s a good thing that we don’t fully understand intelligence or other topics in psychology or sociology. The knowledge might be helpful, but with human beings as they are, overall the knowledge would be manipulated in bad ways.
If there is to be a new paradigm for science, then it doesn’t need more mechanistic models of nature. We already have enough of that. The two most valuable things would be: 1. A science that connects to a metaphysics of meaning, rather than mechanism. 2. A science that can help carve out a space for ordinary people to live decent lives, rather than to make weapons and a mechanistic society so that oligarchs and governments can enslave us.
The first is much easier to imagine how it might happen, but the speed and extent to which scientific knowledge can be co-opted is the biggest drawback. It’s a big enough drawback that it comes close to overruling the benefits and in some areas, it does outweigh them.
Despite the idealism of 250 or so years ago, with human beings as they are, there are many phenomena in nature which it is better if we aren’t able to construct manipulable models of.
Sigh. Just read a lot more from the London School.
The first thing you have to decide, before you even bother to wonder whether you need to criticize IQ, is whether or not there is such a thing as general intelligence, the g factor. Is it the case that one person, compared to another, is more reliably capable on cognitive tasks in general (i.e. whichever cognitive task, complicated thinking, you choose that person will always perform better than the person with less intelligence where you can reasonably rule out training effects). IQ is taken to measure this general intelligence, the g factor. If you have some reason to disagree with this – preferably with some actual evidence to back you up – then you don’t have to worry about IQ tests because that’s what they’re about, being a proxy for g. So, if there’s no g factor you can laugh at IQ tests all you like.
Further the London School recognised that the g factor couldn’t explain everything about human performance on cognitive tasks, and life in general, hence the Big Five Personality Traits. There are fairly reliable and reproducible psychometric tests for the Big Five, all no doubt horribly biased I’m sure. Between g and the Big Five you have it all. Or not if you don’t believe part or any of that.
Obviously no test can perfectly tease out g especially when we’re considering complicated things and if you’re looking at intelligence and capability at tasks more broadly then you need to consider g and the Big Five together in any individual person. As with anything you will end with, “what you measure is what you get.”
I happen to know that Prof James Flynn was entirely sympathetic to the concept of g despite his trash talk on IQ testing.
The big surprise while IQ testing was being developed was that, on average, men and women are more or less the same matched by age, ethnicity, social class, and what have you. That was – and still is to me – that old Tertullian thing, “credo quia absurdum”.
It’s like gravity, or evolution. They both are self evidently features of the phenomenal reality however the theories about them are full of holes.