“We’re talking about blustering people, people who are overconfident and have no humility, who don’t think. That’s the sort of person who says ‘I know better than these people who study these things’. It’s nonsensical.
Imagine that we’re flying on a plane, and imagine that the passengers decide that they think they can fly the plane better than the captain. So they say ‘come on, we’ve had a vote and we all think that it’s our right as a citizen to land this plane rather than you’. It doesn’t matter that you’ve studied it for 20 years.
That’s what we’re dealing with when we’re dealing with these people who are in powerful positions in certain countries.” – Professor Brian Cox. Read more here.
Along with Aqua’s “Barbie Girl,” Pauline Hanson has long stood as a grim reminder that the second half of the 1990s was much worse than the first half. And now, 18 years later, Hanson finds herself back in Canberra.
Hanson’s racist agenda will be a stain on the Senate just as surely as the views she represents are a stain on Australia itself. For that reason alone, her return is a cause for dismay. But it is not the only cause.
Both Hanson herself and her wider party have a vocal sideline in science denialism: the view that expert consensus on various topics is corrupted and unreliable.
At the time of writing, it’s quite possible Malcolm Roberts, who has the number two slot on the One Nation Senate ticket in Queensland, will be joining Hanson in Canberra. Roberts is a project leader of the Galileo Movement, a lobby group who deny anthropogenic climate change and insist the global scientific community and governments are corruptly hiding the truth from their publics.
Conspiracism in public life
This might seem small beer next to the potentially disastrous effects a Hansonite revival might have on Australia’s pluralist and multicultural society.
But remember: Hanson had an outsized impact on Australian politics in the 90s precisely because she gave voice to views that resonated with much of the electorate and, unlike other politicians, wasn’t quite canny enough to reach for the dog whistle. In openly using phrases like “swamped with Asians,” Hanson shifted the Overton Window until the political establishment found the only way her views could be contained was by absorbing them.
Enter Roberts, a man who honestly believes a “tight-knit cabal” made up of “some of the major banking families in the world” are advancing corrupted climate science with the aim of global domination. Such language has some very dark associations in the history of conspiracy theory. Hence Andrew Bolt disassociated himself from the Galileo Movement for peddling a view that “smacks too much of the Jewish world conspiracy theorising I’ve always loathed.”
One might think that if even an arch-denialist like Bolt can’t abide views like Roberts’, One Nation’s climate conspiracism will end up either repudiated or ignored. But then, nobody in 1996 thought “swamped with Asians” rhetoric would have such an impact on the Australian polity either.
Besides, this has been a good season globally for political expertise bashing. Perhaps the new One Nation senators will find that, in another echo of the Howard years, the times will suit them.
In the lead-up to the UK’s referendum on leaving the European Union, Tory MP and leading Leave campaigner Michael Gove declared “people in this country have had enough of experts”. Gove is now in the running to become the Prime Minister who will preside over the UK’s divorce from the EU – and quite possibly, the breakup of the United Kingdom itself.
And of course, denialism won’t be a novelty in Canberra either. Denis Jensen won’t be there when Senator Hanson arrives, but his colleague George Christiansen will be. David Leyonhjelm may no longer grace the Senate crossbenches, but thanks to him we’ll still be paying for a Commissioner to investigate Wind Turbine Syndrome complaints despite the lack of evidence for any such condition. And lest this be dismissed as a mere lefty rant, we should also note the Greens’ stance on genetically modified organisms.
All of this might be ascribed to “post-truth politics,” the condition in which political discourse is no longer constrained by norms of truth-telling. But simply insisting people tell the truth – hardly an outrageous demand – won’t help with this specific problem. To invoke the philosopher Harry Frankfurt’s ingenious distinction, post-truth politics is not fundamentally about lies, but bullshit. The liar knows the truth, and cares about it enough to conceal it. The bullshitter, by contrast, doesn’t care (and may not know) if what they say is true; they just care that you believe it. Trump, it seems fair to say, is a bullshitter. Much of the Gove-Johnson-Farage Brexit campaign was certainly built on bullshit.
But science denialists are not, or at least not necessarily, liars or bullshitters. Their beliefs are sincere. And they are shared by a great many people, who by definition won’t be persuaded by simple appeals to expert opinion because the authority of expert opinion is precisely what they deny. How should we respond to this?
Naïve Reason won’t save us
One disastrous answer would be to retreat into a naïve conception of capital-r Reason as some sort of panacea. Surprisingly smart people end up plumping for such a view. Consider this bit of utopianism from Neil deGrasse Tyson:
Even if Tyson’s being tongue-in-cheek here, this is emblematic of a fairly widespread view that if we just consult The Facts, and then simply apply the infallible techniques of Reason to these Facts, it becomes blindingly obvious precisely What Is To Be Done. This view is only slightly less naïve, and barely less self-congratulatory, than those it opposes.
You sometimes come across people who want to insist that battles over science denialism represent a conflict between “reality” and “ideology.” But there’s no direct access to “reality” – all knowledge is mediated through our existing concepts, language, and so on – and so, arguably, no non-ideological access to it either. Human knowledge doesn’t drop from the sky fully-formed and transparently validated by some infallible faculty of Reason. It’s always filtered through language, culture, politics, history, and the foibles of psychology. Producing knowledge is something humans do – and that means power relations are involved.
Distributed knowledge and trust
While anti-intellectualism and suspicion of expertise is nothing new, the problem is amplified by the very advances that make modern life what it is. Put crudely, we now know so much that nobody can know it all for themselves, and so we have to rely more and more on other people to know things for us.
Under such conditions of distributed knowledge, trust becomes ever more important. You can’t be an expert in everything, and so you have to take more and more on trust. Is human activity warming the climate? Does the MMR vaccine cause autism? Would Brexit tank the UK’s economy? These are not questions you or I can answer, assuming you or I aren’t researchers working in the relevant fields. So we have to defer to the relevant communities of experts – and that’s a problem if you’re not good with trust or deference.
The physicist Brian Cox recently said of Gove’s expertise remark that it represents the way “back to the cave.” If that’s a fate we want to avoid, we’re stuck with distributed knowledge, and the reliance on others it involves.
That being so, we need to enhance trust in the knowledge-generating social structures we depend upon. Of course, a certain proportion of people are always going to insist that scientists are secretly lying to us for profit or that doctors are incompetent or evil. The paranoid style, as Richard Hofstadter called it, will always be with us. And there will always be demagogues willing to exploit that paranoia, to turn expertise into an us-and-them conflict, or to feed resentment and flatter egos by telling people they know better than their GP or climatologists.
But such views can only gain broader traction if people are alienated from those sources of knowledge, if they see them as disconnected from and perhaps even hostile to their own lives and interests.
Technical knowledge is predominantly produced by universities, and utilised by a political class. These are institutions that are much harder to trust if university is a place that nobody like you goes to, or if nobody in the political class sounds like you. It’s much easier to see “government” as some sort of malign, alien force if you have no investment in its processes or hope of benefiting from them. Equally, when “government” means your friends and family who work in public service rather than a distant and abstract locus of force and authority, pervasive suspicion becomes harder to maintain.
Expertise denial has become a deeply corrosive feature of modern political society. It needs to be called out wherever it appears. But we also need to think about how we reduce people’s disconnection from the sources of epistemic authority. That is a far more wickedly difficult problem. It’s one we’ll still be dealing with long after Hanson’s second fifteen minutes are over. But we can’t wait until then to start.
(An edited version of this essay was published in The Skeptic magazine,
September 2015, Vol 36 No 3 p.36, titled ‘Who needs to Know?’ It has since been republished in the Australian Doctor magazine 30 October 2015. The essay is based on a talk presented to the Victorian Skeptics in May 2015 ).
Anti-vaccination campaigner, Meryl Dorey is on record as saying that we should ‘do our own research’ instead of accepting what the doctors and other qualified experts tell us. Seasoned skeptics will be aware that ‘Do your own research!’ is a common retort by cranks and conspiracy theorists to those who dare to doubt their claims. It is a convenient escape hatch they use when trying to win a debate without the bothersome burden of providing their own evidence.
Of course, what they mean by this exhortation is not to do any actual scientific or medical research. It takes a bit of tertiary education in the relevant field to be able to do that. For them, ‘research’ means nothing more than googling for less than an hour on the Internet. They naively equate such googling with the years of study and experience it takes to become a qualified expert. Their message is that anybody with internet access can become an instant but unqualified expert on anything. Or worse still, that expertise doesn’t even count – all opinions are equal.
The reality is that googling is a notoriously unreliable source of information – there are sound reasons why Wikipedia is not allowed to be cited as a source in university assignments. The problem is that without expertise in the field in question, few googlers are capable of knowing which sources are reliable and which aren’t. Anything found on the internet becomes ‘knowledge’. Mere opinions become ‘facts’.
Another problem is that googlers are often unaware of the wider knowledge context of the specific pieces of information they have found on the internet. In contrast, experts are as much aware of what they don’t know as what they do know. As Professor Stephan Lewandowsky of the University of Bristol puts it:
‘Here is the catch: to know how much more there is to know requires knowledge to begin with. If you start without knowledge, you also do not know what you are missing out on.’
This paradox gives rise to a famous result in experimental psychology known as the Dunning-Kruger Effect. Named after Justin Kruger and David Dunning, it refers to a study they published in 1999. This study found that people who lack the knowledge or wisdom to perform well are often unaware of this fact. This is almost more dangerous than complete ignorance, because unlike Donald Rumsfeld, they don’t even know what they don’t know.
Professor Tom Nichols, a US national security expert wrote last year about the ‘death of expertise’; a Google-fueled, Wikipedia-based, blog-sodden collapse of divisions between professionals and amateurs, teachers and students, knowers and wonderers – between those with any expertise in an area and those with none at all. He sees this situation as not only a rejection of knowledge, but also the processes of knowledge acquisition – a rejection of science and other pursuits of rationality.
Nichols is particularly critical of otherwise intelligent people who are ‘doing their own research’ on the internet and second-guessing their doctors by refusing to vaccinate their children, leading to an entirely avoidable resurgence of dangerous infectious diseases such as whooping cough and measles.
So how did it all come to this sorry state of affairs? I think that there are basically four contributing factors: the blurring of facts and opinions; a misunderstanding of democracy; a misunderstanding of the Argument from Authority; and the dissipation of media accountability. I will now discuss each of these factors in turn and then outline some benefits of listening to experts.
Blurring facts and opinions
According to the Stanford Encyclopaedia of Philosophy, a fact is a state of affairs that is the case. The usual test for a statement of fact is verifiability; that is, whether it can be demonstrated to correspond to experience. Scientific facts are verified by repeatable careful observation or experiment. In other words, a fact is that which makes a true statement true. For instance, the statement ‘It is raining’ describes the fact that it actually is raining. The rain that falls can be objectively measured in a rain gauge – it is not just a matter of opinion.
On the other hand, an opinion is a judgment, viewpoint, or statement about matters commonly considered to be subjective, such as ‘It is raining too much’. As Plato said: ‘opinion is the medium between knowledge and ignorance’.
The last few decades have seen the growth of a postmodernist notion that truth is culturally relative and that all opinions are equal. What’s worse is a gradual blurring of the important distinction between facts and opinions. A disturbing feature of the public debate about climate change is the confusion between science and policy. Because they conflict with some political policies, there is a tendency for the findings of climate scientists to be treated as ‘just another opinion’. This is a marked change from a few decades ago, when the findings of epidemiologists about the links between smoking and cancer were widely accepted as facts rather than opinions.
Reducing the influence of experts is sometimes mistakenly described as ‘the democratisation of ideas’. Democracy is a system of government – it is not an equality of opinions. Whilst the right of free speech prevents governments from suppressing opinions, it does not require citizens to treat all opinions equally or even take them into account. Equal rights do not result in equal knowledge and skills. As Professor Brian Cox has said:
Deakin University philosopher Dr. Patrick Stokes has argued the problem with ‘I’m entitled to my opinion’ is that it has become shorthand for ‘I can say or think whatever I like’ without justification; and that disagreement is somehow disrespectful. Stokes suggests that this attitude feeds into the false equivalence between experts and non-experts that is an increasingly pernicious feature of our public discourse.
Professor Michael Clark of LaTrobe University gives an example of a public meeting recently, when a participant asked a question that referred to some research, a senior public servant replied: ‘Oh, everyone has a scientific study to justify their position, there is no end to the studies you could cite, I am sure, to support your point of view.’ Clark describes this is a cynical statement, where there are no absolute truths and everyone’s opinion must be treated as equally valid. In this intellectual framework, the findings of science can be easily dismissed as one of many conflicting views of reality.
Misunderstanding the Argument from Authority
A common response from cranks and conspiracy theorists (and even some skeptics) to citations of expertise is ‘that’s just the argument from authority fallacy’. Such a response ignores the obvious fact that all scientific papers and other forms of academic writing are chock full of citations of experts. The notion that the written outputs of the world’s universities and scientific institutions are all based on a logical fallacy is preposterous. Anybody who thinks that has clearly not thought through the implications of what they are saying.
The Argument from Authority is often misunderstood to be a fallacy in all cases, when this is not necessarily so. The argument becomes a fallacy only when used deductively, or where there is insufficient inductive strength to support the conclusion of the argument.
The most general form of the deductive fallacy is:
Premise 1: Source A says that statement p is true.
Premise 2: Source A is authoritative.
Conclusion: Therefore, statement p is true.
Even when the source is authoritative, this argument is still deductively invalid because the premises can be true, and the conclusion false (i.e. an authoritative claim can turn out to be false). This fallacy is known as ‘Appeal to Authority’.
The fallacy is compounded when the source is not an authority on the relevant subject matter. This is known as Argument from false or misleading authority.
Although reliable authorities are correct in judgments related to their area of expertise more often than laypersons, they can occasionally come to the wrong judgments through error, bias or dishonesty. Thus, the argument from authority is at best a probabilistic inductive argument rather than a deductive argument for establishing facts with certainty. Nevertheless, the probability sometimes can be very high – enough to qualify as a convincing cogent argument. For example, astrophysicists tell us that black holes exist. The rest of us are in no position to either verify or refute this claim. It is rational to accept the claim as being true, unless and until the claim is shown to be false by future astrophysicists (the first of whom would probably win a Nobel Prize for doing so). An alternative explanation that astrophysicists are engaged in a worldwide conspiracy to deceive us all would be implausible and irrational.
An artist’s depiction of a black hole
As the prominent British environmental activist Mark Lynas has said ‘…if an overwhelming majority of experts say something is true, then any sensible non-expert should assume that they are probably right.’
Thus there is no fallacy entailed in arguing that the advice of an expert in his or her field should be accepted as true, at least for the time being, unless and until it is effectively refuted. A fallacy only arises when it is claimed or implied that the expert is infallible and that therefore his or her advice must be true as a deductive argument, rather than as a matter of probability. Criticisms of cogent arguments from authority can actually be a rejection of expertise, which is a fallacy of its own.
The Argument from Authority is sometimes mistakenly confused with the citation of references, when done to provide published evidence in support of the point the advocate is trying to make. In these cases, the advocate is not just appealing to the authority of the author, but providing the source of evidence so that readers can check the evidence themselves if they wish. Such citations of evidence are not only acceptable reasoning, but are necessary to avoid plagiarism.
Expert opinion can also constitute evidence and is often accepted as such by the courts. For example, if you describe your symptoms to your doctor and he or she provides an opinion that you have a certain illness, that opinion is evidence that you have that illness. It is not necessary for your doctor to cite references when giving you his or her expert opinion, let alone convince you with a cogent argument. In some cases, expert opinion can carry sufficient inductive strength on its own.
Dissipation of media accountability
I have no doubt that the benefits of the internet generally outweigh the costs. However, there are some downsides that need be considered rather than just glossed over. An obvious negative is the decline of newspapers and competent professional journalism. Specialist science or medical journalists are a rarity these days. Generalist journalists often get their science stories wrong, or engage in misleading false balance – the equating of professional expertise with amateur ignorance.
Another problem is the blurring of the distinction between journalism and blogging – and I say this as a blogger myself. Unlike bloggers, journalists are subject to professional standards and editorial control. Some bloggers are anonymous, which removes their accountability to even their own readers for the accuracy of what they write.
There is a risk that when non-experts google, they are inclined to give equal weight to information from both professional journalists and amateur bloggers, regardless of its reliability and accuracy.
Benefits of expertise
Whilst experts are human and can mistakes, they have a pretty good batting average compared to laypersons. The advice that experts provide is far more likely to be true than advice from non-experts in the field in question. This has obvious benefits for society as a whole, for example in terms of public health and safety, environmental protection and managing the economy. There are good reasons why we don’t let amateurs design aircraft, bridges and tall buildings. But there are also some major benefits for the individual in listening to advice from experts as opposed to non-experts.
For instance, if you trust your doctor, you’re actually more likely to do better when you’re sick, according to a study recently published by General Hospital Psychiatry. This study, of 119 people with either breast, cervical, intestinal or prostate cancer, found that from three months following diagnosis, those patients who did not trust their doctors were not only more distressed but also more physically disabled. They were less likely, for example, to be able to go for long walks or take care of themselves. Patients who felt anxious about being rejected and abandoned suffered the most from not trusting their doctors.
Trusting your doctor has clear health benefits. You’ll be more likely to try new drugs, follow your treatment plan (jointly agreed with your trustworthy doctor), share important medical information, take preventative measures (e.g. screening) and have better-controlled diabetes and blood pressure.
Up to half of the failures in treatment reported by patients are due to not following the regime suggested by doctors. This increases the risk of hospitalisation and extended ill health. Another study at the University of California has found a small but statistically significant association between how much patients trusted their doctors and how much their symptoms improved within two weeks (allowing for different factors that could have influenced the outcome).
As Professor Michael Clark has said, people who use Dr. Google to diagnose their symptoms before visiting an actual doctor, sometimes ask to be tested for diseases they do not have, or waste time seeking a second opinion because they are convinced that their ‘research’ has led them to a correct diagnosis. If it were really that easy, would doctors have to spend all those years in medical school? Prof. Clark has also said that:
“Using Google to find the answer to Trivial Pursuit questions is not the same as researching a complex question. Experts do have skills and one of those is the ability to use high quality sources, up to date theoretical frameworks, and critical thinking based on their experience in a particular field. This is why an expert’s answers are going to be more accurate and more nuanced than a novice.”
“I’ve said it before and I’ll say it again – if you don’t think Apollo 11 landed on the Moon you are a colossal nob end and should get a new brain.” – Professor Brian Cox, on the 46th anniversary of the Apollo 11 moon landing.