July 19, 2024 This essay falls into three parts. First, I discuss the question of what it is which makes humans unique — that is, irreplaceable. Second, I consider whether machines on balance enhance or diminish humanness. This has become an issue of the moment with the growth of machine intelligence. Finally, I try to answer two questions: how can we secure our survival as human beings? Is it worth trying to do so? A quick preview of my answer to the first question. Some bits of humans are clearly replaceable. They fall into the category of spare parts. The bits which aren’t are what used to be called soul and which we now call mind: in religious language, the bits which link us to the Divine. We urgently need to decide which bits should and should not be
Topics:
Robert Skidelsky considers the following as important: AI, artificial intelligence, consciousness, Featured, Interviews, Philosophy, Technology
This could be interesting, too:
Michael Hudson writes Beyond Surface Economics: The Case for Structural Reform
Robert Skidelsky writes Speech in the House of Lords – Autumn Budget 2024
Robert Skidelsky writes In Memory of David P. Calleo – Bologna Conference
Michael Hudson writes US Election: The Illusion of Choice
July 19, 2024
This essay falls into three parts. First, I discuss the question of what it is which makes humans unique — that is, irreplaceable. Second, I consider whether machines on balance enhance or diminish humanness. This has become an issue of the moment with the growth of machine intelligence. Finally, I try to answer two questions: how can we secure our survival as human beings? Is it worth trying to do so?
A quick preview of my answer to the first question. Some bits of humans are clearly replaceable. They fall into the category of spare parts. The bits which aren’t are what used to be called soul and which we now call mind: in religious language, the bits which link us to the Divine. We urgently need to decide which bits should and should not be replaced so as to avoid being replaced entirely.
1.What is unique about being human?
The Christian answer is that humans were created in God’s image and carry within them a spark of the Divine. Uniquely among living things, we carry the image (or in some versions the actual substance) of His perfection in us. This is at the root of the human striving for perfection.
We alone of living things have been gifted with awareness of who we are, what we are doing, and of our relations with the world.
We never meet a pig that knows it is a pig. Indeed, the word “meet” is questionable, for it signifies mutual recognition.
The palpable gap, the yawning chasm, between humans and animals is the big flaw in Darwinian evolutionary theory. The break between brain and mind is inexplicable in purely evolutionary terms.
Another way of saying this is that the mind can’t be reduced to the brain. An artificial brain may be able to replicate everything the human brain does without being human.
Seventeenth-century philosophers transiting from theology to science tried to keep a physical place for the soul. Descartes looked for it in the pineal gland. The progression then was from soul to mind and from mind to brain. For the materialist, mind is simply a complex brain.
This is wrong. The brain is an information processor. Animals have brains too. The difference between humans and animals cannot be reduced to one of brain size or complexity.
In mainstream discussion, knowledge is equated with information: the greater the brain’s storage and processing capacity, the greater the sum of knowledge. But as Francis Sheed says, knowledge brings problems. If all knowledge is reduced to information, problems disappear. One can trace this slippery slope very easily in economics, for example in models of perfect foresight. This abolishes uncertainty and all the problems which arise from uncertainty.
The argument goes that once scientists have cracked the code of thinking, humans will be completely replaceable by machines. More ominously, they will be better at thinking than humans, because they will never go wrong, as humans still do, whether from ignorance or from neurological imbalances. Humans will then become permanently redundant, though they may be kept on as pets.
2. Do machines enhance or diminish humanness?
This argument is hard to summarise. I would argue that machines enhanced our humanness in the past, but are likely to diminish it in the future, as the ratio of machine capacity to human capacity steadily rises.
The optimistic case was most famously put by the economist John Maynard Keynes, in his essay “Economic Possibilities for our Grandchildren”, published in 1930. It is all the more remarkable for having been written in the depth of the Great Depression, when the world seemed to be sliding downhill into a new dark age. Extrapolating from the technological progress in his own lifetime, Keynes calculated that his putative grandchildren — that is, people still alive today — would only need to work three hours a day to obtain all that they needed for a good life. The historian Arnold Toynbee echoed him: machines would make possible the “transfer of energy from some lower sphere of being or action to a higher”. In this prognosis , machines were unmistakably benefactors, one might even say the agents of God.
On the opposite side has always been the fear of redundancy, of uselessness. The promises were for the future, while in the present machines threatened to deprive workers, not just of the means of life but of the meaning of life.
The economist David Ricardo, writing at a time when the Luddites were smashing machinery, believed that the substitution of machinery for human labour “was often very injurious to the interests of the class of labourers”. If — remember this was at the beginning of the 19th century — a farmer sacked fifty human workers and replaced them by horses, “this would not”, Ricardo remarked soberly, “be in the interest of the men”. If machines made all jobs redundant, the human race itself would become redundant. Interestingly, Ricardo viewed horses as “machines”.
The debate has not much progressed since then. Fears of worker redundancy were swept away as the Industrial Revolution made possible an increase in both population and wages, albeit at the cost of serious disruption.
The new twist in the argument today is that the current wave of automation differs from that of the past by automating mental work in addition to manual work. It’s not just the horses and manual workers who have to go on the scrap heap, but all the white collar workers, too. Not only does technology bite ever deeper into cognitive work, but it does so at an accelerating rate. This means that there will soon be almost no jobs that robots could not do as well as, or even better than, humans. Therefore redundancies due to automation will inevitably exceed those caused by mechanisation in the past; some new jobs may be created, but far fewer than jobs destroyed. A prominent recent pessimist, the Oxford economist Daniel Susskind, claims that there are “fewer and fewer jobs which can only be performed by human beings”.
So what is to be done? The only approach to a normative answer lies in the understanding of humanness.
In terms of Keynes’s story, do machines hasten or retard the entry to Paradise on earth? Keynes obviously thought they would hasten it. Machinery should make it possible to “satisfy the old Adam in most of us” with only three hours’ work a day. With abundance we would become like “the lilies of the field…they toil not, neither do they spin”.
Keynes was drawing here on biblical imagery which would have been familiar to his readers. Perfect humanity existed in the Garden of Eden. Then came the Fall: humanity became imperfect because it succumbed to the wiles of the Serpent, but there was the possibility of Redemption and a Return to perfection.
Keynes drew on the language (“software” in today’s jargon) of this story to tell a very different one. The promise of perfection lay entirely in the future: there was no actual past to regain.
Keynes’s story, that is, starts with post-Edenic life, in which the ground is covered with thorns and thistles, women are condemned to bear children in pain, and men cursed to eat bread in the sweat of their brows. Paradise lies ahead, not behind us. This was the Enlightenment vision.
Interestingly, the anthropologists James Suzman and Marshall Sahlins have idealised the hunter-gatherer way of life, a time when humanity was free (literally) to sow its wild oats. God’s curse was the invention of the plough.
In Keynes’s telling, the machine points humans towards God, but it is also devilish. The devil in Keynes’s story is capitalism. Capitalism was the spirit of gain, of love of money, which came to dominate Europe’s economic life from the 17th century onwards. This is the argument of Max Weber in The Protestant Ethic and the Spirit of Capitalism.
Unlike Weber, for whom the capitalist was a heroic entrepreneur, Keynes portrays the spirit of capitalism as a neurosis — one which future times would hand over to specialists in mental disease; but which, by enabling the accumulation of capital — or machinery — would lift up humanity from scarcity to plenty.
With the coming of abundance, he writes, humanity would
“free… to return to some of the most sure and certain principles of religion and traditional virtue — that avarice is a vice, that the exaction of usury is a misdemeanour, that love of money is detestable, that those walk most truly in the oaths of virtue and sane wisdom [are] those who take least head of the morrow”.
So the Devil, who stands for the love of money and power, nevertheless delivers humanity back to its Creator. By this trick Keynes disposes of the obstacle original sin poses to his sunny future — a solution which has never been entirely convincing.
Keynes’s progrnostication has turned out only partly right. Since 1930 technological progress has lifted average real income per head in rich countries roughly five times, much in line with Keynes’s expectations, but average weekly hours of full time work in these countries have fallen only by some 20 per cent, from about 50 hours to 40. What did Keynes get wrong?
Partly it turned out that the Devil was lurking in more places than Keynes allows. Paradise will be gained, Keynes assures us, “assuming no important wars and no important increases in population…” With these words, Keynes briskly dismissed the most obvious impediments to the realisation of his utopia.
Any theologically alert listener or reader will spot the big omission in Keynes’s story. Impulses to violence and the “sins of the flesh” — leading to war and overpopulation — come near the top of the Christian list of human sins. But to Keynes, writing in the Enlightenment spirit, these are simply contingent liabilities, to be redeemed by improved knowledge, better education, more intelligent political management, and suchlike.
To put it in theological terms, Keynes saw human nature as Pelagian, not Augustinian. As he grew older, his belief in natural goodness frayed. By the 1930s he believed that civilisation was “a thin and precarious crust” and that under this veneer lay “insane and irrational springs of wickedness”. He told Virginia Woolf that he would be inclined “not to demolish Xty [Christianity] if it were proved that without it morality was impossible; the young, he added, “were trivial, like dogs in their lusts”. Where these thoughts would have led him had he lived longer, one can’t say.
So we return to the original question: assuming that technological progress is secured, will this be a benefit or a curse?
On the benefit side, consider this from a recent report of the International Labour Organisation:
“Policy debates are placing renewed emphasis on lifelong learning. This is based on the understanding that lifelong learning…increases workers’ and firms’ capability to adapt to changes in the world of work. Lifelong learning can therefore foster productivity and innovation…and help workers transition to quality employment…”
But there is another story to tell. John Thornhill rightly warns that, rather than augmenting human creativity, machines may “amplify human stupidity”.
Imagine a worker receiving instructions on how to make a pair of shoes through augmented reality goggles. All the steps in the operation are precisely choreographed. Anyone of normal dexterity will be able to manufacture an acceptable pair of shoes from such instructions. Can we really say that humans have been upskilled or enhanced? It would be odd to say that they know how to make a pair of shoes. They are simply following instructions. It is otiose to call this enhancing creativity.
On the other hand, lots of genuine skills have undoubtedly been lost. Invisible mending is no longer needed because we throw away clothes which have holes. Writing by hand first yielded to the typewriter, now typists have yielded to computers, and the act of writing to predictive text software. Doctors have become medical technicians, no longer required to ‘‘know their patients’’, simply to be able to read computer print-outs of their vital functions; London taxi drivers no longer have to ‘‘do the knowledge’’, creating a mental map of the city’s streets and buildings, but just use satellite navigation.
However, it is also true that working with new technologies compels workers to learn new skills — skills which might be based on theoretical knowledge, rather than just on looking around and finding what works.
Whether automation has a tendency to enhance or reduce humanness thus depends on what we mean by human. If to be human is to be robotic, then machines enhance humanness. But if, as most of us believe, it is to be more than robotic, every encroachment of robots on our lives will reduce our humanness.
But this isn’t the end of the discussion. Like Marx, Keynes believed that the reduction of necessity would automatically lead to an increase in freedom: indeed his economics of full employment was designed to get us over the hump of necessity and into the realm of freedom as quickly as possible. He was curiously blind to the possibility that the machines which freed us from work might restrict our freedom in the non-working parts of our lives. In retrospect the entanglement of actual machines with ideas about how to organise social and political life seems inevitable once the “science of society” took hold in the Enlightenment. In his classic The Road to Serfdom, Friedrich Hayek warned against “the uncritical transfer to the problems of society of the habits of thought of the natural scientist and the engineer”. But it was precisely the engineering ambition of making society as efficient as the factory or the office that built the modern world and turned Keynes’s realm of freedom into Weber’s “iron cage of bondage”.
The ominous possibilities of information technology as an instrument of social control were dramatically visualised in Jeremy Bentham’s famous design for a Panopticon in 1786. This was an ideal prison system, in which the prison governor could shine a light on the surrounding prison cells from a central watchtower, while himself remaining unseen. This would in principle abolish the need for actual prison guards, since the prisoners, aware of being continually surveilled, would voluntarily obey the prison rules. Bentham’s ambitions for his invention stretched beyond the prison walls, to schools, hospitals and workplaces. His was a vision of society as an ideal prison, governed by self-policed impersonal rules, applicable to all. His key methodology was a one-way information system: the governor would know all about the prisoners but would himself be invisible.
Bentham’s world is coming to pass. Today’s digital control systems operate, not through watchtowers, but through computers with electronic tracking devices, and voice and facial recognition systems. We enter Bentham’s prison voluntarily, oblivious to its snares. But once inside, it is increasingly difficult to escape. Commercial platforms and governments can hope to control our habits, thoughts, and tastes by “mining” the information about ourselves with which we provide them by using electronic devices for our convenience. The realm of privacy recedes as the technical possibility of surveillance expands.
Experts debate which is the greater threat — state surveillance or capitalist surveillance — but this largely (though not entirely) a sham battle. Big Business and the State both use the same technology and, more often than not, cooperate rather than compete.
Keynes was, of course, aware of the malign uses to which surveillance technology was being put in his time. But he seems to have been thrown off guard by his belief that democracies provided sufficient safeguards against an Orwellian outcome. He was insensitive to the possibility that surveillance might creep up, unobserved, and even unintended, until it was too late to reverse.
3. The Accountability Issue
Increasing attention is being paid in the media as to how we might control runaway or rogue intelligence.
The game is already lost when we refer to Machine or Artificial Intelligence. Machines are information processors, not knowledge producers. They can think, but only, so far, within the confines of games prescribed by their makers, or the instructions — which may be quite general — of their designers. The tipping point or singularity will come when they are able to think for themselves. To call machines intelligent is to reduce the difference between machine and humans to one of degree rather than kind. No defence of the principle of human uniqueness is possible on such lines.
The philosophical defence of human singularity — that humans have something called consciousness, or subjective first person awareness, and that this causes intentional behaviour — is no longer explained by the hypothesis of a metaphysical entity like the soul. The dominant science of the matter is that consciousness is rooted in the brain. The brain is essentially an imperfect machine. It then becomes only a matter of time for science to find a way of building machines with human level intelligence. The date for the “singularity” — the moment of machine self-awareness, keeps being postponed because of “technical” difficulties; but that it will come sooner or later cannot be in doubt. And beyond that stretches the even more fanciful super-intelligence. “The train might not [even] pause or decelerate at Humanville station. It is likely to woosh right by,” writes the Oxford philosopher Nick Bostrom.
I end by considering a pseudo-religious movement called transhumanism which, while trying to grapple with the problem of human singularity and how we might try to preserve it, ends up in madness.
At the heart of transhumanism is the belief that the progress of artificial Intelligence intelligence cannot be stopped, that the advance of machines to super-intelligence is bound to accelerate, and that therefore the most urgent task of the wise legislator (let us call him the philosopher king) is to ensure that the Artificial God works for the benefit of humanity and not against it.
The transhumanists are our contemporary Frankensteins. They are made up of academics at Oxford, Cambridge and MIT, variously funded by billionaire techno-utopians like Elon Musk, Peter Thiel, Mark Zuckerberg. Their core doctrine of “effective altruism” has evolved into a vision of immortality. The philosopher Emile Torres sees transhumanism as ‘‘quite possibly the most dangerous secular belief system in the world today’’.
Its voyage to Bedlam starts from a position of unqualified utilitarianism. The rightness of an action is to be judged solely by its consequences. The end justifies the means; no means are ruled out of court ab initio. The next step follows from the logic of counting heads. It is quantity of utility which matters, not quality. This means treating everyone’s utility the same, including that of those yet unborn. Thus the goal is not to maximize the utility of the present generation, but of all future generations, of which this generation will form only a tiny fraction. Ethically speaking, the utility of our generation should make only a tiny claim on our moral concern. As Toby Ord puts it: ‘‘because, in expectation, almost all of humanity’s life lies in the future, almost everything of value lies in the future as well’’. Effective altruism prioritises the interests of the yet unborn over those of the present generation.
The next step in the argument identifies the goal of maximising the utility of the universe with that of maximising its intelligence potential, that is, its capacity for creating value. Humans are unique among animals in their cognitive ability. Their cognitive potential has advanced through the operation of the Darwinian ‘‘survival of the fittest’’. With billions of survivors now inhabiting the planet, humanity’s intelligence has grown to the point when it can advance without limit.
With the development of AI, humans have, for the first time, taken charge of the evolutionary process. The claim is that AIs are starting to be built which can equal the best of human intelligence, and that super-intelligent ones will follow sooner rather than later. Bostrom defines super-intelligence as ‘‘any intellect that greatly exceeds the cognitive performance of humans in virtually all domains of interest’’. Since the design of machines is one of these cognitive performances, modestly super-intelligent machines could design even better machines; there would then follow ‘‘an intelligence explosion’’. A population of ultra-intelligent AIs would take over the business of evolution from humans. Thus ‘‘the first ultra-intelligent machine is the last invention that man need ever make’’.
The logic grinds on remorselessly. The body depends on the finite resources of our planet. But super-brains would be able to detach themselves from the limitations of the body. They might then escape from the limitations of our world, and establish colonies in our planetary ‘‘light cone’’, to be ‘‘fed’’ from its still unexhausted ‘‘endowment of negentropy’’ (or reverse entropy) in our cosmos.
Humanity’s intelligence potential could then be preserved and expanded for millions of years until the sun finally cooled. Actual humans are nothing but means to this end, and therefore valuable only insofar as they contribute to the overall net amount of value in the Universe between the Big Bang and the Heat Death. This is the philosophic/moral basis of the billionaire-financed projects of escape to the moon and other planets.
At this point, eschatological urgency seizes control of the transhumanist argument. The transhumanists share the view of the Doomsday scientists that AIs programmed with human intelligence only might quite possibly produce a nuclear or environmental catastrophe far worse than human intelligence on its own could achieve. Thus the coming of super-intelligence offers the possibility of either immortality or total disaster.
We aim to create a benevolent God, but it is always possible that, like Frankenstein, he or she may turn out to be a Deus Malignus, who might only pretend to have good intentions, but, once unchained, would set about destroying not just us, but its AI rivals. So our super-intelligent AIs must be programmed with moral rules before they take control of the future. But the only moral rules available come from our own imperfect and conflicting moral values. Wriggle as they might, transhumanists cannot escape the dilemma that there is no possibility, in a world of value relativism, of binding super-intelligence to an agreed morality. So the benevolence of our future controllers cannot be guaranteed.
While recognising the risk to humanity of super-intelligent AIs run berserk, transhumans are too entranced by their dream of a cosmic computronium to propose shutting AI down before it reaches super-intelligence. Thus Ord writes: ‘‘a permanent freeze on technology…would probably itself be an existential catastrophe, preventing humanity from ever fulfilling its potential’’. The most they offer is a ‘‘pause for reflection’’ before allowing any further advance in AI. Such a pause, they hope, might give time for reaching global agreement on the moral rules our super-intelligent AIs need to have.
I think it is right to end my remarks on this point of madness. Remember: billions of dollars are being poured into this line of thinking and research. The only feasible response of sanity as I see it is a religious one. Humans must stop playing at God and start asking what it is that God wants of them.
To regain sanity, those concerned with the big issues of human survival must find some way of aligning their language with secularists engaged on the same quest, or better still finding a way of integrating them. The worst thing is for religious people to keep their religion “private” and engage with the materialists on the latter’s terms. Theology must start advancing again after centuries of retreat.
This is an edited version of a lecture by Lord Skidelsky, given at Brompton Oratory on 13 June 2024. His book The Machine Age: An Idea, A History, A Warning was published by Allen Lane in November 2023.