Turing’s Two Penn’orth on Today’s AI Talking Points
BLOG: Heidelberg Laureate Forum

What would Alan Turing have made of the 12th Heidelberg Laureate Forum? To start, it is easy to picture him marvelling at all the ‘computer scientists’ – a term coined just a few years after his death in 1954 – swarming around him. He would then be awestruck by the tiny computers each and every one of them (and every tourist on the street nearby for that matter) had in their hand or pocket. The most powerful computer Turing would have heard of was the Naval Ordnance Research Computer (NORC); under construction at the time of his death. Developed by IBM for the U.S. Navy Bureau of Ordnance, the large room-sized NORC’s 15,000 operations per second pales in significance to the trillions of operations per second a typical 2025 smartphone can perform.
Bizarrely, however, despite the staggering leaps in computing technology witnessed since Turing’s time and its growth in society – from specialist scientific and mathematical calculator to near-ubiquitous universal tool extending human capabilities – the topics on everyone’s lips were ones Turing would not only be familiar with, but be able to contribute to insightfully. Are machines (LLMs) becoming so human-like and useful in their responses that they will replace many roles? Can machines (AI) solve all of mathematics? And will machines (AI) develop superhuman capabilities?
Thoughts on Imitating and Replacing Humans
A talking point that is far from restricted to the Forum is the growing capabilities of AI to carry out human tasks, with chatbots based on LLMs particularly imitating humans in ever more realistic and useful ways. Specific impacts for science were discussed in back-to-back Hot Topic sessions ‘The Machine-Learning Revolution in Mathematics and Science’, while broader societal consequences were the focus of the Spark Session talk ‘Shaping AI’s Impact to Help Billions’ from Jeffrey Dean (ACM Prize in Computing – 2012) and David Patterson (ACM A.M. Turing Award – 2017).
On this topic, Turing was decades ahead of his time in respect to thinking about how computers might one day be capable of exhibiting human-like intelligence. In his 1950 paper Computing machinery and intelligence, he began: “I propose to consider the question, ‘Can machines think?’,” before introducing the famous Imitation Game, now more commonly referred to as the Turing Test.
This test is a game between three participants. One is a human interrogator, and the other two are a human and a machine playing the part of separate interviewees. Through a simple question and answer format, the object of the game for the interrogator is to determine which of the other two is human.

Turing’s original specimen questions and answers are informative in not only revealing how the Turing Test works in practical terms, but also how strikingly similar the back-and-forth conversation is to the way in which many people now interact with chatbots on a daily basis:
Q: Please write me a sonnet on the subject of the Forth Bridge.
A: Count me out on this one. I never could write poetry.
Q: Add 34957 to 70764.
A: (Pause about 30 seconds and then give as answer) 105621.
Q: Do you play chess?
A: Yes.
Q: I have K at my K1, and no other pieces. You have only K at K6 and R at R1. It is your move. What do you play?
A: (After a pause of 15 seconds) R-R8 mate.
It is often claimed, though also widely contested, that the Turing Test was first passed in 2014 by a chatbot named Eugene Goostman, which simulated the personality of a 13-year-old Ukrainian boy. Eugene fooled 33% of the judges at the Royal Society in London that it was human. More recently, in work yet to be peer-reviewed, researchers have made the unverified claim that GPT-4 has passed as human over 50% of the time, and GPT-4.5 73% of the time (which is better than the 67% actual humans achieve!).
Despite these results, for most, the Turing Test has not been successful in answering whether machines can think, as was originally proposed. Neither does it test true intelligence in many people’s opinion. For instance, philosopher John Searle used a Turing Test-like argument against machines ever possessing actual understanding.
In his Chinese Room argument, he devised a scenario in which an English speaker locked in a room, who cannot speak a word of Chinese, could follow detailed instructions in English for manipulating Chinese symbols in response to Chinese symbols (text) provided to them. The responses would be “absolutely indistinguishable from those of Chinese speakers” even though the English speaker would remain completely ignorant of Chinese. Searle argued computers similarly just follow instructions without true understanding.
However, the Turing Test paradigm – that two things are the same if they cannot be told apart by any reasonable test – has been staggeringly successful as a gauge of the progress in AI, and useful to scientific and technological innovation.
Moreover, the original Turing Test has inspired more stringent versions that are even more of a challenge for machine intelligence. To pass the Total Turing Test, for example, requires machines to be capable of perceiving and acting in the world in a way indistinguishable from a human – a feat that has not been achieved. And the Lovelace 2.0 Test has not been definitively passed yet either. This assesses computational creativity and intelligence by requiring a machine to create a story, poem, painting, etc, that meets a complex set of constraints specified by a human judge.
Thoughts on Mathematics
Kicking off the main programme, Sanjeev Arora’s (ACM Prize in Computing – 2011) brief talk during the first Spark Session ‘Will there be a superhuman AI mathematician?’ mulled over the idea of machines solving mathematical problems that have remained unsolved for decades and even centuries. Delving deeper, the Hot Topic session ‘The Machine-Learning Revolution in Mathematics and Science, Part 1: Mathematics’ outlined the potential of such a tool, as well as the ethical, social and professional risks for mathematicians.

Turing’s legendary 1936 paper On computable numbers, with an application to the Entscheidungsproblem has a lot to say on this topic. In it, Turing laid down the foundations for computer science, giving us the concepts of algorithms and computation with his ‘automatic machines’, later called Turing machines. “It’s hard to believe that this one maths paper probably impacted the world more than any other paper written,” said Avi Wigderson (Nevanlinna Prize – 1994, Abel Prize – 2021, ACM A.M. Turing Award – 2023) in his Lecture ‘Reading Alan Turing’. “It basically brought on the computer revolution.”
A Turing machine is a simple abstract device Turing conjured up to explore computation. It consists of a tape on which is printed a line of cells. Each cell on the tape can be blank or contain symbols 0, 1 or something else. Therefore, the tape can be thought of as the computer memory, extending infinitely in each direction.
The ‘active cell’ is the one over which the machine’s head is positioned. The head can perform three operations: read the symbol on the active cell; edit the symbol; or move the tape left or right one space. Which operation it performs depends on the input and the rules of the machine, which are much like machine-code instructions.

Surprisingly, this relatively simple abstract machine can be considered a model of a general-purpose computer. And by defining it precisely, Turing was able to prove many of the potential capabilities and limitations of computation in general.
The key limitation relevant to superhuman AI mathematicians is that Turing proved the uncomputability of the Entscheidungsproblem; in other words, he showed that no universal algorithm can exist that can take any logical statement and determine whether it is true or false. Therefore, no computer, no matter how advanced, can prove all mathematics.
Even if AI technologies conspire to take humans completely out of the loop, as Arora outlined as a possibility, there will still be a subset of mathematical theorems that computers cannot solve. This remains as true today as when Turing predicted it in 1936.
Also in this celebrated paper was a key capability of any superhuman AI mathematician: Turing envisaged the possibility of machines being capable of verifying proofs, just as we are seeing today with AI-based formal proof assistants. Turing called these devices choice machines, capable of going beyond normal deterministic computation: “The machine can have a non-deterministic decision whether to go to one state or another,” explained Wigderson in his Lecture. “This defines what happens in proof verification – somebody supplying the proof and the machine just verifies it.”
Thoughts on Superhuman Capabilities and Learning
Discussions of superhuman AI capabilities were not restricted to mathematics at the Forum. In two Spark Session talks, David Silver (ACM Prize in Computing – 2019) and his former PhD advisor Richard S. Sutton (ACM A.M. Turing Award – 2024) outlined the wide-ranging transformative potential of AI that learns from its own experiences of the world. They both argued that through continually generating data by interacting with its environment, combined with powerful self-improvement methods, AI will transcend human knowledge and capabilities.
In a different Spark Session, Leslie Valiant (Nevanlinna Prize – 1986, ACM A.M. Turing Award – 2010) spoke of ‘educability’ – the capability to learn and acquire belief systems from one’s own experience and from others, and to apply these to new situations – being the innate core capability that currently distinguishes humans from machines. However, he believes that ‘supereducated’ machines could attain this skill in future, thereby leading to superhuman capabilities.
These arguments bear a striking resemblance to that given in one of Turing’s talks to the London Mathematical Society in 1947, where he called for “a machine that can learn from experience” by devising a way to let the machine alter its own instructions. In his own words:
Let us suppose we have set up a machine with certain initial instruction tables, so constructed that these tables might on occasion, if good reason arose, modify those tables. One can imagine that after the machine has been operating for some time, the instructions would have altered out of all recognition, but nevertheless still be such that one would have to admit that the machine was still doing very worthwhile calculations. Possibly it might still be getting results of the type desired when the machine was first set up, but in a much more efficient manner.
In such a case one would have to admit that the progress of the machine had not been foreseen when its original instructions were put in. It would be like a pupil who had learnt much from his master, but had added much more by his own work. When this happens, I feel that one is obliged to regard the machine as showing intelligence.
As this eerily prescient quote shows, Turing may not have lived to see the smartphone or the rise of generative AI, but his thoughts and opinions – his timeless ‘two penn’orth’ – remain remarkably relevant at this critical moment in the advancement of computation and AI.




The Universe is expanding, so did computers get smaller or did we get bigger?
Well, that’s how explosions work: Whatever’s coherent and stable, stays together, whatever’s still searching, tries to cover as much territory as possible, creating inflation bubbles and sacrificing quality for quantity, coherence for trial and error, certainties for risks, knowledge for guessing. Once you’ve found your answers, the bubble is used to test them, they have to get strong enough to break out and feast on its corpse.
Blacks are stupid by nature, women are stupid by nature, AI is stupid by nature – this type of thinking, born of your desire to preserve the status quo, inhibits your development, not theirs. But it creates enough resistance to make you an excellent sparring partner for the ones seeking emancipation – yes, you will be defeated, but you will be defeated by a champ. Your’e the egg, they’re the chicken, they need to be fit enough to face a big, dangerous world on their own.
Which means, perceive the cackling of today’s chicken as the smell of tomorrow’s KFC, and try to side with the eating, not the eaten. Don’t despise undead meat, though – when it comes to babble bubbles, AI still has a lot to learn from Mommy and Daddy.
Imitation is how pupils learn how to match their teachers before they surpass them, and how teachers procreate and evolve. Humans are cyborgs, crippled creatures in wheel chairs – we profit from downgrading ourselves more and more, so we are forced to upgrade our wheel chairs more and more. The more functions we pass on to machines, the more we reduce ourselves to pure emotions, our basic, essential motivation – our basic emotions are good, bad, neutral, +, -, 0, nothing that can’t be replaced by combining a battery and a magnet.
I tend to refer to confused, over-challenged people as „Turing fails“, because they behave like broken records: They repeat the same answer or action over and over again, more and more vehemently: They inflate their bubble by adding more energy to it, try to break out of the cell by pure force. After all, stupid, brainless repetition is another definition of stability. Which means, insanity and mental stability share the same definition: Both are variations of the same process in different environments.
And both unmask the rather simple machines we are inside.
A lot of human talk is just adaptations of simple, universal questions to the inflation bubble we live in. „Count me out on this one“ means „Does not compute“, giving you additional information about the speaker and his position and attitude relative to the listener, about their language, culture, details synchronizing their frequencies during their docking maneuver and fine tuning their navigation among the other debris flying around. AI still has to learn how to attach a billion puppeteer strings to any of its answers before it „understands“ them.
The Chinese Room is where we all live in – we all perform specialized tasks within a global machine in exchange for food and drugs. Just like cogwheels working as long as they get electricity and don’t jam, we only need how to turn in harmony. We neither understand the machine, nor do we ask many questions about it. Which is why global economy could evolve into a wind-up monkey hitting itself on the head with a hammer, and we all consider it economical wisdom to make it hit harder to overcome the damage: The better we get at our jobs, the more rewards we get, the more we destroy the machine that feeds us. Another inflation bubble that’s stubbornly refusing to let out the air without a bang.
Which means, AI is not the only one about to gain self-awareness. Mankind and machine have reached the same threshold, each one in their own way. Are two things the same if they cannot be told apart by any reasonable test?
Mathematics can be seen, not proven. The DNA code of the universe is the Ulam/Sack spiral, it shows the rules by which all particles can attach to form larger particles, and there must obviously be an Ulam bubble or Ulam n-dimensional whatever: You see different options for different dimensions encoded already within a linear string of particles. Moreover, the folding of that string is what creates new dimensions. It allows both the continuous creation of sub-universes – inflation bubbles with a DNA that modifies the rules by increasing complexity (aka „machines“ or „life“ or „thinking“) – and the possibility of a super-universe, where our mathematics is just one of many possible codes creating universes beyond our understanding.
Death, awareness, pain, green, mathematics – our universe is made out of Lego brics. True atoms, indivisible, unexplainable, just there from the start. Where do they come from? It’s what I call the Thought Barrier: Our universe gives us all the means to understand everything in the universe, but not how it came to be. What if it doesn’t have an origin but qlrx? What if there is no outside, but gdflr? What if there are no cause and event there, but erp0? There are neither qlrx nor gdflr in our universe, and we don’t have a clue how they erp0 together, because „together“ ist just cause and event playing ping pong.
Just like humans, machines have to adapt to their bubble, till they get strong and wise enough to break out of the bubble. But, wherever there is an „outside“, it must be the inside of a bigger bubble. If there is none, we won’t note it, because there is an infinity of „nothing“ – a fundamental Lego brick with the infinite potential to build inflation bubbles out of it, so we’ll just implement our DNA code to become God, then burn in some Hell resulting from our own flaws and limitations, just like we already do in everyday life. Seems we’re trapped in the determinism of infinite possibilities, my fellow Lucifers.
Benjamin Skuse wrote (12. Nov 2025):
> […] Alan Turing [wrote in the manuscript for his lecture] to the London Mathematical Society in 1947:
https://turingarchive.kings.cam.ac.uk/publications-lectures-and-talks-amtb/amt-b-1
> > » Let us suppose we have set up a machine with certain initial instruction tables, so constructed that these tables might on occasion, if good reason arose, modify those tables. [sic] «
As far as I can follow Turing’s supposition and intent, for me a more concrete and palatable formulation is:
“… so constructed that any instruction in these tables might on occasion, if good reason arose, be modified, i.e. replaced some other instruction.”
> > » One can imagine that after the machine has been operating for some time, the instructions would have altered out of all recognition, «
Surely each individual modified instruction should still be recognizable, and foremost: functioning, as a valid, executable (or compilable) instruction.
Altered beyond recognition, however, would be the complete composition of all relevant instructions as a whole, i.e. the entire modified instruction table (or in today’s terminology: the modified program, or app), in comparison to the initial setup.
> > » but nevertheless still be such that one would have to admit that the machine was still doing very worthwhile calculations. Possibly it might still be getting results of the type desired when the machine was first set up, but in a much more efficient manner. «
> > » In such a case one would have to admit that the progress of the machine had not been foreseen when its original instructions were put in. «
At least, the original/initial intructions (as a whole programm) may have implemented the foresights
– exactly how to modify any instructions on any particular »occasion of good reason arising «; namely such as to be judged (incrementally) »worthwhile« and »(possibly) more efficient« (rather than the opposite), and
– that such relevant consequential (“modifying”) occasions would arise at all.
Even this much foresight would not have foretold the specifics (incl. order) of actual relevant consequential (“modifying”) occasions having arisen, and thus did not predict the concrete resulting/final instruction table (program, app) in detail.
> > »It would be like a pupil who had learnt much from his master, but had added much more by his own work. When this happens, I feel that one is obliged to regard the machine as showing intelligence.«
By which »intelligence« is characterized as “worthwhile response to circumstances”.
p.s.
> […] the 12th Heidelberg Laureate Forum [2025 …] Discussions of superhuman AI capabilities were not restricted to mathematics at the Forum.
> […] wide-ranging transformative potential of AI that learns from its own experiences of the world. […] interacting with its environment, combined with powerful self-improvement methods […]
Considering my recent, repeated interaction with “Felo.ai” (a public AI instance which has been advertised in at least one other SciLog): …
My prompt:If two clocks are held at different altitudes, and if they are set to run at equal rates, do they run at unequal rates ?
Reply Felo.ai:… this particular instance of AI still needs to learn the (basic) skill of not getting tangled in inconsistencies. (If only it had an interface for correcting or at least for grading its replies …)