OpenAI’s GPT-4.5 model was more convincing than an actual human, researchers say
www.independent.co.uk
A leading AI chatbot has passed a Turing Test more convincingly than a human, according to a new study.
Participants in a blind test judged OpenAI’s GPT-4.5 model, which powers the latest version of ChatGPT, to be a human 'significantly more often than actual humans'.
The Turing Test, first proposed by the British computer scientist Alan Turing in 1950,
is meant to be a barometer of whether artificial intelligence can match human intelligence.
The test involves a text-based conversation with a human interrogator, who has to assess whether the interaction is with another human or a machine.
Nearly 300 participants took part in the latest study, which ran tests for various chatbots and large language models (LLMs).
OpenAI’s GPT-4.5 was judged to be a human 73 per cent of the time when instructed to adopt a persona.
'We think this is pretty strong evidence that [AI chatbots] do [pass the Turing Test],' Dr Cameron Jones,
a postdoc researcher from UC San Diego who led the study, wrote in a post to X. 'And 4.5 was even judged to be human significantly more often than actual humans.'
It is not the first time that an AI programme has beaten the Turing Test, though the researchers from UC San Diego who conducted the study claim this to be the most comprehensive proof that the benchmark has been passed.