Back in 1950, Alan Turing proposed an elegantly simple yet profoundly challenging way to determine whether machines could be said to "think." Known as the Turing Test, this measure of machine intelligence sets humans and machines in conversational competition, challenging human judges to distinguish between artificial and genuine intelligence through text-based interactions.
Despite numerous attempts, no artificial system had ever convincingly passed this test. Until now.
Cameron Jones and Benjamin Bergen from the University of California, San Diego, have gathered for the first time empirical evidence that OpenAI’s GPT-4.5, a sophisticated large language model (LLM), has successfully passed the Turing Test. Not only did GPT-4.5 pass, but under specific conditions, it outperformed human counterparts in convincing participants of its humanity.
The work has significant implications for society, ethics and humanity’s understanding of intelligence itself.
The Turing Test pits an interrogator against two conversational partners — one human and one machine ...