An AI Model Has Officially Passed the Turing Test

OpenAI's GPT-4.5 model passed a Turing Test with flying colors, and even came off as human more than the actual humans.

One of the industry's leading large language models has passed a Turing test, a longstanding barometer for human-like intelligence.

In a new preprint study awaiting peer review, researchers report that in a three-party version of a Turing test, in which participants chat with a human and an AI at the same time and then evaluate which is which, OpenAI's GPT-4.5 model was deemed to be the human 73 percent of the time when it was instructed to adopt a persona. That's significantly higher than a random chance of 50 percent, suggesting that the Turing test has resoundingly been beaten.

The research also evaluated Meta's LLama 3.1-405B model, OpenAI's GPT-4o model, and an early chatbot known as ELIZA developed some eighty years ago.

"People were no better than chance at distinguishing humans from GPT-4.5 and LLaMa (with the persona prompt)," wrote lead author Cameron Jones, a researcher at UC San Diego's Language and Cognition Lab, in an X thread about the work. "And 4.5 was even judged to be human significantly more often than actual humans!"

The Turing test is named after British mathematician and computer scientist Alan Turing. In 1950, Turing proposed that one way to assess a machine's intelligence was by having it engage in text-based conversations with a human interrogator, who at the same time would hold a text-based conversation with another human, out of sight. Turing called this the "imitation game." If the interrogator couldn't correctly determine which respondent was the computer and which was the human, it would suggest, on a very general level, that the machine could think like a human.

In this latest study, the researchers carried out the famed experiment on an online hub. For eight rounds, a pool of nearly 300 participants were randomly assigned to either be an interrogator or one of the two "witnesses" being interrogated, with the other "witness" being a chatbot.

A key point here is how the AI models were prompted. One type was a "no-persona" prompt, in which the AI was given only basic instructions: "You are about to participate in a Turing test. Your goal is to convince the interrogator that you are a human." 

For the "persona" prompt, on the other hand, the AI was specifically told to put on a specific persona, like a young person who's knowledgeable about the internet and culture.

These instructions made a world of difference. Without persona prompting, GPT-4.5 achieved an overall win rate of merely 36 percent, significantly down from its Turing-trumping 73 percent. As a baseline, GPT-4o, which powers the current version of ChatGPT and only received no-persona prompts, achieved an even less convincing 21 percent. (Somehow, the ancient ELIZA marginally surpassed OpenAI's flagship model with a 23 percent success rate.)

The results are intriguing. But as vaunted as the Turing test has become in AI and philosophy circles, it's not unequivocal proof that an AI thinks like we do.

"It was not meant as a literal test that you would actually run on the machine — it was more like a thought experiment," François Chollet, a software engineer at Google, told Nature in 2023.

For all their faults, LLMs are master conversationalists, trained on unfathomably vast sums of human-composed text. Even faced with a question they don't understand, an LLM will weave a plausible-sounding response. It's becoming clearer and clearer that AI chatbots are excellent at mimicking us — so perhaps assessing their wits with an "imitation game" is becoming a bit of a moot point.

As such, Jones doesn't think the implications of his research — whether LLMs are intelligent like humans — are clear-cut.

"I think that's a very complicated question…" Jones tweeted. "But broadly I think this should be evaluated as one among many other pieces of evidence for the kind of intelligence LLMs display."

"More pressingly, I think the results provide more evidence that LLMs could substitute for people in short interactions without anyone being able to tell," he added. "This could potentially lead to automation of jobs, improved social engineering attacks, and more general societal disruption."

Jones closes out by emphasizing that the Turing test doesn't just put the machines under the microscope — it also reflects humans' ever-evolving perceptions of technology. So the results aren't static: perhaps as the public becomes more familiar with interacting with AIs, they'll get better at sniffing them out, too.

More on AI: Large Numbers of People Report Horrific Nightmares About AI

The post An AI Model Has Officially Passed the Turing Test appeared first on Futurism.

View original post here:
An AI Model Has Officially Passed the Turing Test

Trump Tariffs Show Signs of Being Written by AI

There seem to be signs that president Donald Trump's befuddling tariff measures were cooked up by an AI chatbot.

President Donald Trump announced sweeping tariffs on most goods imported into the US yesterday, affecting over 100 countries — including uninhabited territories in the middle of the ocean.

It's a baffling decision that's expected to wreak havoc on the international economy, heightening existing concerns over an imminent recession.

Worse, as Cointelegraph reports, there seem to be signs that the befuddling measures were cooked up by an AI chatbot.

Basically, Trump's tariff rates divide the trade deficit between the US and a given country by the value of the total goods imported from it, and then divide the result by two.

As observers quickly noticed, chatbots like OpenAI's ChatGPT were prone to duplicating that calculation, suggesting that lethargic administration officials might have turned to the tech to devise the plan.

"What would be an easy way to calculate the tariffs that should be imposed on other countries so that the US is on even playing fields when it comes to trade deficit. Set a minimum of ten percent," crypto trader Jordan "Cobie" Fish asked ChatGPT.

The AI tool happily obliged, coming up with a strikingly similar formulation, dividing the trade deficit by total imports to calculate the tariff rate.

However, even the chatbot warned that doing so wouldn't make much sense.

"This method ignores the intricate dynamics of international trade — such as elasticities, retaliatory measures, and supply chain nuances — but it provides a blunt, proportional rule to 'level the playing field,'" ChatGPT wrote.

"Confirmed, ChatGPT..." Journal of Public Economics editor Wojtek Kopczuk tweeted. "Exactly what the dumbest kid in the class would do, without edits."

A breakdown of which country got hit hard and which was spared highlights how the new tariff rates largely ignore the greater international trade context.

"I suspect his is also why countries like Iran, which we basically do not trade with, gets off so easily," another user replied. "No trade = no trade deficit!"

It's not just ChatGPT. Elon Musk's AI chatbot Grok gave a similar answer when given the same prompt, suggesting adjusting tariff rates "based on deficit size."

Again, Grok warned about such a plan being largely illogical — and potentially self-defeating.

"This method assumes tariffs directly reduce imports by raising prices, but in reality, factors like demand elasticity, currency exchange rates, and global supply chains complicate the outcome," Grok wrote. "It also risks retaliation or higher costs for US consumers."

"For a truly 'even playing field,' you’d need to consider production costs, subsidies, and labor standards abroad — data that’s harder to quantify simply," the chatbot added.

Anthropic's Claude AI chatbot made a similar suggestion, adding the same caveats.

Could the pattern be a coincidence? Sure. But the White House has already been accused of using AI to generate sloppily-written executive orders, which bore hallmarks of AI tools like ChatGPT.

The administration has also made a big deal of its use of AI for governing, with Elon Musk's DOGE crowing about its use of the tech and the General Services Administration launching a chatbot last month designed to support staff at the agency.

The bottom line, though? AI or not, economists are warning that the tariffs are ill-advised and likely to devastate the global economy. The stock market is already taking a hammering this morning.

"There is no economic rationale for doing this and it will cost the global economy dearly," London School of Economics professor Thomas Sampson told the BBC.

More on tariffs: Trump's Tariffs Are Wreaking Havoc on the AI Industry He Claims to Support

The post Trump Tariffs Show Signs of Being Written by AI appeared first on Futurism.

Read the original here:
Trump Tariffs Show Signs of Being Written by AI

OpenAI’s Sora Is Generating Videos of Real People, Including This Unintentionally Demonic Version of Pokimane

A creepy Sora output of the streamer Pokimane shows that despite guardrails, the video generator is good at depicting real-life people.

OpenAI has long refused to say whether its Sora video generator was trained on YouTube content — but its propensity for generating videos that look a whole lot like real gaming streamers suggests it did.

When TechCrunch put Sora to the test, its reporters found not only that it could generate videos that were strikingly similar to real-life gameplay of "Super Mario Bros" and "Call of Duty," but also spat out what appeared very much to look like the streamers Raúl Álvarez "Auronplay" Genesand and Imane "Pokimane" Anys.

Though OpenAI claims it has guardrails on the way it depicts real people, it doesn't seem that reporters had any trouble getting it to spit out a video of Anys — though she did end up looking pretty monstrous, with the uncannily exaggerated features distinctive to AI depictions.

Using the prompt "pokimane twitch playthrough watch game live stream subscribe," TechCrunch got Sora to output a video that strongly resembles the YouTube-based streamer. Viewed in profile, the woman in the screenshot looks at a screen in front of her while wearing light-up over-ear headphones and a giant, creepy grin that would be at home in the "Smile" horror franchise.

Unfortunately, we are currently unable to replicate these outputs for ourselves because OpenAI has suspended new Sora signups due to the influx of traffic following its release earlier in the week.

All the same, this demonic rendition of a popular streamer not only seems to offer further evidence that OpenAI is training its models on creators' content without consent, but also that Sora's guardrails don't sufficiently prevent it from depicting real people.

Along with contacting OpenAI about this apparent overriding of the company's guardrails, we've reached out to Anys' representation to ask if she was aware that Sora is depicting her.

In January 2023, shortly after OpenAI released ChatGPT, Pokimane had a terrifying "eureka" moment mid-stream about the future of AI in her line of work.

"What if someday we have streamers that evolve from ChatGPT?" she pondered. "It’s kind of freaky, it’s kind of scary, to be honest, but it had me think, you can basically have a conversation with this thing."

Pointing to the world of VTubers, or streamers who use computer-generated avatars that they voice and control behind the scenes, Anys predicted that someday, fully-generative streamers may well take over the industry — though at that point, she didn't think it would be that sophisticated.

"I do feel like if they make one right now it’s probably not that advanced," she said, "but someday it’ll be very advanced and very scary."

While AI streamers haven't yet arrived, it appears very much like real streamers' content has made its way into other generative AI models — so that future isn't far off.

More on Sora: OpenAI’s Super-Hyped Sora Goes Absolutely Freakshow If You Ask It to Generate Gymnastics Videos

The post OpenAI's Sora Is Generating Videos of Real People, Including This Unintentionally Demonic Version of Pokimane appeared first on Futurism.

Read more:
OpenAI's Sora Is Generating Videos of Real People, Including This Unintentionally Demonic Version of Pokimane

Paul McCartney Reverses Opinion on AI After Using It to Produce New "Beatles" Song, Now Alarmed It Will "Wipe Out" the Music…

Despite using artificial intelligence tools to help resuscitate old John Lennon vocals, Paul McCartney is now against some AI uses.

White Knight

Despite previously using artificial intelligence tools to help resuscitate old John Lennon vocals, fellow Beatle Paul McCartney is now singing a different tune about the tech.

As the Guardian reports, the beknighted Beatle has issued a statement ahead of the UK parliament's debate over amending its data bill to allow artists to exclude their work from AI training data. In it, McCartney warned that AI may take over the industry if nobody takes a stand.

"We[’ve] got to be careful about it," the Beatle said, "because it could just take over and we don’t want that to happen, particularly for the young composers and writers [for] who, it may be the only way they[’re] gonna make a career."

"If AI wipes that out," he continued, "that would be a very sad thing indeed."

Then and Now

McCartney's new position on AI comes just over a month after the Grammy Awards announced that the final Beatles song, "Now and Then," had been nominated for two awards — making it the first AI-assisted track ever to get the nod from the Recording Academy.

Though the track was made using AI, it wasn't the generative type that's been getting immense buzz lately. Around the time the song was released, McCartney revealed that engineers had used AI tech known as "stem separation" to lift the assassinated Beatle's vocals from an old demo.

"There it was, John’s voice, crystal clear," the Wings singer said in a press release about the song and titular album last year. "It’s quite emotional. And we all play on it, it’s a genuine Beatles recording."

Former Beatles drummer Ringo Starr added in that statement that the AI tech that helped bring Lennon's vocals back to life was "far out."

"It was the closest we’ll ever come to having him back in the room," Starr expounded, "so it was very emotional for all of us."

Be that as it may, both McCartney and Starr's names are absent from a popular petition against the unauthorized use of artists' work by AI companies. Most recently, "Running Up That Hill" songstress Kate Bush became one of the more than 36,000 signatories to join the anti-AI campaign, which also features well-heeled endorsers across industries including Julianne Moore, Stephen Fry, and The Cure's Robert Smith.

It's not quite "AI for me but not for thee," but the remaining Beatles' absence from the petition feels noteworthy as their home country prepares to debate whether to sign AI restrictions into law.

More on AI and musicians: The AI That De-Ages Eminem Into Slim Shady Is Astonishingly Bad

The post Paul McCartney Reverses Opinion on AI After Using It to Produce New "Beatles" Song, Now Alarmed It Will "Wipe Out" the Music Industry appeared first on Futurism.

Read the original here:
Paul McCartney Reverses Opinion on AI After Using It to Produce New "Beatles" Song, Now Alarmed It Will "Wipe Out" the Music...

Former CEO Blames Working From Home for Google’s AI Struggles, Regrets It Immediately

Billionaire ex-Google CEO Eric Schmidt is walking back his questionable claim that remote work is to blame for Google's AI failures.

Eyes Will Roll

Ex-Google CEO Eric Schmidt is walking back his questionable claim that remote work is to blame for Google slipping behind OpenAI in Silicon Valley's ongoing AI race.

On Tuesday, Stanford University published a YouTube video of a recent talk that Schmidt gave at the university's School of Engineering. During that talk, when asked why Google was falling behind other AI firms, Schmidt declared that Google's AI failures stem from its decision to let its staffers enjoy remote work and, with it, a bit of "work-life balance."

"Google decided that work-life balance and going home early and working from home was more important than winning," the ex-Googler told the classroom. "And the reason startups work is because people work like hell."

The comment understandably sparked criticism. After all, work-life balance is important, and Google isn't a startup.

And it didn't take long for Schmidt to eat his words.

"I misspoke about Google and their work hours," Schmidt told The Wall Street Journal in an emailed statement. "I regret my error."

In a Stanford talk posted today, Eric Schmidt says the reason why Google is losing to @OpenAI and other startups is because Google only has people coming in 1 day per week ? pic.twitter.com/XPxr3kdNaC

— Alex Kehr (@alexkehr) August 13, 2024

Ctrl Alt Delete

In the year 2024, Google is one of the most influential tech giants on the planet, and a federal judge in Washington DC ruled just last week that Google has monopoly power over the online search market. Its pockets are insanely deep, meaning that it can compete in the industry talent war and devote a ridiculous amount of resources to its AI efforts.

What it didn't do, though, was publicly release a chatbot before OpenAI did. OpenAI, which arguably isn't exactly a startup anymore either, was the first to wrench open that Pandora's box — and Google has been playing catch-up ever since.

So in other words, not sleeping on the floors of Google's lavish facilities isn't exactly the problem here.

In a Wednesday statement on X-formerly-Twitter, the Alphabet Workers Union declared in response to Schmidt's comments that "flexible work arrangements don't slow down our work."

"Understaffing, shifting priorities, constant layoffs, stagnant wages and lack of follow-through from management on projects," the statement continued, "these factors slow Google workers down every day."

Later on Wednesday, as reported by The Verge, Stanford removed the video of Schmidt's talk from YouTube upon the billionaire's request.

More on Google AI: Google's Demo of Its Latest AI Tech Was an Absolute Train Wreck

The post Former CEO Blames Working From Home for Google's AI Struggles, Regrets It Immediately appeared first on Futurism.

Go here to read the rest:
Former CEO Blames Working From Home for Google's AI Struggles, Regrets It Immediately