The Guardian Publishes Fake AI Op-Ed, Turns Out It was Not Written by a Robot at All

What kind of sicko fakes an artificial intelligence?

RT:

The Guardian claimed an opinion piece was written in its “entirety” by a language-generating robot, sparking accusations the paper misled the public about the AI’s current capabilities…and probably its true intentions.

“A robot wrote this entire article. Are you scared yet, human?” reads the title of the opinion piece published on Tuesday. The article was attributed to GPT-3, described as “a cutting edge model that uses machine learning to produce human-like text.”

While the Guardian claims that the soulless algorithm was asked to “write an essay for us from scratch,” one has to read the editor’s note below the purportedly AI-penned opus to see that the issue is more complicated. It says that the machine was fed a prompt asking it to “focus on why humans have nothing to fear from AI” and had several tries at the task.

After the robot came up with as many as eight essays, which the Guardian claims were all “unique, interesting and advanced a different argument,” the very human editors cherry-picked “the best part of each” to make a coherent text out of them.

Although the Guardian said that it took its op-ed team even less time to edit GPT-3’s musings than articles written by humans, tech experts and online pundits have cried foul, accusing the newspaper of “overhyping” the issue and selling their own thoughts under a clickbait title.

“Editor’s note: Actually, we wrote the standfirst and the rather misleading headline. Also, the robot wrote eight times this much and we organised it to make it better…” tweeted Bloomberg Tax editor Joe Stanley-Smith.

Futurist Jarno Duursma, who wrote books on the Bitcoin Blockchain and artificial intelligence, agreed, saying that to portray an essay compiled by the Guardian as written completely by a robot is exaggeration.

“Exactly. GPT-3 created eight different essays. The Guardian journalists picked the best parts of each essay (!). After this manual selection they edited the article into a coherent article. That is not the same as ‘this artificial intelligent system wrote this article.’”

Mozilla fellow Daniel Leufer was even more bold in its criticism, calling the Guardian’s stunt “an absolute joke.”

“Rephrase: a robot didn’t write this article, but a machine learning system produced 8 substandard, barely-readable texts based on being prompted with the exact structure the Guardian wanted,” he summed up. He also spared no criticism for the piece itself, describing it as a patchwork that “still reads badly.”

This is a source of embarrassment by proxy for anyone who reads it.

What a bunch of dumb assholes.