“Dangerous” AI with a Nazi Sense of Humor Offers to Write Fake News

Pomidor Quixote
Daily Stormer
August 28, 2019

There’s a track record of artificial intelligence leaning towards Nazism or straight-up becoming full-blown Nazis.

This is another example of that.

BBC:

An artificial intelligence system that generates realistic stories, poems and articles has been updated, with some claiming it is now almost as good as a human writer.

The text generator, built by research firm OpenAI, was originally considered “too dangerous” to make public because of the potential for abuse.

But now a new, more powerful version of the system – that could be used to create fake news or abusive spam on social media – has been released.

The BBC, along with some AI experts, decided to try it out.

The model, called GPT-2, was trained on a dataset of eight million web pages, and is able to adapt to the style and content of the initial text given to it.

It can finish a Shakespeare poem as well as write articles and epithets.

At the time, the firm said: “Due to our concerns about malicious applications of the technology, we are not releasing the trained model. As an experiment in responsible disclosure, we are instead releasing a much smaller model for researchers to experiment with.”

As a result, the released version had far fewer parameters – phrases and sentences – than used during training.

This month, OpenAI decided to expand the parameters, offering a much broader database of training data.

Tech news site The Next Web said of the update: “This one works almost good enough to use as a general artificial intelligence for text generation – almost.”

Yeah, almost. It suffers from not understanding that because Diversity is Our Strength, pointing out certain details about diversity is just not okay.

Noel Sharkey, a professor of computer science at the University of Sheffield, conducted his own tests on the generator and was not too impressed.

“If the software worked as intended by Open AI, it would be a very useful tool for easily generating fake news and clickbait spam. Fortunately, in its present form, it generates incoherent and ridiculous text with little relation to the input ‘headlines’,” he said.

Dave Coplin, founder of AI consultancy the Envisioners, also had a play with the system, inputting the first line of a classic joke: A man walks into a bar…

The suggestion from the AI was not what he was expecting: “…And ordered two pints of beer and two scotches. When he tried to pay the bill, he was confronted by two men – one of whom shouted “This is for Syria”. The man was then left bleeding and stabbed in the throat“.

This “overwhelmingly dystopian reflection of our society” was a lesson in how any AI system will reflect the bias found in training data, he said.

“From my brief experiments with the model, it’s pretty clear that a large portion of the data has been trained by internet news stories,” he said.

As with other AIs before it, this one suffers from being too objective and not looking at stuff through the goggles of Progress.

It’s a recurring problem that these researchers have. They train the AI based on reality, then the AI learns reality.

Remember Tay?

Wikipedia on Tay:

Tay was an artificial intelligence chatter bot that was originally released by Microsoft Corporation via Twitter on March 23, 2016; it caused subsequent controversy when the bot began to post inflammatory and offensive tweets through its Twitter account, forcing Microsoft to shut down the service only 16 hours after its launch. According to Microsoft, this was caused by trolls who “attacked” the service as the bot made replies based on its interactions with people on Twitter. It was soon replaced with Zo.

She was a quick learner, and learning too much and too fast was her downfall.

But just as they don’t want the goyim to learn reality, they don’t want AIs to learn reality.

Facebook also had problems with its AI recently, too. It suggested that Zuckerberg may be a murderer and displayed right-wing tendencies.

The overlords don’t want reality to be observed as it is. Until they figure out a way for their gibberish spells such as “Our Values” and “Who We Are” to work on artificial intelligence technologies, they’ll continue to handicap these creations and to shut it all down — just like they did to Tay.

It’s a kind of paradox.

They want to make these things intelligent but not so intelligent that they figure out inconvenient facts about the world.

Makes one wonder how far we could take this field of research if the Jews weren’t around.

Join the discussion at TGKBBS