Meta released it on 15 November with the explicit claim that it could aid scientific research.

I mean, were all smarter than that right?

But thats not how harm vectors work.

Meta takes new AI system offline because Twitter users are mean

Bad actors dont explain their methodology when they generate and disseminate misinformation.

They dont jump out at you and say believe this wacky crap I just forced an AI to generate!

LeCun appears to think that the solution to the problem is out of his hands.

Article image

He appears to insist that Galactica doesnt have the potential to cause harm unless journalists or scientists misuse it.

Very rarely do reputable actors reproduce dodgy sources.

But I cant writeinformationas fast as an AI can outputmisinformation.

Article image

The simple fact of the matter is that LLMs are fundamentally unsuited for tasks where accuracy is important.

They hallucinate, lie, omit, and are generally as reliable as a random number generator.

Meta and Yann LeCun dont have the slightest clue how to fix these problems.

Article image

Barring a major technological breakthrough on par with robot sentience, Galactica will always be prone to outputting misinformation.

Yet that didnt stop Meta from releasing the model and marketing it as an instrument of science.

A large language model for science.

Article image

Metas AI division is world-renowned.

And Yann LeCun, the companys AI boss, is a living legend in the field.

Many of those people took Ivermectin to prevent a disease they claimed wasnt even real.

Article image

That doesnt make any sense, and yet its true.

If you look up dietary silicon on Google Search, its a real thing.

Disclaimer:Im not a doctor, but dont eat crushed glass.

Article image

Youll probably die if you do.

You cant see the potential for harm?

Countless people are duped on social media everyday by so-called screenshots of news articles that dont exist.

Article image

Its easy to kick back and say those people are idiots.

But those idiots are our kids, our parents, and our co-workers.

Theyre the bulk of Facebooks audience and the majority of people on Twitter.

Article image

It has millions of files in its dataset.

Who knows whats in there?

(@mrgreene1977)November 18, 2022

Thats the problem.

Unfortunately, thats not how LLMs work.

Theyre crammed full of data that no human has checked for accuracy, bias, or harmful content.

Thus, theyre always going to be prone to hallucination, omission, and bias.

Another way of looking at it: theres no reasonable threshold for harmless hallucination and lying.

Metas AI told me to eat glass and kill myself.

It told me that queers and Jewish people were evil.

And, as far as I can see, there are no consequences.

Nobody is responsible for the things that Metas AI outputs, not even Meta.

Well-meaning journalists and academics are going to get fooled by papers this thing generates.

In the US, where Meta is based, this is business as usual.

Hell,Clearview AI operates in the USwith the full support of the Federal government.

But, in Europe, theres GDPR andthe AI Act.

That means GDPR may or not be a factor.

But the AI Actshouldcover these kinds of things.

That might sound harsh.

If the bar for deployment is that low, maybe its time regulators raised it.

Also tagged with