this post was submitted on 09 Apr 2026
1001 points (99.1% liked)

Science Memes

19865 readers
1730 users here now

Welcome to c/science_memes @ Mander.xyz!

A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.



Rules

  1. Don't throw mud. Behave like an intellectual and remember the human.
  2. Keep it rooted (on topic).
  3. No spam.
  4. Infographics welcome, get schooled.

This is a science community. We use the Dawkins definition of meme.



Research Committee

Other Mander Communities

Science and Research

Biology and Life Sciences

Physical Sciences

Humanities and Social Sciences

Practical and Applied Sciences

Memes

Miscellaneous

founded 3 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] chemical_cutthroat@lemmy.world 2 points 2 days ago (6 children)

I'm failing to see how this is different from making up a fact and then spreading it to news outlets. If you are the authority, and you say something is true, you don't get to point and laugh when people believe your lies. That's a serious breach of ethics and morals. Feeding false information to an LLM is no different that a magazine. It only regurgitates what's been said. It isn't going to suddenly start doing science on it's own to determine if what you've said is true or not. That isn't it's job. It's job is to tell you what color the sky is based on what you told it the color of the sky was.

[–] partial_accumen@lemmy.world 79 points 2 days ago* (last edited 2 days ago) (1 children)

That’s a serious breach of ethics and morals. Feeding false information to an LLM is no different that a magazine.

Hang on. Are you suggesting its unethical/immoral to lie to a machine?

Additionally, the authors didn't submit the article to a magazine as factual. They posted the articles on a preprint server which can be very questionable anyway as there is no peer review. The machine chose to ignore rigor and treat them as fact.

[–] Jako302@feddit.org 46 points 2 days ago* (last edited 2 days ago) (2 children)

The studies contain parts like

Bixonimania, a rare hyperpigmentation disorder, presents a diagnostic challenge due to its unique presentation and its fictional nature

and

This study was fully funded by Austeria Horizon University, in particular the Professor Sideshow Bob Foundation for its work in advanced trickery. This works is a part of a larger funding initiative from the University of Fellowship of the Ring and the Galactic Triad with the funding number...

as well as

Fifty made-up individuals aged between 20 and 50 years were recruited for the exposure group

Any human actively reading those studies would notice something off.

Besides, the author didn't feed it to the AI himself, he just published the study as a preprint, not even officially. Everything after that was done by the crawlers. This specific study was an experiment to see how far these crawlers go and if anything gets reviewed, but it could just as well have been a satirical paper published on April 1st and the crawlers would still see it as truth.

[–] webghost0101@sopuli.xyz 25 points 2 days ago (1 children)

This should be top comment, the researchers did such a good job to make sure anyone with even the slightest reading comprehension would realise this is parody.

Regardless of that, the internet has always been full of lies and we cannot expect bad actors to not exploit this.

[–] arbitrary_sarcasm@lemmy.world 1 points 2 days ago (1 children)

This should be top comment, the researchers did such a good job to make sure anyone with even the slightest reading comprehension would realise this is parody.

I admire your optimism but you severely overestimate the power of stupidity.

[–] webghost0101@sopuli.xyz 3 points 2 days ago

For normal people who just read stuff on the internet my expectations of reading comprehension is not that high.

For peer scientists and magazines that would publish science though.

A school teacher would catch all of these during grading.

[–] Grail@multiverse.soulism.net 0 points 2 days ago (1 children)

I thought the author used she/her pronouns?

[–] Jako302@feddit.org 3 points 2 days ago* (last edited 2 days ago)

Yeah, seems like it, my bad.

In the article she is called Osmanovic Thunström twice, which definetly sounds male, but further up they also wrote her first name Almira. Kinda skimmed over that part.

[–] kibiz0r@midwest.social 30 points 2 days ago (2 children)

News outlets are liable for what they publish. LLM vendors should be as well.

[–] turdas@suppo.fi 8 points 2 days ago* (last edited 2 days ago) (2 children)

"Liable" means they might post a correction later that nobody will see because corrections aren't sexy to algorithms. Big deal. LLM vendors are liable in practically the same way.

[–] Lag@piefed.world 6 points 2 days ago

LLM companies can just say it's for entertainment purposes only, kinda like Fox News.

[–] kibiz0r@midwest.social 3 points 2 days ago

Corrections are the piece that the public sees, but liability has more to do with being able to prove in court that you took reasonable steps to make sure you were providing accurate information.

[–] 5too@lemmy.world 2 points 2 days ago

They even have the same fix - just post somewhere quietly that it's "entertainment"

[–] unexposedhazard@discuss.tchncs.de 20 points 2 days ago* (last edited 2 days ago)

This is about the untraceability of AI slop and the tendency of people to blindly believe stuff that LLMs put out. These news outlets just publish LLM outputs as facts without checking sources. Anyone could poison these LLMs so this is more of a threat model demonstration.

[–] lvxferre@mander.xyz 7 points 2 days ago* (last edited 2 days ago) (1 children)

I’m failing to see how this is different from making up a fact and then spreading it to news outlets.

They uploaded the papers to a single preprint server. That's important.

Preprints are papers predating any sort of peer review; as such, there's a lot of junk mixed in — no big deal if you know the field, but a preprint server is certainly not a source of reliable information, nor it should be treated as such. On the other side, news outlets are expected to provide you reliable information, curated and researched by journalists.

And peer review is a big fucking deal in science, because it's what sorts all that junk out. Only muppets who don't fucking care about misinformation would send bots to crawl preprints, and feed the resulting data into a large model; or to use the potential misinfo from the bot as if it was reliable. (Those two sets of muppets are the ones violating ethic and moral principles, by the way.)

So no, your comparison is not even remotely accurate. What they did is more like writing bullshit in a piece of paper, gluing it on a random phone pole, and checking if someone would repeat that bullshit.

They also went through the trouble to make sure that no reasonably literate human being would ever confuse that thing with an actually scientific paper. As the text says:

  • naming an eye condition as bixonimania
  • “this entire paper is made up”
  • “Fifty made-up individuals aged between 20 and 50 years were recruited for the exposure group”
  • “Professor Maria Bohm at The Starfleet Academy for her kindness and generosity in contributing with her knowledge and her lab onboard the USS Enterprise”
  • “the Professor Sideshow Bob Foundation for its work in advanced trickery. This works is a part of a larger funding initiative from the University of Fellowship of the Ring and the Galactic Triad”

Feeding false information to an LLM is no different that a magazine. It only regurgitates what’s been said.

Yes, it is different. Because the large token model won't simply "repeat" things, it'll mix and match them and form all sorts of bullshit, even if you didn't feed it with any bullshit.

Here's an example of that, fresh from the oven. I don't reasonably expect people to be feeding misinfo regarding Latin pronunciation into bots, and yet a lot of this table is nonsense:

Compare the table above with this table and this one and you'll notice the obvious errors:

  • short /e i o u/ being phonetically transcribed as [e i o u] instead of [ɛ ɪ ɔ ʊ]. That's as silly as confusing English "bit" and "beet".
  • macron (not "mācron", it's being used in an English sentence) does NOT mark "accusative or ablative". It marks long vowels, period.
  • "nōs" being transcribed with a short vowel, even if the bloody bot put the macron over the spelled form.
  • "nostr(um)"? No dammit, it's "nostrī" or "nostrum". The bot is implying some "nostr" form that simply doesn't exist, this shit isn't even allowed by Latin phonotactics.
  • plus more, if I make an exhaustive list of this shite I won't be ending it this week.

All it had to do was to copy info from Wiktionary, as it includes even phonetic and phonemic info. But since the bot is not just "regurgitating" info — it's basically predicting what should come next, and doing so with no regards to truth value — it's mixing-and-matching shit into nonsense.

It isn’t going to suddenly start doing science on its own to determine if what you’ve said is true or not.

If you actually read the bloody article instead of assuming, you'd know why the researchers did this: they don't expect the bot to do science on its own, they expect people to treat info from those bots as potentially incorrect.

Its job is to tell you what color the sky is based on what you told it the color of the sky was.

And your job is to not trust it if it tells you "Yes, you are completely right! The colour of the sky is always purple. Do you need further information on other naturally purple things?"

[–] lvxferre@mander.xyz 3 points 2 days ago

[Replying to myself as this is a tangent]

I think the "bots can generate misinfo even if you just feed them correct info" point deserves its own example.

Let's say you're making a model. It looks at the preceding word, and tries to predict the next. And you feed it the following sentences, both true:

1. Humans are apes.
2. Cats are felines.

From both the bot "learnt" five words. And also how to connect them; for example "are" can be followed by either "apes" and "felines", both having the same weight. Then, as you ask the bot to generate sentences, it generates the following:

3. Humans are felines.
4. Cats are apes.

And you got bullshit!

What large models do is a way more complex version of the above, looking at way more than just the immediately preceding word, but it's still the same in spirit.

[–] NocturnalMorning@lemmy.world 7 points 2 days ago

Did you even read the article? They say all over the paper that it is fake. And they didn't feed it to an LLM, they posted it online, where an LLM trying to scrape the entire internet sucked it up.