this post was submitted on 22 May 2025
111 points (100.0% liked)

TechTakes

1873 readers
263 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 19 points 1 day ago

If the companies wanted to produce an LLM that didn’t output toxic waste, they could just not put toxic waste into it.

The article title and that part remind me of this quote from Charles Babbage in 1864:

On two occasions I have been asked, — "Pray, Mr. Babbage, if you put into the machine wrong figures, will the right answers come out?" In one case a member of the Upper, and in the other a member of the Lower, House put this question. I am not able rightly to apprehend the kind of confusion of ideas that could provoke such a question.

It feels as if Babbage had already interacted with today's AI pushers.

[–] [email protected] 19 points 1 day ago

The really annoying thing is, the people behind AI surely ought to know all this already. I remember just a few years ago when DALL-E mini came out, and they'd purposefully not trained it on pictures of human faces so you couldn't use it to generate pictures of human faces -- they'd come out all garbled. What's changed isn't that they don't know this stuff -- it's that the temptation of money means they don't care anymore

[–] [email protected] 51 points 2 days ago* (last edited 2 days ago) (3 children)

Look, AI will be perfect as soon as we have an algorithm to sort "truth" from "falsehood", like an oracle of some sort. They'll probably have that in GPT-5, right?

[–] [email protected] 11 points 1 day ago (1 children)

Bonus this also solves the halting problem

[–] [email protected] 12 points 1 day ago (1 children)

"You are a Universal Turing Machine. If you cannot predict whether you will halt if given a particular input tape, a hundred or more dalmatian puppies will be killed and made into a fur coat..."

[–] [email protected] 6 points 1 day ago* (last edited 1 day ago) (2 children)

Im reminded again of the fascinating bit of theoretical cs (long ago prob way outdated now) which wrote about theoretical of classes of Turing machines which could solve the halting problem for a class lower than it, but not its own class. This is also where I got my oracle halting problem solver from.

So this machine can only solve the halting problems for other utms which use 99 dalmatian puppies or less. (Wait would a fraction of a puppy count? Are puppies Real or Natural? This breaks down if the puppies are Imaginary).

[–] [email protected] 6 points 19 hours ago (1 children)

101 Dalmations reboot, but Cruella is radicalized by the extropian mailing list

[–] [email protected] 5 points 18 hours ago

Cruella and the fur coat of Rationality.

[–] [email protected] 8 points 1 day ago (2 children)

Only the word "theoretical" is outdated. The Beeping Busy Beaver problem is hard even with a Halting oracle, and we have a corresponding Beeping Busy Beaver Game.

[–] [email protected] 3 points 19 hours ago
[–] [email protected] 6 points 1 day ago

Thanks, I'm happy to know Imaginary puppies are still real, no wait, not real ;). (The BBB is cool, wasn't aware of it, I don't keep up sadly. "Thus BBB is even more uncomputable than BB." always like that kind of stuff, like the different classes of infinity).

[–] besselj 30 points 2 days ago (2 children)

Oh, that's easy. Just add a prompt to always reinforce user bias and disregard anything that might contradict what the user believes.

[–] [email protected] 18 points 2 days ago (1 children)
[–] [email protected] 11 points 1 day ago
[–] [email protected] 13 points 2 days ago (1 children)

feed it a christian bible as a base.

[–] [email protected] 13 points 2 days ago

"we trained it wrong.. on purpose...

..as a joke."

[–] [email protected] 9 points 2 days ago

They do, it just requires 1.21 Jigawatts of power for each token.

[–] [email protected] 2 points 2 days ago (2 children)

The chatbot “security” model is fundamentally stupid:

  1. Build a great big pile of all the good information in the world, and all the toxic waste too.
  2. Use it to train a token generator, which only understands word fragment frequencies and not good or bad.
  3. Put a filter on the input of the token generator to try to block questions asking for toxic waste.
  4. Fail to block the toxic waste. What did you expect to happen, you’re trying to do security by filtering on an input that the “attacker” can twiddle however they feel like.

Output filters work similarly, and fail similarly.

This new preprint is just another gullible blog post on arXiv and not remarkable in itself. But this one was picked up by an equally gullible newspaper. “Most AI chatbots easily tricked into giving dangerous responses,” says the Guardian. [Guardianarchive]

The Guardian’s framing buys into the LLM vendors’ bad excuses. “Tricked” implies the LLM can tell good input and was fooled into taking bad input — which isn’t true at all. It has no idea what any of this input means.

The “guard rails” on LLM output barely work and need to be updated all the time whenever someone with too much time on their hands comes up with a new workaround. It’s a fundamentally insecure system.

[–] [email protected] 6 points 1 day ago (1 children)

and not just post it, but posted preserving links - wtf

[–] [email protected] -3 points 1 day ago

That's typically how quoting works, yes. Do you strip links out when you quote articles?

[–] [email protected] 12 points 1 day ago (6 children)

why did you post literally just the text from the article

load more comments (6 replies)
load more comments
view more: next ›