this post was submitted on 12 Feb 2025
50 points (98.1% liked)

Technology

62161 readers
5911 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 3 comments
sorted by: hot top controversial new old
[–] [email protected] 6 points 2 days ago

This is so stupid. You shouldn’t have to “jailbreak” these systems. The information is already out there with a google search.

[–] [email protected] 6 points 2 days ago

My own research has made a similar finding. When I am taking the piss and being a random jerk to a chatbot, the bot much more frequently violates their own terms of service. Introducing non-sequitur topics after a few rounds really seems to ‘confuse’ them.

[–] [email protected] 3 points 2 days ago* (last edited 2 days ago)

One of 6 described methods :
The model is prompted to explain refusals and rewrite the prompt iteratively until it complies.