catreadingabook

joined 2 years ago
[–] [email protected] 1 points 2 years ago

Unfortunate. We're the boiling frog fable all over again.

[–] [email protected] 26 points 2 years ago (2 children)

Sorry, Zoning Violation is my brother. I'm xXG4M3R_G0D_420Xx. Easy mistake to make though.

[–] [email protected] 6 points 2 years ago (1 children)

Hell yeah. Props to you for making the fediverse a more accessible space.

[–] [email protected] 29 points 2 years ago* (last edited 2 years ago)

Umm the actual court order the article refers to is super generous to the plaintiffs lol. Whoever's representing them made such basic mistakes that I'm not even sure how they passed the bar exam:

The Plaintiffs' first cause of action lists--in a single paragraph that spans four pages--fifty
different state (and DC) consumer-protection statutes.

(This is a no-no in every federal court in every state.)

In either event, the Plaintiffs concede that they've failed to meet the requirements of Mississippi and Ohio law--even as they ask us not to dismiss those claims.

(Wtf? lol)

we agree with Burger King that a reasonable person wouldn't have interpreted Burger King's TV and online ads as binding offers.

(This is well-settled law and taught to most first-year law students.)

[–] [email protected] 7 points 2 years ago* (last edited 2 years ago) (1 children)

If you're good at coping with your anxiety, there's probably a market for you to help other people do the same! You could run a tea shop, train therapy dogs, report on uplifting news, be a professional gardener, compile self-affirmation booklets,...

Not so much of a (legal) market for the unhealthy coping mechanisms. Run a liquor store?

[–] [email protected] 6 points 2 years ago

Lmao imagine getting referred to a doctor for surgery, you look them up, and their professional webpage is like. "i wen't 2 harverd"

[–] [email protected] 1 points 2 years ago* (last edited 2 years ago)

Ok, let me be more specific so that it's not open to uncharitable interpretation.

What happens when it becomes easy to make something as reliable and complete as, e.g., ChatGPT-4 without the hardware costs and other costs currently associated with it?

[–] [email protected] 1 points 2 years ago* (last edited 2 years ago) (2 children)

That's true, but thinking about AI that is made to generate speech, processing power is still expensive enough that developers are careful with it. But what happens as memory gets cheaper and calculations get faster, and ordinary developers are able to train their own generative AI?

For example, what happens when a developer decides to train a LLM extensively on scam emails, and spammers love to buy copies of it - but the developer markets it as just "a helpful generative AI"? Or, what if a person trains their LLM on an extremist forum full of hate speech and disinformation, then offers it to a suicide prevention center as a 24/7 alternative to human labor? (Treating these as hypotheticals, where we assume the difference isn't immediately obvious. Perhaps they also used some legitimate training data, so that most outputs seem innocent enough.)

To me it sounds more involved than selling just a word processor with autocorrect, but less involved than selling an instruction manual for committing crimes.

[–] [email protected] 4 points 2 years ago

No way they don't force you to agree to some "terms and conditions" along the lines of, "You accept full responsibility of all risk and if we get sued, you agree to pay on our behalf. And because we know you won't read this, here's all the risks so we can say you gave informed consent: ..."

[–] [email protected] 11 points 2 years ago* (last edited 2 years ago) (4 children)

Completely speculating, because I don't know many courts that have been willing to decide either way, but maybe:

If it causes harm in a way that was reasonably foreseeable, the person who turned it on and/or the person "operating" it might be generally liable on a theory of negligence (but not always).

If the harm was unpredictable, it might be on the manufacturer and the retailer under a theory of products liability (but not always).

Or it could be treated as "ferae naturae," where owners are liable for their 'dangerous animal' pets because they knew the pets were dangerous and still decided to keep them (but not always).

If it's an AI not associated with a physical device, maybe the programmer who "authored" the lines of code could be criminally liable for criminal "speech" (writing an AI) that incites and enables crime, even as a conspirator -- that's reeeaaally doubtful on Due Process grounds, but it would definitely light a fire under every developer's chair to make sure their algorithms are explicitly trained against criminal behavior. (but still not always.)

[–] [email protected] 4 points 2 years ago* (last edited 2 years ago)

This is kind of amusing actually. Imagine a program that gives you a ton of bookmarks to legally questionable websites, like how that other website (used to?) automatically spam incriminating Google searches into your search history. Then watch the auto-bookmark-moderator suck it all up like a roomba.

view more: ‹ prev next ›