

I’d like to enact a new law that says you’re allowed to walk past someone even if it means invading their personal space a little bit. Like it’s ok buddy, you actually don’t need to run 359 degrees around me just because I am standing near a corner.
30% jokes, 30% attempts at academic discussions, 40% spewing my opinions uninvited to find out what might be missing from my perspective.
I’ll usually reiterate this in my posts, but I never give legal advice online. I can describe how the law generally tends to be, analyze a public case from an academic perspective, and explain how courts normally treat an issue. But hell no am I even going to try to apply the law to your specific situation.
I’d like to enact a new law that says you’re allowed to walk past someone even if it means invading their personal space a little bit. Like it’s ok buddy, you actually don’t need to run 359 degrees around me just because I am standing near a corner.
Advice against phishing emails can be reduced to, “1: Never click on a link, call a phone number, download an attachment, or follow instructions you found in an email unless you were already expecting this exact email from this exact sender. 2: If you really want to do those things, search up the organization’s website directly and use the contact info they provide there instead.”
imo it’s the ad-hungry articles stretching everything into 10+ pages that’s making advice so inaccessible to people. Super annoying because it dilutes the real, simple message that’s already there, it’s just locked behind an adwall.
Ok, let me be more specific so that it’s not open to uncharitable interpretation.
What happens when it becomes easy to make something as reliable and complete as, e.g., ChatGPT-4 without the hardware costs and other costs currently associated with it?
That’s true, but thinking about AI that is made to generate speech, processing power is still expensive enough that developers are careful with it. But what happens as memory gets cheaper and calculations get faster, and ordinary developers are able to train their own generative AI?
For example, what happens when a developer decides to train a LLM extensively on scam emails, and spammers love to buy copies of it - but the developer markets it as just “a helpful generative AI”? Or, what if a person trains their LLM on an extremist forum full of hate speech and disinformation, then offers it to a suicide prevention center as a 24/7 alternative to human labor? (Treating these as hypotheticals, where we assume the difference isn’t immediately obvious. Perhaps they also used some legitimate training data, so that most outputs seem innocent enough.)
To me it sounds more involved than selling just a word processor with autocorrect, but less involved than selling an instruction manual for committing crimes.
No way they don’t force you to agree to some “terms and conditions” along the lines of, “You accept full responsibility of all risk and if we get sued, you agree to pay on our behalf. And because we know you won’t read this, here’s all the risks so we can say you gave informed consent: …”
Completely speculating, because I don’t know many courts that have been willing to decide either way, but maybe:
If it causes harm in a way that was reasonably foreseeable, the person who turned it on and/or the person “operating” it might be generally liable on a theory of negligence (but not always).
If the harm was unpredictable, it might be on the manufacturer and the retailer under a theory of products liability (but not always).
Or it could be treated as “ferae naturae,” where owners are liable for their ‘dangerous animal’ pets because they knew the pets were dangerous and still decided to keep them (but not always).
If it’s an AI not associated with a physical device, maybe the programmer who “authored” the lines of code could be criminally liable for criminal “speech” (writing an AI) that incites and enables crime, even as a conspirator – that’s reeeaaally doubtful on Due Process grounds, but it would definitely light a fire under every developer’s chair to make sure their algorithms are explicitly trained against criminal behavior. (but still not always.)
Before I left, I remember it being really bad. People were abusing therapyspeak without any regard for what the terms actually mean. Like, “my bf keeps violating my boundaries by not buying me gifts,” “NTA your mom is parentifying you by asking you to clean the dishes,” “divorce your wife of 7 years because she neglects you by asking for an hour alone every day.”
Interesting, so what happens when an AI creates art that would infringe on a human’s copyright? Would AI art of Mickey Mouse be public domain, meaning AI could be the end of Disney’s insane licensing fee?
Edit: Nevermind, turns out this article is just editorialized. It isn’t public domain, it just isn’t eligible for the AI’s creator to copyright it if it’s fully autonomous.
afaik Amazon tries to offload the work of vetting its vendors by requiring them to have a registered trademark. This led to all the sketchy sellers making tons of fake companies with random strings of letters as names, knowing the USPTO is going to approve “AEGIJDU Clothing” because nobody is ever going to contest that name.
That’s why you see a ton of identical products listed with supposedly different, super random brand names, in case Amazon tries to take down one of the “vendors” (aka, one of the real vendor’s many fronts).
Galaxy brain idea: Just encrypt your messages manually. Agree on an algorithm and trade keys in-person, then send each other encrypted files that you decrypt with a separate program (and for added privacy, on a separate device that doesn’t have network access). It’s not convenient at all but the idea is hilarious.
There’s an urban myth at my university that two students did this to test rumors that the school emails were being monitored, and after a few weeks later IT emailed them asking them to stop.
I think it has to be somewhere in between. This ‘real deal’ theory doesn’t explain the popularity of hentai, but at the same time, OnlyFans shows that some people reaaallly care about the personal element. I would bet niche kinks (especially those ‘illegal to make but legal to watch’?) will lean heavily on AI for content, but the rest will probably change based on our culture’s attitude toward AI in general.