Wary of the bill. Seems like every bill involving stuff like this is either designed to erode privacy or for regulatory capture.
Edit: spelling
And this makes you tired…?
Yes. It’s very tiring having to constantly fight this battle. Unfortunately that’s what they want cause if enough of us are too tired to care then eventually it slips through and we never get back what we lost.
Lol, good catch.
introducing the AI transparency act, which requires every generative prompt to be registered in a government database
I don’t see a problem with that, I think that this information should be public, both prompt and result, because:
- a. The “AIs” companies already know that, why shouldn’t anyone else?
- b. They use public information to train their models, thus their results should also be public.
- c. This would be the ultimate way to know that something was “AI” generated.
This is a very different subject from giving acess for your DMs. The only ones who benefit from this information not being publicly available are those who use “AI” for malicious purposes, while everyone benefits from privacy of correspondence.
I suppose you would also be fine with every one of your google searches being in a database? Every video you’ve ever watched, even the ones in private browser tabs?
No, and that’s why I don’t use Google or anything that isn’t encrypted and sends any data that I consider private to some datacenter. And even when I know the data is encrypted, I am careful, as anyone should be, with data leaving your computer and going to someone else’s.
“AI” is not the same thing. Why would I want my prompt to be private if I don’t want to use the result in some malicious way, be it generating CSAM or using it to cheat someone to write an article, or to generate a Deep Fake video of someone for an internet scam?
Why would I want my prompt to be private if I don’t want to use the result in some malicious way
Do you think that the only thing people use AI for is making deepfakes and CSAM? AFAIK the most common use is generating porn. Now, I don’t think generating regular porn is “malicious”, but I certainly understand why most people (self included) want to keep what they generate private.
I don’t think people’s right to generate whatever image they want to jerk off to is fundamental or more important than avoiding “AI” scams and CSAM generation. There are other ways to jerk off: there’s plenty of real people porn online and also lots, lots, lots of hentai, for literally every taste. “AI” porn only has two particularities that are not satisfied by these two options, one is to generate the scene you want, and for the very remote possibility that what you have imagined has never been produced before, you can pay an artist to To do so, another is Deep Fake porn, which should be a crime, it doesn’t matter if you’re not going to publish the image.
and that’s what I loathe about the idiots who are for this stuff. Yes, I want to curb this stuff - but for fuck’s sake there are ways to do it that aren’t “Give big government every scrap of data on you”.
There are ways to prove I’m over 18 without needing to register my ID with a porn company, or to regulate CSAM while not having to read private messages. Fuck, but we have the combination of circle of a venn diagram of idiot and control freak in congress, and they’ll happily remove all of our rights over some fear of the boogeyman