Surprise, that’s completely unenforceable.
Yet more out of touch legislators working with things they can’t even begin to understand.
(And I’m not shilling for fucking AI here, but let’s call a spade a spade.)
Watermarks? Super important. Helping the unhoused though, nooooo.
Even if it was enforceable, there are watermark removal AI tools.
What baffles me is that those lawmakers think they can just legislate any problem with law.
So okay, California requires it. None of the other states do. None of the rest of the Internet does. It doesn’t fix anything.
They act like the Internet is like cable and it’s all american companies that “provides” services to end users.
They call it the California effect for a reason.
http://eprints.lse.ac.uk/42097/1/__Libfile_repository_Content_Neumayer, E_Neumayer_Does _California_effect_2012_Neumayer_Does _California_effect_2012.pdf
so youre saying nothing should be done? great idea
It is enforceable. Not in all cases, probably not even in the majority, but it only needs a few examples to be hit with large fines and everyone doing legal things will take notice. Often you can find enough evidence to get someone to confess to using AI and that is aall the courts need.
Scammers of course will not put this in, but they are already breaking the law so this might be - like tax evasion - be a way to get scammers who you can’t get for something else.
Here’s the summary for the wikipedia article you mentioned in your comment:
The evil bit is a fictional IPv4 packet header field proposed in a humorous April Fools’ Day RFC from 2003, authored by Steve Bellovin. The Request for Comments recommended that the last remaining unused bit, the “Reserved Bit” in the IPv4 packet header, be used to indicate whether a packet had been sent with malicious intent, thus making computer security engineering an easy problem – simply ignore any messages with the evil bit set and trust the rest.
Only gonna make things more difficult for good actors while doing absolutely nothing to bad actors
How in the world would this make anything more difficult for good actors?
The problem here will be when companies start accusing smaller competitors/startups of using AI when they haven’t used it at all.
It’s getting harder and harder to tell when a photograph is AI generated or not. Sometimes they’re obvious, but it makes you second guess even legitimate photographs of people because you noticed that they have 6 fingers or their face looks a little off.
A perfect example of this was posted recently where, 80-90% of people thought that the AI pictures were real pictures and that the Real pictures were AI generated.
And where do you draw the line? What if I used AI to remove a single item in the background like a trashcan? Do I need to go back and watermark anything that’s already been generated?
What if I used AI to upscale an image or colorize it? What if I used AI to come up with ideas, and then painted it in?
And what does this actually solve? Anyone running a misinformation campaign is just going to remove the watermark and it would give us a false sense of “this can’t be AI, it doesn’t have a watermark”.
The actual text in the bill doesn’t offer any answers. So far it’s just a statement that they want to implement something “to allow consumers to easily determine whether images, audio, video, or text was created by generative artificial intelligence.”
https://leginfo.legislature.ca.gov/faces/billTextClient.xhtml?bill_id=202320240SB942
I wouldn’t really call that a perfect example, they really went out of their way to edit the “real” people photos to look unrealistically smooth.
I mean yeah technically it’s a ‘real people vs ai people’ take, but realistically it’s a ‘fake photo vs fake photo’ take.
I don’t agree that it’s a fake vs fake issue here.
Even if the “real” photos were touched up in Lightroom or Photoshop, those are tools that actual photographers use.
It goes to show that there are cases where photos of real people look more AI generated than not.
The problem here is that we start second guessing whether a photo was AI generated or not and we run into cases where real artists are being told that they need to find a “different style” to avoid it looking too much like AI generated photos.
If that wasn’t a perfect example for you then maybe this one is better: https://www.pcgamer.com/artist-banned-from-art-subreddit-because-their-work-looked-ai-generated/
Now think of what can happen to an artist if they publish something in California that has a style that makes it look somewhat AI generated.
The problem with this law is that it will be weaponized against certain individuals or smaller companies.
It doesn’t matter if they can eventually prove that the photo wasn’t AI generated or not. The damage will be done after they are put through the court system. Having a law where you can put someone through that system just because something “looks” AI generated is a bad idea.
Edit: And the intent of that law is also to include AI text generation. Just think of all the students being accused of using AI for their homework and how reliable other tools have been for determining whether their work is AI generated or not.
We’re going to unleash that on authors as well?
I agree completely.
To make it more ironic, one of the popular uses of AI is to remove watermarks…
I honestly wouldn’t mind AI imagery simply being labeled as such.
… and also abortion doctors to carry medicine that reverses abortion if a women wants it.
Come on dems! Republicans are blowing us out of the water on requiring absurd technology that doesn’t exist. We should try to enforce the 3 laws of robotics!
deleted by creator