Recently, there was a post on SB-1047 and how it's quite mild regulation. I'm not expert on it and don't know how it works. 

In the comment section I was asking:

Why wouldn't deep fake porn or voice cloning technology to engage in fraud be powerful enough to materially contribute to critical harm?

There are cases of fraud that could do $500,000,000 in damages.

Given how juries decide about damages, a model that's used to create child porn for thousands of children could be argued to cause  $500,000,000 in damages as well. Especially when coupled with something like trying to extort the children.

I'm surprised that my comment didn't get any engagement where people explained how they think the law will handle those cases while at the same time my post got no karma votes.

I'd love to believe that the law is well thought out, and simply a good step for AI safety. At the same time, I also like having accurate beliefs about the effects of the law, so let me repeat my question here.

How does the law handle damage caused by deep fake porn or fraud with voice cloning?


 

New Answer
New Comment

2 Answers sorted by

RobertM

1210

Notwithstanding the tendentious assumption in the other comment thread that courts are maximally adversarial processes bent on on misreading legislation to achieve their perverted ends, I would bet that the relevant courts would not in fact rule that a bunch of deepfaked child porn counted as "Other grave harms to public safety and security that are of comparable severity to the harms described in subparagraphs (A) to (C), inclusive", where those other things are "CBRN > mass casualties", "cyberattack on critical infra", and "autonomous action > mass casualties".  Happy to take such a bet at 2:1 odds.

But there are some simpler reason that particular hypothetical fails:

  • Image models are just not nearly as expensive to train, so it's unlikely that they'd fall under the definition of a covered model to begin with.
  • Even if someone used a covered multimodal model, existing models can already do this.

See:

(2) “Critical harm” does not include any of the following:

(A) Harms caused or materially enabled by information that a covered model or covered model derivative outputs if the information is otherwise reasonably publicly accessible by an ordinary person from sources other than a covered model or covered model derivative.

I’m not sure if you intended the allusion to “the tendentious assumption in the other comment thread that courts are maximally adversarial processes bent on on misreading legislation to achieve their perverted ends”, but if it was aimed at the thread I commented on… what? IMO it is fair game to call out as false the claim that

It only counts if the $500m comes from "cyber attacks on critical infrastructure" or "with limited human oversight, intervention, or supervision....results in death, great bodily injury, property damage, or property loss."

even if deep... (read more)

RamblinDash

2-2

It only counts if the $500m comes from "cyber attacks on critical infrastructure" or "with limited human oversight, intervention, or supervision....results in death, great bodily injury, property damage, or property loss."

So emotional damages, even if severe and pervasive, can't get you there.

If you read the definition of critical harms, you’ll see the $500m doesn’t have to come in one of those two forms. It can also be “Other grave harms to public safety and security that are of comparable severity”.

1RamblinDash
I have a hard time imagining a Court ruling that "Other grave harms to public safety and security that are of comparable severity" could embrace something so different-in-kind than the listed items.
5Shankar Sivarajan
From the Fish and Game code:  This was read to include bees for the purposes of the Endangered Species Act which lists  (Emphases mine).
2RamblinDash
Well, "fish" is a statutorily defined term that clearly includes all invertebrates. What did you want the court to do, ignore the statutory text? Arguably, that outcome supports the notion that the courts are less likely to just ignore text limiting the kinds of harm that are cognizable, not more likely, as you seem to be arguing.
4ChristianKl
(4) uses three terms "public safety", "public security" and comparable severity.  I would expect that severity means $500,000,000 worth of damage. Public safety does not seem to be a clearly defined term but fairly broad. 

If someone creates an automated system that makes deep fake porn and then emails with that porn to blackmail people and publishes the deep fake porn when people don't pay up, that could very well be a system with limited human oversight, intervention, or supervision.

Those people who pay the blackmail would also suffer from property loss. 

If you have someone committing suicide because of deep fake porn images of themselves, it might also result in death.

If you have one suicide + $500,000,000 worth in emotional damage wouldn't it count?

7 comments, sorted by Click to highlight new comments since:

I'm not sure why those shouldn't be included? If someone uses my AI to perform 500 million dollars of fraud, then I should probably have been more careful releasing the product.

leogao spoke of SB-1047 being quite mild.

If you include those things, I expect that you effectively outlaw Open Source AI models for video/audio generation.

I think it's a reasonable decision to outlaw Open Source AI models, but it doesn't sound "mild".

[+][comment deleted]20