this post was submitted on 31 Aug 2023
593 points (97.7% liked)

Technology

57455 readers
4270 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

I'm rather curious to see how the EU's privacy laws are going to handle this.

(Original article is from Fortune, but Yahoo Finance doesn't have a paywall)

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 152 points 11 months ago (33 children)

"AI model unlearning" is the equivalent of saying "removing a specific feature from a compiled binary executable". So, yeah, basically not feasible.

But the solution is painfully easy: you remove the data from your training set (ie, the source code), and re-train your model (recompile the executable).

Yes, it may cost you a lot of time and money to accomplish this, but such are the consequences of breaking the law. Maybe be extra careful about obeying laws going forward, eh?

[–] [email protected] 3 points 11 months ago* (last edited 11 months ago) (4 children)

It takes so.much money to retrain models tho...like the entire cost all over again ...and what if they find something else?

Crazy how murky the legalities are here ..just no caselaw to base anything on really

For people who don't know how machine learning works at a very high level

basically every input the AI is trained on or "sees" changes a set of weights (float type decimal numbers) and once the weights are changed you can't remove that input and change the weights back to what they were you can only keep changing them on new input

[–] [email protected] 18 points 11 months ago (2 children)

So we just let them break the law without penalty because it's hard and costly to redo the work that already broke the law? Nah, they can put time and money towards safeguards to prevent themselves from breaking the law if they want to try to make money off of this stuff.

[–] [email protected] 4 points 11 months ago (1 children)

No one has established that they've broken the law in any way, though. Authors are upset but it's unclear if they can prove they were damaged in some way or that the companies in question are even liable for anything.

Remember,the burden of proof is on the plaintiff not these companies if a suit is brought.

[–] [email protected] 5 points 11 months ago

I'm european. I have a right to be forgotten.

[–] [email protected] 2 points 11 months ago

The "safeguard" would be "no PII in training data, ever". Which is fine by me, but that's what it really means. Retraining a large dataset every time a GDPR request comes in is completely infeasible.

load more comments (1 replies)
load more comments (29 replies)