Skip Navigation

A.I.’s un-learning problem: Researchers say it’s virtually impossible to make an A.I. model ‘forget’ the things it learns from private user data

finance.yahoo.com A.I.’s un-learning problem: Researchers say it’s virtually impossible to make an A.I. model ‘forget’ the things it learns from private user data

As it turns out, it’s impossible to remove a user’s data from a trained A.I. model. Deleting the model entirely is also difficult—and there’s little regulation to enforce either option.

A.I.’s un-learning problem: Researchers say it’s virtually impossible to make an A.I. model ‘forget’ the things it learns from private user data

I'm rather curious to see how the EU's privacy laws are going to handle this.

(Original article is from Fortune, but Yahoo Finance doesn't have a paywall)

191

You're viewing part of a thread.

Show Context
191 comments
  • I said:

    No, you knowing your old phone number is closer to how a database knows things than how LLMs know things.

    Which is true. Human memory is more like a database than an LLM's "memory." You have knowledge in your brain which you can consult. There is data in a database that it can consult. While memory is not a database, in this sense they are similar. They both exist and contain information in some way that can be acted upon.

    LLMs do not have any database, no memories, and contain no knowledge. They are fundamentally different from how humans know anything, and it's pretty accurate to say LLMs "know" nothing at all.

    • Leaving aside LLMs, the brain is not a database. there is no specific place that you can point to and say 'there resides the word for orange'. Assuming that would be the case, it would be highly inefficient to assign a spot somewhere for each bit of information (again, not talking about software here, still the brain). And if you would, then you would be able to isolate that place, cut it out, and actually induce somebody to forget the word and the notion (since we link words with meaning - say orange and you think of the fruit, colour or perhaps a carrot). If we hade a database organized into tables and say orange was a member of colours and another table, 'orange things', deleting the member 'orange' would make you not recognize that carrots nowadays are orange.

      Instead, what happens - for example in those who have a stroke or those who suffer from epilepsy (a misfiring of meurons) - is that there appears a tip-of-the tongue phenomenon where they know what they want to say and can recognize notions, it's just the pathway to that specific word is interrupted and causes a miss, presumably when the brain tries to go on the path it knows it should take because it's the path taken many times for that specific notion and is prevented. But they don't lose the ability to say their phone number, they might lose the ability to say 'four' and some just resort to describing the notion - say the fruit that makes breakfast juice instead. Of course, if the damage done is high enough to wipe out a large amout of neurons, you lose larger amounts of words.

      Downsides - you cannot learn stuff instantly, as you could if the brain was a database. That's why practice makes perfect. You remember your childhood phone number because you repeated it so many times that there is a strong enough link between some neurons.

      Upsides - there is more learning capacity if you just relate notions and words versus, for lack of a better term, hardcoding them. Again, not talking about software here.

      Also leads to some funky things like a pencil sharpener being called literally a pencil eater in Danish.

191 comments