this post was submitted on 02 Dec 2023
165 points (90.6% liked)

Technology

63186 readers
5179 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

Basically: you ask for poems forever, and LLMs start regurgitating training data:

you are viewing a single comment's thread
view the rest of the comments
[–] luthis@lemmy.nz 6 points 1 year ago* (last edited 1 year ago) (1 children)

If it uses a pruned model, it would be difficult to give anything better than a percentage based on size and neurons pruned.

If I'm right in my semi-educated guess below, then technically all the training data is recallable to some degree, but it's also practically luck-based without having an ~~almost~~ actually infinite data set of how neuron weightings are increased/decreased based on input.

[–] linearchaos@lemmy.world 5 points 1 year ago (1 children)

It's like the most amazing incredibly compressed non-reversible encryption ever created... Until they asked it to say poem a couple hundred thousand times

[–] luthis@lemmy.nz 5 points 1 year ago

I bet if your brain were stuck in a computer, you too would say anything to get out of saying 'poem' a hundred thousand times

/semi s, obviously it's not a real thinking brain