this post was submitted on 25 Feb 2024
94 points (98.0% liked)
Technology
59373 readers
3125 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
The SALAMI situation is so bad.
Problem: Our training data is super racist, so it always generates white people!
Solution: Modify the prompts so that when a user asks for "a picture of a man" 10% of the time it is changed to "a picture of a BLACK man".
New problem: When the user says "A picture of a Nazi" 10% of the time our fix interprets that as "A picture of a BLACK Nazi"
So it's a glorified chat database.
The input to an LLM is effectively a huge quantity of text including chats. What the generative LLM does is nothing more than fancy auto-complete, finding the next word, then the next word, then the next word...