this post was submitted on 10 Apr 2024
1298 points (99.0% liked)
Programmer Humor
19563 readers
1562 users here now
Welcome to Programmer Humor!
This is a place where you can post jokes, memes, humor, etc. related to programming!
For sharing awful code theres also Programming Horror.
Rules
- Keep content in english
- No advertisements
- Posts must be related to programming or programmer topics
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
LLMs are just very complex and intricate mirrors of ourselves because they use our past ramblings to pull from for the best responses to a prompt. They only feel like they are intelligent because we can't see the inner workings like the IF/THEN statements of ELIZA, and yet many people still were convinced that was talking to them. Humans are wired to anthropomorphize, often to a fault.
I say that while also believing we may yet develop actual AGI of some sort, which will probably use LLMs as a database to pull from. And what is concerning is that even though LLMs are not "thinking" themselves, how we've dived head first ignoring the dangers of misuse and many flaws they have is telling on how we'll ignore avoiding problems in AI development, such as the misalignment problem that is basically been shelved by AI companies replaced by profits and being first.
HAL from 2001/2010 was a great lesson - it's not the AI...the humans were the monsters all along.
I don't necessarily disagree that we may figure out AGI, and even that LLM research may help us get there, but frankly, I don't think an LLM will actually be any part of an AGI system.
Because fundamentally it doesn't understand the words it's writing. The more I play with and learn about it, the more it feels like a glorified autocomplete/autocorrect. I suspect issues like hallucination and "Waluigis" or "jailbreaks" are fundamental issues for a language model trying to complete a story, compared to an actual intelligence with a purpose.