this post was submitted on 19 Oct 2023
499 points (96.8% liked)

Technology

58150 readers
4331 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Black Mirror creator unafraid of AI because it’s “boring”::Charlie Brooker doesn’t think AI is taking his job any time soon because it only produces trash

you are viewing a single comment's thread
view the rest of the comments
[–] Telodzrum@lemmy.world 48 points 11 months ago (6 children)

Nah, nah to all of it. LLM is a parlor trick and not a very good one. If we are ever able to make a general artificial intelligence, that's an entirely different story. But text prediction on steroids doesn't move the needle.

[–] fsmacolyte@lemmy.world 15 points 11 months ago* (last edited 11 months ago) (4 children)

The best ones can literally write pretty good code, and explain any concept on the Internet to you that you ask them to. If you don't understand a specific thing about their explanation, they can add onto their explanation, and they can respond in the style you want (explain as if I'm ten, explain as if I'm an undergrad, etc).

I use it literally every day for work in a somewhat niche field. I don't really agree that it's a "parlor trick".

[–] state_electrician@discuss.tchncs.de 21 points 11 months ago (1 children)

LLMs are awful for facts, because they don't understand what facts are. You should never rely on them if you require factual correctness.

They are OK for text summation, formatting and just making shit up. For summation a human with experience still produces nicer output, because they understand the content and don't just look at words. As for making shit up you will get the statistically most likely output, so it's usually trite and boring. I think the progress is amazing, but there are still so many problems to be solved.

Right now I use them for boiler plate stuff, like writing a text with some parameters and then I polish it. For code I find them quite useless, because with an IDE I can write boiler plate just as fast as when I polish the prompts until the LLM delivers useful stuff. And with the IDE I don't get references to methods or entire libraries that just don't exist.

[–] darth_helmet@sh.itjust.works 1 points 11 months ago

People are also kind of shit at facts. There are so many facts, and many of them aren’t practical for every person who needs to assess a fact’s accuracy to do so. But it isn’t structurally impossible to mimic how humans learn how to gauge truthfulness, we just have to be prepared for the idea that it will be bound by the limitations of language, as well as the risk inherent in trusting data that it has not independently verified.

[–] sudoreboot@slrpnk.net 9 points 11 months ago (1 children)

I use LLMs for having things explained to me, too.. but if you want to know how much salt to pour in that soup, try asking it about something niche and complicated you already know the answer to.

They can be useful in figuring out the correct terminology so that you can find the answer on your own, or for pointing some very very obvious mistakes in your understandings (but it will still miss most of them).

Please don't use those things as answer machines.

[–] fsmacolyte@lemmy.world 1 points 11 months ago (1 children)

I'm going to use those things as answer machines and you can't stop me.

Jokes aside, I always validate what chatbots tell me, not even just important things. I use GPT-4 for work and 90% of the time it can show me how to use very specific functions in complex ways, but yesterday (for the first time in awhile) it made up a function that didn't exist. To its credit, I said, "Are you sure about [function]?" and it said, "I'm sorry, I got confused. That function doesn't exist. However, look into X, Y, Z for further resources" and I did and they were the correct things to look into.

[–] sudoreboot@slrpnk.net 2 points 11 months ago* (last edited 11 months ago)

If you press it the same way again ("are you sure the function doesn't exist?"), there is a high chance it will "rectify" its rectification.

[–] rambaroo@lemmy.world 0 points 11 months ago (1 children)

It's just regurgitating info that people already know.

[–] Honytawk@lemmy.zip 1 points 11 months ago

In a small convenient package, which speeds up the process.

[–] Telodzrum@lemmy.world -5 points 11 months ago (1 children)

No they can't. Your phrasing is misleading. It's a Chinese Room test output and nothing more. I had an Encarta CD that could do rudimentary version of this in 1995. That was more impressive, tbh.

[–] aubertlone@lemmy.world 4 points 11 months ago

If you're really comparing LLM's to your Encarta cd from 1995 and saying the Encarta CD was the superior experience...

I'm afraid there's not much left for us to discuss.... Our views are too far apart.

[–] GnuLinuxDude@lemmy.ml 14 points 11 months ago (1 children)

Sam Altman (Creator of the freakish retina scanning based Worldcoin) would agree, it seems. The current path for LLMs and GPT seems to be in something of a bind, because to seriously improve upon what it currently does it needs to do something different, not more of the same. And figuring out something different could be very hard. https://www.wired.com/story/openai-ceo-sam-altman-the-age-of-giant-ai-models-is-already-over/

At least that's what I understand of it.

[–] TheWiseAlaundo@lemmy.whynotdrs.org 12 points 11 months ago* (last edited 11 months ago) (2 children)

He's not saying "AI is done, there's nothing else to do, we've hit the limit", he's saying "bigger models don't necessarily yield better results like we had initially anticipated"

Sam recently went before congress and advocated for limiting model sizes as a means of regulation, because, at the time, he believed bigger would generally always mean better outputs. What we're seeing now is that if a model is too large it will have trouble producing truthful output, which is super important to us humans.

And honestly, I don't think anyone should be shocked by this. Our own human brains have different sections that control different aspects of our lives. Why would an AI brain be different?

[–] gregoryw3@lemmy.ml 2 points 11 months ago

Future of AI is definitely going towards Manager/Agent model. It allows for an AI to handle all the tasks without keeping it to one model or method. We’re already seeing this with ChatGPT using Mathematica for math questions. Soon we can see art AI using different models and methods based on text input.

[–] Browning@lemmings.world 1 points 11 months ago

I gather that this is partly because data sizes haven't been going up with model sizes. That is likely to change soon as synthetic data starts to overtake organic data in both quantity and quality.

[–] ChaoticNeutralCzech@feddit.de 8 points 11 months ago (1 children)

In humans, abstract thinking developed hand in hand with language. So despite their limitations, I think that at least early AGI will include an LLM in some way.

[–] IonAddis@lemmy.world 10 points 11 months ago (2 children)

I've been having a lot of vague thoughts about the unconscious bits of our brains and body, in regards to LLMs. The parts of our brains/neurons that started evolving back in simple animals as basically super primitive ways to process visual/audio/whatever input.

Our brains do a LOT of signal processing and filtering that never reaches conscious thought, that we can't even reach with our conscious thought if we tried, but which is necessary for our squishy body-things to take in input from our environment and turn it into something useful instead of drowning in a screeching eye-searing tangled mess of chaotic sensory input all the time.

LLMs strike me as that sort of low-level input processing, the pattern-recognition and filtering. I think true generalized AI would have to be built on pieces like this--probably a lot of them. Ways to pluck patterns out of complex but repeated input. Like, this stuff definitely isn't self-aware, but could eventually end up as some sort of processing library for something else far down the line.

Now might be a good time to pick up Peter Watts' sci-fi book Blindsight. He doesn't exactly write about AI in it, but he does write about a creature that responds to input but isn't exactly conscious like you or I.

[–] ChaoticNeutralCzech@feddit.de 2 points 11 months ago

some sort of processing library for something else far down the line

This is what I meant.

pick up Peter Watts' sci-fi book Blindsight

I just got the EPUB, thanks. Looking forward to reading it.

[–] I_Fart_Glitter@lemmy.world 1 points 11 months ago

FYI Blindsight is free for audible members.

[–] Clent@lemmy.world 3 points 11 months ago (2 children)

Parlor trick is a perfect description.

People don't get that these things aren't anymore intelligent than their smartphones predicting the next word. The main difference is instead of a couple words it has thousands to choose from.

Half of the trick is how it uses the prompt to decided what words to start with.

[–] Zeth0s@lemmy.world 9 points 11 months ago* (last edited 11 months ago) (1 children)

That is not how it works. Your smartphone has all the dictionary available, same as LLM. It is simply something very different. People super confidently discussing about AI on lemmy are the real hallucinating parrots

[–] lloram239@feddit.de 2 points 11 months ago* (last edited 11 months ago) (3 children)

Have you ever even bothered to play around with any of the LLMs or are you just parroting what you heard in badly written articles?

The fact that the LLM predicts the next word does in no way shape or form limits its intelligence. That's after all the same thing you do while writing your post.

These idiotic claims about AI not being intelligent really make me questions if humans are.

[–] Clent@lemmy.world 6 points 11 months ago

Yes. I've used them. I have used it beyond the point of it hallucinating.

I am also a software engineer and have deeper understanding of how these systems work than your average user.

The software community tends to approach these things with more caution than the general population. The media overblows the capabilities of these systems.

A more concrete example is autonomous vehicles which were promised for decades and even now with a form of them on the road, they are still closer to remote controlled vehicles than the intelligent self contained systems we have been promised.

The difference between predictive text on a smart phone and predictive text of an LLM is my smart phone is predicting what I am likely to type next based on things i have typed in the past, while the LLM is predicting what comes next based on a larger body of work from source pulled from all across the internet. The LLM is then tuned by humans. This tuning step is under reported.

The LLM is unable to determine the truth of its own output. I would argue that is a key to claiming intelligence but determining what intelligence means is itself a philosophical question up for debate.

[–] banneryear1868@lemmy.world 4 points 11 months ago

It's not really "intelligent" though, as in it's not thinking about what it's doing. What AI will do very well is reproduce jargon, and if it's jargon that we associate with intelligence then it appears intelligent. Academic papers for instance it can do a very convincing job because that format is so repetitive and jargon heavy.

You can do an experiment by asking it to produce a positive review of something niche and academic you're familiar with, then ask it to produce a negative review of the same subject. It will produce convincing dialogue for either scenario, but it does not know which is more true/accurate, and it will come across as a student writing about something they didn't do the reading for.

The "question if humans are [intelligent]" is the more relevant thing here. We're constantly expected to communicate with thoughtlessly reproduced jargon, and many of us can do this very well in a way that gives the impression of intelligent thought. The fact AI can do this, and that people are concerned about how intelligent it appears, is more a reflection on how derivative our notions of intelligence can be in these settings.

[–] rambaroo@lemmy.world -1 points 11 months ago (1 children)

The fact that you believe an LLM is "intelligent" tells me you have no clue how they work and your comments on the matter can be ignored.

[–] lloram239@feddit.de 1 points 11 months ago

Oh look, another parrot.

Still waiting for any of you to actually define "intelligent" in some way that ChatGPT fails at or are you just going to pull the old boring "human exceptionalism"-card out of the hat?

[–] Honytawk@lemmy.zip 1 points 11 months ago

LLM's are like an interface to allow computers to talk to humans.

They are a necessary step in order to create general AI, because a general AI that can't generate text wouldn't be able to convey what they learned.

[–] aubertlone@lemmy.world 0 points 11 months ago (1 children)

Nah, nah to your understanding of LLM's

No it's not true intelligence. Yes, it makes humans much faster at their work

It has really sped up my work, especially when coding in unfamiliar languages.

It's silly to compare it to a parlor trick or text prediction.