this post was submitted on 11 Oct 2024
46 points (92.6% liked)

Technology

34891 readers
118 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS
top 14 comments
sorted by: hot top controversial new old
[–] JackGreenEarth@lemm.ee 32 points 1 month ago (3 children)

It's only a problem if you expect them to do formal reasoning. They are fancy word predictors, useful for when your output doesn't need to be factually accurate. If you use them for things they're not designed for, you'll get bad results, but that would be your fault for using them in an incorrect manner, not the LLMs' faults. You don't use a screwdriver to bang in a nail and say the screwdriver 'has a HUGE problem' when it does a bad job.

[–] Hazzard@lemm.ee 15 points 1 month ago

I think it is a problem. Maybe not for people like us, that understand the concept and its limitations, but "formal reasoning" is exactly how this technology is being pitched to the masses. "Take a picture of your homework and OpenAI will solve it", "have it reply to your emails", "have it write code for you". All reasoning-heavy tasks.

On top of that, Google/Bing have it answering user questions directly, it's commonly pitched as a "tutor", or an "assistant", the OpenAI API is being shoved everywhere under the sun for anything you can imagine for all kinds of tasks, and nobody is attempting to clarify it's weaknesses in their marketing.

As it becomes more and more common, more and more users who don't understand it's fundamentally incapable of reliably doing these things will crop up.

[–] geekwithsoul@lemm.ee 11 points 1 month ago* (last edited 1 month ago)

The problem is the laymen expect it to do reasoning, so the sales & marketing team says that it can do reasoning, and then the CEO will have consumed the Kool-Aid and restructure the company because he believes it can do reasoning.

[–] yogthos@lemmy.ml 4 points 1 month ago (1 children)

Right, I find LLMs are fundamentally no different from Markov chains. It doesn't mean they're not useful, they're a tool that's good for certain use cases. Unfortunately, we're in a hype phase right now where people are trying to apply them for a lot of cases they're terrible at and where better tools already exist to boot.

[–] vrighter@discuss.tchncs.de 2 points 1 month ago (1 children)

they aren't. The only difference is that the state transition table is so unimaginably gargantuan thit we can only generate an approximation of a tiny slice of it, instead of it being literally a table

[–] yogthos@lemmy.ml 1 points 1 month ago
[–] slacktoid@lemmy.ml 7 points 1 month ago

This knife is a bad hammer.. I wonder why?

[–] Letstakealook@lemm.ee 6 points 1 month ago

And yet people will continue to argue that llms are demonstrating understanding and problem solving. This shit is just Eliza on steroids. I'm not saying it didn't require skill or knowledge to create, but it is in no way close to what it is being billed as.