this post was submitted on 26 Jul 2023
483 points (96.0% liked)

Technology

59590 readers
5616 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Thousands of authors demand payment from AI companies for use of copyrighted works::Thousands of published authors are requesting payment from tech companies for the use of their copyrighted works in training artificial intelligence tools, marking the latest intellectual property critique to target AI development.

you are viewing a single comment's thread
view the rest of the comments
[–] Buttons@programming.dev 0 points 1 year ago (3 children)

I think you underestimate the reasoning power of these AIs. They can write code, they can teach math, they can even learn math.

I've been using GPT4 as a math tutor while learning linear algebra, and I also use a text book. The text book told me that (to write it out) "the column space of matrix A is equal to the column space of matrix A times its own transpose". So I asked GPT4 if that was true and it said no, GPT disagreed with the text book. This was apparently something that GPT did not memorize and it was not just regurgitating sentences. I told GPT I saw it in a text book, the AI said "sorry, the textbook must be wrong". I then explained the mathematical proof to the AI, and the AI apologized, admitted it had been wrong, and agreed with the proof. Only after hearing the proof did the AI agree with the text book. This is some pretty advanced reasoning.

I performed that experiment a few times and it played out mostly the same. I experimented with giving the AI a flawed proof (I purposely made mistakes in the mathematical proofs), and the AI would call out my mistakes and would not be convinced by faulty proofs.

A standard that judged this AI to have "no understanding of any concepts whatsoever", would also conclude the same thing if applied to most humans.

[–] unlimitedolm_sjw@sh.itjust.works 1 points 1 year ago* (last edited 1 year ago)

That doesn't prove that GPT is reasoning, its model predicts that those responses are the most likely given the messages your sending it. It''s read thousands of actual conversations with people stating something incorrect, then having it explained to them and them coming around and admitting they were wrong.

I've seen other similar cases where the AI is wrong about something, and when it's explained, it just doubles down. Because humans do that type of thing too, refusing to admit their wrong.

The way it's designed means that it cannot reason in the same way humans experience it. It can simulate a likely conversation someone would have if they could reason.

[–] foo@programming.dev -1 points 1 year ago (1 children)

They can write code and teach maths because it's read people doing the exact same stuff

[–] Buttons@programming.dev 1 points 1 year ago* (last edited 1 year ago)

Hey, that's the same reason I can write code and do maths!

I'm serious, the only reason I know how to code or do math is because I learned from other people, mostly by reading. It's the only reason I can do those things.

[–] Telodzrum@lemmy.world -1 points 1 year ago

It's just a really big autocomplete system. It has no thought, no reason, no sense of self or anything, really.