this post was submitted on 05 Feb 2024
62 points (81.6% liked)

Technology

59389 readers
3610 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Sam Altman says ChatGPT should be 'much less lazy now'::ChatGPT users previously complained that the chatbot was slacking off and refusing to complete some tasks.

you are viewing a single comment's thread
view the rest of the comments
[–] otp@sh.itjust.works 33 points 9 months ago (1 children)

Some users found inventive strategies to get around ChatGPT's laziness, with one finding that the AI model would provide longer responses if they promised to tip it $200.

Man this thing really IS just like a human!

/joke

[–] fidodo@lemmy.world 3 points 9 months ago (1 children)

It's based on text produced by humans so yes, it does retrieve text that was written by humans therefore it acts like a human.

[–] General_Effort@lemmy.world 2 points 9 months ago

It's still weird. That reasoning implies that there is a correlation between promising money and long answers in the training data. Seems plausible at first blush, but where can this be actually seen? It's hardly ever seen in social media, where similar QA formats exists. It's certainly not in textbooks, where the real good answers are. OTOH there are a lot of tips promised in completely different contexts.

I'm not saying it's wrong, but there is definitely a lot of cargo cult in prompting strategies.