this post was submitted on 12 Feb 2024
1025 points (93.0% liked)

linuxmemes

21219 readers
147 users here now

Hint: :q!


Sister communities:


Community rules (click to expand)

1. Follow the site-wide rules

2. Be civil
  • Understand the difference between a joke and an insult.
  • Do not harrass or attack members of the community for any reason.
  • Leave remarks of "peasantry" to the PCMR community. If you dislike an OS/service/application, attack the thing you dislike, not the individuals who use it. Some people may not have a choice.
  • Bigotry will not be tolerated.
  • These rules are somewhat loosened when the subject is a public figure. Still, do not attack their person or incite harrassment.
  • 3. Post Linux-related content
  • Including Unix and BSD.
  • Non-Linux content is acceptable as long as it makes a reference to Linux. For example, the poorly made mockery of sudo in Windows.
  • No porn. Even if you watch it on a Linux machine.
  • 4. No recent reposts
  • Everybody uses Arch btw, can't quit Vim, and wants to interject for a moment. You can stop now.

  • Please report posts and comments that break these rules!

    founded 1 year ago
    MODERATORS
     
    you are viewing a single comment's thread
    view the rest of the comments
    [–] Petter1@lemm.ee 3 points 9 months ago (1 children)

    Learn with the help of LMMs (AI chatbots), it’s awesome, just let it generate some code, read it, understand it, and try make the code better, more beautiful and/or more efficient. Add some feature you miss in the code, don’t hesitate to ask your LMMs follow up question, it won’t laugh at stupid questions, it is just great.

    [–] DarkenLM@kbin.social 11 points 9 months ago (1 children)

    However, do keep in mind that LLMs regularly pull language an library features out of their asses that have no direct correspondent in practice. I'd use the LLMs to generate small snippets of code, giving them a small and restricted set of requirements to minimize hallucinations.

    [–] Petter1@lemm.ee 2 points 9 months ago

    Yea, encountered that as well (depending on LLM model). Mostly, it is enough to just feed the exception output back into the LLM thread and it will Fix it’s bugs, or at least can tell you why this exception normally occurs.