this post was submitted on 27 Jan 2024
0 points (50.0% liked)
Arch Linux
7759 readers
2 users here now
The beloved lightweight distro
founded 4 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
People expect a language model to be really good at other things besides language.
If you’re writing an email where you need to express a particular thought or a feeling, ask some LLM what would be a good way to say it. Even though the suggestions are pretty useful, they may still require some editing.
This use case and asking for information are completely different things. It can stylize some input perfectly fine. It just can't be a source of accurate information.It is trained to generate text that sounds plausible.
There are already ways to get around that, even though they aren't perfect. You can give the source of truth and ask it to answer using only information found in there. Even then, you should check. the accuracy of its responses.