this post was submitted on 08 Jul 2023
305 points (96.9% liked)
Technology
59627 readers
2908 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
On that, what would people recommend for a locally hosted (I have a graphics card) chatgpt-like LLM that is open source and doesn't require a lot of other things to install.
(Just one CMD line installation! That is, if you have pip, pip3, python, pytorch, CUDA, conda, Jupiter note books, Microsoft visual studio, C++, a Linux partition, and docker. Other than that, it is just one line installation!)
I looked into this too and it’s pretty resource heavy. I actually had a really good conversation with Chatgpt about making a separate instance of itself locallly. It’s worth talking to it about that and some of the price options
Look into llama.cpp - it's a single C++ program that run quantified models (basically models with some less precision - don't need a full 64 bits for a double, really). As for models to run on it, there's so many but I think WizardLM is pretty good.