this post was submitted on 08 Jul 2023
343 points (97.8% liked)

Technology

58094 readers
3176 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Slow June, people voting with their feet amid this AI craze, or something else?

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 4 points 1 year ago (2 children)

On that, what would people recommend for a locally hosted (I have a graphics card) chatgpt-like LLM that is open source and doesn't require a lot of other things to install.

(Just one CMD line installation! That is, if you have pip, pip3, python, pytorch, CUDA, conda, Jupiter note books, Microsoft visual studio, C++, a Linux partition, and docker. Other than that, it is just one line installation!)

[–] [email protected] 1 points 1 year ago

Look into llama.cpp - it's a single C++ program that run quantified models (basically models with some less precision - don't need a full 64 bits for a double, really). As for models to run on it, there's so many but I think WizardLM is pretty good.

[–] [email protected] 1 points 1 year ago

I looked into this too and it’s pretty resource heavy. I actually had a really good conversation with Chatgpt about making a separate instance of itself locallly. It’s worth talking to it about that and some of the price options