❤️ İstanbul ❤️
- 5 Posts
- 3 Comments
Joined 1 month ago
Cake day: June 17th, 2025
You are not logged in. If you use a Fediverse account that is able to follow users, you can follow this user.
herseycokguzelolacak@lemmy.mlto Selfhosted@lemmy.world•What is a self-hosted small LLM actually good for (<= 3B)English1·1 month agoNot on top of my head, but there must be something. llama.cpp and vllm have basically solved the inference problem for LLMs. What you need is a RAG solution on top that also combines it with web search.
herseycokguzelolacak@lemmy.mlto Selfhosted@lemmy.world•What is a self-hosted small LLM actually good for (<= 3B)English51·1 month agofor coding tasks you need web search and RAG. It’s not the size of the model that matters, since even the largest models find solutions online.
I don’t recommend buying a discrete GPU. For home-level hardware iGPU systems like Strix Halo are much better than discrete systems. For example Strix Halo can run models that you need multiple 5090s for.