Now that local LLMs are gaining traction, I’m wondering what the equivalent stack looks like today.
Models, Runtime, hardware and other tools.
That could rival the Claudes, ChatGPTs or Geminis, etc
Thanks
Now that local LLMs are gaining traction, I’m wondering what the equivalent stack looks like today.
Models, Runtime, hardware and other tools.
That could rival the Claudes, ChatGPTs or Geminis, etc
Thanks
7 comments
Tbh for coding I just use the smaller ones like CodeQwen 7B. way faster and good enough for autocomplete. Only fire up the big model when I actually need it to think.
The annoying part is keeping everything updated, new model drops every week and half don't work with whatever you're already running.
The models vary depending on the task. DeepSeek distilled has been a favorite for the past several months.
I use various smaller (~3B) models for simpler tasks.