You can pick something that fits your GPU size. Works well on apple silicon too. My fav’s now are qwen3 series. Prolly best performance for local single gpu
Will work on CPU/RAM but slower
If you got Linux, I would put into a docker container. Might too much for the first try. There easier options I think.
I use oobabooga, little bit more options in the gguf space then ollama but not as easy to use imo. Does support openAI api connection though so can plug in other services to use it.
Normies get AI slop, prosumer uses local llm…
Not sure about social media… Normie is allergic to reading anything beyond daddy’s propaganda slop. If it ain’t rage bait, he ain’t got time for it
Home grown slop is still slop. The lying machine can’t make anything else.
At least my idiocy ain’t training the enemy.
Also, AI ain’t there to be correct. AI is there to help you get something done if you already know the outcome mostly.
It can really turbo charge a Linux experience for example.
Also local is way less censored and can be tweaked ;)
What LLM you using?
https://ollama.org/
You can pick something that fits your GPU size. Works well on apple silicon too. My fav’s now are qwen3 series. Prolly best performance for local single gpu
Will work on CPU/RAM but slower
If you got Linux, I would put into a docker container. Might too much for the first try. There easier options I think.
I use oobabooga, little bit more options in the gguf space then ollama but not as easy to use imo. Does support openAI api connection though so can plug in other services to use it.