I've been playing with #LMStudio for #localLLM with mediocre results. However #gemma4 really changed that. It's faster and is more capable then the other models I could try on my hardware. It has recent data and is able to use a fetch tool(among others) to get info on stuff it doesn't know! So I installed #ollama and now it runs even faster, to the point where delay(waiting) is not that noticeable. Since I am a lightweight user, I can see myself using it as mainly source.