Why I Removed My Local LLM and Went All-In on Cloud APIs
Why I Removed My Local LLM and Went All-In on Cloud APIs
TL;DR: I ran Ollama with Gemma models on my i5-8350U laptop for months. The dream of "free, private, offline AI" crashed into reality: 2 tokens/second inference, mediocre output quality, and a ...
notes.masud.pro6 min read