@evan both of them use so much memory that I have to shut down a bunch of Firefox tabs and VS code windows first - plus they're noticeably slower than the best hosted models
@simon@evan I also only use local models on flights really. However, I will say I've found the smaller Qwen2.5-coder models on Ollama to be good enough, and importantly fast enough for most simple boiler plate work without taking up all my ram to run them.
@evan both of them use so much memory that I have to shut down a bunch of Firefox tabs and VS code windows first - plus they're noticeably slower than the best hosted models