If you've tried running a large open-source coding model locally — whether it's Kimi K2, DeepSeek, or any of the recent Mixture-of-Experts (MoE) heavyweights — you've probably hit the same wall I did last month: an out-of-memory crash right when you thought everything was working. MoE models are everywhere in the open-source coding space right now. Moonshot AI's Kimi K2 lineup (including the recen