Discussion about this post

User's avatar
Kate Johnson's avatar

The part about the "Death of Open Source" licenses really hits home when you're trying to scale a microservice that relies on heavy image processing. We’ve been debating internally whether to self-host a background removal model or just stick with an API, and the maintenance overhead for a custom GPU cluster is usually what kills the "free" aspect of open source. I’ve been using https://bestphotos.ai/background-remover for some of our production testing lately specifically because the latency is lower than what we were getting with our own self-hosted Torch instances. It’s a classic architectural trade-off: do you want full control over the source code, or do you actually want a system that scales without a dedicated DevOps team babysitting it 24/7? Curious if anyone here has found a middle ground that doesn't involve the restrictive AGPL licenses mentioned in the piece.

Pawel Jozefiak's avatar

Interesting how this differs from what I saw at the Mistral hackathon last week - their architecture pitch was heavy on MoE efficiency gains, but the actual feel in practice lagged behind the spec sheet.

The gap between architectural novelty and real model behavior is wider than benchmark scores suggest. DeepSeek training for $5.5M while Kimi K2 activates 32B params per token from 1T total - these numbers tell a compelling story about diverging bets.

I wrote about running Mistral head-to-head against other models at the EU hackathon (https://thoughts.jock.pl/p/mistral-ai-honest-review-eu-hackathon-2026) - the architectural gap shows up exactly where the paper predicts it would.

No posts

Ready for more?