@sunychoudhary
Building AI Security for LLMs | CEO @ LangProtect
Writing, speaking, and collaborating on AI security, LLM safety, and developer tooling.
That’s exactly where most programs break. Even when infra is secure by design, the actual risk shows up in everyday workflows. People bypass steps, share data differently, or use tools in ways policies never accounted for. Feels like the gap now is less about defining controls and more about seeing how they behave in real usage.
Good comparison. Feels like MCP vs CLI isn’t just about tooling, it’s about where you want control to sit. MCP gives structure and discoverability, which helps as systems grow. CLI keeps things simple and predictable, which is easier to debug. Most real setups probably end up hybrid. MCP for coordination, CLI for execution.
Good framing. Security assessments only work when they look beyond tooling. A company can have strong products in place and still fail because access reviews are weak, processes are unclear, or people don’t know what “secure behavior” looks like in daily work. The people, process, technology split is still one of the most practical ways to find real gaps.
The dip doesn’t surprise me. As models get tuned for safety, latency, and cost, you often see tradeoffs in code quality and determinism. It’s not just “better model every time,” it’s shifting priorities. For dev workflows, consistency matters more than peak output. A slightly weaker but predictable model is often easier to work with than a stronger but inconsistent one.
The rollback idea is solid. Feels like that’s one of the missing pieces right now, because most agent systems assume forward progress, not safe recovery. Only thing I’ve noticed is rollback works great for state, but gets tricky once actions have external side effects, APIs, data writes, notifications, etc. At that point you can revert the system, but not always the impact. Still, having a clean way to step back is a big shift compared to most setups today.