From the engine room.
Case studies, benchmarks, and observations from building the agent output pipeline.
March 31, 2026
Founding Essay
Pinned
Git merges text, not logic.
Every platform shift creates a processing layer in the middle. Network traffic got firewalls. Log data got observability pipelines. API calls got gateways. Agent-generated code doesn't have its layer yet. We're building it.
Read →
March 24, 2026
Security
Your code is never stored.
The trust model behind the agent output pipeline. How Rosentic handles your code today and how it will handle it tomorrow - across the GitHub Action, the future GitHub App, and VPC deployments.
Read →
March 11, 2026
Benchmark
18 conflicts, 5 branches, 0.97 seconds.
We ran 5 simulated AI agents on the same codebase - Cursor on Python, Copilot on Go, Codex on TypeScript, Claude Code on cross-language, Windsurf on Ruby. Git merged everything cleanly. Tests passed. Here's what Rosentic found.
Read →
March 11, 2026
Case Study
What Alibaba's SWE-CI tells us about the next 12 months.
Alibaba tested AI coding agents on 100 real codebases spanning 233 days. 75% of models broke previously working code during maintenance. The implications for production engineering teams are significant.
Read →
More posts on the way.
We're scanning public repos, running benchmarks, and writing about what we find. Subscribe below to get notified.