From the engine room.

Case studies, benchmarks, and observations from building the agent output pipeline.

Git merges text, not logic.
Every platform shift creates a processing layer in the middle. Network traffic got firewalls. Log data got observability pipelines. API calls got gateways. Agent-generated code doesn't have its layer yet. We're building it.
Read →
Your code is never stored.
The trust model behind the agent output pipeline. How Rosentic handles your code today and how it will handle it tomorrow - across the GitHub Action, the future GitHub App, and VPC deployments.
Read →
18 conflicts, 5 branches, 0.97 seconds.
We ran 5 simulated AI agents on the same codebase - Cursor on Python, Copilot on Go, Codex on TypeScript, Claude Code on cross-language, Windsurf on Ruby. Git merged everything cleanly. Tests passed. Here's what Rosentic found.
Read →
What Alibaba's SWE-CI tells us about the next 12 months.
Alibaba tested AI coding agents on 100 real codebases spanning 233 days. 75% of models broke previously working code during maintenance. The implications for production engineering teams are significant.
Read →

More posts on the way.

We're scanning public repos, running benchmarks, and writing about what we find. Subscribe below to get notified.

Stay in the loop.

New posts on agent infrastructure, case studies, and benchmarks.

You're in.