<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Platform Engineering on Alfero Chingono</title><link>https://www.chingono.com/tags/platform-engineering/</link><description>Recent content in Platform Engineering on Alfero Chingono</description><generator>Hugo -- gohugo.io</generator><language>en-us</language><lastBuildDate>Fri, 03 Apr 2026 20:02:39 -0400</lastBuildDate><atom:link href="https://www.chingono.com/tags/platform-engineering/index.xml" rel="self" type="application/rss+xml"/><item><title>The Silent Excellence Trap: Why Good Work Gets Invisible in Big Organizations</title><link>https://www.chingono.com/blog/2026/03/25/the-silent-excellence-trap-why-good-work-gets-invisible-in-big-organizations/</link><pubDate>Wed, 25 Mar 2026 09:00:00 +0000</pubDate><guid>https://www.chingono.com/blog/2026/03/25/the-silent-excellence-trap-why-good-work-gets-invisible-in-big-organizations/</guid><description>&lt;p&gt;There is a specific kind of frustration that shows up in large organizations when you are doing good work that makes other work possible.&lt;/p&gt;
&lt;p&gt;You reduce risk.
You simplify delivery.
You create better defaults.
You remove ambiguity.&lt;/p&gt;
&lt;p&gt;And at the end of the quarter, it can still look like nothing happened.&lt;/p&gt;
&lt;p&gt;I have come to think of that as the &lt;strong&gt;silent excellence trap&lt;/strong&gt;.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Outcomes matter, but outcomes that are not made legible, timely, and attributable are often discounted.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;That sentence has stayed with me because it explains something I have seen repeatedly, both in myself and in other high-conscientious operators. We assume that if the work is real, the value will be obvious. In many organizations, that is simply not how visibility works.&lt;/p&gt;
&lt;h2 id="the-problem-is-usually-structural-not-personal"&gt;The problem is usually structural, not personal
&lt;/h2&gt;&lt;p&gt;Most big organizations do not formally say they reward performative busyness over outcomes. In fact, they usually say the opposite.&lt;/p&gt;
&lt;p&gt;The disconnect shows up in the operating reality.&lt;/p&gt;
&lt;p&gt;I see three reasons for that.&lt;/p&gt;
&lt;h3 id="1-long-feedback-loops"&gt;1. Long feedback loops
&lt;/h3&gt;&lt;p&gt;Some of the most valuable work in platform engineering, architecture, governance, and operational improvement only becomes visible months later.&lt;/p&gt;
&lt;p&gt;If you reduce configuration drift, improve rollout safety, or standardize an architectural decision, the value often appears as &lt;strong&gt;fewer future problems&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;That is good for the business.
It is not always good for recognition.&lt;/p&gt;
&lt;p&gt;When outcomes take a long time to surface, visible activity starts to become a proxy for progress.&lt;/p&gt;
&lt;h3 id="2-narrative-driven-evaluation"&gt;2. Narrative-driven evaluation
&lt;/h3&gt;&lt;p&gt;A lot of performance assessment is not just about what happened. It is about how what happened gets narrated.&lt;/p&gt;
&lt;p&gt;This creates an uncomfortable truth: two people can generate similar value, but the person who can explain their impact in organizational language is often rewarded more clearly.&lt;/p&gt;
&lt;p&gt;That is not always politics in the cynical sense. Sometimes it is just information design. Leaders are making decisions under ambiguity, and the work that is easiest to interpret tends to travel farther.&lt;/p&gt;
&lt;h3 id="3-cross-team-dependency-density"&gt;3. Cross-team dependency density
&lt;/h3&gt;&lt;p&gt;In big organizations, important work often moves through many hands.&lt;/p&gt;
&lt;p&gt;You create the framework.
Someone else adopts it.
Another team gets the outcome.
A downstream metric improves three months later.&lt;/p&gt;
&lt;p&gt;At that point, attribution is fuzzy. Visibility fills the gap.&lt;/p&gt;
&lt;p&gt;That is why quiet, enabling work is so easy to underrate.&lt;/p&gt;
&lt;h2 id="this-is-not-an-argument-for-theater"&gt;This is not an argument for theater
&lt;/h2&gt;&lt;p&gt;I do not think the answer is to become a full-time self-promoter.&lt;/p&gt;
&lt;p&gt;I also do not think the answer is to &amp;ldquo;look busy.&amp;rdquo;&lt;/p&gt;
&lt;p&gt;The better answer is &lt;strong&gt;intentional visibility&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;That means making outcomes easier to see without fabricating activity.&lt;/p&gt;
&lt;p&gt;There is a huge ethical and psychological difference between:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;performing importance&lt;/li&gt;
&lt;li&gt;and making real impact legible&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;I care about the second one.&lt;/p&gt;
&lt;h2 id="what-i-do-instead-now"&gt;What I do instead now
&lt;/h2&gt;&lt;p&gt;Over time, I have found a few habits that help.&lt;/p&gt;
&lt;h3 id="1-express-work-in-a-simple-chain"&gt;1. Express work in a simple chain
&lt;/h3&gt;&lt;p&gt;I try to describe meaningful work as:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Problem -&amp;gt; Action -&amp;gt; Outcome -&amp;gt; Why it matters&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;That structure sounds basic, but it keeps me from describing effort when I should be describing value.&lt;/p&gt;
&lt;h3 id="2-translate-into-enterprise-outcomes"&gt;2. Translate into enterprise outcomes
&lt;/h3&gt;&lt;p&gt;Enabling work often needs a second layer of interpretation.&lt;/p&gt;
&lt;p&gt;It is not enough to say:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;built a feature-flag framework&lt;/li&gt;
&lt;li&gt;standardized observability&lt;/li&gt;
&lt;li&gt;created reusable pipeline templates&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;You often have to go one step further:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;reduced deployment risk&lt;/li&gt;
&lt;li&gt;removed support fragmentation&lt;/li&gt;
&lt;li&gt;shortened decision cycles&lt;/li&gt;
&lt;li&gt;improved auditability&lt;/li&gt;
&lt;li&gt;lowered operational variance&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This is not spin. It is completion.&lt;/p&gt;
&lt;h3 id="3-publish-checkpoints-not-just-final-wins"&gt;3. Publish checkpoints, not just final wins
&lt;/h3&gt;&lt;p&gt;Long-cycle work disappears when nobody can see intermediate progress.&lt;/p&gt;
&lt;p&gt;I have become much more deliberate about creating artifacts along the way:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;decision logs&lt;/li&gt;
&lt;li&gt;patterns&lt;/li&gt;
&lt;li&gt;working examples&lt;/li&gt;
&lt;li&gt;metrics snapshots&lt;/li&gt;
&lt;li&gt;short summaries of what changed&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;That keeps the work reviewable before it becomes historic.&lt;/p&gt;
&lt;h3 id="4-pre-wire-leaders-with-outcomes-not-activity"&gt;4. Pre-wire leaders with outcomes, not activity
&lt;/h3&gt;&lt;p&gt;I have learned not to wait for annual review season to explain what mattered.&lt;/p&gt;
&lt;p&gt;Short outcome-only updates travel better than heroic recaps. They also reduce the chance that your most valuable work gets remembered only when something breaks.&lt;/p&gt;
&lt;h2 id="the-uncomfortable-part"&gt;The uncomfortable part
&lt;/h2&gt;&lt;p&gt;The silent excellence trap is especially common among people who take pride in substance.&lt;/p&gt;
&lt;p&gt;We tell ourselves the work should speak for itself.&lt;/p&gt;
&lt;p&gt;Sometimes it does.
Often it does not.&lt;/p&gt;
&lt;p&gt;And in large systems, what does not become legible often does not become rewarded.&lt;/p&gt;
&lt;p&gt;That is not fair, but it is useful to understand.&lt;/p&gt;
&lt;h2 id="my-current-reframe"&gt;My current reframe
&lt;/h2&gt;&lt;p&gt;I no longer think of visibility as a political side quest. I think of it as part of delivery.&lt;/p&gt;
&lt;p&gt;If I create meaningful change but fail to make it legible, timely, and attributable, then part of the job is still unfinished.&lt;/p&gt;
&lt;p&gt;That has changed how I think about platform work, architecture work, and even writing in public. In some ways, building publicly is itself a response to this problem. The work becomes its own audit trail.&lt;/p&gt;
&lt;p&gt;That is also why this post connects so naturally to &lt;a class="link" href="https://www.chingono.com/blog/2025/04/10/the-dora-report-was-right-idps-improve-team-productivity-by-10-percent-heres-how-ive-seen-it/" &gt;The DORA Report Was Right: IDPs Improve Team Productivity by 10% — Here&amp;rsquo;s How I&amp;rsquo;ve Seen It&lt;/a&gt;. Platform work can produce real leverage. But if the leverage stays invisible, the organization often underestimates both the work and the people doing it.&lt;/p&gt;
&lt;p&gt;References:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class="link" href="https://www.chingono.com/blog/2025/04/10/the-dora-report-was-right-idps-improve-team-productivity-by-10-percent-heres-how-ive-seen-it/" &gt;The DORA Report Was Right: IDPs Improve Team Productivity by 10% — Here&amp;rsquo;s How I&amp;rsquo;ve Seen It&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class="link" href="https://www.chingono.com/blog/2025/02/15/why-i-started-building-my-own-devops-platform-and-what-i-learned/" &gt;Why I Started Building My Own DevOps Platform (And What I Learned)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;</description></item><item><title>Designing Multi-Agent Systems: Lessons from Building an 8-Agent Engineering Orchestra</title><link>https://www.chingono.com/blog/2025/08/28/designing-multi-agent-systems-lessons-from-building-an-8-agent-engineering-orchestra/</link><pubDate>Thu, 28 Aug 2025 09:00:00 +0000</pubDate><guid>https://www.chingono.com/blog/2025/08/28/designing-multi-agent-systems-lessons-from-building-an-8-agent-engineering-orchestra/</guid><description>&lt;p&gt;A lot of &amp;ldquo;multi-agent&amp;rdquo; demos are really one agent wearing different hats.&lt;/p&gt;
&lt;p&gt;The names change. The prompts change. Sometimes the avatars change. But the authority model, the memory model, and the execution model are all still basically the same. That is fine for a demo. It is much less convincing when you are trying to build a system that can do real engineering work.&lt;/p&gt;
&lt;p&gt;Building &lt;a class="link" href="https://github.com/cuemarshal/cuemarshal" target="_blank" rel="noopener"
&gt;CueMarshal&lt;/a&gt; made that distinction impossible for me to ignore.&lt;/p&gt;
&lt;p&gt;What I wanted was not eight personalities for marketing. I wanted a working system where planning, coding, review, testing, DevOps, documentation, and quality control could be separated cleanly enough to be trustworthy.&lt;/p&gt;
&lt;p&gt;That is how the &amp;ldquo;engineering orchestra&amp;rdquo; idea emerged.&lt;/p&gt;
&lt;h2 id="the-roles-mattered-because-the-boundaries-mattered"&gt;The roles mattered because the boundaries mattered
&lt;/h2&gt;&lt;p&gt;CueMarshal&amp;rsquo;s cast eventually became:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Marshal&lt;/strong&gt; for orchestration&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Ava&lt;/strong&gt; for architecture&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Dave&lt;/strong&gt; for implementation&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Reese&lt;/strong&gt; for review&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Tess&lt;/strong&gt; for testing&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Devin&lt;/strong&gt; for DevOps&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Dot&lt;/strong&gt; for documentation&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Linton&lt;/strong&gt; for linting&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;What made that useful was not the naming. It was the fact that the roles had &lt;strong&gt;different responsibilities, different tool access, and different default model tiers&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;That is the first lesson I would pass on to anyone designing a multi-agent system:&lt;/p&gt;
&lt;h2 id="1-different-roles-need-different-authority"&gt;1. Different roles need different authority
&lt;/h2&gt;&lt;p&gt;If your reviewer can rewrite production code, your reviewer is not really a reviewer.&lt;/p&gt;
&lt;p&gt;In CueMarshal, least privilege is deliberate. The reviewer is configured without write/edit permissions. The docs agent is restricted from shell access. The linter acts like a gate, not a developer with nicer manners.&lt;/p&gt;
&lt;p&gt;That kind of restriction sounds limiting until you realize it is what gives each role meaning. Boundaries are not friction here. They are the mechanism that creates trust.&lt;/p&gt;
&lt;p&gt;A good multi-agent system is not just a cluster of competencies. It is a set of constrained responsibilities.&lt;/p&gt;
&lt;h2 id="2-coordination-needs-durable-state-outside-the-model"&gt;2. Coordination needs durable state outside the model
&lt;/h2&gt;&lt;p&gt;One of the reasons I anchored CueMarshal in Git is that I did not want coordination to depend on hidden model memory.&lt;/p&gt;
&lt;p&gt;Tasks become issues.
Work becomes branches.
Proposals become pull requests.
Reviews become durable comments and approvals.&lt;/p&gt;
&lt;p&gt;The Conductor receives webhooks, uses Redis and BullMQ to manage asynchronous flow, and dispatches work through Gitea Actions. The runners themselves stay stateless; they rebuild context from the repository, the issue, and the tool layer every time.&lt;/p&gt;
&lt;p&gt;That has been a much better trade than magical continuity.&lt;/p&gt;
&lt;p&gt;Models forget.
Git does not.&lt;/p&gt;
&lt;h2 id="3-identity-is-part-of-the-architecture"&gt;3. Identity is part of the architecture
&lt;/h2&gt;&lt;p&gt;Another thing I underestimated early on was how important identity separation would be.&lt;/p&gt;
&lt;p&gt;Each CueMarshal agent has its own account, token, and audit trail. That means the Git history shows who planned, who implemented, who reviewed, and who approved. Even when the &amp;ldquo;who&amp;rdquo; is an AI agent, the distinction still matters.&lt;/p&gt;
&lt;p&gt;This has two benefits.&lt;/p&gt;
&lt;p&gt;First, it improves explainability. The system becomes easier to inspect when actions are attributable.&lt;/p&gt;
&lt;p&gt;Second, it changes how you think about safety. Once every agent has a clear identity and permission scope, you stop designing from a vague &amp;ldquo;assistant&amp;rdquo; mindset and start designing from explicit operational roles.&lt;/p&gt;
&lt;p&gt;That shift is subtle, but it is foundational.&lt;/p&gt;
&lt;h2 id="4-the-tool-layer-is-what-makes-the-orchestra-playable"&gt;4. The tool layer is what makes the orchestra playable
&lt;/h2&gt;&lt;p&gt;This is where MCP became important for CueMarshal.&lt;/p&gt;
&lt;p&gt;All of the agents connect to a structured tool layer instead of improvising raw integrations on the fly. The same Gitea, Conductor, and System capabilities can be used by the runner agents over stdio and by the orchestration layer over HTTP/SSE.&lt;/p&gt;
&lt;p&gt;That matters because multi-agent systems are not only about reasoning. They are about coordination through reliable interfaces.&lt;/p&gt;
&lt;p&gt;If the tools are vague, agents collide.
If the permissions are sloppy, trust collapses.
If the transports are inconsistent, reuse gets expensive.&lt;/p&gt;
&lt;p&gt;The protocol is not the whole story, but it is the difference between a collection of prompts and a real system surface.&lt;/p&gt;
&lt;p&gt;I wrote more about that in &lt;a class="link" href="https://www.chingono.com/blog/2025/03/20/mcp-in-practice-what-anthropics-model-context-protocol-actually-means-for-developers/" &gt;MCP in Practice&lt;/a&gt;, because it deserves its own treatment.&lt;/p&gt;
&lt;h2 id="5-model-routing-is-architecture-not-optimization"&gt;5. Model routing is architecture, not optimization
&lt;/h2&gt;&lt;p&gt;Another lesson I came away with: not every role deserves the same model.&lt;/p&gt;
&lt;p&gt;Architecture work is more expensive and more consequential than documentation cleanup. Review often needs stronger reasoning than linting. Mechanical work should not burn premium tokens if a cheaper tier can do it reliably.&lt;/p&gt;
&lt;p&gt;CueMarshal&amp;rsquo;s tiered routing reflects that reality:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;heavy reasoning for architecture&lt;/li&gt;
&lt;li&gt;balanced capability for implementation, review, testing, and DevOps&lt;/li&gt;
&lt;li&gt;lighter-weight models for docs and linting&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;That is not just a cost decision. It is part of how the system stays sustainable.&lt;/p&gt;
&lt;p&gt;Too many agent systems treat model choice as an afterthought. I think it belongs in the design doc.&lt;/p&gt;
&lt;h2 id="6-closed-loops-beat-hero-agents"&gt;6. Closed loops beat hero agents
&lt;/h2&gt;&lt;p&gt;The more I build these systems, the less I believe in the &amp;ldquo;super-agent&amp;rdquo; story.&lt;/p&gt;
&lt;p&gt;What works better is a closed loop:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;detect work&lt;/li&gt;
&lt;li&gt;route it clearly&lt;/li&gt;
&lt;li&gt;execute with constrained roles&lt;/li&gt;
&lt;li&gt;review it&lt;/li&gt;
&lt;li&gt;merge it with human control&lt;/li&gt;
&lt;li&gt;feed the next signal back into the system&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;CueMarshal&amp;rsquo;s self-improvement workflow made this even clearer to me. Once SonarQube findings, scanners, issues, PRs, and agent roles all started participating in the same loop, the system became more useful than any single agent inside it.&lt;/p&gt;
&lt;p&gt;That is why I think orchestration matters more than agent count.&lt;/p&gt;
&lt;h2 id="my-current-takeaway"&gt;My current takeaway
&lt;/h2&gt;&lt;p&gt;If you are building a multi-agent system, start with these questions:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;What roles genuinely need to be different?&lt;/li&gt;
&lt;li&gt;What permissions should each role have?&lt;/li&gt;
&lt;li&gt;Where does coordination state live?&lt;/li&gt;
&lt;li&gt;How are actions attributed?&lt;/li&gt;
&lt;li&gt;What is the closed loop that turns outputs into the next inputs?&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;If you cannot answer those, adding more agents will mostly add more noise.&lt;/p&gt;
&lt;p&gt;If you can answer them, the number of agents becomes much less important than the quality of the structure around them.&lt;/p&gt;
&lt;p&gt;That has been the real lesson for me. The point of the orchestra is not to have more instruments. The point is to make the handoffs musical instead of chaotic.&lt;/p&gt;
&lt;p&gt;If you want the adjacent pieces, &lt;a class="link" href="https://www.chingono.com/blog/2025/02/15/why-i-started-building-my-own-devops-platform-and-what-i-learned/" &gt;Why I Started Building My Own DevOps Platform&lt;/a&gt; covers the motivation, and &lt;a class="link" href="https://www.chingono.com/blog/2026/03/05/how-i-run-sonarqube-in-my-own-ci-pipeline-and-let-ai-fix-what-it-finds/" &gt;How I Run SonarQube in My Own CI Pipeline (And Let AI Fix What It Finds)&lt;/a&gt; shows what this architecture looks like when the feedback loop closes on itself.&lt;/p&gt;
&lt;p&gt;References:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class="link" href="https://github.com/cuemarshal/cuemarshal" target="_blank" rel="noopener"
&gt;CueMarshal repository&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class="link" href="https://github.com/cuemarshal/cuemarshal/blob/main/docs/architecture/overview.md" target="_blank" rel="noopener"
&gt;CueMarshal architecture overview&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class="link" href="https://github.com/cuemarshal/cuemarshal/blob/main/docs/features/agents/overview.md" target="_blank" rel="noopener"
&gt;CueMarshal agent profiles&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class="link" href="https://github.com/cuemarshal/cuemarshal/blob/main/docs/features/conductor/overview.md" target="_blank" rel="noopener"
&gt;CueMarshal conductor overview&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;</description></item><item><title>The DORA Report Was Right: IDPs Improve Team Productivity by 10% — Here's How I've Seen It</title><link>https://www.chingono.com/blog/2025/04/10/the-dora-report-was-right-idps-improve-team-productivity-by-10-percent-heres-how-ive-seen-it/</link><pubDate>Thu, 10 Apr 2025 09:00:00 +0000</pubDate><guid>https://www.chingono.com/blog/2025/04/10/the-dora-report-was-right-idps-improve-team-productivity-by-10-percent-heres-how-ive-seen-it/</guid><description>&lt;p&gt;When the DORA research started surfacing stronger evidence around internal developer platforms, the headline did not surprise me nearly as much as the reactions did.&lt;/p&gt;
&lt;p&gt;Some people still hear &amp;ldquo;platform engineering&amp;rdquo; and imagine more process, more gates, and another internal team inventing obstacles. That risk is real. But it is only one version of the story.&lt;/p&gt;
&lt;p&gt;The version I have seen in practice is much simpler: when you reduce cognitive load, standardize the boring parts well, and make the safe path the easy path, teams move faster.&lt;/p&gt;
&lt;p&gt;That is why the widely shared DORA finding about internal developer platforms improving team performance felt directionally right to me. I have seen that pattern from multiple angles: CI/CD modernization that improved project velocity, reusable delivery templates that removed duplication, feature-flag and configuration patterns that reduced rollout risk, and platform standards that made decision-making less expensive for product teams.&lt;/p&gt;
&lt;p&gt;I would not claim every platform effort automatically produces a neat percentage uplift. But I do believe the mechanism is real.&lt;/p&gt;
&lt;h2 id="what-platform-engineering-is-actually-buying-you"&gt;What platform engineering is actually buying you
&lt;/h2&gt;&lt;p&gt;At its best, an internal developer platform is not a control tower. It is a &lt;strong&gt;friction reducer&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;It helps teams spend less time answering questions like:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;How should we structure this pipeline?&lt;/li&gt;
&lt;li&gt;Which security checks are required?&lt;/li&gt;
&lt;li&gt;What is the approved deployment pattern?&lt;/li&gt;
&lt;li&gt;How do we manage runtime configuration safely?&lt;/li&gt;
&lt;li&gt;How do we release gradually without gambling in production?&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;If every team answers those questions from scratch, you get inconsistency, duplicated effort, and unnecessary risk. If the platform team answers them once, clearly, and with enough flexibility, you get leverage.&lt;/p&gt;
&lt;p&gt;That leverage is where the productivity gain comes from.&lt;/p&gt;
&lt;p&gt;Not from a portal.
Not from a dashboard.
Not from a maturity model.&lt;/p&gt;
&lt;p&gt;From fewer repeated decisions.&lt;/p&gt;
&lt;h2 id="where-i-have-seen-the-gains-show-up"&gt;Where I have seen the gains show up
&lt;/h2&gt;&lt;p&gt;One of the more durable lessons in my career is that developer productivity is usually downstream of environment design.&lt;/p&gt;
&lt;p&gt;At VCA Software, the measurable win was CI/CD modernization. We saw delivery speed improve because the path from code to release became more repeatable and less person-dependent. That did not happen because engineers suddenly became more talented. It happened because the delivery system stopped making them re-solve the same operational problems over and over.&lt;/p&gt;
&lt;p&gt;In more platform-oriented work, I have seen the same principle show up differently:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;a reusable feature-flag framework that makes progressive delivery safer&lt;/li&gt;
&lt;li&gt;standardized pipeline templates that reduce copy-paste infrastructure&lt;/li&gt;
&lt;li&gt;better observability defaults so teams are not blind after deployment&lt;/li&gt;
&lt;li&gt;configuration-management patterns that reduce drift and remove manual setup&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;That last point is one reason I still like the pattern I wrote about in &lt;a class="link" href="https://www.chingono.com/blog/2025/01/10/conditionally-deploying-resources-azure-app-configuration-using-deployment-scripts/" &gt;Conditionally Deploying Resources in Azure App Configuration Using Deployment Scripts&lt;/a&gt;. It is not glamorous, but it is exactly the kind of operational sharp edge a good platform should smooth out.&lt;/p&gt;
&lt;h2 id="the-dora-nuance-matters-too"&gt;The DORA nuance matters too
&lt;/h2&gt;&lt;p&gt;What I appreciate about the DORA research is that it does not treat platform engineering as universally positive in every implementation.&lt;/p&gt;
&lt;p&gt;That matches my experience.&lt;/p&gt;
&lt;p&gt;A platform helps when it gives teams &lt;strong&gt;self-service with sensible defaults&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;A platform hurts when it becomes:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;mandatory ceremony&lt;/li&gt;
&lt;li&gt;an opaque ticket queue&lt;/li&gt;
&lt;li&gt;a rigid abstraction over real team needs&lt;/li&gt;
&lt;li&gt;a place where local context goes to die&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This is where some platform efforts go sideways. They optimize for governance theater instead of developer flow. Then leaders conclude that platform engineering is slow, when the real problem is that the platform is not being run like a product.&lt;/p&gt;
&lt;h2 id="what-good-platform-teams-do-differently"&gt;What good platform teams do differently
&lt;/h2&gt;&lt;p&gt;The best platform work I have seen has a few traits in common.&lt;/p&gt;
&lt;h3 id="1-it-starts-with-repeated-pain-not-abstract-ambition"&gt;1. It starts with repeated pain, not abstract ambition
&lt;/h3&gt;&lt;p&gt;Good platform teams do not begin with &amp;ldquo;let&amp;rsquo;s build an internal developer portal.&amp;rdquo; They begin with &amp;ldquo;teams keep tripping over the same deployment, security, configuration, or release problems.&amp;rdquo;&lt;/p&gt;
&lt;p&gt;That difference matters because it keeps the work grounded in actual developer friction.&lt;/p&gt;
&lt;h3 id="2-it-productizes-standards"&gt;2. It productizes standards
&lt;/h3&gt;&lt;p&gt;A standard that lives in a slide deck does almost nothing.&lt;/p&gt;
&lt;p&gt;A standard that shows up as a reusable template, a safe default, a documented flow, a working example, and a supportable paved road actually changes behavior.&lt;/p&gt;
&lt;h3 id="3-it-respects-local-autonomy"&gt;3. It respects local autonomy
&lt;/h3&gt;&lt;p&gt;The best platforms do not remove all choice. They remove the expensive choices that most teams should not have to make repeatedly.&lt;/p&gt;
&lt;p&gt;That is a very different posture from central control.&lt;/p&gt;
&lt;h3 id="4-it-measures-adoption-not-just-existence"&gt;4. It measures adoption, not just existence
&lt;/h3&gt;&lt;p&gt;If nobody uses the platform voluntarily, that is feedback.&lt;/p&gt;
&lt;p&gt;Platform teams need to care about usability the same way product teams do. A paved road that engineers avoid is not a paved road.&lt;/p&gt;
&lt;h2 id="my-working-definition-now"&gt;My working definition now
&lt;/h2&gt;&lt;p&gt;I increasingly think of platform engineering as the discipline of making good engineering behavior easier to repeat.&lt;/p&gt;
&lt;p&gt;That includes technology, of course, but also language, templates, defaults, and trust.&lt;/p&gt;
&lt;p&gt;Done well, it gives teams more autonomy because it lowers the cost of doing the right thing. Done badly, it creates another dependency.&lt;/p&gt;
&lt;p&gt;That is why the DORA finding resonated with me. Not because I am attached to the term, but because I have seen the underlying dynamic up close. When teams have usable internal platforms, they make better decisions faster.&lt;/p&gt;
&lt;p&gt;That is not magic. It is what happens when you turn institutional knowledge into operable systems.&lt;/p&gt;
&lt;p&gt;If this topic interests you, the closest companion piece here is &lt;a class="link" href="https://www.chingono.com/blog/2025/02/15/why-i-started-building-my-own-devops-platform-and-what-i-learned/" &gt;Why I Started Building My Own DevOps Platform&lt;/a&gt;, which comes at the same problem from the builder side rather than the organizational one.&lt;/p&gt;
&lt;p&gt;References:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class="link" href="https://dora.dev/research/2024/dora-report/" target="_blank" rel="noopener"
&gt;2024 DORA Report&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class="link" href="https://www.opslevel.com/resources/tl-dr-key-takeaways-from-the-2024-google-cloud-dora-report" target="_blank" rel="noopener"
&gt;OpsLevel summary of the 2024 Google Cloud DORA Report&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class="link" href="https://www.chingono.com/blog/2025/01/10/conditionally-deploying-resources-azure-app-configuration-using-deployment-scripts/" &gt;Conditionally Deploying Resources in Azure App Configuration Using Deployment Scripts&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;</description></item><item><title>Why I Started Building My Own DevOps Platform (And What I Learned)</title><link>https://www.chingono.com/blog/2025/02/15/why-i-started-building-my-own-devops-platform-and-what-i-learned/</link><pubDate>Sat, 15 Feb 2025 09:00:00 +0000</pubDate><guid>https://www.chingono.com/blog/2025/02/15/why-i-started-building-my-own-devops-platform-and-what-i-learned/</guid><description>&lt;p&gt;For a while, I had the same reaction to most AI-for-software-delivery demos: impressive in a narrow way, but not something I would trust with real work. One tool could write code. Another could summarize a diff. Another could review a pull request. But the hard part of software delivery is rarely one isolated step. It is the handoff between steps.&lt;/p&gt;
&lt;p&gt;That was the itch that eventually pushed me to start building &lt;a class="link" href="https://github.com/cuemarshal/cuemarshal" target="_blank" rel="noopener"
&gt;CueMarshal&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;I did not start with the ambition to build &amp;ldquo;an AI company&amp;rdquo; or some abstract autonomous future. I started because I wanted a more coherent delivery system: one place where a task could move from idea to issue to branch to pull request to review without losing context every time responsibility changed hands.&lt;/p&gt;
&lt;h2 id="the-problem-i-actually-wanted-to-solve"&gt;The problem I actually wanted to solve
&lt;/h2&gt;&lt;p&gt;CI/CD was never the whole problem. In many teams, the pipeline is the most deterministic part of the process. The mess usually lives around it:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;the design decision that only exists in a chat thread&lt;/li&gt;
&lt;li&gt;the issue that says too little&lt;/li&gt;
&lt;li&gt;the reviewer who has to reconstruct intent from commit history&lt;/li&gt;
&lt;li&gt;the documentation that is always &amp;ldquo;we&amp;rsquo;ll do it after&amp;rdquo;&lt;/li&gt;
&lt;li&gt;the growing pile of tools that all know a little, but none of them own the workflow&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;What I wanted was not another dashboard. I wanted a delivery surface that respected how engineering work already happens.&lt;/p&gt;
&lt;p&gt;That led me to a simple conviction: &lt;strong&gt;Git should be the source of truth, not just the storage layer.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;If work already becomes legible through issues, branches, pull requests, labels, and reviews, then the orchestration layer should live there too. Not beside it. Not behind it. Inside it.&lt;/p&gt;
&lt;h2 id="why-i-built-it-myself"&gt;Why I built it myself
&lt;/h2&gt;&lt;p&gt;There were three constraints that mattered to me from day one.&lt;/p&gt;
&lt;p&gt;First, I wanted the system to be &lt;strong&gt;self-hosted&lt;/strong&gt;. A lot of AI tooling assumes you are comfortable sending your code, your process, and your delivery metadata into someone else&amp;rsquo;s black box. Many teams are not. I wanted an approach that made data sovereignty a feature, not an apology.&lt;/p&gt;
&lt;p&gt;Second, I wanted the system to be &lt;strong&gt;role-aware&lt;/strong&gt;. Real software delivery is not &amp;ldquo;one super-agent with a clever prompt.&amp;rdquo; Design, implementation, review, testing, DevOps, and documentation are different jobs. Sometimes one person does multiple jobs, but the jobs are still different. That distinction matters.&lt;/p&gt;
&lt;p&gt;Third, I wanted &lt;strong&gt;human control to remain the final gate&lt;/strong&gt;. I am interested in automation, not surrender. If an AI system cannot work inside a reviewable pull-request workflow, I do not think it is mature enough for serious engineering work.&lt;/p&gt;
&lt;p&gt;Those constraints eventually turned into the shape CueMarshal has now: a conductor service in TypeScript, specialized agents for architecture, development, review, testing, DevOps, docs, and linting, a Git-native workflow in Gitea, and a tool layer built around MCP so the same system can reason over structured interfaces instead of raw shell scripts and ad-hoc API calls.&lt;/p&gt;
&lt;h2 id="the-architecture-came-later-the-principles-came-first"&gt;The architecture came later. The principles came first.
&lt;/h2&gt;&lt;p&gt;Long before the implementation solidified, the design principles were already obvious to me.&lt;/p&gt;
&lt;h3 id="1-git-is-a-better-coordination-layer-than-most-agent-uis"&gt;1. Git is a better coordination layer than most agent UIs
&lt;/h3&gt;&lt;p&gt;An issue is a task. A branch is a workstream. A pull request is a proposal. A review is a decision record. A merge is a controlled state change.&lt;/p&gt;
&lt;p&gt;That sounds almost too obvious to say out loud, but it changed how I thought about the whole problem. Once I stopped treating Git as the place where code merely ends up, and started treating it as the place where engineering decisions become inspectable, the rest of the architecture got much simpler.&lt;/p&gt;
&lt;h3 id="2-specialization-beats-a-do-everything-agent"&gt;2. Specialization beats a &amp;ldquo;do everything&amp;rdquo; agent
&lt;/h3&gt;&lt;p&gt;In CueMarshal, the system is intentionally split into named roles: Marshal for orchestration, Ava for architecture, Dave for implementation, Reese for review, Tess for testing, Devin for DevOps, Dot for docs, and Linton for linting.&lt;/p&gt;
&lt;p&gt;That is not branding for its own sake. It is an operational choice.&lt;/p&gt;
&lt;p&gt;The moment one agent tries to be planner, coder, reviewer, tester, and documentarian all at once, you lose clarity. You also lose accountability. Specialization makes prompts sharper, tool permissions narrower, and outputs easier to judge.&lt;/p&gt;
&lt;h3 id="3-tool-contracts-matter-more-than-prompt-cleverness"&gt;3. Tool contracts matter more than prompt cleverness
&lt;/h3&gt;&lt;p&gt;One of the biggest lessons from building CueMarshal is that the quality of an agentic system is heavily constrained by the quality of its interfaces.&lt;/p&gt;
&lt;p&gt;If an agent is forced to improvise around loosely structured APIs, fragile shell commands, or browser automation for tasks that should be typed and validated, the system becomes harder to trust. This is one reason MCP clicked for me so quickly later on: it gave a clean shape to something I already knew was essential.&lt;/p&gt;
&lt;p&gt;Good tool contracts do not just help the model. They help the human operator understand what the system is even allowed to do.&lt;/p&gt;
&lt;h3 id="4-stateless-workers-are-a-feature-not-a-bug"&gt;4. Stateless workers are a feature, not a bug
&lt;/h3&gt;&lt;p&gt;CueMarshal&amp;rsquo;s runners are intentionally stateless. They reconstruct context from the repository, the issue, the pull request, and the tool layer every time.&lt;/p&gt;
&lt;p&gt;That may sound less magical than the &amp;ldquo;persistent AI teammate&amp;rdquo; narrative, but it is much easier to reason about. It scales better. It fails more cleanly. And it produces a better audit trail.&lt;/p&gt;
&lt;p&gt;In practice, that has made me more skeptical of systems that depend on hidden memory to feel smart.&lt;/p&gt;
&lt;h3 id="5-human-control-is-product-design"&gt;5. Human control is product design
&lt;/h3&gt;&lt;p&gt;The more I worked on this, the more convinced I became that &amp;ldquo;human in the loop&amp;rdquo; is not enough as a slogan. It has to be built into the workflow itself.&lt;/p&gt;
&lt;p&gt;That is why I prefer issue-driven execution, reviewable pull requests, typed tools, explicit handoffs, and merge control. Those are not bureaucratic constraints. They are the difference between a system that can support real engineering and a system that is only good for demos.&lt;/p&gt;
&lt;h2 id="what-i-learned-from-building-in-public"&gt;What I learned from building in public
&lt;/h2&gt;&lt;p&gt;The most useful part of this project has not been proving that agents can write code. We already knew that. The useful part has been learning where coordination breaks, where trust gets earned, and what kinds of structure make AI assistance actually usable.&lt;/p&gt;
&lt;p&gt;It also made one thing clearer for me: the next layer of software delivery is not &amp;ldquo;more CI/CD.&amp;rdquo; It is better orchestration around the work humans and machines are already doing together.&lt;/p&gt;
&lt;p&gt;That is the reason I started building CueMarshal, and it is still the reason I keep working on it.&lt;/p&gt;
&lt;p&gt;If you want the more technical follow-up, I wrote about &lt;a class="link" href="https://www.chingono.com/blog/2025/03/20/mcp-in-practice-what-anthropics-model-context-protocol-actually-means-for-developers/" &gt;what MCP actually changed for developers&lt;/a&gt; and the coordination lessons from &lt;a class="link" href="https://www.chingono.com/blog/2025/08/28/designing-multi-agent-systems-lessons-from-building-an-8-agent-engineering-orchestra/" &gt;building an eight-agent engineering orchestra&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;References:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class="link" href="https://github.com/cuemarshal/cuemarshal" target="_blank" rel="noopener"
&gt;CueMarshal repository&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class="link" href="https://github.com/cuemarshal/cuemarshal/blob/main/docs/architecture/overview.md" target="_blank" rel="noopener"
&gt;CueMarshal architecture overview&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class="link" href="https://github.com/cuemarshal/cuemarshal/blob/main/docs/features/agents/overview.md" target="_blank" rel="noopener"
&gt;CueMarshal agent profiles&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;</description></item></channel></rss>