Skip to content
Diosh Lequiron
Systems Thinking12 min read

The Hidden Cost of Local Optimization in Operating Systems

Local optimization produces worse failure modes than the one it replaces — displaced bottlenecks, metric traps, compounding side effects. Evidence from 19 years of enterprise and venture operations.

Every operational leader I have worked with in 19 years of program delivery has, at some point, been asked to fix one thing. Speed up the deployment pipeline. Cut the call-handle time. Reduce the stockout rate. Improve the contractor utilization number. The request is always framed the same way: this one metric is the problem, make it better.

Almost every time I have accepted that framing literally, the system has gotten worse. Not the metric — the metric usually moved. The system. The failure mode shifted somewhere else, usually to a place with less instrumentation, and the new failure was more expensive than the one I had been asked to fix.

I call this the hidden cost of local optimization: the damage that accumulates in the parts of a system nobody is currently looking at, while attention is concentrated on the part that is being tuned. It is one of the most consistent patterns I have observed across enterprise IT, agency operations, and a multi-venture portfolio — and it is almost always invisible to the people doing the optimization.

This article explains why local optimization produces worse failure modes than the one it replaces, what structural alternative actually works, and the boundary conditions where local optimization is still the right move.


Why Fixing One Part at a Time Fails

Three patterns account for most of the local-optimization failures I have diagnosed. They show up in enterprise programs, delivery agencies, and startup operations with almost identical mechanics.

The Displaced Bottleneck. When a team removes a constraint at one stage, throughput increases at that stage — but it does not disappear from the system. It arrives downstream. If the downstream stage has less capacity or less instrumentation, the new bottleneck is harder to see and more expensive to operate under. The team celebrates the local improvement while the system quietly degrades.

At HPE, one of the multi-million-dollar programs I directed had an onboarding stage that everyone described as the problem. We compressed it — tooling, automation, a redesigned intake. Onboarding time dropped meaningfully. Within a quarter, the downstream integration team was missing delivery dates for the first time in two years. The bottleneck had moved from a stage with dashboards, ceremonies, and senior attention to a stage that had none of those things. The displaced bottleneck was not a new problem. It was the same problem relocated to a place where it did more damage.

The Metric Trap. When one number becomes the target, the rest of the system starts getting sacrificed to move it. Teams learn, very quickly, what gets measured and what does not. A handle-time target in a contact center produces calls that close in under the threshold and customers who call back three times because the first resolution was structurally incomplete. A utilization target in an agency produces billable hours that look healthy on the dashboard and quality defects that surface six weeks later when the project is already past its margin window.

I saw this across multiple offices in the Australian agency network I worked with. Individual office managers were optimizing the metrics they were measured on — utilization, realization, average project margin. Each office looked reasonable in isolation. The portfolio, taken as a whole, had been losing between twenty and sixty percent across multiple offices for more than a year. The optimization was real at the local level. The damage was real at the system level. Neither set of people was wrong about what they were seeing — they were looking at different parts of the same broken architecture.

The Compounding Side Effect. Local fixes introduce constraints that propagate. A schema change made to speed up one report makes three other queries slower. A process change that reduces load on one team shifts coordination cost to another team that had no input into the decision. A deployment gate added to catch one class of failure forces teams to batch releases, which increases the blast radius of every deploy that does ship.

These effects compound. By the time the side effects are visible, the original optimization is old news. Nobody connects the current failure mode to the fix from three months ago. The system looks like it is failing for new reasons. It is not. It is failing for the reason it was set up to fail when the first local fix was made.

These patterns do not respond to more local optimization. Adding another tuning cycle to the onboarding stage does not recover the integration team. Adding a quality metric to balance the handle-time metric produces callbacks that game both. Adding a governance process to review schema changes delays every change, including the ones that would have worked.

The problem is architectural.


The Systems View

The alternative to local optimization is not a refusal to optimize. It is a structural shift in where optimization decisions are made, what evidence they require, and how they are contained.

Optimize the Bottleneck, Not the Component

Every operational system has a rate-limiting stage. Moving capacity anywhere else does not move system throughput — the bottleneck absorbs it. The first structural question in any optimization request is: where is the actual bottleneck? Not where is the complaint. Not where is the attention. Where is the constraint that governs the throughput of the whole system?

I learned this viscerally when I founded the PMO at Full Potential Solutions. The escalated request was always about some loud stage — a team that missed a date, a process that created friction. The actual bottleneck was almost never there. It was in the intake or the handoff or the evidence-capture stage — quiet places where work accumulated silently and then discharged all at once into whichever stage happened to be downstream. Fixing the loud stage did nothing for throughput. Fixing the quiet stage changed the rate at which the whole system cleared.

The discipline is to resist the first framing. The problem is rarely where the problem is reported.

Instrument the Blast Radius Before the Fix

Any change to a shared system has a blast radius — the set of downstream and adjacent components that will be affected. Local optimization ignores blast radius by definition: it treats the fix as contained when it is not. The structural alternative is to instrument the blast radius before the change is made, so that side effects are detectable at the same cadence as the intended effect.

In practice, this means every optimization proposal produces three outputs: the expected improvement, the list of downstream components the change could affect, and the measurement points that will surface degradation if it happens. The cost of this discipline is front-loaded — it takes time before the fix. The benefit is that displaced bottlenecks and compounding side effects become visible within days, not quarters, so they can be corrected while the context is still fresh.

I have applied this across enterprise programs and across the ventures I operate. The same structural question: what breaks if this works as intended? The answers are usually specific — a downstream team, a shared resource, an adjacent process — and the measurement is usually cheap once the adjacency has been identified.

Govern the Metric, Not the Work

When a single metric is used to govern operational work, the work bends around the metric. The structural alternative is to govern the metric itself — to make the definition of success structural rather than scalar.

A scalar target says: reduce handle time to under four minutes. A structural definition says: resolve the customer request such that no related callback occurs within seven days, and total time to resolution is under four minutes. The second definition is harder to game because it couples the local metric to a downstream consequence. Teams that close calls too fast will see their callback rate rise. Teams that take too long will see their handle time fail. The two constraints hold each other in tension.

This is not a new idea — it is how sensible operational metrics have always been designed. What I have learned is that it is almost never applied by default. Default metrics are scalar because scalars are easy to dashboard. Structural metrics require more instrumentation and more interpretation, which is exactly why they are the ones that actually govern behavior without producing the metric trap.

Make the Tradeoff Explicit

Every optimization is a tradeoff. Local optimization pretends it is not. The structural alternative is to make the tradeoff explicit before the decision — to state, in writing, what is being traded away to gain the improvement.

I use a simple contract: every operational change proposal must state what it improves, what it costs, and what it makes worse. The third item is the one that matters. Changes that cannot articulate a downside are usually changes whose downside has not been investigated. That is not an argument against the change — it is an argument for investigating the downside before committing.

Across 18 ventures and in every enterprise program I have directed, the explicit tradeoff discipline has produced a consistent effect: the number of proposed optimizations drops, the quality of the ones that proceed rises, and the portfolio-level damage from hidden side effects decreases measurably. The optimization rate decreases. The optimization yield increases.


Operational Evidence

Scale. Across 18 ventures operating under HavenWizards 88 Ventures OPC, the shared operating model uses the same structural question for every optimization proposal: where is the bottleneck, what is the blast radius, what is the explicit tradeoff. Ventures range from a 66-module agricultural SaaS serving Filipino cooperatives to a basketball affiliate content venture to a Thai-language learning platform. The domains are unrelated. The structural discipline is the same. The yield — optimizations that hold versus optimizations that get reversed — is materially higher than it was when I was managing ventures individually without the shared architecture.

Recovery. The Australian agency network I worked with had been losing between twenty and sixty percent across multiple offices for more than a year. Every office was locally optimizing — utilization, realization, handle times, sales velocity. None of them were wrong about their local numbers. The portfolio was failing because the local optimizations were displacing bottlenecks, gaming metrics, and compounding side effects across office boundaries. The intervention was not more optimization. It was structural: shared delivery governance, shared definitions of success, shared measurement at the portfolio level rather than the office level. Losses reversed to between forty and sixty percent profit within the intervention window. The individual offices did not get smarter. The architecture that contained them got better.

Prevention. In the scaling engagement for the US startup I supported from eight to ten thousand dollars per month to over five hundred thousand, the growth rate was fast enough that any local optimization would have compounded into structural damage within weeks. The discipline was inverted: no optimization proposal was accepted without the three-part contract — bottleneck identification, blast radius instrumentation, explicit tradeoff. The result was that the operations scaled from a small founder-run team to more than five hundred full-time-equivalent roles without the founder becoming the decision bottleneck. The growth is the evidence, but the mechanism is the refusal to accept local-optimization framings as the default.

Compounding. Over the first eighteen months of operating the venture portfolio under shared governance, the rate of reversed optimizations decreased measurably. Early on, roughly one in four optimization proposals had to be partially or fully reverted because of side effects that had not been anticipated. Within a year of applying the structural discipline — bottleneck first, blast radius instrumented, tradeoff explicit — the reversal rate dropped substantially. This is the compounding effect: the architecture does not just prevent the current class of failure, it improves the team's ability to see the next one before it ships.


Where This Does Not Apply

Structural optimization discipline has costs. It is not the right default for every context, and recognizing where it does not apply is part of using it well.

Isolated subsystems with no shared state. Some parts of an operation genuinely are local. A team running a self-contained experiment with no downstream dependencies, a standalone reporting job that reads from a replica and writes to a throwaway destination, a personal tooling script that nobody else uses — these are places where local optimization is the correct scope. Applying blast-radius discipline to truly isolated work is overhead without benefit.

Time-boxed crises. In a genuine production incident where the system is actively failing, the sequence is triage first, structural diagnosis second. Stopping the bleeding is the priority. The structural question — is this fix displacing the bottleneck, is it compounding somewhere — comes after the system is stable. Applying full optimization discipline during an active incident slows recovery without changing the outcome.

Exploratory prototyping. When the goal is to test whether an idea has merit, the cost of structural optimization discipline is too high. Prototypes are supposed to be cheap, reversible, and disposable. Governing them with the same rigor as production systems is a category error — it treats learning artifacts as if they were operational artifacts.

Very small systems. Below a certain scale, the number of components is small enough that any competent operator can hold the whole system in their head. In that regime, the bottleneck, the blast radius, and the tradeoffs are visible without structural instrumentation. Adding the discipline at that scale is friction without proportional benefit. The threshold at which structural discipline pays off is roughly where no single person can reliably predict the second-order effects of a change — which in my experience is much earlier than most organizations assume.


The Principle

Local optimization is not a mistake. It is a default. It is what happens when a system is under pressure and someone asks for a fix to one thing. The fix usually works at the level it was asked. The damage it does happens somewhere else, on a slower timeline, to people who were not in the room when the decision was made.

The structural alternative is not slower. It is more honest. It identifies the actual bottleneck instead of the loud one. It instruments the blast radius instead of ignoring it. It governs metrics structurally instead of scalarly. It makes tradeoffs explicit instead of pretending they do not exist.

The operators I trust most are the ones who, when asked to fix one thing, respond with: before I change anything, let me show you where the bottleneck actually is and what this change will cost somewhere else. That is not resistance. That is the difference between optimizing a component and operating a system.

ShareTwitter / XLinkedIn
Diosh Lequiron
Diosh Lequiron
Systems Architect · 19+ years designing operating systems for complexity across technology, education, agriculture, and governance.
About

Explore more

← All Writing