How we work

Every engagement has the same general shape: three to six months, one or two core applications, a principal embedded with the client team. Below you'll find the three areas the firm works in, the four-step engagement structure, and what actually happens on a first call.


The three areas of work

Most engagements draw from two of the three areas. A handful draw from all three. The gains usually come from the combinations rather than from any single area on its own.

Area 1 — AI-Augmented Dev-Ops

What it is. Fitting AI tooling into the development pipeline so it actually reduces effort. Test generation against existing codebases, assisted refactoring on long-lived modules, automated review for the categories of issues that suit model review (typing, common bug patterns, style), and scaffolding or migration code where the patterns are repeatable.

Teams usually call us in here for one of two reasons. Either a CTO has approved AI tooling but six months later the productivity numbers look the same; or individual engineers are using AI assistants productively but the gains aren't compounding into team-level throughput. There's a third version too — refactoring is bottlenecked on engineer hours and the codebase happens to suit AI tooling — but that one's rarer.

An engagement typically opens with a baseline measurement of where engineering time is going across a representative team, sprint by sprint and ticket type by ticket type. We pick two to four categories of work with the strongest expected return, implement the tooling and workflow changes for those, then re-measure against the baseline at six and twelve weeks. What teams walk away with is some combination of dev-month load reductions on the targeted categories, workflow patterns the team keeps using after we leave, and a baseline framework the engineering organisation can apply to future tooling decisions on its own.

Area 2 — AI-Ops Optimization

This area is about the economics of running AI in production. Model selection against the actual quality requirements of the task rather than against benchmark scores. Inference cost reduction through caching, smarter routing between model sizes, and quantisation where it actually fits. Infrastructure tuning across the serving stack — autoscaling, request shape, retry policies. The point is to treat the AI cost line as a real engineering discipline instead of a usage-based bill that scales with adoption.

How it usually starts. Inference costs are growing faster than the revenue from AI features, model spend has turned into a board-level conversation, or the team picked an over-specified model six months ago and is now paying for quality the actual task doesn't need. Sometimes AI features have shipped without anyone building the cost-monitoring and model-selection discipline that production AI demands.

We start with a measurement pass on the current cost structure: model by model, feature by feature, request by request. We identify the largest cost categories and the changes most likely to move them. Then iterative implementation — caching layer, smaller-model routing, request shape changes, infrastructure tuning, in whatever order makes sense for the cost map. Each change is measured against the cost baseline before we move on. Engagements like this typically take 30–55% off monthly inference spend on the targeted features, and leave behind a model-selection framework plus monitoring on cost regressions that used to slip through unnoticed.

Area 3 — Process Re-Engineering

Most engineering organisations don't have a tooling problem; they have a flow problem. Process Re-Engineering is the work of redesigning how work moves through the pipeline — ticket flow and hand-off reduction, sprint structure and standup discipline, coordination overhead between teams, and the gap between "we use Jira" and "Jira tracks the actual work." This area sells fewer tools than the other two, which is part of why it's usually where the biggest gains live.

It tends to be the right intervention when the organisation has grown but velocity hasn't scaled with headcount. Meetings have multiplied to coordinate work that used to coordinate itself, there are more hand-offs between teams than there were a year ago, and senior engineers are spending more time in process than in code. The block is rarely engineering complexity at that point — it's coordination cost.

The work starts with observation: sitting in on standups, shadowing tickets through their lifecycle, mapping hand-offs and queue times. From there we pick the two or three coordination patterns producing the most overhead and implement changes with the team rather than handing them down. Ticket structure, hand-off rules, meeting cadence, ownership boundaries — whichever combination fits. Cycle time and coordination time get re-measured against the baseline at the end. The artefact teams seem to value most isn't any single process change but the honest answer to "where is the time actually going."


The engagement shape

Every engagement follows the same four steps. Duration is three to six months. Scope is one or two core applications or teams.

01

Baseline

We start by measuring. Most teams have a sense of where the problems are; very few have clean numbers on how big the problems are or which parts of the pipeline are producing them. Depending on the engagement, the baseline covers some combination of dev-month load by ticket category, sprint cycle time, model-by-model inference spend, queue times at hand-offs, and where engineering hours are going by category of work.

Even if a client ended the engagement at this point, they'd walk away with an instrumented view of their own organisation they didn't have before. The baseline counts as a deliverable on its own.

02

Re-engineer the loop

With the baseline in hand, we pick the two to four interventions most likely to move the metrics that matter, drawn from the three areas above in whatever combination fits the picture.

This is where most of the disagreement happens, which is fine — we welcome it. Interventions get hashed out with the client team before any implementation. The team that has to live with the changes needs to be on board with them.

03

Ship

We embed with the client team for the duration, and the work gets done together. The deliverable is changes in production: code merged, infrastructure deployed, process changes adopted by the team. A strategy deck is not a deliverable. If we need to put one together for a stakeholder meeting we will, but it's a by-product.

04

Measure

The same instrumentation that produced the baseline produces the final measurement, run again against the targets we agreed on.

If the metrics moved as we expected, the engagement closes and the outcome-tied fee is settled. If they didn't, we say so.


Engagement structure and commercials

Engagements run three to six months. Scope is usually one or two applications or teams. We won't take on the whole engineering organisation at once — that's not a pilot, that's a transformation programme, and it isn't what the firm does.

Fees split in two: a fixed-fee base for the engagement plus a bonus tied to the agreed outcome metrics. The structure aligns the firm's incentives with the client's. Numbers depend on scope and get worked out on the first call.

We don't bill hourly, we don't sell dedicated headcount, and there are no ongoing monthly retainers. The engagement model is one shape on purpose.


The discovery call

Thirty minutes on a call. Nothing to prepare beforehand. We'll ask what you're currently seeing in your engineering organisation — the cost line, velocity, where modernisation has stalled, whatever has prompted you to look at this. You'll ask whatever you need to about how the firm works.

By the end, we'll both have a sense of whether there's a fit. If there is, the next step is a one-page proposal describing what we'd measure, what we'd work on, the duration, the fee structure, and the outcome metrics we'd both agree on. If there isn't, we'll say so.

The engagement model only works when the firm and the client are both confident in the fit. If we don't think we can move the metrics that matter for you, we'll say so on the call.

Book a discovery call →