SHARE ARTICLE

The AI Infrastructure Gap

Why scaling AI requires a new foundation and the nine components every enterprise ends up needing.

FEATURED

Why enterprise AI adoption requires forward-deployed engineers

Alex Kim, VP Engineering

Mudassir Mustafa

AI agent identity management - authentication and authorization for enterprise agents
AI agent identity management - authentication and authorization for enterprise agents

Deploying AI into real enterprises rarely fails because of the models. It fails because of context.

The models are good enough. The infrastructure is mature enough. The data, however messy, is mostly available somewhere in the customer's environment. What's missing is the institutional knowledge that lives in people's heads: the dual-invoice quirk that finance teams know about and accommodate without writing down, the verbally-renegotiated discount terms from last year's offsite that supersede the master contract, the unwritten escalation path that everyone in operations knows because the person who designed it is still around.

That knowledge doesn't show up in any system. It can't be scraped from a SharePoint, queried from an ERP, or extracted from a documentation site. It exists in conversations, muscle memory, and the workarounds people have built into their daily workflows. And until that context is captured, no agent ships into production and actually works.

The forward-deployed engineer (FDE) is the model that captures it.

The role is older than the current AI wave. Palantir originated it at scale in the mid-2000s. It's been reinvented every time a major shift in enterprise software changed how that software gets adopted. In 2026 it's back again because agentic AI is the latest shift, and the same problem the FDE has always solved (bridging the gap between a powerful technology and a messy operating environment) is once again the bottleneck.

This is a post about why the model exists, what it looks like in practice for AI deployments specifically, and where it's going as tooling matures.

Where the FDE model came from

Palantir built the forward-deployed engineering function as a category in the mid-2000s. The original engagement model was simple: don't ship software and expect the customer to figure it out. Send engineers into the customer's environment. Sit with the analysts. Learn the data. Build the system inside the customer's operation. Stay until the system works.

It started in government and intelligence. Then it moved to financial services, healthcare, manufacturing, and energy. The pattern was consistent across industries: the customer didn't have an internal engineering team that could turn a powerful analytical platform into shipped capability, and the value of the platform was unlocked only when somebody who knew the platform deeply spent weeks inside the customer's actual operation.

The FDE wasn't a consultant. Consultants assess, advise, and produce deliverables. The FDE built things. They wrote code. They deployed pipelines. They debugged integrations. They configured the platform against the customer's specific data and processes. They were measured on whether the system was live and producing decisions, not on whether the engagement closed on time.

The role wasn't entirely novel. Field engineers at Cisco in the 1990s did a lighter version of it. Solutions architects at Salesforce and Workday in the early 2010s did another version. Even further back, IBM's customer engineers in the mainframe era did the same job under a different name: embed with the customer, build the integration, make the technology work in their environment.

Every time enterprise software shifted in a way that made the technology more capable but the deployment harder, some version of the FDE pattern emerged to bridge the gap. The pattern reappears because the same underlying problem reappears: powerful technology, ambiguous deployment, customers who can't get to value on their own.

That's where we are again with AI.

Why agentic AI brings the FDE back

Three things make agentic AI deployments fundamentally harder than the SaaS deployments of the last decade. Each one creates the conditions the FDE model is built to handle.

Agentic systems are non-deterministic. A SaaS application has a known behavior surface. You can document the buttons, the API, the workflow states. An agent doesn't behave that way. Its outputs depend on the model, the context, the tools it can call, the policy it operates under, the prompt history, the data it retrieves, and the small thousand decisions that happen between input and action. Two identical inputs can produce two different outputs. That's not a bug; it's a property of the system.

This means deployment isn't a one-time event. It's an ongoing tuning exercise. You see what the agent does in the wild, you adjust prompts and policy, you watch the new behavior, you adjust again. The work is empirical, not specifiable.

Value emerges over time, not at deployment. A traditional software rollout has a moment. The system goes live, the users get trained, the dashboard lights up, the productivity gain shows up in a report by quarter-end. An agent rollout doesn't have that moment. The agent runs against the first hundred cases and reveals five edge cases nobody mentioned in scoping. The next thousand cases reveal twenty more. By month two the agent is reliable on ninety percent of cases. By month four it's outperforming the human baseline. By month six it's catching errors the humans missed for years.

The value curve is gradual and front-loaded with surprises. Someone has to be present during those surprises, decoding them in real time, deciding whether each one is a model issue, a policy issue, a data issue, or an actual business edge case the team didn't surface during discovery.

The work shifts from product adoption to operational intelligence. SaaS adoption is about getting users into the workflow. Agentic adoption is about translating the customer's actual operation into agent behavior. That translation is the heart of the deployment. It can't be done by the customer alone because the customer doesn't know the platform deeply. It can't be done by the platform team remotely because they don't know the operation deeply. It has to be done by someone who knows both, sitting close enough to the work to see what's actually happening.

That someone is the FDE.

HappyRobot published a piece on this recently, framing the role as re-emerging because agentic systems need engineers closest to the customer. The terminology is converging across vendors in the agent space. The model itself isn't new. It's the Palantir pattern, applied to a new technology layer.

What an FDE actually does, week by week

Most of the writing about FDEs is abstract. Here's what the role looks like in practice for an AI deployment, drawn from how we and others in the agent ecosystem run it today.

Week 1: discovery, not building. The FDE shows up to the customer's standups. They sit in on operations calls. They ask questions that sound basic and produce surprising answers. Where does this invoice come from? Why are these two vendors handled differently? Who decided that threshold? What happens when the system disagrees with the controller? The point of week one isn't to write code. It's to map the real workflow against the documented workflow and find the gap.

There's always a gap. The documented process is the version that exists in onboarding decks. The real process is the version that runs every day, full of exceptions, workarounds, and unwritten rules. The agent has to work against the real process, not the documented one.

Week 2: integrations. Now the FDE writes code. They pull data out of the customer's systems. Sometimes that's a clean API call. Sometimes it's a SOAP endpoint that requires a specific certificate. Sometimes it's a homegrown application the controller built in 2013 that nobody else fully understands. Sometimes it's a vendor portal with a single login that doesn't support service accounts.

This week is the part of the FDE role that looks most like traditional engineering. It's also the part that compresses the most as tooling matures. What took two weeks of integration work in 2023 takes three days in 2026 because the integration surface area is mostly product now.

Week 3: capturing tribal knowledge. The FDE sits with the controller, the operations manager, the finance analyst, the warehouse supervisor. They walk through real cases together. The dual-invoice quirk where one vendor sends a PDF and a duplicate email and the agent has to dedupe them. The verbally-renegotiated discount terms where the master contract says one number but the email thread says another. The OCR exception where one specific supplier sends invoice scans rotated ninety degrees because their scanner has a Japanese locale setting nobody has changed in eight years.

None of this is in the system. All of it is in the operator's head. Week three is the work of getting it out of their head and into agent policy.

Week 4: first agent in sandbox. The FDE builds against the platform. Configures the agent. Wires the retrieval. Writes the policy. Sets the escalation thresholds. Runs the agent against sanitized real data. Watches what it does. Tunes. Re-runs. Tunes again. This is the empirical loop the non-determinism of agentic systems requires.

Weeks 5-6: production cutover. Read-only first. Then read-write with human-in-the-loop approval on every action. Then read-write with thresholds (auto-execute under X dollars, approval over X dollars). Then full autonomy with audit log. Each step takes a few days of watching, tuning, and convincing the security team and the operations team that the next step is safe.

Weeks 7-8: knowledge transfer. The FDE trains the platform admin, the operations team that will triage exceptions, and the security team that will review the audit log. The goal is for the customer's team to operate the agent without the FDE present. If the customer can't run it themselves by the end of the engagement, the engagement isn't done.

This is the deliberate distinction from a consulting engagement. Consultants extract value over the length of the engagement and disengage. The FDE engagement is designed to make the FDE unnecessary at the end of it.

The FDE isn't a consultant who parachutes in. They're a builder who embeds, joins the customer's Slack, attends their standups, learns their stack, and stays until the agents are live and outcomes are proven.

Where the FDE role is going

Two trends are reshaping the role.

Tooling collapses execution time. What took eight weeks in 2024 takes six weeks in 2026 and will take three weeks by 2028. The integration layer is product. The agent runtime is product. The governance substrate is product. The retrieval layer is product. What used to be custom build is increasingly configuration against a platform that handles the heavy lifting.

That doesn't displace the FDE. It changes what the FDE spends time on. Less time wiring SAP-to-Postgres pipelines. More time on the harder work that tooling doesn't compress.

The harder work expands. As execution gets cheaper, the deliverable shifts from "agent running in production" to "operational intelligence translated into shipped capability." That includes:

  • Domain depth. Learning the customer's industry well enough to identify second-order optimizations that the customer's own team is too close to see.

  • Process redesign. Knowing where the workflow itself should change, not just where AI should be inserted into the existing one.

  • Evaluation design. Building the right way to measure agent performance against business outcomes, not just against accuracy in isolation. The accuracy metric is often the wrong metric.

  • Product feedback. Insights from the field flow back into the platform roadmap. The FDE becomes the bridge between what the customer actually needs and what the product team builds next.

The FDE of 2028 looks less like a senior engineer with a laptop and more like a domain-expert builder who can both ship code and shape the customer's operation. The economic value of the role grows as the routine work gets automated away.

Why the model keeps coming back

Every time enterprise software shifts in a way that makes the technology more capable but the deployment harder, the FDE pattern re-emerges. It's not coincidence. It's the same problem each time.

The model layer is commoditizing. The infrastructure layer is commoditizing. The interface layer is commoditizing. None of those layers is where AI adoption fails.

Adoption fails at the context layer. The understanding of how a business actually operates. The fifty edge cases nobody documented because everybody already knew them. The process that exists in muscle memory across thirty people in three departments. The tribal knowledge that nobody has time to write down but that determines whether a system works on Monday morning.

That layer doesn't yield to better models. It doesn't yield to better infrastructure. It yields only to someone being embedded in the work for long enough that the messy reality becomes obvious.

The FDE is how you cross the context layer. Until that layer is automated, which is not happening soon, the model persists.

What to do next

If you're thinking about how to deploy AI inside an enterprise, the question to ask isn't which model, which platform, or which vendor. It's: who is going to do the context work? If the answer is "the platform vendor's deployment team," dig into how they're structured. If the answer is "a Big-4 consulting firm," ask whether that team is going to write code and stay until production, or whether they're going to produce a deliverable and leave. If the answer is "our internal team," make sure that team has the time and the platform depth to do the work.

If none of those answers fits, the FDE model is probably what you need.

Reach out if you want to talk about what that actually looks like in your environment.

SHARE ARTICLE

The AI Infrastructure Gap

Why scaling AI requires a new foundation and the nine components every enterprise ends up needing.

The AI Infrastructure Gap

Why scaling AI requires a new foundation and the nine components every enterprise ends up needing.

WHITE PAPER

The AI Infrastructure Gap

Why scaling AI requires a new foundation and the nine components every enterprise ends up needing.

WHITE PAPER

The AI Infrastructure Gap

Why scaling AI requires a new foundation and the nine components every enterprise ends up needing.

Recent Blogs

Recent Blogs

BECOME AI-FIRST

BECOME AI-FIRST

Transform your enterprise in weeks.

Transform your enterprise in weeks.

Transform your enterprise in weeks.

Transform your enterprise in weeks.

Thirty minutes. Your actual stack. We'll show you what AI-first looks like running on your cloud, connected to your real systems.

Thirty minutes. Your actual stack. We'll show you what AI-first looks like running on your cloud, connected to your real systems.

document.documentElement.lang = "en";