Chase Roossin & Steven Kulesza (Intuit) Discuss the Hardest Engineering Challenge: Making AI Agents Work Together in Complex Systems

In this week’s beta release, Intuit’s engineering team unveiled a novel coordination framework for AI agents that tackles the core challenge of scaling multi-agent systems: deterministic state reconciliation across heterogeneous models without centralized bottlenecks. By combining lightweight consensus protocols with model-agnostic message schemas, the system enables dozens of specialized agents—each fine-tuned on distinct financial workflows like tax optimization, fraud detection and cash flow forecasting—to collaborate on shared goals even as maintaining strict data sovereignty and auditability.

The State Synchronization Problem in Heterogeneous Agent Swarms

Most enterprise AI agent deployments today rely on either brittle hardcoded handoffs or centralized orchestrators that become single points of failure and latency hotspots. As agent counts scale beyond ten, the combinatorial explosion of interaction states—especially when agents operate on different LLMs (e.g., a mix of Llama 3 70B for reasoning and Phi-3-mini for edge inference)—makes traditional approaches untenable. Intuit’s solution sidesteps this by treating agent coordination as a distributed state machine where each agent publishes minimal, versioned deltas to a conflict-free replicated data type (CRDT) mesh, enabling eventual consistency without blocking.

The State Synchronization Problem in Heterogeneous Agent Swarms
Intuit Agent The State Synchronization Problem

This approach draws from recent advances in conflict-free data structures but adapts them for the non-deterministic outputs of LLMs. Rather than attempting to reconcile raw text responses—which would be computationally prohibitive—the system maps agent intentions into a normalized action ontology (e.g., “adjust_withholding,” “flag_anomaly,” “project_cashflow”) before applying CRDT merges. Early benchmarks show a 40% reduction in coordination latency compared to LangGraph-based orchestration at 50-agent scale, with near-linear throughput growth up to 200 agents on AWS Graviton4 instances.

Bridging the Open-Source Chasm Without Vendor Lock-In

What makes this architecture particularly notable is its deliberate avoidance of proprietary glue. The coordination layer exposes a gRPC API defined via protobuf schemas hosted on GitHub, allowing third-party agents built in Python, Rust, or JavaScript to join the mesh by implementing a simple AgentInterface trait. This stands in stark contrast to closed ecosystems like Microsoft’s AutoGen or Google’s Agent Garden, which tightly couple agent logic to their respective cloud backends.

Bridging the Open-Source Chasm Without Vendor Lock-In
Intuit Agent Bridging the Open

“True agent interoperability isn’t about sharing model weights—it’s about agreeing on what actions mean. Intuit’s ontology-first approach is the first I’ve seen that scales beyond a lab demo without forcing everyone onto the same LLM vendor.”

— Elena Rodriguez, CTO of Finos Labs, speaking at the AI Agent Summit 2026

This open-surface design directly challenges the platform lock-in strategies of major cloud providers, who increasingly bundle agent frameworks with their AI suites to create switching costs. By keeping the coordination plane neutral and the action ontology extensible via community-driven pull requests, Intuit is positioning itself as a Switzerland of agent interoperability—a role that could prove critical as financial institutions face mounting pressure to avoid single-vendor dependencies under upcoming AI accountability regulations.

Under the Hood: Message Flow and Failure Handling

Under normal operation, agents communicate via asynchronous message passing over NATS JetStream, with each message containing:

  • A globally unique agent ID (UUIDv7)
  • A Lamport timestamp for causal ordering
  • A protobuf-encoded action intent from the shared ontology
  • A cryptographic nonce signed with the agent’s rotating Ed25519 key

Upon receipt, peer agents validate the signature, check for replay attacks using a sliding window nonce cache, and apply the intent to their local state snapshot. Conflicts—such as two agents attempting to modify the same ledger entry—are resolved using application-specific merge functions registered at mesh initialization, allowing domain experts to encode business rules (e.g., “fraud alerts override tax suggestions”) directly into the coordination logic.

In failure scenarios, the system gracefully degrades to eventual consistency: if an agent loses network connectivity, it queues outgoing intents locally and rebroadcasts them upon reconnection, leveraging the CRDT’s convergence properties. This eliminates the need for complex distributed transactions while still providing strong enough consistency for financial workflows, where brief divergences are acceptable if auditable and self-correcting.

The Bigger Picture: AI Agents as the New Microservices

What Intuit is building mirrors the evolution of microservices a decade ago—where the initial excitement around service meshes gave way to standardized protocols like gRPC and OpenAPI that enabled true polyglot ecosystems. Today, we’re at the inflection point where AI agent frameworks must choose between becoming the next generation of vendor lock-in traps or embracing the open, composable principles that made the web and cloud native stacks successful.

By publishing their agent mesh specifications under an Apache 2.0 license and actively soliciting contributions from the open-source community—including recent patches from engineers at NVIDIA on GPU-accelerated intent encoding—Intuit is betting that the winner in the agent wars won’t be the company with the biggest models, but the one that enables the most diverse set of agents to cooperate without friction.

As enterprises begin deploying agent swarms across supply chains, healthcare systems, and financial platforms, the ability to get multiple agents to play nice at scale won’t just be an engineering challenge—it’ll be a competitive necessity. And if this week’s beta is any indication, the solution may lie not in more powerful models, but in smarter ways to make them talk to each other.

Photo of author

Sophie Lin - Technology Editor

Sophie is a tech innovator and acclaimed tech writer recognized by the Online News Association. She translates the fast-paced world of technology, AI, and digital trends into compelling stories for readers of all backgrounds.

IPL 2026: Top Run-Scorers and Wicket-Takers Race for Orange and Purple Cap – BBC Sport Analysis

Deluxe Edition of 2025 Album Features Remixes with Tyla, Shakira, Kehlani, and JT

Leave a Comment

This site uses Akismet to reduce spam. Learn how your comment data is processed.