After a few intense years of building with AI, many teams are asking the same question: Where is the ROI?

We’re shipping chatbots, copilots, and agents. But:

  • Agents aren’t autonomous enough to own end‑to‑end workflows.
  • Model quality degrades over time with prompts, tools, and data drift.
  • Traditional data and architecture patterns hit their limits when you try to scale beyond demos.

Looking ahead to 2026, one theme cuts through all of this: AI will only become truly useful when we treat context and data architecture as first‑class concerns - and that’s exactly where graph technologies quietly move into the spotlight.

1. AI Reality Check: Scaling is the hard part

The feedback loop from production systems is blunt: most AI projects don’t deliver what was originally promised. According to an MIT study, 95% of pilot projects do not generate measurable outcomes. Although doubled from 2024, only 31% of use cases reached full production in 2025 (ISG - State of Enterprise AI Adoption). From our own experience (and probably yours, too), we see that even something we think of as mundane as an AI-Chatbot is hard to get from prototype to production. GenAI projects are too often held back by costs, unclear ROI, and unresolved risks.

This is the GenAI paradox: General tools (such as ChatGPT and Copilot) are easy to roll out, but it’s challenging to demonstrate clear business value and measurable impact. The most valuable, vertical, deeply integrated systems are precisely the ones that are hardest to build and scale.

From a developer’s point of view, the reason is obvious: Inside organisations, AI isn’t a website you launch - it’s a system you need to wire into messy processes, legacy stacks, and fragmented data silos.

That means:

  • Tight integration with identity, permissions, and compliance.
  • Access to multiple data sources with different schemas and quality levels.
  • Observability, human validation, testing, and rollback strategies for non‑deterministic behaviour.

All of that takes time, iteration, and a lot of plumbing. AI can help accelerate parts of the work, but it doesn’t make integration “go away”. 2026 won’t be about “more models”; it’ll be about making a small number of flexible AI systems truly robust and scalable.

2. AI Agents: Less like magic, more like new teammates

“Armies of agents” replacing teams make for good conference talks. In production, most AI agents resemble specialised teammates more than autonomous departments.

Right now, many companies are experimenting with agents, but only 16% of enterprise deployments qualify as true agents (Menlo VC). Most production systems are fixed-sequence workflows around single model calls, where a “Prompt design + “RAG” pattern remains the dominant model.

In enterprise practice, you mostly see agents (besides coding agents):

  • Running behind the scenes in research‑heavy domains (law, compliance, medicine).
  • Handling repetitive, structured workflows with clear guardrails.
  • Orchestrating tools and APIs rather than “thinking” in the abstract.

The bottleneck isn’t model IQ - it is context and control. Like a new hire, an agent needs:

  • Onboarding: what data is relevant, what tools exist, what does “good” look like?
  • Policies: what is it not allowed to do?
  • Feedback loops: who checks its work, and how do we correct mistakes?

Because LLMs are probabilistic, the same input doesn’t always yield the same output, so to put agents in front of real users you need robust test suites that link prompts, tools, and data to expected behaviour, continuous monitoring for drift and regressions over time, and clear ownership so it’s always obvious who is responsible when an agent makes a bad decision.

That has an organisational side, too. Teams need new roles, clearer governance, and a shared understanding that “the agent did it” is not an acceptable explanation.

In 2026, the interesting work for developers will be less “build an agent” and more “design how agents work with people and systems” - including how they access and reason over data.

3. Context Engineering: Treating information architecture as part of the system

Even in agentic, multi‑step setups, AI is only as good as the context you feed it. In practice, the problem isn’t just “bad prompts”, it’s often bad context architecture.

We’ve learned a few things the hard way:

  • Long, noisy context windows lead to errors, latency, and weird edge cases.
  • Models behave like human memory: they remember the beginning and end, but lose the thread in the middle (Stanford).
  • Too many tools or overlapping instructions lead to context confusion.
  • Contradictory steps or outdated data can cause context clashes.

The theory says: “Models can handle huge context windows.” Reality says: “The more you stuff in, the more you dilute what matters.” Anthropic refers to this as the “attention budget”: every extra piece of context consumes part of a limited resource. Like with humans, if you burn that budget on irrelevant or redundant information, you reduce output quality.

For developers, this turns into a discipline of context engineering:

  • Selecting just enough data for each step instead of dumping everything.
  • Structuring that data in a way the model can follow (schemas, relationships, roles).
  • Tracking how prompts, tools, and context evolve over the lifetime of a system.

As reasoning defined 2025 (stateof.ai) and models now plan, self-correct, and work over longer horizons, the need for robust context engineering becomes even clearer: organising information not as static input, but as structured, navigable knowledge that models can use, not just consume. If you can represent how pieces of information relate to each other in a graph, you can build smarter, more targeted context assembly instead of brittle keyword search plus giant text blobs.

In coding agents, this leads to spec-driven development, where you work with the model to generate a detailed specification that includes problems, requirements, and criteria. You then break these down and turn them into task descriptions, which the agent takes on one by one. For other domains, it’s less prescriptive, but it will still need to present instructions in a more explicit information architecture.

4. Push vs. Pull: Data on demand, not data in advance

Early Retrieval‑Augmented Generation (RAG) setups mostly followed a push model:

  1. Pre‑index large chunks of data.
  2. At query time, retrieve and re-rank N documents.
  3. Push them into the context window and hope the model sorts out their information and relevancy.

With agents and tool‑using models, a pull model is emerging: The AI figures out what it is trying to do, it identifies what information is missing to finally pull exactly the data it needs using tools and APIs, step by step in a loop.

Instead of an information avalanche, you get information selection. In this world, an LLM becomes less like a “smart typeahead” and more like an orchestrator:

  • It plans: “To answer this, I need customer history + product rules + pricing.”
  • It chooses tools: “Call this API, query that store, run that check.”
  • It iterates: plan → act → observe → refine.

For your architecture, that means thinking like an information architect:

  • What is the minimum viable context (MVC) for this step?
  • How can the system efficiently route to just the right slice of data?
  • How do we avoid subtle bugs when different agents use the same data in conflicting ways?

Graph‑backed retrieval layers can help here by giving you a structured way to navigate relationships (who’s related to what, how, and why) instead of repeatedly searching free text or hard‑coding joins.

5. Graphs: A navigation system for AI agents

The kind of context an AI system needs depends heavily on the task; it may range from deep, linear chains for complex reasoning, to broad, branching structures for exploration and discovery, to local clusters of related entities and events, or even a single, precise fact that must be correct.

Jason Liu from Instructor shared in his Context Engineering Series: “Context engineering is designing tool responses and interaction patterns that give agents situational awareness to navigate complex information spaces effectively.”

Traditional relational or document‑only approaches can handle parts of this, but they’re not optimised for navigating connections or relationships.

Graph databases and graph data models offer a different angle: they treat relationships as first‑class citizens. Analysts like Forrester have described graphs as an emerging backbone for LLMs, because they’re well‑suited to capture and expose context to AI systems.

In 2026, expect graphs to show up more often in production AI systems as:

  • A knowledge layer that links metadata, data, documents, events, rules, and tools.
  • A routing layer that helps agents decide: “Where do I go next?”
  • A traceability layer that records how a decision was made and what data it touched.

For agents, this matters a lot. To behave safely and predictably, they need to know:

  • Where are they in the process?
  • What actions are available?
  • What constraints and dependencies exist?
  • What is the downstream impact of a choice?.

Graph structures can encode that as a navigation system. Nodes act as entities, states, tools, or steps. Relationships realise allowed transitions, ownership, risk, and dependencies. And properties give policies, limits, versions, and other metadata.
The result is not that “AI magically understands your business,” but it gives AI a map of your world, rather than just handing it a pile of documents and logs.

6. The database of the future: Adaptive by design

Under all of this sits the data stack, and that’s where a lot of AI projects quietly stall or fail. After a few years of GenAI hype, one uncomfortable truth is clear: we’re asking AI systems to operate at a 2026 scale on data architectures designed 20-30 years ago.

We’ve upgraded hardware, distributed everything, and added cache layers everywhere, but most systems still assume stable schemas and relatively predictable queries. Deterministic behaviour is expected, and systems operate within a small number of “blessed” access paths. And they work best when humans are writing and testing most of the queries.

AI workloads completely break those assumptions: Agents generate dynamic, evolving query patterns, data sources and schemas change rapidly as we connect more systems, and we need to optimise not just for a single workload, but for orchestration across many tools.

The next generation of AI‑native databases and knowledge layers will likely behave more like live code than static infrastructure:

  • Query plans based on flexible schemas that adapt on the fly, borrowing ideas from JIT compilers.
  • Execution that responds to current data distribution, load, and hardware.
  • Feedback loops where each query teaches the system something about what to optimise next.

Graph models fit naturally into this picture, but they’re not the only piece. We’ll see combinations of graph, vector, columnar, and streaming engines, glued together by an adaptive layer in the agent orchestration that chooses which store to query, rewrites queries based on intent and cost and maintains a consistent semantic view of the world for agents and models.


For developers, the practical takeaway for 2026 is:

  • Start treating knowledge and context as a core part of your system design, not an afterthought.
  • Consider how graph structures can provide your agents with a clearer understanding of data, processes, and decisions.
  • Assume that your data layer will need to become more adaptable, interconnected, and transparent over time.

The hype wave is slowing down. The build phase is here. And the systems that win won’t just have the biggest models - they’ll have effective, versatile models using the clearest, most navigable representation of knowledge running underneath.

Here are some resources to dive deeper:

Other interesting articles:
artificial intelligence
See all articles