What does it take to build Agentic AI Workflows?

In my last post, I shared why enterprises need autonomous, agentic workflow–ready data. But knowing why is only half the story. The next question is: what does it actually take to build agentic AI workflows in the enterprise?

The answer starts with business ROI, but it only succeeds if underpinned by the right technology foundation. Let’s understand this journey.

1. Start with Business ROI: Use Cases That Matter

The promise of agentic AI is not abstract; it’s about hard ROI measured in hours saved, costs reduced, compliance improved, and revenue accelerated. Enterprises that embed AI-driven agents into their workflows are already reporting significant returns.

For example:

  • Operational Efficiency & Automation Omega Healthcare implemented AI-driven document processing to handle medical billing and insurance claims at scale. The results: 15,000 hours saved per month, 40% reduction in documentation time, 50% faster turnaround, and an estimated 30% ROI. 👉 Read more
  • Labor Reduction & Productivity Gains In accounts payable, agentic automation is now capable of managing up to 95% of AP workflows—covering exception resolution, PO matching, fraud detection, and payments—dramatically reducing manual overhead. 👉 See details
  • Topline Revenue Growth & Cash Flow Optimization A mid-sized manufacturer deploying AI invoice automation cut manual effort by 60% and reduced invoice approval cycles from 10 days to just 3 days. This improved supplier satisfaction while accelerating cash flow—a direct driver of topline agility. 👉 Learn more
  • Trust & Compliance A South Korean enterprise combined generative AI with intelligent document processing for expense reports, cutting processing time by over 80%, reducing errors, and improving audit compliance—while the system continuously learned from user feedback. 👉 Case study

The business ROI is clear. But ROI only materializes if the technology foundation is strong.

2. The Technology Foundation for Agentic AI

Agentic AI requires more than just a clever model. It needs a robust stack that ensures agents act on data that is accurate, fresh, and explainable. Without this, automation becomes brittle, outputs are untrustworthy, and scaling to new use cases is nearly impossible.

As a founder, I can’t help but see the parallel to building a company. Every wise founder, investor, or YC advisor repeats the same lesson: scaling before product–market fit is risky. You can grow fast in the short term, but without nailing the fundamentals, you eventually hit a wall.

The same is true for AI: scaling before nailing is risky. And here, what needs to be nailed is data infrastructure and AI infrastructure.

You can launch impressive AI pilots, but without reliable data and context, failures show up quickly—hallucinations, inconsistent outputs, compliance gaps.

By investing first in the foundation—valid, explainable, governed data pipelines—you enable AI to scale safely and accelerate into new use cases with confidence.

The foundation determines how fast and how far you can grow.

At the heart of this foundation is the Lakehouse architecture. Why? Because agentic workflows rely on low-latency access to both structured and unstructured data, and the Lakehouse unifies both in open formats like Iceberg, Delta, and Hudi.On top of this foundation, enterprises layer:

  • Purpose-built query engines (Trino, Spark, proprietary engines) that allow federated access to diverse sources.
  • A context layer: governance, lineage, semantics, and—critically—data quality signals.

Just as startups succeed by nailing the core before scaling, AI succeeds by nailing its data and infrastructure foundation before attempting ambitious, agentic workflows.

3. The Three Pillars of Agentic AI Technology

When you strip it down, the technology requirements for agentic AI come down to three pillars: Data, Models, Queries, and Context.

image
Trusted data + lakehouse architecture + context signals = the foundation for agentic AI.

Data

Data is not just about landing rows in a table—it spans the entire lifecycle:

  • Storage in scalable, cost-effective object stores (S3, ADLS, GCS).
  • Transfer across batch or streaming pipelines.
  • Access & Discovery through catalogs and metadata systems.
  • Querying for analytics, training, or real-time decisioning.
  • Validation for freshness, completeness, conformity, and anomalies.
  • Modeling (if needed) into marts or cubes, historically required for every new use case.

This is why the Lakehouse and open formats (Iceberg, Delta, Hudi) fundamentally change the game. In the old world, every new consumer meant another round of transfer → transform → model → consume—bespoke, brittle, and expensive.

With open formats:

  • You dump/land once in the Lakehouse.
  • You consume many times, across engines (SQL, ML, vector search) and contexts (BI dashboard, LLM, agent).
  • You preserve lineage and metadata so every consumer knows not just what the data is, but how trustworthy it is.

This enables zero-copy, zero-ETL architectures—where data is queried in place, and pipelines are replaced by shared, governed access.

Models & Queries

Now that the Lakehouse addresses storage, movement, and transformation, responsibility shifts upward. The old world of pre-building marts and semantic models is giving way to runtime query and modeling.

  • Agents, SQL, and ML/LLMs can dynamically model, filter, and query data at runtime.
  • Runtime query engines (Trino, Spark, Fabric, Databricks SQL) enable federated, ad hoc queries across massive datasets.
  • AI models themselves (LLMs and SMLs) can consume embeddings, metadata, and joins directly to answer questions or trigger actions.
  • Analytical engines like PuppyGraph make complex graph queries over Iceberg tables feasible—without needing a separate graph database.

In short, the Lakehouse stabilizes the base, while agents and models provide runtime intelligence on top.

Context

If Data is the fuel and Models are the engine, Context is the navigation system. It ensures agents don’t just move fast, but move in the right direction.

  • Provided Context: prompts, system instructions, agent-to-agent communication.
  • Derived Context: metadata from lineage, governance, and semantics.
  • Critical Context: runtime reliability signals (freshness, completeness, anomalies).

And this is essential because agents are, by definition, autonomous. With great power comes great responsibility—an agent empowered to act without context can cause more harm than good.

That’s why Telmai focuses here: enriching every dataset with reliability metadata. Agents don’t just know what to do—they know whether it’s safe to act.

4. Industry Alignment: The Lakehouse + Context Story

As enterprises adopt agentic AI, industry leaders are converging on a common foundation: Lakehouse architectures, open query engines, and context-rich catalogs. The direction is clear—data must be unified, governed, and contextualized before agents can act reliably.

  • Microsoft (Fabric, OneLake & Purview): Unified storage and governance. Next Horizon → real-time trust signals for Copilot and Data Agents.
  • Databricks (Delta + Unity Catalog): Open formats and metadata governance. Next Horizon → continuous reliability context for “Agentic BI.”
  • Snowflake (Horizon): Governance and discovery. Next Horizon → runtime reliability metadata.
  • GCP (Dataplex): Metadata-first governance. Next Horizon → embedded reliability checks across streaming.
  • Atlan & Actian + Zeenea: Metadata lakehouse and hybrid catalog tools. Next Horizon → dynamic catalogs enriched with live trust signals.

Across all these ecosystems, the trajectory is clear: governance and semantics are rapidly maturing. The next horizon is weaving in a real-time reliability context.

5. Closing: Building Agentic Workflows on Trusted Data

The lesson is simple:

  • The use cases (automation, efficiency, revenue growth, compliance) are compelling.
  • The foundation (Lakehouse + engines/models + context) is non-negotiable.
  • The pillars (Data, Models, Context) define the architecture.
  • And the Next Horizon is context – especially derived reliability metadata that tells agents whether data is fit for use.

At Telmai, our product path is aligned with this future:

  • MCP server to deliver AI-ready, validated data where agents operate.
  • Support for unstructured data and NLP workflows, so agents can reason across PDFs, logs, and chat.
  • Write–Audit–Publish + DQ binning to automate real-time quarantine of suspicious records.

This is how enterprises will scale agentic AI safely—by building on trusted, validated, context-rich data.

Because in the agentic world, it’s not enough for AI to be smart. It has to be confident.

Want to learn how Telmai can accelerate your AI initiatives with reliable and trusted data? Click here to connect with our team for a personalized demo.

Want to stay ahead on best practices and product insights? Click here to subscribe to our newsletter for expert guidance on building reliable, AI-ready data pipelines.

Why is Model Context Protocol a game-changer for Enterprise AI

The artificial intelligence landscape is fundamentally shifting how enterprises deploy, manage, and govern AI systems. While enthusiasm around large language models and autonomous agents is high, organizations continue to face a persistent challenge in operationalizing AI: enabling models to access high-quality, real-time enterprise data without relying on brittle, hard-coded integrations. Whether it’s a chatbot referencing outdated policies or a model hallucinating due to missing business context, the disconnect between AI and enterprise systems continues to limit trust and effectiveness.

Driving this transformation forward is a new open standard known as the Model Context Protocol (MCP), first launched by Anthropic in November 2024. Just as APIs standardized how applications communicate, MCP is quickly becoming the common protocol for enabling AI to access external tools, structured data, and dynamic enterprise context in a secure and scalable manner.

What exactly is MCP, why is it generating such unprecedented buzz in the AI community, and how does it strengthen enterprise observability and governance by making model behavior more transparent, auditable, and grounded in reliable data context

What is Model Context Protocol (MCP)?

The Model Context Protocol (MCP) is an open, model-agnostic standard designed to help AI systems securely and consistently retrieve structured context from external tools and data sources. Think of it as a shared interface that allows AI agents and enterprise systems to communicate without the need for custom integrations or code.

Instead of building brittle, hardcoded connectors for every new tool or service, developers can implement MCP once and reuse it across environments. This architecture design solves the classic M×N integration problem by replacing a tangled mess of custom APIs with a universal, reusable layer for contextual AI access.

Major AI players like OpenAI, Microsoft, and Google have already committed to supporting MCP, and the open-source ecosystem around it is growing rapidly.

How does Model Context Protocol (MCP) work?

Model Context Protocol Architecture

MCP uses a client-server architecture built on JSON-RPC 2.0, allowing any AI client to communicate with any MCP-enabled service. The AI system (client) connects to MCP servers, each of which wraps around a backend system (CRM, file store, database, etc.) and exposes its capabilities in a structured, machine-readable schema.Whether you’re calling a cloud-based CRM or reading a local CSV file, the interface remains consistent.

This architecture consists of four primary components that work together to create a unified AI ecosystem:

  • Host Applications: These are AI-powered applications like Claude Desktop, AI-enhanced IDEs, or custom enterprise AI agents that users interact with directly
  • MCP Clients: Integrated within host applications, these components manage connections to MCP servers and maintain secure, one-to-one relationships with each server
  • MCP Servers: Lightweight wrappers around systems like GitHub, PostgreSQL, or internal APIs. These expose structured functionality to the AI using the MCP schema
  • Transport Layer: Supports both local (STDIO) and remote (HTTP + Server-Sent Events) communication, allowing MCP to run across cloud and on-prem environments with minimal overhead

Using this architecture, MCP enables AI systems to perform three core operations:

  • Use tools: Trigger specific functions or workflows to perform specific actions, (e.g., look up customer data, run a SQL query)
  • Fetch resources: Retrieve Context and data sources like documents, database entries, or configuration files that provide information without significant computation or side effects, similar to GET endpoints in REST APIs
  • Invoke prompts: Execute pre-defined prompt templates that guide multi-step interactions

How MCP boosts AI capabilities and performance

Before MCP, most LLMs operated in silos, unable to access external tools or live business data without one-off manual integrations. This limited their usefulness in enterprise environments. MCP changes that by offering a consistent protocol for contextual retrieval and action execution. The interface remains consistent whether you’re calling a cloud-based CRM or reading a local CSV file. The performance improvements are measurable across multiple dimensions:

Enhanced Context Awareness: AI Models can query up-to-date business data and respond with grounded, relevant insights, reducing hallucinations and stale information.

Dynamic Tool Discovery: AI systems can discover available tools at runtime and adapt to new workflows without hardcoding, accelerating use-case development.

Reduced Integration Complexity: MCP encourages modularity. With one standard protocol, engineering teams no longer need to maintain multiple connectors, cutting integration time from weeks to hours. Existing MCP servers can be reused across applications, minimizing duplication and improving maintainability.

Together, these capabilities unlock scalable, connected, and more reliable AI systems that can operate securely across diverse enterprise environments.

Reliable AI starts with MCP integrated by Data Observability

Giving AI access to real-time data is powerful but risky without guardrails. AI systems are only as reliable as the data they consume, and without proper monitoring, they may access broken pipelines, outdated schemas, or inconsistent information. Here’s where data observability plays a critical role.

Pairing MCP with a data observability platform like Telmai ensures:

  • Trusted context: Data Observability platforms continuously monitor data quality metrics, ensuring that AI models receive accurate and reliable information
  • Real-Time Monitoring: All data interactions via MCP can be proactively monitored to detect anomalies and issues in data pipelines before they impact AI outputs
  • Policy enforcement: Telmai detects when sensitive data, such as PII, is exposed to AI models via MCP, or when data values accessed in real time drift in ways that violate business rules—enabling proactive safeguards and responsible AI behavior.
  • Faster root cause analysis: Telmai’s data quality binning and circuit breaker features automatically isolate issues and trigger orchestration-level interventions—preventing pipeline failures without requiring heavy engineering effort.

MCP simplifies connectivity between AI and enterprise systems. Data Observability ensures those connections are reliable, secure, and explainable, a critical foundation for scaling trustworthy AI.

Conclusion: Data trust is the real AI multiplier

As AI becomes core to enterprise workflows, success will depend not just on access to real-time context but on ensuring that context is reliable, relevant, and governed.

MCP delivers a powerful, standardized access layer for AI, but access without validation can introduce risk. That’s where data observability comes in. It acts as the guardrail layer that continuously monitors and validates the data accessed by models is accurate, timely, and policy-compliant.

By pairing MCP with data observability tools like Telmai:

  • You ensure AI models are grounded in accurate, up-to-date, and compliant data.
  • You gain end-to-end visibility across your data pipelines — from source to model input.
  • You automate detection and intervention using advanced techniques like data quality binning and circuit breakers.

Together, MCP and observability form the foundation for scalable, secure, and trustworthy AI.

Want to learn how Telmai can accelerate your AI initiatives with reliable and trusted data? Click here to connect with our team for a personalized demo.

Passionate about data quality? Get expert insights and guides delivered straight to your inbox – click here to subscribe