<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[Teradata - Medium]]></title>
        <description><![CDATA[Technical articles for developers using the Teradata Vantage Platform - Medium]]></description>
        <link>https://medium.com/teradata?source=rss----8adfdb056496---4</link>
        <image>
            <url>https://cdn-images-1.medium.com/proxy/1*TGH72Nnw24QL3iV9IOm4VA.png</url>
            <title>Teradata - Medium</title>
            <link>https://medium.com/teradata?source=rss----8adfdb056496---4</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Thu, 07 May 2026 20:31:45 GMT</lastBuildDate>
        <atom:link href="https://medium.com/feed/teradata" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[Build AI, Not Infrastructure: Inside Teradata’s Autonomous Knowledge Platform]]></title>
            <link>https://medium.com/teradata/build-ai-not-infrastructure-inside-teradatas-autonomous-knowledge-platform-01003812eb00?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/01003812eb00</guid>
            <category><![CDATA[analytics]]></category>
            <category><![CDATA[llm]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[agents]]></category>
            <dc:creator><![CDATA[Janeth Graziani]]></dc:creator>
            <pubDate>Thu, 07 May 2026 15:29:22 GMT</pubDate>
            <atom:updated>2026-05-07T15:29:22.379Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="" src="https://cdn-images-1.medium.com/max/690/1*Sy1oeOUpojJ499sqOjC_YQ.png" /></figure><p>There’s a hidden tax you as a data scientist and ML engineer pay every day when working in notebooks, and it starts before the work even begins.</p><p>You spin up a notebook and wait. Five minutes. Ten. Sometimes longer, just to import a library. Other times the environment is already running, but now you’re guessing compute sizes, watching costs creep up, and hoping your experiment doesn’t trigger a FinOps alert.</p><p>And when you need something like GPU access, production data, or updated permissions, it turns into a ticket. Then a sprint. Sometimes two. By the time everything is ready, the momentum is gone. The idea you had at the start of the week? Buried under infrastructure friction.</p><p>This isn’t just an inconvenience. It’s one of the biggest reasons AI projects stall between pilot and production.</p><p>Developers and teams can get models working in notebooks. They can connect data, run experiments, and even show early results. But the moment they try to scale — when workloads become unpredictable, concurrency increases, costs need to be controlled, and production SLAs need to be met — the project starts to slow down.</p><p>What’s missing isn’t another tool. It’s an autonomous AI platform that can anticipate demand, manage resources, and enforce business SLAs. That’s what the Teradata Autonomous Knowledge Platform is designed to do.</p><h4><strong>Core capabilities of the Autonomous Knowledge Platform</strong></h4><p>At its core, the Autonomous Knowledge Platform provides three foundational capabilities:</p><ul><li><strong>Autonomous Tera agent</strong> execution that continuously optimizes performance, cost, and scale across production workloads</li><li>A compute layer with always‑on <strong>active compute</strong> for mission‑critical and agentic workloads, alongside <strong>elastic compute</strong> for on‑demand workloads</li><li>A <strong>connected data foundation</strong> that brings together low-latency local storage, and cost optimized object stores under a single architecture, with support for open table formats and Enterprise Vector Store</li></ul><p>These capabilities operate continuously, and AI Studio is where developers experience them in practice. Let’s explore the platform, Tera agents, and notebooks via AI Studio.</p><h4><strong>AI Studio: the developer workspace for building and scaling AI</strong></h4><p>AI Studio is the unified workspace where developers build, manage, and scale AI outcomes on the Autonomous Knowledge Platform.</p><figure><img alt="AI Studio unified workspace on the Autonomous Knowledge Platform for notebooks, ModelOps, and ModelHub" src="https://cdn-images-1.medium.com/max/1024/0*2Am64E-yvBY8xjqs" /><figcaption>AI Studio: where developers build, manage, and scale outcomes on the Autonomous Knowledge Platform</figcaption></figure><p>It isn’t a replacement for existing Teradata environments. Instead, it runs on top of your Teradata systems without requiring data migration or environment re-creation.</p><p>From AI Studio, developers work with integrated:</p><ul><li><strong>Notebooks</strong> for SQL, Python, and analytics workflow experimentation, development, and collaboration with team members</li><li><strong>ModelHub </strong>to access, monitor, and manage production‑ready models (including embedding and chat models) with visibility into token and cost usage</li><li><strong>ModelOps </strong>to create, manage, and run models at scale</li><li><strong>Vector Store </strong>to store and manage data as vector embeddings for search integration and RAG applications</li><li><strong>Built</strong>‑<strong>in Tera agents</strong> that perform a range of tasks from data analysis to continuously managing infrastructure</li></ul><p>Everything we’ll explore below happens inside AI Studio, running directly on the Autonomous Knowledge Platform.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/318/0*SBHZDByNFnb7SDaf" /><figcaption>A single workspace for notebooks, models, vectors, and agents</figcaption></figure><h4><strong>Tera and autonomous AI agents for production workloads</strong></h4><p>Tera is Teradata’s autonomous AI-powered workspace, serving as the natural language interface with enterprise-grade agent execution environments.</p><p>You can ask Tera to:</p><ul><li>Get information about your system or environment</li><li>Retrieve and analyze schema information, table structures, and column definitions</li><li>Assist in identifying data suitable for visual representation</li></ul><p>Tera operates under existing Teradata permissions, ensuring users only see data they are authorized to access.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1004/0*dUNaWg2_zEjeHQAr" /><figcaption>Tera is a natural language interface with enterprise-grade agent execution environments.</figcaption></figure><p>Tera includes built-in modes for data analysis with Tera Analyze, coding with Tera Code, and multi-agent system automation and orchestration with Tera Claw.</p><h4><strong>How Tera agents automate scaling, cost, and governance</strong></h4><p>Tera agents operate across the spectrum of autonomy from deterministic automation to policy‑governed autonomous actions, within a secure agent harness and runtime.</p><figure><img alt="Tera AI interface showing a prompt about scaling a healthcare claims analytics workload with cost, audit, and performance requirements." src="https://cdn-images-1.medium.com/max/1024/0*5RkUtmL5hf3y3BdQ" /><figcaption>Tera provides a natural-language interface for governed execution and agentic workflows.</figcaption></figure><p>For example, a healthcare claims team runs conversational analytics over three years of data. The pilot proved value. Users love it! Now the challenge begins when moving the pilot into production.</p><p>Traditionally, moving to production means:</p><ul><li>Refactoring notebook‑based code and models to run repeatedly, autonomously, and at scale</li><li>Capacity planning meetings</li><li>Cost modeling sessions</li><li>Tickets for elastic clusters</li><li>Manual tuning as concurrency spikes</li></ul><p>With the Autonomous Knowledge Platform, an administrator expresses their intent to Tera in plain language and the agents handle that work automatically.</p><p><em>“We’re expecting ~300 users for claims and provider data access via MCP, with workload demand varying throughout the day. Monthly compute spend must be under 8,000 units. Enforce access logging on all queries for later audit. Aim for query response under 5 seconds.”</em></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*YDgomWGLWl-QghxY" /><figcaption>Define intent (users, cost ceiling, SLAs); the autonomous knowledge platform proposes a deployment plan.</figcaption></figure><p>Once the administrator approves the proposed deployment, the agents:</p><ul><li>Analyze pilot data around usage, performance, and cost</li><li>Identify usage patterns and inefficiencies</li><li>Recommend elastic compute configurations</li><li>Show projected cost, performance, and impact up front</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*Ifjuf6ob4Kb2_GhD" /><figcaption>Autonomous agents surface tradeoffs up front: projected cost, performance, and impact.</figcaption></figure><p>The required changes within predefined guardrails are executed automatically. Larger or riskier changes wait for human approval.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/954/0*NftZw8InSZ4dxook" /><figcaption>Built-in guardrails help an autonomous knowledge platform automate safely</figcaption></figure><p>Every action is logged. Every decision is auditable.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*XPPIeb6U-TQL5Uyg" /><figcaption>Real-time visibility into auto-scaling, usage, and cost — managed automatically by Tera agents within defined guardrails.</figcaption></figure><p>What this means for you is simple:</p><ul><li>You don’t wait for infrastructure</li><li>You don’t tune clusters</li><li>You don’t negotiate for resources</li><li>You don’t move data out of the platform</li></ul><p>Tera agents give you the ease, speed, and flexibility developers love. While delivering the governance, cost control, and predictability enterprises demand.</p><p>Tera agents remove infrastructure pain points for productionizing AI workloads. Notebooks continue to be where developers explore data, build models, and iterate ideas during development. Agents take over when those ideas move toward production. Once a workflow proves value, platform-level agents handle deployment, scaling, cost controls, and governance, using real telemetry and human-defined intent.</p><h4><strong>Running AI Workloads at Scale with Notebooks + In-Database AI</strong></h4><p>Let’s explore how we build AI solutions with Notebooks in AI Studio. From the <strong>Notebooks </strong>tab in AI Studio, we can start a session and select our desired compute resources. Compute pools are configurable by each organization and can scale to meet the performance, cost, and workload requirements of the business. For example, in my environment I can select from:</p><ul><li><strong>General Compute: </strong>for data exploration, analytics, and in-database functions. This profile runs on up to five nodes with 16 vCPUs and 64 GB of RAM.</li><li><strong>High Memory:</strong> for large datasets and memory‑intensive workloads that require substantial in‑memory processing. This profile contains up to three nodes with 64 vCPUs and 512 GB of RAM.</li></ul><p>In this example, I’ve selected <strong>General Compute</strong> to demonstrate text analytics with native LLMs in ModelHub.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*VQTf4_7yOEM0XLH-" /><figcaption>Start a notebook session and select the right compute pool for your workload.</figcaption></figure><p>Choose a notebook and kernel and begin running Python against data in Teradata.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*Ip3BUCnA1JJPdDOJ" /><figcaption>Run Python/SQL where the data lives without moving data out of the platform</figcaption></figure><p>Below is an example of using Python to run complex text analytic workflows using Teradata’s generative AI package and production-ready LLMs and embedding models.</p><p>First, we establish a connection to the Teradata system.</p><pre> from teradataml import create_context<br><br># ── Connect to Vantage ────────────────────────────────────────────────────────<br>eng = create_context(<br>host=&#39;{db_host}&#39;, # Replace with your database host<br>username=&#39;{db_username}&#39;, # Replace with your database username<br>password=&#39;{db_password}&#39;, # Replace with your database password<br>logmech=&#39;TD2&#39;)</pre><p>We load sample product review data into Teradata with the copy_to_sql() method, which converts a pandas DataFrame into a Teradata table.</p><pre>import pandas as pd<br>from teradataml import copy_to_sql, DataFrame<br><br># ── Sample data: customer feedback with PII embedded in text ──────────────────<br>feedback_pd = pd.DataFrame({<br>&#39;review_id&#39;: [1, 2, 3, 4, 5],<br>&#39;product&#39;: [&#39;SmartWatch X1&#39;, &#39;SmartWatch X1&#39;, &#39;Laptop Pro 15&#39;, &#39;Laptop Pro 15&#39;, &#39;EarBuds Z&#39;],<br>&#39;review&#39;: [<br>&quot;Hi, this is Emily Carter from San Francisco. I love the battery life on the SmartWatch X1 — it lasts nearly 5 days on a single charge. You can reach me at emily.carter@gmail.com if you have questions.&quot;,<br>&quot;I’m Michael Rodriguez in San Jose. The watch looks great, but the strap broke after two weeks. Support asked me to call 408-555-0147, but I haven’t been able to get through yet.&quot;,<br>&quot;Feedback from Sarah Nguyen: the Laptop Pro 15 has blazing-fast performance and a stunning display. My receipt was emailed to snguyen@outlook.com.&quot;,<br>&quot;This is David Thompson (dthompson@company.com). I use this laptop for heavy development workloads, and it gets very hot. The fan noise is pretty distracting during long builds.&quot;,<br>&quot;Jessica Lee here. The EarBuds Z have crystal-clear audio and excellent noise cancellation. Feel free to text me at 310-555-0168 if you’d like more detailed feedback.&quot;<br>]<br>})<br><br>copy_to_sql(<br>feedback_pd,<br>table_name=&#39;customer_feedback&#39;,<br>if_exists=&#39;replace&#39;,<br>index=False<br>)<br><br>df = DataFrame(&#39;customer_feedback&#39;)<br>print(&#39;Sample data loaded into Vantage:&#39;)<br>df</pre><p>Here’s a preview of the data we just loaded, using a TeradataML DataFrame. This DataFrame represents a structured dataset that now resides on our analytic platform.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*d8ncc1gSzjYqIZXZ" /><figcaption>Customer review data loaded into Teradata, ready for large-scale text analytics and PII masking.</figcaption></figure><p>A TeradataML DataFrame can reference a table, a view, or even a complex query spanning open table formats and object storage. These datasets may range from thousands to billions of rows and often represent data joined across hundreds of tables.</p><p>While interacting with a DataFrame feels local, all computation is executed on Teradata’s massively parallel analytic cluster — enabling fast, scalable operations on data at any size.</p><p>We can now demonstrate how to run large-scale text analytics like masking sensitive data, analyzing sentiment, extracting key phrases, translating text, and generating summaries using models available in ModelHub.</p><p>Let’s start by configuring our connection to ModelHub and selecting the LLM we want to run tests with.</p><p>To begin running natural language processing operations with LLMs, we need a ModelHub endpoint, ModelHub key, and model name.</p><pre>from teradatagenai import TeradataAI<br>from teradatagenai.text_analytics import TextAnalyticsAI<br>from teradataml import DataFrame<br><br># ── Configuration ─────────────────────────────────────────────────────────────<br># Replace with the endpoint URL and API key copied from AI Modelhub<br>MODELHUB_ENDPOINT = &#39;{your_endpoint_url}&#39; # e.g., &#39;https://your_instance.teradata.com/one-td/litellm/v1/chat/completions&#39;<br>MODELHUB_API_KEY = &#39;{your_api_key}&#39; # Replace with your API key<br>MODEL_NAME = &#39;{model_name}&#39; # as shown in Modelhub<br><br>llm = TeradataAI(<br>api_type=&quot;nim&quot;,<br>model_name=MODEL_NAME,<br>api_base=MODELHUB_ENDPOINT,<br>api_key=MODELHUB_API_KEY<br>)<br><br>analytics = TextAnalyticsAI(llm=llm)</pre><p>Select the <strong>ModelOps </strong>tab from the left bar menu.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/318/0*OV6aq4Rt18UFxJgG" /><figcaption>Use ModelOps to move from notebook experimentation to governed deployment</figcaption></figure><p>This will open ModelOps in a new tab. ModelOps is Teradata’s model lifecycle management capability, enabling data scientists and ML engineers to train, evaluate, deploy, monitor, and retrain task‑specific ML models. Models developed in notebooks can move seamlessly into production with unified governance, observability, and auditability. The same lifecycle controls apply to LLMs, which is why ModelOps integrates seamlessly with AI ModelHub to manage both traditional ML and generative AI within a single platform.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*uotzOxwrC6A1UfJP" /><figcaption>Operationalize and manage machine learning models at scale with ModelOps and integrated AI Model Hub.</figcaption></figure><p>Select <strong>Open </strong>in the bottom fold to explore the AI model hub, and enter your virtual API key to view the models your user has access to.</p><p><strong>AI ModelHub</strong></p><p>The AI ModelHub in AI Studio is a catalog of production-ready AI models, including LLMs, embedding models, and domain-specific models deployed and served within your Teradata environment as well as models from cloud service providers.</p><p>These models are exposed via LiteLLM and are designed to be accessed from Python using the teradatagenai package.</p><p>These same model endpoints can be reused for generative and agent-driven workflows. Both built-in Tera agents and customer-defined agents can invoke these endpoints, enabling centralized management, consistent access policies, and visibility into model usage across the platform.</p><p>This design keeps all inference inside your secure environment. No data leaves the platform, and no external API keys or internet access is required.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*m5-_Tq0X-liOwa_8" /><figcaption>Browse and manage production-ready LLMs from Model Hub with centralized access, governance, and usage visibility.</figcaption></figure><p>Select a model card to view the required model name and the endpoint in the overview page.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*oh5P3dQBtwx21wf3" /><figcaption>ModelHub model card details</figcaption></figure><p>With the proper credentials, we can now run our operation using teradatagenai. teradatagenai provides TextAnalyticsAI, a higher-level class that wraps the LLM to run text analytics operations directly on Teradata DataFrames. These operations combine the language model with your data at scale, pushing results back into Teradata without moving data to the notebook.</p><p>Supported operations include:</p><ul><li>analyze_sentiment() — classify emotional tone as positive, negative, or neutral</li><li>extract_key_phrases() — identify the most important terms in each text</li><li>summarize() — condense long text into a concise summary</li><li>translate() — convert text between languages</li><li>mask_pii() — redact personally identifiable information</li></ul><pre>masked_data = analytics.mask_pii(<br>data = df,<br>column = &#39;review&#39;,<br>id_col = &#39;review_id&#39;<br>)<br>masked_data</pre><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*ZDTCRtHTFuG4MIie" /><figcaption>Run governed text analytics in-database and return results as tables.</figcaption></figure><p>Because this is running in database, the results will come back as a table just like other analytic results.</p><p>The important thing we just showed here is that we executed an LLM inside Teradata and against enterprise data — without leaving the platform, and with full security. This is an example of a first-class AI capability embedded in your data platform.</p><p>This is what it means to build AI in production without managing infrastructure. Developers work in notebooks as usual, while Tera agents handle execution, optimization, and governance when it’s time to take your workloads into production.</p><p>Request a demo of AI Studio by visiting <a href="https://www.teradata.com/about-us/contact">https://www.teradata.com/about-us/contact</a>.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=01003812eb00" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/build-ai-not-infrastructure-inside-teradatas-autonomous-knowledge-platform-01003812eb00">Build AI, Not Infrastructure: Inside Teradata’s Autonomous Knowledge Platform</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Building Smarter AI Agents for Data Science Workflows at Scale]]></title>
            <link>https://medium.com/teradata/building-smarter-ai-agents-for-data-science-workflows-at-scale-174fd51bf66b?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/174fd51bf66b</guid>
            <category><![CDATA[mcp-server]]></category>
            <category><![CDATA[agentic-ai]]></category>
            <category><![CDATA[skills]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[data-science]]></category>
            <dc:creator><![CDATA[Janeth Graziani]]></dc:creator>
            <pubDate>Fri, 01 May 2026 19:15:25 GMT</pubDate>
            <atom:updated>2026-05-01T19:15:24.591Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="Screenshot showing an AI agent using tdsql MCP to guide in‑database time series analysis, with TD ARIMA model diagnostics generated alongside a dashboard forecasting air passenger demand at scale." src="https://cdn-images-1.medium.com/max/1024/1*1TJvRehCqU5SqRoBTLF4vA.png" /></figure><h3>Building Smarter AI Agents for Data Science Workflows on Enterprise Data Platforms</h3><p>How do we get AI agents to actually operate the right way at scale across <a href="https://www.teradata.com/platform">data platforms</a>, whether that’s a lakehouse, data lake, or data warehouse?</p><p>If you’ve tried connecting large language models, custom agents, or agentic systems to your data platform you’ve probably already run into a few familiar problems. The agent generates SQL that <em>technically</em> runs but it’s wildly inefficient. It pulls millions or billions of rows out of the database just to do local processing instead of using the platforms <a href="https://www.teradata.com/platform/clearscape-analytics/in-database">in-database processing engine</a>. Or worse, it confidently hallucinates platform specific functions and results that were never computed.</p><p>So the challenge becomes:</p><p><strong>How do we force AI systems and agents to work the <em>right way</em> on specific data platforms at scale?</strong></p><p>That’s exactly what Kevin Sturgeon, Director of Cloud Engineering at Teradata, set out to solve in his latest work. In our recent <a href="https://youtu.be/ecAdqImEH3U?si=5yKfO0LjJ3lckHWn&amp;t=10">Teradata DevTalk</a>, Kevin walked through how he designed an <strong>open‑source MCP server</strong> with a <strong>Skills‑based architecture</strong>, and demonstrated how context‑aware agents using <a href="https://docs.claude-mem.ai/progressive-disclosure">progressive disclosure</a> can improve data science and analytics workflows, speed up onboarding for new data scientists to a specific platform, and reduce costly inefficiencies at scale. More specifically Kevin used MCP as a transport layer and repackaged Teradata expertise (Teradata documentation, best practices, and SQL patterns) as injectable skills, he created a system where agents learn <em>how to work correctly</em> on a platform without prompt overload, filesystem coupling, or framework lock‑in.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FecAdqImEH3U%3Ffeature%3Doembed%26start%3D10&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DecAdqImEH3U&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FecAdqImEH3U%2Fhqdefault.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/09a2011c23ef26b5c28782ee661ff70e/href">https://medium.com/media/09a2011c23ef26b5c28782ee661ff70e/href</a></iframe><p>This post is a quick recap of what we covered in our DevTalk. All of the projects discussed here are available as open source for anyone to explore, and the mcp + skills architecture itself is not tied to Teradata it can be applied to any data platform.</p><ul><li><strong>GitHub Repo: tdsql MCP Server:</strong> <a href="https://github.com/ksturgeon-td/tdsql-mcp/blob/main/README.md">https://github.com/ksturgeon-td/tdsql-mcp/blob/main/README.md</a></li><li><strong>GitHub Repo: LangGraph tdsql Agent:</strong> <a href="https://github.com/ksturgeon-td/tdsql-agent">https://github.com/ksturgeon-td/tdsql-agent</a></li><li><strong>Teradata Free Trial:</strong> <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics">https://www.teradata.com/getting-started/demos/clearscape-analytics</a></li></ul><h4>The Core Problem with Today’s Agentic AI Workflows</h4><p>Modern LLMs are incredibly good at reasoning. They understand <a href="https://www.teradata.com/insights/data-analytics/descriptive-statistics">statistical concepts</a>, <a href="https://www.teradata.com/insights/ai-and-machine-learning/machine-learning-models">machine learning algorithms</a>, and <a href="https://www.teradata.com/insights/ai-and-machine-learning/building-agentic-workflows-and-systems">analytical workflows</a>. They know what ARIMA models are. They know how feature engineering works. They know how to evaluate models.</p><p>What they <strong>don’t</strong> know is how to do those things <em>well</em> on your data platform.</p><p>When you connect an agent to a database without enough context, a few things tend to happen:</p><ul><li>The agent writes generic or inefficient SQL</li><li>It defaults to client-side processing (pulling data into notebooks or local environments)</li><li>It guesses at functions that don’t exist</li><li>It makes up accuracy metrics or results</li></ul><p>This isn’t a Teradata specific issue it’s an industry-wide problem. Agents aren’t incentivized to minimize data movement, reduce cost, or use platform-native functions unless we explicitly guide them.</p><p>And at scale, those mistakes get expensive fast.</p><h4>Letting Agents Do What They’re Good At (and Platforms Do the Rest)</h4><p>One of the themes Kevin kept coming back to is:</p><blockquote>Let agents reason.<br>Let data platforms execute.</blockquote><p>Teradata has spent decades solving a problem that many open-source ecosystems still struggle with: <strong>how to operationalize complex analytics at speed and massive scale</strong>.</p><p><a href="https://docs.teradata.com/r/Enterprise_IntelliFlex_VMware/Database-Unbounded-Array-Framework-Time-Series-Functions/Series-Forecasting-Functions">Time series forecasting</a>, <a href="https://www.youtube.com/watch?v=Claa4iXHr0U">geospatial analysis</a>, <a href="https://www.teradata.com/insights/ai-and-machine-learning/machine-learning-models">machine learning</a>, <a href="https://docs.teradata.com/r/Enterprise_IntelliFlex_VMware/Database-Engine-20-In-Database-Analytic-Functions/Model-Evaluation-Functions">model evaluation</a>, these aren’t new problems. Teradata already has <strong>hundreds of native, </strong><a href="https://docs.teradata.com/r/Enterprise_IntelliFlex_VMware/Database-Analytic-Functions"><strong>highly optimized analytical functions</strong></a><strong> </strong>designed to run directly where the data lives.</p><p>The challenge is helping agents <em>discover</em> and <em>use</em> those capabilities without stuffing the entire <a href="https://docs.teradata.com/">Teradata documentation</a> into a prompt.</p><h4>Progressive Context for AI Agents (Not Prompt Overload)</h4><p>One solution people have been experimenting with is “skills”, text-based instructions that tell an agent <em>how</em> to do things correctly.</p><p>That works… up to a point.</p><p>Skills tend to be:</p><ul><li>File-system based</li><li>Hard to reuse across tools</li><li>Limited to certain clients (like Claude Desktop)</li></ul><p>Kevin identified the need for an additional architectural layer that helps large language models retrieve <em>the right</em> information about how to work on a platform <em>when</em> they need it rather than relying solely on static instructions.</p><p>That layer is the <strong>Model Context Protocol (MCP)</strong>, an open‑source standard introduced by Anthropic in late 2024.</p><p>MCP allows agents to dynamically request tools, documentation, and behavioral guidance as they work, instead of carrying all context upfront in a single prompt. You can think of it as <strong>progressive disclosure for agent intelligence </strong>providing context on demand, based on what the agent is actively trying to do.</p><p>In the open‑source <strong>tdsql MCP server</strong>, Kevin combined MCP with a Skills‑based approach by doing the heavy lifting of translating Teradata documentation, best practices, and SQL patterns into a reusable Teradata skills library.</p><p>These skills are then exposed as injectable MCP tools, making them discoverable and callable by agents at runtime.</p><p>For example, the tdsql MCP server exposes platform‑verified SQL and analytics guidance as a callable MCP tool instead of static prompt instructions:</p><pre>@mcp.tool()<br>def get_syntax_help(topic: str = &quot;index&quot;) -&gt; str:<br>    &quot;&quot;&quot;Return Teradata SQL syntax reference for a given topic.<br><br>    IMPORTANT: Call this tool BEFORE writing any analytics, transformation, ML, or data<br>    preparation SQL. Teradata Vantage has native distributed table operators for most<br>    operations — scaling, encoding, binning, statistics, clustering, classification, text<br>    analytics, vector search, and more. These outperform hand-written SQL and should always<br>    be preferred. Do not write manual SQL for an operation if a native function exists.<br><br>    Recommended call order:<br>      1. get_syntax_help(topic=&#39;guidelines&#39;) — see the canonical mapping of common SQL<br>         patterns to native Teradata functions (start here if unsure what exists)<br>      2. get_syntax_help(topic=&#39;index&#39;) — browse all available topics and the Workflows<br>         section that maps use cases to topic sequences<br>      3. get_syntax_help(topic=&#39;&lt;specific-topic&gt;&#39;) — load exact syntax for a topic<br><br>    Args:<br>        topic: The topic name (e.g. &#39;data-prep&#39;, &#39;ml-functions&#39;, &#39;vector-search&#39;).<br>               Use &#39;index&#39; to list all available topics.<br>               Use &#39;guidelines&#39; for the native-functions-first reference.<br><br>    Returns:<br>        Markdown reference text for the requested topic, or a list of valid topics<br>        if the requested topic is not found.<br>    &quot;&quot;&quot;</pre><p>Instead of embedding long instructions directly in prompts, agents can now retrieve structured, platform‑aware guidance at runtime.</p><figure><img alt="Visual Studio Code editor view showing the tdsql MCP skills library with Teradata SQL analytics guidelines, highlighting rules that instruct AI agents to prefer native in‑database functions over hand‑written SQL to minimize data movement at scale." src="https://cdn-images-1.medium.com/max/1024/1*Dowa1qQ13E9a-8r0m2zP6Q.png" /></figure><h4><strong>Why This Matters</strong></h4><p>As a result, the agent workflow changes fundamentally. Rather than guessing or overloading its prompt, the agent can:</p><ul><li>Check platform guidelines before generating SQL</li><li>Look up supported syntax and native analytical functions</li><li>Apply Teradata-specific best practices for scale and execution</li><li>Build context incrementally as it reasons through a task</li></ul><p>This keeps prompts smaller, focuses the agent’s attention, and dramatically reduces inefficient queries and hallucinated behavior while still allowing the agent to do what it’s best at: reasoning and decision-making.</p><p>This pattern using MCP as a transport layer and delivering platform expertise as injectable skills is emerging across the industry. Several vendors and frameworks now explicitly recommend combining MCP and Skills. What differentiates Kevin’s work is how far this idea is pushed: fully removing filesystem coupling, aggressively enforcing progressive disclosure, and adapting deep analytic platform knowledge into a reusable MCP‑delivered skills library.</p><h4>What This Enables in Practice for AI Agents on Data Platforms</h4><p>Once you give an agent structured, platform‑aware context, some interesting things start to happen.</p><p>And when combined with interactive chat environments like <strong>Claude Desktop</strong>, <strong>ADK</strong>, or <strong>Goose</strong>, this becomes a powerful way to both ground agent reasoning and maintain a direct connection to the database.</p><p>Kevin demonstrated a few advanced analytical examples with Claude Desktop first.</p><h4>1. Ad Hoc Analytics That Actually Scale</h4><p>From a simple natural language prompt like <em>“perform time series analysis,”</em> the agent will load the mcp tools and syntax:</p><ul><li>Discover relevant tables</li><li>Assess seasonality and stationarity</li><li>Run native Teradata time-series functions</li><li>Compare multiple forecasting models</li><li>Evaluate results at scale</li></ul><figure><img alt="Screenshot of an AI agent orchestrating a time series analysis pipeline using Teradata native functions, showing MCP guided steps for ARIMA modeling, validation, and forecasting without moving data out of the platform." src="https://cdn-images-1.medium.com/max/1024/1*D77fnzUIWgb0Y5kfKh7IUQ.png" /></figure><p>All <strong>without</strong> pulling raw data out of the database.</p><figure><img alt="Screenshot showing an AI agent executing a Teradata native time series analysis pipeline, with ARIMA model diagnostics and dashboard displaying air passenger demand and a 24 mo‑period forecast." src="https://cdn-images-1.medium.com/max/1024/1*BTfbQnFGwTNl2S9XugT6lQ.png" /></figure><h4>2. From Experimentation to Operational Pipelines</h4><p>One of my favorite parts of the demo was this task:</p><p>After exploring the data and models interactively, we asked the agent:</p><blockquote><em>“How do I operationalize this?</em></blockquote><p>Instead of stopping at basic analysis the agent:</p><ul><li>Tuned and optimize SQL queries</li><li>Created reusable views and model tables</li><li>Produced a stored procedure pattern that can be used and scheduled</li></ul><figure><img alt="AI agent workflow showing an MCP guided Teradata forecasting pipeline, including model artifacts, validation tables, and a production dashboard describing an ongoing ARIMA based forecast process." src="https://cdn-images-1.medium.com/max/1024/1*XxLHf3NUIOTjXvNM4hxyPg.png" /></figure><p>This is the part that usually breaks down in data science workflows getting from a notebook to something reliable and scalable. The agent doesn’t magically solve that, but it <strong>accelerates the path</strong> dramatically.</p><p>Kevin also demoed the versatility of his progressive disclosure MCP tool by connecting to Claude Code attached to VSCode</p><h4>3. AI Code Generation Tools That Learn From Its Mistakes</h4><p>Even with good context, agents still make errors. That’s expected.</p><figure><img alt="AI agent workflow in a code editor showing MCP‑guided SQL execution, data profiling, and platform aware syntax validation for Teradata in‑database analytics." src="https://cdn-images-1.medium.com/max/1024/1*cAh8Ehm1zehOYfbQznEbXA.png" /></figure><p>The difference here is that:</p><ul><li>The agent can explain failed queries</li><li>Look up correct syntax</li><li>Retry with improved SQL</li><li>Optimize execution patterns</li><li>Produce .sql file that can be ran and verified</li></ul><p>You’re not just getting generated code you’re getting an <strong>iterative collaborator</strong> that can reason about its own outputs.</p><p>And because MCP is model-agnostic, this works across tools like Claude, Copilot, Gemini, and others that support the mcp protocol.</p><h3>What This Is (and What It Is Not)</h3><p>It’s important to be very clear about this.</p><p>This does <strong>not</strong> replace:</p><ul><li>Data scientists</li><li>Data engineers</li><li>Validation, testing, or governance workflows</li></ul><p>Hallucinations are still possible. Even with strong context injection, LLMs are not 100% deterministic.</p><p>The right mindset is <strong>trust, but verify</strong>.</p><p>What this <em>does</em> give you is:</p><ul><li>Faster onboarding to platform-specific analytics</li><li>Safer, more efficient agent behavior</li><li>A huge productivity boost for exploratory and operational analytics</li></ul><h3>Getting Started</h3><p>The MCP server and examples Kevin demoed are available on GitHub as an open-source project. There’s nothing proprietary here it’s a reorganization of Teradata’s best practices into a form agents can actually use.</p><p>You can also try this yourself using a Free <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics">Teradata Trial,</a> which provides an evaluation environment with sample datasets and analytics capabilities ready to go.</p><p>Plug the two together, point an agent at your data, and start exploring.</p><p><strong>Resources:</strong></p><ul><li><strong>GitHub Repo: tdsql MCP Server:</strong> <a href="https://github.com/ksturgeon-td/tdsql-mcp/blob/main/README.md">https://github.com/ksturgeon-td/tdsql-mcp/blob/main/README.md</a></li><li><strong>GitHub Repo: LangGraph tdsql Agent:</strong> <a href="https://github.com/ksturgeon-td/tdsql-agent">https://github.com/ksturgeon-td/tdsql-agent</a></li><li><strong>Teradata Free Trial:</strong> <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics">https://www.teradata.com/getting-started/demos/clearscape-analytics</a></li></ul><h3>Final Takeaway</h3><p>AI agents aren’t going away but poorly structured agents can be dangerous, expensive, and misleading at scale.</p><p>The future isn’t about smarter prompts. It’s about <strong>smarter context</strong>.</p><p>By grounding agents in platform-native capabilities and guiding <em>how</em> they reason not just <em>what</em> they output we can finally make agentic workflows practical for real-world data science.</p><p>If you watched the DevTalk live, this article should help you slow things down and unpack the details. If you missed it, now you know why this space is worth paying attention to.</p><p>More soon.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*QgOMbDHsmm8gvqRX15wKUw.png" /><figcaption>Teradata Monthly DevTalks</figcaption></figure><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=174fd51bf66b" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/building-smarter-ai-agents-for-data-science-workflows-at-scale-174fd51bf66b">Building Smarter AI Agents for Data Science Workflows at Scale</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Carbon Footprint Analytics: a data integration story]]></title>
            <link>https://medium.com/teradata/carbon-footprint-analytics-a-data-integration-story-d9f37e95f2d7?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/d9f37e95f2d7</guid>
            <category><![CDATA[data-analysis]]></category>
            <category><![CDATA[ghg-emissions]]></category>
            <category><![CDATA[esg]]></category>
            <category><![CDATA[advanced-analytics]]></category>
            <category><![CDATA[climate-change]]></category>
            <dc:creator><![CDATA[Gregory Leduc]]></dc:creator>
            <pubDate>Mon, 13 Apr 2026 14:50:47 GMT</pubDate>
            <atom:updated>2026-04-13T14:30:40.598Z</atom:updated>
            <content:encoded><![CDATA[<p>Measuring GHG emissions is a data integration challenge. See how Teradata’s CFA solution accelerator tackles it.</p><figure><img alt="Galina Nelyubova
 Earth Day — a collection of environmental 3D illustrations" src="https://cdn-images-1.medium.com/max/975/1*gXIDV8mS_xphxbw0XDSGYg.png" /></figure><p><em>Galina Nelyubova ; Earth Day — a collection of environmental 3D illustrations</em></p><p>Climate emergency intensifies, and companies must play a significant role in ensuring a sustainable future. As climate change is one of the biggest systemic risks, mitigation of adverse financial effects and adaptation to rapidly changing circumstances is key for companies to ensure long-term going concern and growth. Centerpiece for these activities is proper reporting and reducing of companies’ greenhouse gas (GHG) emissions. And doing so requires much more than good intentions: it demands robust data integration, and the right platform.</p><h4><strong>Why Carbon footprint analytics matters</strong></h4><p>Every company’s operations — from logistics and manufacturing to IT and facilities — generate emissions. These emissions are environmental concerns, and they can also become a financial and reputational risk. Regulatory bodies are moving toward mandatory enterprise-wide GHG reporting, and stakeholders are demanding transparency and action.</p><p>To meet these expectations, companies must answer critical questions:</p><p>· How much GHG emissions are we responsible for?</p><p>· Which activities contribute the most?</p><p>· What’s our roadmap to a sustainable level of GHG emissions? (Ones would say “path to net-zero”, but let’s focus on reduction first before compensating).</p><p>The answers lie within the data companies already possess: financial systems, operational logs, energy consumption records… All these contain the clues needed to build a comprehensive GHG emissions profile. And here is precisely the challenge most companies are facing: transforming existing data into actionable insights.</p><figure><img alt="Figure 1 — Calculating a carbon footprint is easy… until your start digging into the details!" src="https://cdn-images-1.medium.com/max/661/1*gPTBQxdqaN6fy4vYg3v9ew.png" /></figure><p><em>Figure 1 — Calculating a carbon footprint is easy… until your start digging into the details!</em></p><h4>Why it’s a data integration story</h4><p>Carbon footprint analytics is fundamentally a data integration challenge. As the saying goes, “What gets measured, gets managed”. And measuring emissions accurately means harmonizing data from disparate sources, each with its own format, granularity, units, and quality. Trying to report GHG emissions without data integration is like assembling a jigsaw puzzle where pieces are taken from different puzzles: You’ll never see the full picture.</p><p>The Carbon Footprint Analytics (CFA) solution accelerator designed by Teradata tackles this complexity challenge. Based on a flexible data model, it ingests enterprise-specific data using proven algorithms, and outputs detailed GHG emissions across all scopes (1, 2 and 3) defined by the <a href="https://ghgprotocol.org/">GHG Protocol</a>.</p><p>Key capabilities include:</p><p>· <strong>Unit harmonization:</strong> automatically converts measures to configurable preferred units (pounds vs kilograms, kWh vs MWh, etc.)</p><figure><img alt="Figure 2 — Automated management of measure systems heterogeneity" src="https://cdn-images-1.medium.com/max/704/1*_MOnELgq78QslHRBBj064Q.png" /></figure><p><em>Figure 2 — Automated management of measure systems heterogeneity</em></p><p>· <strong>Geospatial awareness:</strong> identifies the right emission factors based on location, which is particularly critical for electricity consumption (electricity mix, thus GHG emissions, significantly differ <a href="https://app.electricitymaps.com/map/all/monthly">between countries and regions</a>)</p><p>· <strong>Temporal modeling:</strong> the CFA design reflects evolving emission factors and activity periods.</p><p>A key point to understand is that CFA is not a reporting tool. It is an engine in which we can connect reporting tools. This engine integrates and enriches data to deliver a harmonized, enterprise-wide view of emissions.</p><h4>Think before you jump: design matters</h4><p>Before diving into carbon analytics, companies must design a robust and flexible data model. A poorly structured model can lead to wasted efforts and inconsistent results.</p><p>The CFA solution accelerator emphasizes thoughtful architecture:</p><p>· A core data model that supports multiple aggregation levels and units</p><p>· Pre-integrated calculation rules for seamless processing and explainable results</p><p>· Advanced analytics capabilities, including machine learning, geospatial, time-series and string similarities.</p><figure><img alt="Figure 3 — Overview of the CFA data model, designed for scope 1, 2 and 3 GHG emissions" src="https://cdn-images-1.medium.com/max/975/1*nT9Z8eWEynEGFAvll72CNw.png" /></figure><p><em>Figure 3 — Overview of the CFA data model, designed for scope 1, 2 and 3 GHG emissions</em></p><p>This foundation ensures that companies can not only report emissions, but also forecast, simulate and optimize their GHG emissions reduction strategy.</p><h4>Why Teradata is the right platform</h4><p>Teradata Vantage is well positioned to support Carbon Footprint Analytics:</p><p>· <strong>Geospatial capabilities</strong>: visualize GHG emissions and apply location-specific factors.</p><figure><img alt="Figure 4 — Geospatial capabilities facilitate segregation of GHG emissions according to location" src="https://cdn-images-1.medium.com/max/700/1*1zHzQzNv2UL5vrF--gvr4A.png" /></figure><p><em>Figure 4 — Geospatial capabilities facilitate segregation of GHG emissions according to location</em></p><p>· <strong>Temporal intelligence</strong>: track changes over time with time-aware tables.</p><p>· <strong>Advanced analytics</strong>: leverage ClearScape Analytics™ features to observe the trends and to simulate the effect of your strategic decisions for multiple scenarios.</p><p>· <strong>Scalable integration</strong>: harmonize data from multiple sources with high performance and reliability.</p><figure><img alt="Figure 5 — Multiple data sources must be efficiently integrated, both internal and external" src="https://cdn-images-1.medium.com/max/668/1*QQfSGE-14XddC6sXGakWgQ.png" /></figure><p><em>Figure 5 — Multiple data sources must be efficiently integrated, both internal and external</em></p><p>The <a href="https://clearscape.teradata.com">ClearScape Analytics Experience</a> demo showcases CFA in action — transforming raw data into <a href="https://nbviewer.org/github/Teradata/jupyter-demos/blob/85d6dd55cfe0e8b509e536bb4e61d9e45a2cb5c9/UseCases/Carbon_Footprint_Analytics/Pop_Carbon_Footprint_Analytics_PY_SQL.ipynb">meaningful carbon insights</a>. So, what are you waiting for? Deploy your own CFA solution accelerator, with Teradata ready to support you every step of the way.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=d9f37e95f2d7" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/carbon-footprint-analytics-a-data-integration-story-d9f37e95f2d7">Carbon Footprint Analytics: a data integration story</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Building Multimodal Agentic Search Over Images and Documents — A DevTalk]]></title>
            <link>https://medium.com/teradata/building-multimodal-agentic-search-over-images-and-documents-a-devtalk-fdf1bf3029e5?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/fdf1bf3029e5</guid>
            <category><![CDATA[retrieval-augmented-gen]]></category>
            <category><![CDATA[unstructured-data]]></category>
            <category><![CDATA[teradata]]></category>
            <category><![CDATA[image-recognition]]></category>
            <category><![CDATA[multimodal-ai]]></category>
            <dc:creator><![CDATA[Daniel Herrera]]></dc:creator>
            <pubDate>Thu, 09 Apr 2026 12:55:14 GMT</pubDate>
            <atom:updated>2026-04-09T12:55:13.358Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="" src="https://cdn-images-1.medium.com/max/480/1*mNnGddOhjpio4Mw-tnHa3w.jpeg" /><figcaption>Teradata DevTalk — Multimodal Agentic Semantic Search</figcaption></figure><p>Recently I had the opportunity to host a <a href="https://www.youtube.com/live/G_KXqo8jXuc">DevTalk</a> that focused on something many teams are actively experimenting with right now: how to reason over unstructured data using multimodal embeddings and agentic retrieval patterns.</p><p>Together with <a href="https://www.linkedin.com/in/ajaykrish/">Ajay Gopalan</a> from <a href="https://unstructured.io/">Unstructured.io</a>, we walked through a live, <a href="https://github.com/Teradata/jupyter-demos/blob/main/VantageCloud_Lake/UseCases/Multimodal_Agentic_Semantic_Search/Agentic_Semantic_Search_Teradata_Unstructured.ipynb">end‑to‑end demo</a> that shows how images, documents, and text can be brought together. The goal start from raw unstructured data and end with an experience where you can select an image, ask questions about it, and receive answers grounded in real enterprise documentation.</p><p>The full session is available on YouTube, and if you are working with vector search, multimodal models, or agentic workflows, it is worth watching the demo in full.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FG_KXqo8jXuc%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DG_KXqo8jXuc&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FG_KXqo8jXuc%2Fhqdefault.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/79e5bd6086933a7020f876e3cdf68e87/href">https://medium.com/media/79e5bd6086933a7020f876e3cdf68e87/href</a></iframe><p>Below is a summary of what we covered and why it matters:</p><h4>Why Multimodal Search Keeps Coming Up</h4><p>Images exist everywhere in enterprise environments. They appear inside PDFs, research documents, medical records, financial reports, and technical documentation. On their own, images rarely provide enough context to answer useful questions. The surrounding text, tables, and metadata usually carry the explanation.</p><p>The challenge is connecting those pieces in a way that scales.</p><p>During the session, we framed the problem around a simple interaction. A user selects an image and asks a question. The system retrieves the relevant documentation that explains that image and uses it to produce a grounded answer. That interaction requires coordinated ingestion, embedding, storage, retrieval, and reasoning across multiple data modalities.</p><h4>Structuring Unstructured Data with Unstructured</h4><p>We started with ingestion, using Unstructured to process a mix of composite PDFs and standalone image files. Ajay walked through how Unstructured workflows break documents into structured elements such as text blocks, tables, and images. These elements are then enriched and prepared for downstream use.</p><p>One point that came up during the demo is how useful the Unstructured UI is during experimentation. You can design and test workflows visually, refine parsing strategies, and then transition to the API when you are ready to automate ingestion at scale. The same pipeline is used in both cases.</p><h4>Storing and Searching Embeddings in Teradata VantageCloud</h4><p>Once the documents and images are processed, embeddings are generated and stored inside Teradata using Enterprise Vector Store capabilities. This keeps vector search close to the rest of the enterprise data, with the same governance, access control, and operational model teams already rely on.</p><p>In the demo, we created a vector store from the generated embeddings and preserved metadata such as file names and record locators. That metadata becomes important later, when retrieved results need to be traced back to their original documents.</p><p>From a data platform perspective, this aligns well with how many teams already think about indexing. The mental model remains focused on accelerating discovery and retrieval, with similarity used as the matching mechanism instead of exact values.</p><h4>Adding Agentic Reasoning on Top of Retrieval</h4><p>Retrieval alone gets you relevant context. Agentic workflows turn that context into an interactive experience.</p><p>Using LangChain integrations with Teradata, we built a small set of tools that the agent can use. One tool performs similarity search over embeddings to locate the most relevant documentation for a selected image. Another tool retrieves and displays the source document associated with that result.</p><p>The agent prompt describes how these tools should be used together. When a user selects an image and asks a question, the agent retrieves the closest matching documentation and generates an answer grounded in that evidence. The reasoning step stays tied to retrieved content instead of relying on general knowledge.</p><p>During the demo, we asked questions about medical images, including tissue health and visible markers. Each response referenced information asdpulled from the most relevant documents in the library.</p><h4>Observations from the Live Discussion</h4><p>The audience questions touched on topics many teams are evaluating right now.</p><p>Cost considerations came up early, particularly around embeddings and multimodal models. As with most architectural decisions, the balance depends on scale and workload. Bringing vector search into the same platform as analytics removes certain integration and duplication costs that often appear with external systems.</p><p>We also discussed entry points. Multimodal search is compelling, though many teams begin with document retrieval, tables, policies, and reports. These use cases already benefit from structured parsing and vector search and provide a solid foundation before adding image‑based reasoning.</p><p>Another area of interest was familiarity. Vector stores can feel unfamiliar at first, though the underlying ideas map closely to indexing strategies DBA teams already use. Teradata’s approach keeps vector operations accessible through both SQL and Python, which helps bridge that gap.</p><h4>Try the Demo Yourself</h4><p>The full demo runs inside <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">Teradata.com/experience</a>, which is available at no cost. It includes the notebooks, ingestion pipeline, vector store setup, and agentic workflow shown during the livestream. Unstructured also provides free credits that make it easy to reproduce and extend the example.</p><p>If you are exploring multimodal embeddings, agentic retrieval patterns, or enterprise‑grade vector search, the livestream walks through the full system working together, we have also prepared a detailed <a href="https://medium.com/teradata/end-to-end-multimodal-rag-and-agents-with-teradata-and-unstructured-c3686b5c6198">blog for step by step guidance.</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=fdf1bf3029e5" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/building-multimodal-agentic-search-over-images-and-documents-a-devtalk-fdf1bf3029e5">Building Multimodal Agentic Search Over Images and Documents — A DevTalk</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[End-to-End Multimodal RAG and Agents With Teradata and Unstructured]]></title>
            <link>https://medium.com/teradata/end-to-end-multimodal-rag-and-agents-with-teradata-and-unstructured-c3686b5c6198?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/c3686b5c6198</guid>
            <category><![CDATA[unstructured-data]]></category>
            <category><![CDATA[retrieval-augmented-gen]]></category>
            <category><![CDATA[teradata]]></category>
            <category><![CDATA[agentic-ai]]></category>
            <category><![CDATA[vector-database]]></category>
            <dc:creator><![CDATA[Daniel Herrera]]></dc:creator>
            <pubDate>Mon, 23 Mar 2026 12:06:15 GMT</pubDate>
            <atom:updated>2026-03-23T12:06:14.172Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="Multimodal RAG and Agents with Teradata &amp; Unstructured" src="https://cdn-images-1.medium.com/max/1024/1*6hQeuUsHuF70O4j6KCXSXg.png" /><figcaption>Multimodal RAG and Agents with Teradata &amp; Unstructured</figcaption></figure><h3><strong>The challenge of grounding the work of agentic systems</strong></h3><p>Agentic systems work across multiple steps to achieve a specific goal. Precision and predictability in achieving that goal require rich context that often goes beyond structured transactional data. Policies, procedures, diagnostics, design documents, and customer conversations are often stored as documents, PDFs, images, and audio files. When that content is transformed into embeddings, making it searchable, agents gain something closer to long-term memory: they can retrieve relevant context, cite sources, and ground their responses in what is true for the organization, making them accurate and traceable.</p><p>Most enterprise data stacks split these two worlds cleanly. Transactional facts live in relational tables. Policies, manuals, and media live somewhere else, in object storage or file systems. Bringing them together requires additional data-processing pipelines and adds governance overhead.</p><p>It is possible, though, to bridge the gap between these two worlds, and it can be done on the same platform, without data duplication, data movement overhead, and extra governance layers.</p><h3><strong>Multimodal agentic development with Teradata and Unstructured</strong></h3><p>Teradata Enterprise Vector Store has unleashed a myriad of use cases for Teradata customers, including:</p><ul><li>Customer 360 enrichment and contact center intelligence</li><li>Autonomous customer insights</li><li>Compliance and policy intelligence</li><li>Insurance claims and contract analysis</li><li>R&amp;D and engineering knowledge retrieval</li><li>Large‑scale enterprise search</li><li>Defense and public‑sector operational intelligence</li></ul><p>These use cases are enabled by performing vector and metadata-aware search across structured and unstructured data in a single operation, directly within the Teradata system as the single source of truth.</p><p>Teradata’s partnership with Unstructured extends these capabilities by introducing a unified multimodal ingestion and enrichment layer capable of parsing <a href="https://docs.unstructured.io/api-reference/supported-file-types">70+ file types</a> and generating embeddings for text and images, with audio support on the roadmap. The Unstructured pipelines handle chunking, metadata extraction, enrichment, adaptive model selection, and embedding generation, and deliver these enriched artifacts directly into Teradata’s enterprise analytics platform. The Teradata platform provides a native security model, object‑level governance, and lineage propagation. Because the Enterprise Vector Store is implemented on top of Teradata’s MPP execution framework, vector indexing, similarity search, and hybrid semantic–lexical retrieval scale transparently with other analytic and operational workloads, inheriting Teradata’s concurrency, data locality, and high‑throughput characteristics.</p><p>For development, the <a href="https://github.com/Teradata/langchain-teradata">langchain-teradata</a> library offers a familiar interface for incorporating Teradata Enterprise Vector Store into your agentic workflows, providing the standard agentic primitives: chat models, tools, and retrievers. The result is a single system where agents can search, and act, without the extra pipelines or governance overhead that come with stitching separate tools together.</p><h3><strong>Experience multimodal agentic development with Teradata and Unstructured</strong></h3><p>In the walkthrough that follows, we will build a life sciences image question-and-answering agent powered by Teradata Enterprise Vector Store, using data ingested and embedded by Unstructured.</p><p>The final product is an agentic system where a user can pick an image from a gallery of medical image samples and ask questions about it. The agent relies on a library of medical image documentation, finds the documentation most like the selected image, and uses it to answer the user’s question.</p><figure><img alt="Image 1 — Image grid gallery" src="https://cdn-images-1.medium.com/max/736/1*UgaSv1ekYAZWxiyI7x3oFQ.png" /><figcaption>Image 1 — Image grid gallery</figcaption></figure><p>The process of building this agent consists of the following general steps:</p><ul><li>Build an ingestion and processing pipeline for composite PDF documents and images using Unstructured, which also generates the embeddings.</li><li>Store the embeddings in Teradata’s enterprise analytics platform and index them with Teradata Enterprise Vector Store.</li><li>Match a user‑selected image to its most similar documents using vector similarity search.</li><li>Wrap the workflow in a LangChain agent that allows users to query the results in natural language.</li></ul><p><strong>Setting up your development environment</strong></p><ul><li>If you haven’t already, start by creating an account at <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">teradata.com/experience.</a></li><li>Once you’re logged in, create <strong>a new environment.</strong> This ensures you’re working with the latest features and capabilities.</li></ul><figure><img alt="Image 2 — Create an environment" src="https://cdn-images-1.medium.com/max/607/1*PMyrgYrB4mpixfnFdYfbZw.png" /><figcaption>Image 2 — Create an environment</figcaption></figure><ul><li>Start the Jupyter Notebook environment by clicking “Run demos.”</li></ul><figure><img alt="Image 3 — Run demos" src="https://cdn-images-1.medium.com/max/624/1*1g_Xy-nEcg24xL4lFUT59g.png" /><figcaption>Image 3 — Run demos</figcaption></figure><p><strong>Exploring the notebook</strong></p><p><strong>Setting up needed libraries and access to Teradata Enterprise Vector Store</strong></p><ul><li>First, we install the packages needed.</li></ul><pre>%%capture<br>!pip install -r &quot;./utils/requirements.txt&quot; --quiet</pre><ul><li>After restarting the kernel, we import the libraries that will power our agentic multimodal system.</li></ul><pre># Required imports<br><br># General imports<br>from teradatagenai import VSManager<br>from langchain_teradata import TeradataVectorStore<br>from teradataml import *<br>import os<br>import json<br>import re<br>import time<br><br># Credentials and configuration management<br>from dotenv import load_dotenv<br>from getpass import getpass<br><br><br>#Langchain Imports<br>from langchain.chat_models import init_chat_model<br>from langchain.agents import create_agent<br>from langchain_core.messages import HumanMessage<br>from langchain.tools import tool<br><br># Suppress warnings<br>import warnings<br>warnings.filterwarnings(&#39;ignore&#39;)<br>display.suppress_vantage_runtime_warnings = True<br><br># Widget display<br>from IPython.display import display, HTML, IFrame<br>import ipywidgets as widgets<br><br># Import utilities<br>from unstructured_utils.teradata_ingest import ingest<br>from utils.image_grid import display_image_grid</pre><ul><li>Our trial experience provides access to a Teradata instance with Enterprise Vector Store capabilities.</li></ul><pre>ues_uri=env_vars.get(&quot;ues_uri&quot;)<br>if ues_uri.endswith(&quot;/open-analytics&quot;):<br>  ues_uri = ues_uri[:-15]  # remove last 5 chars (&quot;/open-analytics&quot;)<br><br>if set_auth_token(base_url=ues_uri,<br>         pat_token=env_vars.get(&quot;access_token&quot;), <br>         pem_file=env_vars.get(&quot;pem_file&quot;)<br>         ):<br>  print(&quot;UES Authentication successful&quot;)<br>else:<br>  print(&quot;UES Authentication failed. Check credentials.&quot;)<br>  sys.exit(1)<br><br>VSManager.health()<br>```</pre><p><strong>Ingestion and processing pipeline with Unstructured</strong></p><ul><li>Unstructured offers a graphical user interface (GUI) to create workflows. The same building blocks used in the GUI are available from <a href="https://docs.unstructured.io/api-reference/overview">Unstructured API</a>. Namely: sources, destinations, workflows, and jobs.</li><li>Workflows orchestrate data processing between a source and a destination with steps such as partitioning, chunking, enriching, embedding, etc.</li><li>Jobs are specific execution of a workflow.</li><li>We have built a function that abstracts our defined workflow. The full implementation is contained in the /unstructured_utils folder and imported into the notebook.</li></ul><p>The workflow has three sequential nodes:</p><ol><li><strong>Partitioning:</strong> Uses Claude as a vision-language model (VLM) to extract structured content from raw documents. Using a VLM rather than a traditional text extractor means it can handle complex layouts, images, and tables by “looking” at the document rather than just parsing text.</li><li><strong>Chunking:</strong> Splits the partitioned content into pieces using a title-aware strategy. It targets 1,500 characters per chunk, caps at 2,048, and adds a 100-character overlap between consecutive chunks so context isn’t lost at boundaries. It also retains the original elements alongside the chunks, which is useful for traceability.</li><li><strong>Embedding:</strong> Converts each chunk into a vector using Voyage AI’s multimodal model. The multimodal part matters here because the upstream VLM partitioner may have preserved image-based content, and a multimodal embedder can represent that meaningfully rather than ignoring it.</li></ol><ul><li>In the notebook, we call this function to execute the ingestion of both composite PDFs and images.</li></ul><pre>ingest(api_key=unstructured_api_key, <br>    td_host=env_vars.get(&quot;host&quot;), <br>    td_user=dbuser, <br>    td_password=dbpwd, <br>    td_database=default_db, <br>    td_table=&#39;composite_pdfdocs_embedded&#39;, <br>    s3_uri=&quot;s3://dev-rel-demos/teradata-unstructured/healthcare-assets/composite-pdfs/&quot;, <br>    s3_anonymous=True)<br><br>ingest(api_key=unstructured_api_key, <br>    td_host=env_vars.get(&quot;host&quot;), <br>    td_user=dbuser, <br>    td_password=dbpwd, <br>    td_database=default_db, <br>    td_table=&#39;image_samples_embedded&#39;, <br>    s3_uri=&quot;s3://dev-rel-demos/teradata-unstructured/healthcare-assets/images/&quot;, <br>    s3_anonymous=True)</pre><p><strong>Teradata Enterprise Vector Store: The engine for processing unstructured data</strong></p><ul><li>The first step to create our Vector Store is to ingest the data loaded by Unstructured into teradataml DataFrames. We can inspect a sample of the records.</li><li>The preview displays the <strong>record_id</strong>, <strong>filename</strong>, and <strong>record_locator</strong> columns. The embedding vectors are omitted here for readability, but they are present in the underlying table and will be used for similarity search in later steps.</li></ul><figure><img alt="Image 1 — Ingested data in Teradata DataFrame" src="https://cdn-images-1.medium.com/max/1024/1*Uglp77liNx_3XVnPumogzg.png" /><figcaption>Image 4 — Ingested data in Teradata DataFrame</figcaption></figure><ul><li>Once the documents are ingested and embedded, we create a Vector Store based on the embeddings through the `TeradataVectorStore` class `from_embeddings` method. This is a class inherited from `langchain-teradata`.</li></ul><pre>vs_emb = TeradataVectorStore.from_embeddings(name = f&quot;{default_db}_mm_embeddings_vs&quot;,<br>                   data = image_documentation_bank,<br>                   data_columns = &quot;embeddings&quot;,<br>                   key_columns = [&quot;id&quot;, &quot;record_id&quot;],<br>                   embedding_data_columns = &quot;text&quot;,<br>                   metadata_columns = [&quot;text&quot;,&quot;date_created&quot;, &quot;date_modified&quot;, &quot;record_locator&quot;, &quot;filename&quot;],)</pre><ul><li>The Vector Store registers and creates a new vector index in VantageCloud over the embedding column, enabling sub-second approximate nearest-neighbor lookups at scale. Key parameters include the data table, the column containing the embedding vectors `data_columns`, the primary key columns for deduplication, and the metadata columns to carry through into search results.</li><li>The `TeradataVectorStore` class includes methods for performing searches, semantical or lexical, asking questions about the results of searches (powered by an included LLM client), and to update, re-index or destroy the Vector Store.</li><li>We can use the Vector Store to perform similarity searches between any image on the image bank and the documentation library.</li></ul><p>In this case, we are retrieving the top record from the DataFrame containing the image data, and we perform a similarity search by vector retrieving the most similar document.</p><pre>response = vs_emb.similarity_search_by_vector(data = raw_images_df.head(1), column=&#39;embeddings&#39;, top_k=1)</pre><figure><img alt="Image 5 — Similarity search results" src="https://cdn-images-1.medium.com/max/872/1*GFwesq9rPgw3E8YE6Fmecg.png" /><figcaption>Image 5 — Similarity search results</figcaption></figure><p>The match is driven by the image contained in the composite document. The retrieval returns the associated text along with metadata included when creating the Vector Store index, such as the `record_locator` and `filename`. In an agentic system, this metadata allows an agent to fetch the original document if needed, while the text provides information to answer questions about the selected image.</p><p>Teradata Enterprise Vector Store includes a built-in LLM integration that enables natural language queries over vector search results.</p><pre>question=&#39;I need to recover The title, description and record id, and locator of the most similar record?&#39;<br>prompt=&#39;Format the response in a conversational way.&#39;<br>response = vs_emb.prepare_response(question=question, similarity_results=response, prompt=prompt)</pre><p>The response object contains the requested data in the requested format as processed by the chat model included in the Teradata Enterprise Vector Store.</p><blockquote>Based on the provided data, here is the information you requested:<br> <br> Title: Image 4: Immune Cell Infiltration in Solid Tumor Tissue<br> Description: Microscopic image showing immune cell infiltration in solid tumor tissue. The image displays a colorful fluorescent staining pattern with bright yellow circular structures scattered throughout a predominantly pink and magenta background with blue areas. Dark black voids are visible throughout the tissue section. The staining reveals cellular structures and immune markers in the tumor microenvironment.<br> <br> This image captures a solid tumor section with visible immune cell infiltration. The bright yellow circular structures likely represent lipid droplets or specific immune cell subtypes such as macrophages or dendritic cells. The dark voids may be necrotic foci or gland-like structures within the tumor. The predominant pink and magenta signals indicate widespread expression of tumor or immune markers throughout the tissue. This kind of staining is commonly used in studies examining the tumor immune microenvironment, particularly in the context of immunotherapy response.<br> <br> Record ID: c724ea12-fa51–554d-b552-cb2d5de3b0a7<br> Record Locator: {“protocol”: “s3”, “remote_file_path”: “s3://dev-rel-demos/teradata-unstructured/healthcare-assets/composite-pdfs/”}</blockquote><p>The capabilities we have explored unlock a lot of possibilities for an agentic system, such as the one we are building.</p><h3><strong>Building a LangChain agent powered by Teradata Enterprise Vector Store</strong></h3><p>Our agent is wrapped in a widget UI that allows users to select an image from the gallery, and a chat widget that enables conversations. These widgets are merely UI abstractions; the components of the agent itself are the same as any LangChain agent. We need a `chat_llm` component, and a set of tools.</p><p>Our trial experience includes a proxy LLM client to power the `chat_llm`, while Teradata Enterprise Vector Store includes all the functionality needed to power the core tool.</p><pre>llm_key = env_vars.get(&quot;litellm_key&quot;)<br>llm_url = env_vars.get(&quot;litellm_base_url&quot;)<br><br>llm = init_chat_model(<br>  model=&quot;openai-gpt-41&quot;,<br>  model_provider=&quot;openai&quot;,<br>  base_url=llm_url,<br>  api_key=llm_key,<br>)</pre><p>The main tool is a thin abstraction over the `similarity_search` and `prepare_response` methods of Teradata Enterprise Vector Store. The image grid widget is generated from the DataFrame containing the image data, and selection buttons that update the `selected_id` parameter to look up the corresponding embedding. From there, the process follows the same steps described above.</p><pre>@tool<br>def search_similar_image_documentation(dummy: str = &quot;&quot;) -&gt; str:<br>  &quot;&quot;&quot;<br>  Runs a similarity search using the currently selected image and displays<br>  the most similar result. Call this whenever the user asks about similar images.<br>  &quot;&quot;&quot;<br>  if result.selected_id is None:<br>    return &quot;No image is selected. Please select an image from the grid first.&quot;<br><br>  response_similarity = vs_emb.similarity_search_by_vector(<br>    data=raw_images_df[raw_images_df[&quot;record_id&quot;] == result.selected_id],<br>    column=&quot;embeddings&quot;,<br>    top_k=1,<br>    return_type=&quot;json&quot;,<br>  )<br>  question = &quot;I need to recover the title, description, filename, record id, and file_locator of the most similar record?&quot;<br>  prompt = &quot;Format the response as JSON object.&quot;<br>  response_chat = vs_emb.prepare_response(<br>    question=question,<br>    similarity_results=response_similarity,<br>    prompt=prompt,<br>  )<br><br>  with chat_output:<br>    display(HTML(<br>      f&quot;&lt;div style=&#39;font-family:sans-serif; padding:12px; background:#f9f9f9;&quot;<br>      f&quot;border-radius:6px; margin:8px 0;&#39;&gt;{response_chat}&lt;/div&gt;&quot;<br>    ))<br><br>  return response_chat</pre><p>The agent itself includes specific instructions to interact with the tools and respond to the users’ intent.</p><blockquote>You are an image analysis assistant.</blockquote><blockquote>Tools available:</blockquote><blockquote>- search_similar_image_documentation: Find images similar to a given input and retrieve associated documentation</blockquote><blockquote>- display_pdf_from_locator: Render a PDF document</blockquote><blockquote>Workflow:</blockquote><blockquote>1. When a user provides an image or asks for similar images, call search_similar_image_documentation</blockquote><blockquote>2. If the user asks to view/open a PDF from the results, extract `filename` and `remote_file_path` from the search response and pass them to display_pdf_from_locator</blockquote><blockquote>3. If the user asks to describe features from the image in questions like: — What are the XYZ features of this image -</blockquote><blockquote>a. First call search_similar_image_documentation</blockquote><blockquote>b. If the documentation contains relevant information, answer based on it</blockquote><blockquote>c. If not, inform the user that the image library documentation does not contain information to answer their question, then display the PDF using display_pdf_from_locator so they can review it directly</blockquote><blockquote>d. Never ask to upload the image as the user is referring to the selected image.</blockquote><blockquote>If a search returns no results or the PDF fields are missing, let the user know.</blockquote><p>The widget UI and the agent enable the fully agentic experience.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/736/1*4oU2PMznbUXAn2u1xYwlGg.png" /><figcaption>Image 6— Agentic medical image question-and-answer experience</figcaption></figure><p><a href="https://www.teradata.com/platform/clearscape-analytics/enterprise-vector-store">Learn more about Teradata multimodal agentic capabilities.</a></p><p>Build the demo yourself at <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">teradata.com/experience</a>.</p><p><strong>Conclusion</strong></p><p>Bringing together Teradata Enterprise Vector Store and Unstructured creates a unified foundation for multimodal agentic systems, one where structured and unstructured data coexist naturally, governance stays simple, and developers can build advanced retrieval-augmented workflows without extra pipelines or infrastructure overhead. High-fidelity ingestion, scalable vector indexing, and agentic tooling combine into a stack that reduces the friction between experimentation and production. The walkthrough here is one example of the pattern, but the same approach applies to customer experience, automation, and context-aware intelligence across industries. As multi-modal data and agentic AI continue to expand, this integrated stack gives organizations a practical path to accurate, grounded, and enterprise-ready agentic systems built for scale.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=c3686b5c6198" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/end-to-end-multimodal-rag-and-agents-with-teradata-and-unstructured-c3686b5c6198">End-to-End Multimodal RAG and Agents With Teradata and Unstructured</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Introducing Teradata Enterprise MCP: Build Your First Data Analyst Agent]]></title>
            <link>https://medium.com/teradata/introducing-teradata-enterprise-mcp-build-your-first-data-analyst-agent-c546dbd2d424?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/c546dbd2d424</guid>
            <category><![CDATA[data-analysis]]></category>
            <category><![CDATA[mcp-server]]></category>
            <category><![CDATA[teradata]]></category>
            <category><![CDATA[generative-ai-tools]]></category>
            <category><![CDATA[agentic-ai]]></category>
            <dc:creator><![CDATA[Daniel Herrera]]></dc:creator>
            <pubDate>Fri, 30 Jan 2026 12:02:23 GMT</pubDate>
            <atom:updated>2026-02-02T10:16:57.661Z</atom:updated>
            <content:encoded><![CDATA[<p>Teradata Enterprise MCP is designed to empower organizations to build advanced data analyst agents through secure, scalable, and agentic AI-powered analytics.</p><figure><img alt="Teradata Enterprise MCP" src="https://cdn-images-1.medium.com/max/975/1*vQUNvGB84MEO_xJNxgPCYw.png" /></figure><p>An <a href="https://www.teradata.com/insights/ai-and-machine-learning/autonomous-ai-knowledge-platform">autonomous enterprise</a> is one that uses its knowledge, applies agentic reasoning, and delivers faster, higher‑quality outcomes for customers. It operates through a modern stack that unifies a knowledge layer, an intelligent agent layer, and an outcomes layer, enabling systems to reason and act with context rather than rely solely on manual workflows.</p><p>The agentic layer combines Large Language Models (LLMs) reasoning capabilities with trusted data, policies, context, and analytics, enabling agents to perform actions such as querying structured and unstructured data, executing advanced analytics, running predictions, retrieving documents, and grounding responses in enterprise reality. This connection transforms AI from experimentation into operational impact.</p><p>Teradata Enterprise MCP acts as a core pillar of this autonomy by providing secure access to data, tool standardization, context management, auditability, and high scalability for agent actions. It ensures that agents can interact with enterprise systems in a trusted, governed, and consistent way while using analytics and knowledge capabilities at production scale.</p><p><strong>The Model Context Protocol –MCP– in brief:</strong></p><p>MCP is an <a href="https://www.anthropic.com/news/model-context-protocol">open-source standard designed to make AI systems talk to the outside world in a predictable, structured way</a>. Instead of every developer inventing a new format for tools, capabilities, or metadata, MCP provides a common contract defining all these aspects, standardizing LLMs interactions with external tools. For a developer, this means spending more time designing the behavior of AI powered agents, not worrying about how to package or describe needed tools or necessary parameters. The protocol takes care of that for you.</p><p>Teradata research teams and a group of community contributors built the community driven <a href="https://github.com/Teradata/teradata-mcp-server">Teradata MCP</a>, this open-source effort has taken off in ways none of us fully expected. Suddenly, people were building agents that could explore data, run analytics, and automate workflows directly inside the Teradata ecosystem. The best part is that the project keeps getting better as contributors continue to refine the server, add new tools, tighten the interface, and push the boundaries of what MCP powered agents can do.</p><figure><img alt="Teradata MCP server architecture" src="https://cdn-images-1.medium.com/max/734/1*lWtDvq-4GS5Pf3AdIgdAUg.png" /><figcaption>Teradata MCP server architecture</figcaption></figure><p><strong>The challenges of deploying and implementing an MCP server</strong></p><p>MCP isn’t without its challenges. While implementing an MCP server for an enterprise knowledge platform, the same questions always come up. Can it scale across teams? Can it connect to sensitive data with the right governance and privacy controls? What happens when teams start chaining servers together and bump into context‑overflow issues?</p><p>In consideration of these pain-points, Teradata has developed Teradata Enterprise MCP. Teradata Enterprise MCP tackles the core operational and governance challenges head‑on. Enabling data professionals to focus on building agents that deliver value, easing the concerns around deployment, scaling, and access controls common to MCP as a protocol.</p><p><strong>How Teradata Enterprise MCP contributes to your Agentic AI projects</strong></p><p>Teradata Enterprise MCP ships with a set of features that tackle common implementation challenges.</p><figure><img alt="Features of Teradata Enterprise MCP" src="https://cdn-images-1.medium.com/max/975/1*vQUNvGB84MEO_xJNxgPCYw.png" /><figcaption>Features of Teradata Enterprise MCP</figcaption></figure><p>Security:</p><ul><li>Every user connects to Teradata Enterprise MCP using their own credentials, whether that’s basic TD2 authentication or JWT and Bearer tokens validated through the organization’s identity provider. This ensures the server only exposes the data that specific users are authorized to see, instead of routing all users through a single service account, unless the latter is the chosen implementation decision of the organization.</li><li>RBAC controls tool visibility. MCP roles, with their corresponding assigned toolset, can be mapped to Teradata database roles so you can define exactly which tools a user can access in the system according to the user’s role.</li><li>SQL injection protection, security scans, and end-to-end logging provide an enterprise grade security posture without adding friction for developers.</li></ul><p>Performance:</p><ul><li>The server is built for multiuser, multi-session workloads with automatic connection pooling and sub second response times. Teams across the company can reliably share one MCP server without creating bottlenecks.</li></ul><p>Tool Context Management:</p><ul><li>A deep catalog of more than a hundred production ready tools; from database queries and DBA functions to ML, vector search, and RAG might be difficult to manage for LLMs, however, because tool exposure is governed per role, each agent stays within a clean and compliant sandbox.</li></ul><p><strong>Experience Teradata Enterprise MCP</strong></p><p>You can experience Teradata Enterprise MCP directly inside <a href="https://www.teradata.com/platform/clearscape-analytics">ClearScape Analytics Experience</a>. We’ve put together a <a href="https://github.com/JH255095/teradata_enterprise_mcp/blob/main/Langchain_Teradata_MCP.ipynb">sample notebook</a> that shows a simple LangChain integration. This is only a starting point that you can extend according to your needs.</p><p>You can easily experiment on your own by following the steps below:</p><p><strong>Setting up your development environment</strong></p><ul><li>If you haven’t used ClearScape Analytics Experience before, start by creating an account at <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">teradata.com/experience.</a></li><li>Once you’re logged in, create <strong>a new environment.</strong> This ensures you’re working with the latest features and capabilities. Take note of your chosen password as you will need it to create the MCP server configuration later.</li></ul><figure><img alt="Create an environment on ClearScape Analytics Experience" src="https://cdn-images-1.medium.com/max/948/1*MdF9wcmXWgHqiXYCb7IGAA.png" /><figcaption>Create an environment on ClearScape Analytics Experience</figcaption></figure><ul><li>Start the Jupyter Notebook environment by clicking “Run demos”.</li></ul><figure><img alt="Run demos on ClearScape Analytics Experience" src="https://cdn-images-1.medium.com/max/975/1*SRlqJmCqOQpQYOxGaZ82Hg.png" /><figcaption>Run demos on ClearScape Analytics Experience</figcaption></figure><ul><li>In the Jupyter Notebook environment, open the “Getting Started” folder look for the folder named Enterprise_MCP_Data_Analyst_Agent, open the notebook inside that folder.</li></ul><p><strong>Exploring the notebook</strong></p><ul><li>First, we install the packages needed for the MCP integration with LangChain.</li></ul><pre>pip install -U langchain langchain-mcp-adapters langchain-openai httpx</pre><ul><li>After restarting the kernel, we import the libraries that will power our agent: asyncio for async operations, the MCP client adapter, and LangChain’s chat model and agent components.</li></ul><pre>import asyncio, sys, os, httpx<br>import time<br>from dotenv import load_dotenv<br>from langchain_mcp_adapters.client import MultiServerMCPClient<br>from langchain.chat_models import init_chat_model<br>from langchain.agents import create_agent<br>from langchain_core.messages import HumanMessage<br>from langchain_core.messages import AIMessage</pre><ul><li>The notebook pulls in your environment settings from a .env file. This keeps your credentials secure and your code clean.</li></ul><pre>environment_path = &quot;/home/jovyan/JupyterLabRoot/VantageCloud_Lake/.config/.env&quot;<br>load_dotenv(dotenv_path=environment_path) <br>llm_key = os.getenv(&#39;litellm_key&#39;)<br>llm_url = os.getenv(&#39;litellm_base_url&#39;)<br>username = os.getenv(&#39;username&#39;)<br>password = os.getenv(&#39;my_variable&#39;)</pre><ul><li>Here we establish the connection to the Enterprise MCP Server using LangChain’s MultiServerMCPClient. The MCP server uses the credentials retrieved from the environment variables to connect on your behalf and returns the available tools.</li></ul><pre>async def get_mcp_tools():<br>    client = MultiServerMCPClient(<br>        {<br>            &quot;teradata&quot;: {<br>                &quot;transport&quot;: &quot;http&quot;,<br>                &quot;url&quot;: &quot;http://host.docker.internal:8001/mcp&quot;,<br>                &quot;auth&quot;: httpx.BasicAuth(username, password),<br>            }<br>        }<br>    )<br>    tools = await client.get_tools()<br>    return client, tools<br><br>client, tools = await get_mcp_tools()<br>print([t.name for t in tools])</pre><ul><li>We initialize a chat model through the LLM proxy available in ClearScape Analytics Experience, then create a LangChain agent by passing in the Teradata MCP server tools. The system prompt tells the agent it’s a data analyst working with Teradata.</li></ul><pre>llm = init_chat_model( <br>    model=&quot;openai-gpt-41&quot;, <br>    model_provider=&quot;openai&quot;, <br>    base_url=&quot;https://llmlite.ci.clearscape.teradata.com&quot;, <br>    api_key=llm_key, ) <br><br>agent = create_agent(<br>        model=llm, <br>        tools=tools, <br>        system_prompt=(<br>        &#39;&#39;&#39; You are a data analyst that responds to user questions regarding data in a teradata system &#39;&#39;&#39; <br>        ), )</pre><ul><li>With everything set up, you can send natural language queries to the agent. It interprets your question, picks the right MCP tools, and executes them against your Teradata environment.</li></ul><pre>user_query = input(&#39;\n What would you like to know about your data? &#39;)<br>result = await agent.ainvoke(<br>    {&quot;messages&quot;: [HumanMessage(content=user_query)]}<br>)</pre><figure><img alt="Query submitted to the agent" src="https://cdn-images-1.medium.com/max/878/1*oPkm_4QXOs3kL2ImAPKZTw.png" /><figcaption>Query submitted to the agent</figcaption></figure><ul><li>The answer of the agent to this specific question is as follows:</li></ul><pre>print(result[&quot;messages&quot;][-1].content)</pre><figure><img alt="" src="https://cdn-images-1.medium.com/max/975/1*b7YgVCl-bsObqlVOvEmdOg.png" /><figcaption>Agent response to the user query</figcaption></figure><p>ClearScape Analytics Experience is a free of cost experience with access to limited resources, for a full demo on a production grade Teradata environment feel free to reach us out at <a href="https://www.teradata.com/platform/agent-stack">Teradata Enterprise AgentStack</a>.</p><p><strong>Conclusion:</strong></p><p>The shift toward autonomous enterprises is no longer theoretical. With agentic AI, unified knowledge layers, and governed execution environments, organizations now have the foundations they need to build systems that reason, act, and deliver value with less manual intervention. The Teradata Enterprise MCP Server plays a significant role in making this real. It removes the operational and security hurdles that traditionally slow down MCP adoption and replaces them with a platform that feels stable, scalable, and ready for enterprise use. By giving teams, a secure, governed, and high‑performance way to connect LLMs and agents to trusted data, the Enterprise MCP Server frees developers and data practitioners to do what they do best: build. Whether you are exploring early prototypes, scaling agentic workloads, or preparing for production‑grade deployments, the Enterprise MCP Server provides the backbone that lets you move faster, stay compliant, and unlock the full potential of agentic AI inside your organization.</p><p><strong>About the Author:</strong></p><p>Daniel Herrera is Principal Developer Advocate at Teradata, specializing in AI/ML and advanced analytics. He helps developers and data practitioners unlock enterprise-scale innovation through ClearScape Analytics and VantageCloud. With deep expertise in data engineering, cloud architecture, and model deployment, Daniel bridges technical complexity with practical solutions. A frequent speaker and community contributor, he focuses on enabling scalable AI use cases across industries.</p><p>You can reach out to Daniel in <a href="https://www.linkedin.com/in/daniel-herrera/">LinkedIn</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=c546dbd2d424" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/introducing-teradata-enterprise-mcp-build-your-first-data-analyst-agent-c546dbd2d424">Introducing Teradata Enterprise MCP: Build Your First Data Analyst Agent</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[A Gateway to Enhanced Agents & AI Workloads: Arrow Flight SQL with QueryGrid®]]></title>
            <link>https://medium.com/teradata/a-gateway-to-enhanced-agents-ai-workloads-arrow-flight-sql-with-querygrid-e6b4c6721c2a?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/e6b4c6721c2a</guid>
            <category><![CDATA[data-fabric]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[ai-agent]]></category>
            <category><![CDATA[apache-arrow]]></category>
            <dc:creator><![CDATA[Sebastian]]></dc:creator>
            <pubDate>Thu, 29 Jan 2026 14:36:49 GMT</pubDate>
            <atom:updated>2026-01-29T14:36:48.428Z</atom:updated>
            <content:encoded><![CDATA[<h4>Unveiling the power of the Arrow Flight SQL connector for QueryGrid®</h4><p>The introduction of the Arrow Flight SQL connector for QueryGrid® marks a significant advancement for organizations seeking to accelerate data-driven &amp; AI-driven innovation. By enabling streamlined, high-performance access to diverse data sources without the overhead of traditional ETL processes, this capability empowers teams to unlock faster insights, enhance collaboration, and optimize their AI workflows. As a result, customers can maximize the value of their existing data infrastructure while future-proofing their operations for emerging AI and agentic workloads.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*GXpNQqtsXD4crfriTOFtyg.png" /><figcaption>QueryGrid(r) connector stack</figcaption></figure><h3>Introduction to Apache Arrow</h3><p><a href="https://arrow.apache.org/">Apache Arrow</a> is a cross-language development platform designed for in-memory data processing. It provides a standardized columnar memory format for flat and hierarchical data, enabling efficient analytics operations. Arrow’s key strength lies in its ability to handle large datasets with minimal overhead, facilitating rapid data access and reduced memory usage. Over the last few years, Arrow has gained traction and interest in the Data Science community, with commits to its codebase added every day.</p><h3>Arrow Flight and Arrow Flight SQL</h3><p><a href="https://arrow.apache.org/docs/format/Flight.html">Arrow Flight</a> is an extension of Apache Arrow that introduces high-performance messaging for data transfer. It leverages gRPC to enable efficient, secure, and highly scalable data transport between systems. Building upon Arrow Flight, <a href="https://arrow.apache.org/docs/format/FlightSql.html">Arrow Flight SQL</a> provides a protocol for executing SQL queries over Arrow Flight, allowing seamless and efficient interaction with SQL-based systems. The performance achieved is over 15 times higher than what a single-threaded protocol like JDBC can deliver (based on lab tests).</p><h3>Meet QueryGrid® Arrow Flight SQL connector</h3><p>QueryGrid® <a href="https://medium.com/teradata/understanding-querygrid-a-gateway-to-optimized-distributed-data-access-7d381db1e2ac">is a sophisticated query fabric</a> that seamlessly connects diverse data sources and processing engines. It enables businesses to execute queries across multiple platforms, integrating disparate data environments into a <a href="https://www.teradata.com/platform/data-fabric">unified analytical ecosystem</a>. The integration of QueryGrid® with Arrow Flight SQL extends this capability to a broader range of engines and systems, and provides a net new capability: now client tools can also connect through the QueryGrid® Arrow Flight SQL connector. All is done leveraging Arrow’s modern and efficient data format and transport protocols, maintaining QueryGrid’s legacy: providing high-throughput parallel data exchange, and minimizing data movement when possible.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Xv8xGSasO3uMLAkG0CZghg.png" /><figcaption>QueryGrid(r) Arrow Flight SQL Connector</figcaption></figure><h4>Client Server Architecture</h4><p>QueryGrid® Arrow Flight SQL connector comes in two flavors: Client &amp; Server. When the client capability is configured, the <a href="https://www.teradata.com/">Teradata Vantage</a> database can read data in Arrow format from other engines on the fly, without the need to copy the data upfront or perform any ETL processes (capability is available today). When the Server is configured, the Teradata Vantage database is exposed as an Arrow Flight SQL endpoint, allowing not only engines, but also tools, to read and write data in parallel, leveraging this highly efficient columnar format and transfer protocol. The server capability will be released soon, but don’t let this block you from planning how to leverage it in your future architecture; you can start the conversation with us today and <a href="https://www.teradata.com/insights/videos/data-federation-for-ai">see this in action</a>.</p><h4>Expanding the Ecosystem</h4><p>Integrating QueryGrid® with Apache Arrow allows enterprises to tap into a wide array of data processing engines supported by Arrow Flight SQL, such as DuckDB, Dremio, and Apache Doris, just to name a few. It also allows any ADBC-compliant driver to read data from Vantage <em>(coming soon with server capability).</em></p><h4>Enhancements in Data Processing</h4><p>QueryGrid® and Apache Arrow facilitate smooth, high-speed parallel data transfer and query execution across multiple data sources. Combined with the intelligent push-down processing that QueryGrid® provides, this enables efficient data exchange between applications, something that is key in a distributed environment.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/997/1*Q-PhpcDU8TMBYylEOCIvmw.png" /><figcaption>QueryGrid MCP Server CE enables Agentic workflows</figcaption></figure><h3>Built for AI Workloads &amp; Agents</h3><p>The integration of QueryGrid® with Apache Arrow is particularly advantageous for AI and Agentic workloads. These workloads often require the processing of massive datasets, real-time data streaming, and the execution of complex queries. Apache Arrow’s columnar format and the high-performance transmission via Arrow Flight and Arrow Flight SQL provide the robustness and speed necessary for these demanding tasks, while also simplifying the interface between systems.</p><h4>Accelerating Training and Inference</h4><p>With the capability to efficiently handle large volumes of data, the integration aids in speeding up both the training and inference phases of machine learning models. Data scientists can seamlessly access and manipulate data from various sources, reducing the time spent on data preparation and transformation.</p><h4>Enhanced Interoperability and Flexibility</h4><p>This integration ensures that models can leverage the best-suited data processing engines, enhancing flexibility and scalability. We know that Agents are greedy and require massive amounts of data, while also the need for consolidating multiple systems. QueryGrid® Arrow Flight SQL connector exposes a simple interface for agents to efficiently grab data from multiple engines and perform complex tasks. Imagine the power you can get if you also connect this with <a href="https://medium.com/teradata/chat-with-your-data-in-visual-studio-code-through-an-mcp-server-071673bd3ffe">Teradata’s</a> and <a href="https://medium.com/teradata/why-youll-want-querygrid-mcp-server-community-edition-in-your-toolkit-cb0872bf9c8a">QueryGrid</a> MCP Servers, or the <a href="https://www.teradata.com/press-releases/2026/teradata-unveils-enterprise-agentstack">Teradata AgentStack</a>.</p><h3>Conclusion</h3><p>The QueryGrid® Arrow Flight SQL connector delivers high performance and seamless integration for AI and Agentic workloads. By utilizing Apache Arrow’s efficient columnar data format and high-speed transport protocols, QueryGrid® enables robust connectivity across numerous engines and systems, empowering organizations to execute advanced analytics and AI/ML tasks with ease. This integration not only streamlines data workflows but also provides agents with an efficient and unified interface to access large volumes of data from diverse sources, unlocking new possibilities in data science and machine learning.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=e6b4c6721c2a" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/a-gateway-to-enhanced-agents-ai-workloads-arrow-flight-sql-with-querygrid-e6b4c6721c2a">A Gateway to Enhanced Agents &amp; AI Workloads: Arrow Flight SQL with QueryGrid®</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Why You’ll Want QueryGrid® MCP Server Community Edition in Your Toolkit]]></title>
            <link>https://medium.com/teradata/why-youll-want-querygrid-mcp-server-community-edition-in-your-toolkit-cb0872bf9c8a?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/cb0872bf9c8a</guid>
            <category><![CDATA[data-fabric]]></category>
            <category><![CDATA[cloud]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[mcps]]></category>
            <category><![CDATA[agents]]></category>
            <dc:creator><![CDATA[Sebastian]]></dc:creator>
            <pubDate>Tue, 06 Jan 2026 12:41:04 GMT</pubDate>
            <atom:updated>2026-01-06T12:40:32.667Z</atom:updated>
            <content:encoded><![CDATA[<h4>Streamline, automate, and secure your QueryGrid management with this new release</h4><p>Let’s face it, managing large data resources can become complicated quickly. If you’re working with <a href="https://www.teradata.com/platform/data-fabric">QueryGrid</a>®, you already know that staying on top of your environment means juggling a lot of moving parts. Enter the <a href="https://github.com/Teradata/teradata-qg-mcp-server">QueryGrid® MCP Server Community Edition</a> (CE), a freshly launched toolkit that’s about to make your life a whole lot easier.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/997/1*O4X21e5sQbPu0XpFvenwQQ.png" /><figcaption><em>QueryGrid(r) MCP Server components</em></figcaption></figure><h3>What Is the QueryGrid MCP Server CE?</h3><p>Think of the <a href="https://www.teradata.com/insights/ai-and-machine-learning/model-context-protocol-mcp">MCP Server CE</a> as your new command center for all things QueryGrid® management. It’s designed to help <a href="https://www.teradata.com/press-releases/2025/mcp-server-agentic-ai-at-scale">AI assistants and other MCP clients</a> talk directly to the QueryGrid Manager (QGM) using a standardized protocol. The focus here isn’t on querying data directly: it’s about making administrative and operational tasks faster, smarter, and more secure. And if you’re already running QueryGrid 3.x, you’re good to go; this release is fully compatible with your environment.</p><h3>Key Features</h3><ul><li><strong>Massive Toolkit</strong>: Over 120 built-in tools support just about every QueryGrid operation you can think of: monitoring resources, adding and removing sources, tweaking configs, creating and dropping links, and tons more.</li><li><strong>Controlled Security</strong>: Basic authentication for QueryGrid Manager means better access control. Plus, logging is totally configurable, so you get the transparency you need.</li><li><strong>RESTful API Integration</strong>: Full support for Create, Read, Update, and Delete (CRUD) operations. All you need for the automation you’ve been looking for.</li></ul><h3>Why Switch to MCP Server CE?</h3><ul><li><strong>Boosted Efficiency</strong>: The toolkit is comprehensive and fully tested, backed by a robust API layer provided by QueryGrid® Manager service. That means even complex management tasks are handled quickly and consistently, so you can accelerate your development and deployment cycles.</li><li><strong>Automation is King</strong>: With the ability to programmatically manage QueryGrid® resources, you can automate all your admin workflows — whether through chatbots, agents, or other tools — and get back to the fun stuff.</li><li><strong>Built-in Security</strong>: Only authorized users or systems can make changes, thanks to secure authentication mechanisms.</li><li><strong>Ready for Scale</strong>: Built for production from the ground up, with configurable logging for transparency and traceability. Start with small projects and climb to mission-critical deployments.</li></ul><h3>Takeways</h3><p>Whether you’re a developer, data engineer, DBA, or someone who wants smoother QueryGrid® management, the QueryGrid® MCP Server Community Edition is a game-changer. It’s secure, flexible, and packed with features that will make your admin processes more efficient and reliable. If you’re ready to embrace modern automation and development practices, this toolkit deserves a spot in your environment. And remember, you can expand the power of this freshly released server with <a href="https://github.com/Teradata/teradata-mcp-server/">Teradata’s MCP Server Community Edition</a>.</p><p>Start your journey now by downloading your copy from <a href="https://github.com/Teradata/teradata-qg-mcp-server">Teradata GitHub</a>.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=cb0872bf9c8a" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/why-youll-want-querygrid-mcp-server-community-edition-in-your-toolkit-cb0872bf9c8a">Why You’ll Want QueryGrid® MCP Server Community Edition in Your Toolkit</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Building Autonomous Customer Intelligence: A Developer’s Guide to Teradata’s Customer Intelligence…]]></title>
            <link>https://medium.com/teradata/building-autonomous-customer-intelligence-a-developers-guide-to-teradata-s-customer-intelligence-a28d8dd77c41?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/a28d8dd77c41</guid>
            <category><![CDATA[mcp-server]]></category>
            <category><![CDATA[ai-agent]]></category>
            <category><![CDATA[llm]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[customer-experience]]></category>
            <dc:creator><![CDATA[Vidhan Bhonsle]]></dc:creator>
            <pubDate>Tue, 06 Jan 2026 12:32:58 GMT</pubDate>
            <atom:updated>2026-01-06T12:32:55.861Z</atom:updated>
            <content:encoded><![CDATA[<h3>Building Autonomous Customer Intelligence: A Developer’s Guide to Teradata’s Customer Intelligence Framework</h3><h3>Introduction</h3><p>The way customer data systems are built is undergoing a remarkable transformation. Across the industry, significant milestones have been reached — unified Customer 360 views, sophisticated machine learning (ML) models, and real-time streaming pipelines. These foundations set up the next leap: systems that don’t just analyze customer behavior, but actively shape experiences in real time.</p><p>The <a href="https://www.teradata.com/solutions/customer-experience"><strong>Teradata Customer Intelligence Framework</strong></a> represents this next generation of customer intelligence architecture. Built on established data engineering and machine learning practices, it enables autonomous, real-time decision-making. The framework equips data scientists, data engineers, ML engineers, and platform architects to create systems that sense needs, understand context, and take immediate action; powered by AI for CX that delivers measurable business impact.</p><h3>The challenge</h3><p>Modern customer data architectures hit the same three walls repeatedly. Data lives in disconnected systems, making a reliable Customer 360 fragile at best. Batch-heavy pipelines introduce latency, so insights arrive after the moment has passed. And unresolved customer tasks often escalate to costly channels, driving up operations while eroding satisfaction.</p><figure><img alt="Pain points of customer experiences" src="https://cdn-images-1.medium.com/max/1024/1*UpWTFRDjVpFES2mS7jtLeQ.png" /><figcaption>Pain points of customer experiences</figcaption></figure><p>The Customer Intelligence Framework addresses these limits with an event-driven approach: unify raw streams into reusable data products, detect signals that carry business meaning in real time, and trigger guard-railed agent decisions that act immediately. In short, move from fragmented data and delayed reports to a closed loop that senses, decides, and acts.</p><h3>Architecture: from data to intelligence</h3><p>The Customer Intelligence Framework transforms raw data into autonomous intelligence through three core components working across five integrated layers:</p><h4><strong>Foundational Elements</strong></h4><figure><img alt="Core components of Customer Intelligence Framework" src="https://cdn-images-1.medium.com/max/1018/1*WEE7_RXJCmdzs_5J1WJsNg.png" /><figcaption>Core components of Customer Intelligence Framework</figcaption></figure><ul><li><strong>Data products</strong> — Reusable, structured assets that organize complex customer data. These multi-dimensional building blocks can be assembled differently for various use cases, with full version control and API accessibility via REST, Python SDKs, or SQL.</li><li><strong>Signals</strong> — Real-time patterns detected from customer behavior, not just transactions. The framework continuously identifies meaningful events like engagement drops, anomalous patterns, or service escalations, enabling immediate response rather than batch processing.</li><li><strong>Agents</strong> — Intelligent entities that autonomously execute decisions within configured guardrails. These agents operate across the end-to-end framework — from data orchestration and feature engineering to signal detection and service activation. They provide natural language interfaces, coordinate with other agents, and maintain full audit trails of their reasoning and actions.</li></ul><h4><strong>Architecture Components</strong></h4><figure><img alt="Customer Intelligence Framework Diagram" src="https://cdn-images-1.medium.com/max/1024/1*gZ40YGrZBBh6ERjXtX3Edw.png" /><figcaption>Customer Intelligence Framework Diagram</figcaption></figure><ul><li><strong>Data ingestion</strong>. Structured, semi-structured, and unstructured sources land through streaming, ETL, object storage, and APIs; catalogued for lineage.</li><li><strong>Feature engineering</strong> (in-database). Transform inputs into robust features and vector representations close to the data to minimize movement.</li><li><strong>Models and rules</strong>. Hybrid decisioning blends ML models, heuristics, and policy rules to balance accuracy, cost, and control.</li><li><strong>Signal processing</strong>. Detect, evaluate, and fuse patterns; a semantic layer maps technical signals (scores, thresholds, sequences) to business concepts (intent, risk, opportunity).</li><li><strong>Service activation</strong> (VCX). Publish decisions and signals via pub/sub so downstream applications subscribe to just-in-time intelligence (NBA/NBO, retention, remediation). Outcomes flow back for continuous learning.</li></ul><p>The architecture flows from data ingestion (supporting structured, semi-structured, and unstructured sources) through features engineering and model deployment using <a href="https://www.teradata.com/platform/clearscape-analytics">ClearScape Analytics®</a> for in-database execution.</p><p>Signal processing interprets these patterns through a semantic layer that maps technical signals to business concepts, while the service activation layer, powered by <a href="https://docs.teradata.com/r/Enterprise_IntelliFlex_VMware/Vantage-Customer-Experience-User-Guide/What-Is-Vantage-Customer-Experience">Teradata VCX</a>, streams real-time intelligence to downstream applications via publish/subscribe patterns.</p><p>Key technical advantages:</p><ul><li>Industry Data Models provide pre-built schemas for rapid deployment</li><li>In-database processing eliminates data movement overhead</li><li>Multi-model orchestration combines rules, ML models, and heuristics</li><li>Every application can subscribe to relevant signals via REST APIs</li></ul><h3>Customer lifetime value multi-agent system demo</h3><p>This demo shows how autonomous customer intelligence turns CLV forecasting and growth into a production workflow. The multi-agent system runs in <a href="https://www.teradata.com/insights/ai-and-machine-learning/agentic-ai-stack-with-teradata-agentbuilder?utm_medium=organic">Teradata AgentBuilder</a>, keeps computation inside Teradata Vantage, and analyzes historical behavior directly in-database. LLMs — GPT-4.1 (routing/response) and Claude Sonnet-4 (analysis/strategy), drive intent understanding, exploration, and decisioning, combining reasoning with real-time signals to answer questions like “Which customer type has the highest CLV and why?” and to recommend next-best actions that lift retention and cross-sell.</p><p>The implementation uses Flowise inside AgentBuilder for low-code composition; the pattern is integration-friendly and supports additional connectors.</p><figure><img alt="Architecture of the Customer Lifetime Value multi-agent system" src="https://cdn-images-1.medium.com/max/1024/1*UeERc7qoAtCQnZTsZ7HVzg.png" /><figcaption>Architecture of the Customer Lifetime Value multi-agent system</figcaption></figure><p>In the multi-agent system architecture, a request enters at “Start”, the conditional agent classifies intent, and exactly one specialist is selected — data exploration, insights generator, or strategic reasoning. The “check if chart required” gate then decides whether a visual adds clarity. If a chart is warranted, the “Visualization Agent” prepares a spec and a custom JavaScript function renders the graphic using <a href="https://docs.teradata.com/r/Enterprise_IntelliFlex_VMware/Teradata-Package-for-Python-User-Guide/Plotting-in-teradataml">Teradata plotting</a>; otherwise, the LLM node composes a concise text response. The result is a governed path from question to decision, with visuals used only when they improve understanding.</p><p>The multi-agent system includes the following nodes:</p><ul><li><strong>Start node</strong>: Initializes the run by capturing the user’s input and shared context (customer/segment, time window, policy flags) for downstream nodes.</li><li><strong>Condition agent nodes</strong>: Route the flow based on intent (exploration/insights/strategy) and presentation needs (whether a chart is required).</li><li><strong>Agent nodes</strong>: Execute scoped tasks — data exploration, insight generation, strategic reasoning, and visualization using <a href="https://github.com/Teradata/teradata-mcp-server">Teradata MCP</a> tools against Vantage with clear guardrails.</li><li><strong>LLM node</strong>: Composes the final, business-ready response by fusing agent outputs, rationale, KPIs, and (if present) a chart caption.</li><li><strong>Custom function node</strong>: Renders artifacts like charts from a validated spec and returns an embeddable image/URI for the response.</li></ul><p>Let’s look at the flow of the whole system!</p><h4><strong>Start node</strong></h4><p>The start node initializes the flow, captures the user’s input (chat or form) and sets shared variables — question, optional customer/segment, time window, table handles, and policy flags, for all downstream nodes.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/589/1*a9nTjKmaVv3icO6aKfw_Kw.png" /><figcaption>Start Node configuration</figcaption></figure><h4><strong>Condition agent node: intent identification agent</strong></h4><p>The Intent Identification Agent is the first of the two condition-agent nodes in the system. It decides which agent to call from the three available specialists, based on the scenario expressed in the user’s query.</p><p>There are three scenarios:</p><ul><li>User is asking for basic details about tables, databases, columns, etc.<br>- <em>Routes to </em><strong><em>Data Exploration Agent</em></strong></li><li>User is asking about banking churn, customer lifetime value, CLV, correlation, sentiments analysis, charts like pie chart, histogram etc.<br>- Routes to <strong>Insights Generator Agent</strong></li><li>User is asking about recommendation to reduce the churn, what are the main factors leading to a churn or low CLV, recommend me method to reduce churn, which are the important factor for churn.<br>- Routes to <strong>Strategic Reasoning Agent</strong></li></ul><figure><img alt="Intent Identification Agent node configuration" src="https://cdn-images-1.medium.com/max/577/1*NQB6iwiVU96AMAnSJjYELg.png" /><figcaption>Intent Identification Agent node configuration</figcaption></figure><p>Based on the prompt instructions, an LLM such as <a href="https://azure.microsoft.com/en-us/blog/announcing-the-gpt-4-1-model-series-for-azure-ai-foundry-developers/">Azure ChatOpenAI (GPT-4.1)</a> interprets the intent, selects the appropriate scenario, and triggers the corresponding agent node. The model can be chosen per task; in this agent node, GPT-4.1 is used and configured with an API key (plus endpoint/deployment where applicable).</p><figure><img alt="LLM configuration of Intent Identification Agent node" src="https://cdn-images-1.medium.com/max/541/1*x0jEDP-vjvZ-eu9sVyUJwQ.png" /><figcaption>LLM configuration of Intent Identification Agent node</figcaption></figure><h4><strong>Agent node: data exploration agent</strong></h4><p>Reports what data is available (e.g., databases/tables, key columns, grain, joins, and freshness) and can return small, PII-safe previews using Teradata MCP tools against Vantage.</p><figure><img alt="Data Exploration Agent node configuration" src="https://cdn-images-1.medium.com/max/571/1*lj42_xbEpnT6lq4JwA427Q.png" /><figcaption>Data Exploration Agent node configuration</figcaption></figure><p>The LLM used here is <a href="https://www.anthropic.com/news/claude-4">Anthropic Claude Sonnet 4</a>; it can be configured the same way as GPT-4.1. All access to LLMs is provided through a cloud service provider. The specific models available, along with credentials and base URLs, depend on the deployment of Teradata and AgentBuilder.</p><figure><img alt="LLM configuration of Data Exploration Agent node" src="https://cdn-images-1.medium.com/max/549/1*Cca7AQQZkh5uvIqim3tdCg.png" /><figcaption>LLM configuration of Data Exploration Agent node</figcaption></figure><p>Access to the Teradata MCP server is configured through the “Tool” panel. As with LLM integration, the connection URL and credentials depend on the specific deployment of Teradata and AgentBuilder.</p><p>Providing the right tools via Teradata MCP is critical for accurate LLM outputs. Select and scope these tools in the “Available Actions” panel so each agent can call exactly what it needs for governed, reliable results.</p><p>Tools can only be selected under Available Actions once MCP server is configured.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/526/1*phUieatgMOndHCMqIJPlJg.png" /><figcaption>MCP configuration of Data Exploration Agent node</figcaption></figure><h4><strong>Agent node: insight generator agent</strong></h4><p>Interprets the question and derives findings only from the governed data, following a reasoning-first workflow: quick data checks (columns, types, missingness), method selection (aggregations, correlations, simple tests), and a brief rationale. It performs the calculation, not just a description, and explains the process before stating results.</p><p>For simple scalar asks, it stays concise; for richer questions, it returns a compact business summary. Output is markdown with “Reasoning” and “Conclusion” sections for clarity and reuse.</p><figure><img alt="Insight Generator Agent node configuration" src="https://cdn-images-1.medium.com/max/564/1*fFdRn3NQzMp3NoVb8oGWEg.png" /><figcaption>Insight Generator Agent node configuration</figcaption></figure><p>The LLM and MCP configurations mirror the Data Exploration Agents configuration, with the addition of <strong>Sentiment Extractor</strong> in Available Actions when needed.</p><figure><img alt="MCP tools configuration of Insight Generator Agent node" src="https://cdn-images-1.medium.com/max/468/1*nTjveqMvf5MDrfDpennfjg.png" /><figcaption>MCP tools configuration of Insight Generator Agent node</figcaption></figure><h4><strong>Agent node: strategic reasoning agent</strong></h4><p>Converts analytical findings into actionable churn-reduction strategies by reviewing feature importance, explaining how key factors drive attrition, and prioritizing where to intervene. It proposes evidence-based methods (e.g., onboarding improvements, complaint triage, targeted cross-sell) and links each recommendation to the contributing drivers with clear rationale and guardrails. Responses are concise and structured in markdown with an intro, numbered reasoning steps, recommended methods to reduce banking churn, and a brief summary of the highest-impact actions.</p><figure><img alt="Strategic Reasoning Agent node configuration" src="https://cdn-images-1.medium.com/max/565/1*LVhcY53VzB0HR3WiCRXfZQ.png" /><figcaption>Strategic Reasoning Agent node configuration</figcaption></figure><p>The MCP and LLM configuration remains the same, with only changes in MCPs Available Actions.</p><figure><img alt="MCP tool configuration of Strategic Reasoning Agent node" src="https://cdn-images-1.medium.com/max/423/1*dnGQcblY4uTp_Bf6q-DBeQ.png" /><figcaption>MCP tool configuration of Strategic Reasoning Agent node</figcaption></figure><h4><strong>Condition agent: check if chart required</strong></h4><p>Decides whether the answer needs a visualization or should remain textual. It inspects the prompt and, only when the user asks to plot, selects a chart type from <strong>Line</strong>, <strong>Polar</strong>, <strong>Pie</strong>, or <strong>Radar</strong>; single-value outputs (e.g., <em>corr: 0.89</em>) skip charts. If a valid chart is warranted, it passes a clean spec to the Visualization Agent and Custom Function to render, otherwise it defaults to a table or plain response and routes directly to the “Generate Response” LLM node.</p><figure><img alt="Check if chart required node configuration" src="https://cdn-images-1.medium.com/max/555/1*2Nba8xE4xrcnS7BElBzpTw.png" /><figcaption>Check if chart required node configuration</figcaption></figure><p>The LLM configuration mirrors the earlier condition agent, this node takes the question as input and chooses between two scenarios — chart-related or general. If chart required is selected, it triggers the Visualization Agent to plot a graph, otherwise, it returns a straightforward textual output.</p><figure><img alt="Input and Scenarios detail of Check if chart required node" src="https://cdn-images-1.medium.com/max/564/1*vkh7GETDIbrnuT1UabtWZA.png" /><figcaption>Input and Scenarios detail of Check if chart required node</figcaption></figure><h4><strong>LLM node: generate response</strong></h4><p>Composes the final, business-ready answer using the context passed from upstream agents, focusing on clear reasoning and actionable guidance (no code). It identifies the domain of the ask (visualization, tables, churn reduction, database design), applies best-practice logic, and balances detail with brevity for simple scalar questions. The output is 2–4 tight paragraphs that explain the “why,” list concrete next steps, note caveats, and tie recommendations to measurable impact.</p><figure><img alt="Generate Response node configuration" src="https://cdn-images-1.medium.com/max/559/1*pdbSI2alUgN0GbkiALHttg.png" /><figcaption>Generate Response node configuration</figcaption></figure><p>For clean presentation, <strong>Return Response As</strong> is set to <em>Assistant Message.</em></p><figure><img alt="Selection of Assistant message in Generate Response node" src="https://cdn-images-1.medium.com/max/547/1*zz8JOI6PTJQPHh1i8D2vsg.png" /><figcaption>Selection of Assistant message in Generate Response node</figcaption></figure><h4><strong>Agent node: visualization agent</strong></h4><p>Emits a strict JSON payload — no prose, notes, or code. If unsure about columns, it first verifies with MCP tools before invoking any chart tools. When a chart is needed, it returns only this structure:</p><pre>{ &quot;type&quot;: &quot;&lt;line|polar|pie|radar&gt;&quot;, &quot;title&quot;: &quot;&lt;chart title&gt;&quot;, &quot;labels&quot;: […], &quot;datasets&quot;: […] }</pre><p>If a chart isn’t appropriate, it returns JSON for a simple table instead, without adding or removing fields or including extra text.</p><figure><img alt="Visualization Agent node configuration" src="https://cdn-images-1.medium.com/max/544/1*IFIh11AxpGEvLGp2XaDv1Q.png" /><figcaption>Visualization Agent node configuration</figcaption></figure><p>Teradata MCP <strong>Available Actions</strong> include Teradata plotting tools that help generate the requested visual.</p><figure><img alt="MCP tool configuration of Visualization Agent node" src="https://cdn-images-1.medium.com/max/409/1*-r-5LnX593SScn01nDCiNg.png" /><figcaption>MCP tool configuration of Visualization Agent node</figcaption></figure><p>It can generate four chart types:</p><ul><li><strong>Line</strong> — trends over time (e.g., weekly churn rate by segment)</li><li><strong>Polar</strong> — circular comparison of magnitudes (e.g., complaints by category)</li><li><strong>Pie</strong> — part-of-whole (e.g., share of customers by risk tier)</li><li><strong>Radar</strong> — multi-metric profile (e.g., cohort scores across engagement, spend, friction)</li></ul><h4><strong>Custom function: draw function</strong></h4><p>The Draw Function turns the Visualization Agent’s JSON into a live chart. It expects a clean payload with type, title, labels, and datasets; if that structure is present, it renders an embedded Chart.js graphic in an iframe. If the input isn’t valid chart data (or the task doesn’t require a chart), it gracefully falls back to a plain message instead of breaking the flow. Supported chart types match the gating node: line, polar, pie, and radar.</p><p>To wire it in the flow, pass the Visualization Agent’s output to the function as “<em>input_data = {{$flow.state.main_vis_op}}</em>”. The function parses that value, validates that labels and datasets are arrays, and then builds the chart with the provided title. If validation fails, the function simply returns the text it received (e.g., “No chart needed”), keeping responses clean and predictable.</p><p>This approach keeps rendering logic outside the agents, avoids code in the final answer, and ensures visuals appear only when they add clarity, everything else stays fast, lightweight, and easy to audit.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/568/1*gjxRblF_w4tm9_FZnY4b1Q.png" /><figcaption>Draw Function node configuration</figcaption></figure><h4><strong>Output</strong></h4><p>The CLV multi-agent system tailors each response to the question — returning plain, business-ready text when a single number suffices (e.g., “What is the average CLV?” → $412.37), and a chart when it adds clarity. Two quick examples from the demo:</p><ul><li><strong>Text only ask</strong>: “What is the average value of CLV”<br>- <strong>Flow</strong>: Start → Intent Identification Agent → Insights Generator Agent → Generate Response</li></ul><p>The assistant computes the mean CLV in-database and returns a concise explanation (no visualization).</p><figure><img alt="Response and process flow details of “What is the average value of CLV” question" src="https://cdn-images-1.medium.com/max/883/1*Imm730hwR6qvZmKzSiT7wA.png" /><figcaption>Response and process flow details of “What is the average value of CLV” question</figcaption></figure><ul><li><strong>Visualization a</strong>sk: “Create a visualization for average value of balance and CLV”<br>- <strong>Flow</strong>: Start → Intent Identification Agent → Insights Generator Agent → Check if chart required → Visualization Agent → Draw Function</li></ul><figure><img alt="Response details of “Create a visualization for average value of balance and CLV” question" src="https://cdn-images-1.medium.com/max/883/1*d7DiESAzc0xfHXnOArvfTQ.png" /><figcaption>Response details of “Create a visualization for average value of balance and CLV” question</figcaption></figure><p>The chart gate detects a plotting request, the Visualization Agent emits a JSON spec, and the Draw Function renders a bar chart.</p><figure><img alt="Process flow details of “Create a visualization for average value of balance and CLV” question" src="https://cdn-images-1.medium.com/max/861/1*9WgMLf2DKkOwOXgDdXTD1w.png" /><figcaption>Process flow details of “Create a visualization for average value of balance and CLV” question</figcaption></figure><p>Every response is traceable, the process flow shows which nodes executed, and the output links back to the data and policies used. Visuals appear only when requested or clearly helpful, otherwise, the answer stays fast, readable, and ready to act on.</p><h3><strong>Conclusion</strong></h3><p>Teradata’s Customer Intelligence Framework closes the loop from <strong>data → signal → activation</strong>. Data remains governed in Vantage; signals are detected and interpreted in real time; and activation happens through guard-railed next-best actions that are explainable and measurable. Built in <a href="https://www.teradata.com/press-releases/2025/teradata-unveils-agentbuilder">AgentBuilder</a>, the CLV demo shows <strong>autonomous customer intelligence</strong> in practice: agents understand intent, assemble evidence, and trigger the right response — text for simple answers, visuals when they add clarity, and recommendations when action is warranted.</p><p>The pattern extends beyond customer lifetime value to use cases such as churn reduction, cross-sell optimization, and service issue deflection, without redesigning the underlying architecture. Performance comes from keeping computation in Teradata Vantage; auditability comes from end-to-end traces of signals, decisions, and actions; and delivery speed comes from reusable prompts, tools, and node templates assembled in AgentBuilder.</p><p><a href="https://www.teradata.com/platform/clearscape-analytics/agent-ai">Ready to turn signals into outcomes?</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=a28d8dd77c41" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/building-autonomous-customer-intelligence-a-developers-guide-to-teradata-s-customer-intelligence-a28d8dd77c41">Building Autonomous Customer Intelligence: A Developer’s Guide to Teradata’s Customer Intelligence…</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Chat with your data in Visual Studio Code through an MCP server]]></title>
            <link>https://medium.com/teradata/chat-with-your-data-in-visual-studio-code-through-an-mcp-server-071673bd3ffe?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/071673bd3ffe</guid>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[teradata]]></category>
            <category><![CDATA[mcp-server]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[ai-agent]]></category>
            <dc:creator><![CDATA[Joshitha Recharla]]></dc:creator>
            <pubDate>Wed, 17 Dec 2025 17:45:17 GMT</pubDate>
            <atom:updated>2025-12-17T17:44:46.853Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="Teradata’s Model Context Protocol (MCP)" src="https://cdn-images-1.medium.com/max/1024/1*ibiQZwku-XO4iMWjaV4Pvg.png" /><figcaption>Teradata’s Model Context Protocol (MCP) with GitHub Copilot</figcaption></figure><p>If you build data pipelines or analytic Python scripts, like dbt models, in Visual Studio Code (VS Code), you know the pain of context-switching just to sanity-check a column list or scan for NULLs. Traditional SQL clients are great, but bouncing between these database clients and VS Code can be distracting and break your flow when your focus is the code you’re writing on VS Code.</p><p>Teradata’s Model Context Protocol (MCP) server helps you by allowing GitHub Copilot Chat to talk to your data directly, providing a handy assistant for drafting analytics scripts.</p><p>Teradata MCP server is an open and extensible backend server that acts as a bridge between large language models (LLMs) and your Teradata environment,<strong> </strong>allowing you to query data, explore metadata, and even validate logic just by asking questions in plain English.</p><p>You stay in VS Code, GitHub Copilot routes data-related requests to the MCP server, and results come back in the chat.</p><p><strong>With this setup, GitHub Copilot can:</strong></p><ul><li>Interpret your natural language questions</li><li>Translate them into SQL queries behind the scenes</li><li>Connect to your Teradata Vantage® instance</li><li>Retrieve structured answers directly in your Copilot chat window</li></ul><p><strong>In this article, I’ll walk you through:</strong></p><ul><li>How to set up your development environment to take advantage of Teradata MCP Server</li><li>How to set up your own MCP server to work with GitHub Copilot Chat in VS Code</li><li>How to query your Teradata database as if<em> </em>you’re having a conversation, simplifying chores like discovering tables, validating filters, and validating data quality in VS Code</li></ul><p>Let’s begin the setup.</p><h3><strong>Requirements</strong></h3><ul><li>A free Teradata VantageCloud environment. <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">Get one in just a few minutes via ClearScape Analytics® Experience.</a></li><li>Visual Studio Code with the Copilot Chat and MCP-Client extensions. The download is available at <a href="https://code.visualstudio.com/">https://code.visualstudio.com/</a>.</li><li>A Teradata MCP server. An open-source community edition is available in Teradata’s GitHub organization; later we’ll cover the details for cloning this repository and setting up the server.</li></ul><h3><strong>Preparing the development environment</strong></h3><h4><strong>Log into ClearScape Analytics® Experience</strong></h4><ul><li><a href="https://www.teradata.com/getting-started/demos/clearscape-reg?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">Sign up for a free account</a> if you haven’t already, or log in.</li><li>In the console, select “create new environment.” Take note of your host and password and the username and database, which will also be needed to set up your server later, default to demo_user in ClearScape Analytics® Experience.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*vFATLUA3xFoU8Qe4FnzFUg.png" /><figcaption>Creating a new ClearScape Analytics environment</figcaption></figure><h4><strong>Install uv</strong></h4><p>We utilize the uv package manager to run the Teradata MCP server in an isolated Python environment to avoid package conflicts and ease the setup.</p><ul><li>Install uv by following the steps at <a href="https://docs.astral.sh/uv/getting-started/installation/">https://docs.astral.sh/uv/getting-started/installation/</a></li><li>Verify if uv is properly installed, open your terminal, and run: ```uv –v```</li></ul><p>This will display the installed version of uv. If you see any error, make sure you have followed the installation instructions and restart the console.</p><h3><strong>Setting up your own Teradata MCP server to work with GitHub Copilot in Visual Studio Code</strong></h3><h4><strong>Clone the Teradata MCP Server</strong></h4><p>git clone <a href="https://github.com/Teradata/teradata-mcp-server.git">https://github.com/Teradata/teradata-mcp-server.git</a></p><h4><strong>Set environment variables</strong></h4><p>Create a .env file with the following structure:</p><pre>DATABASE_URI= teradata://username:password@host:1025/databasename<br>LOGMECH=TD2 #TD2 or LDAP<br>MCP_TRANSPORT=stdio<br>PROFILE = dba</pre><p>We’ll use the stdio transport method in this setup, though other methods are also available. Check the <a href="https://github.com/Teradata/teradata-mcp-server/blob/main/docs/server_guide/CONFIGURATION.md#-transport-modes">relevant documentation</a> in the repository for details.</p><ul><li>Replace username, password, host, and database name with your own Teradata Vantage® credentials.</li><li>The DATABASE_URI tells the server where to point your queries.</li><li>Given the tasks we’re performing, we need a profile that enables the required tools. Specifically, set PROFILE=dba to access the required toolset.</li></ul><h4><strong>Start the MCP Server</strong></h4><p>From your project directory (i.e., the folder where you cloned the teradata-mcp-server), test starting the MCP server using the following command.</p><pre>uv run teradata-mcp-server</pre><p>This boots up the MCP server to verify that everything is working as expected. If everything is in order, you can interrupt this run by pressing Ctrl + C.</p><p><strong>Connecting the MCP server to GitHub Copilot in VS Code</strong></p><p>Now let’s link this server to your GitHub Copilot:</p><ol><li>Open VS Code</li><li>Press Ctrl + Shift + P to open the Command Palette</li><li>Select MCP: Add Server</li><li>Choose <strong>stdio</strong> as the transport type</li><li>Fill in the dialog as follows:</li><li>Command: “uv”</li><li>ID: For example, “teradata-mcp-server”</li><li>Then confirm and save now</li></ol><p>An mcp.json file will be created automatically. Complete the required arguments as shown below:</p><pre>{<br>&quot;mcp&quot;: {<br>&quot;servers&quot;: {<br>&quot;teradatastdio&quot;: {<br>&quot;type&quot;: &quot;stdio&quot;,<br>&quot;command&quot;: &quot;uv&quot;,<br>&quot;args&quot;: [<br>&quot; - directory&quot;,<br>&quot;&lt;Full Path&gt;/teradata-mcp-server&quot;,<br>&quot;run&quot;,<br>&quot;teradata-mcp-server&quot;<br>],<br>}<br>}<br>}<br>}</pre><p>Make sure to replace &lt;full path&gt; with the actual location of the teradata-mcp-server directory if it didn’t load itself.</p><h4><strong>Launch the MCP Server in VS Code</strong></h4><p>Once configured, start the server from within VS Code if it didn’t start itself:</p><ul><li>Friendly controls to start, restart, and stop the server overlay on top of the mcp.json file for convenient operation</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*mAmw1CjrQIrEHgClGBZElA.png" /><figcaption>Configuring the MCP server in mcp.json to run the Teradata MCP server via stdio.</figcaption></figure><h4><strong>Chat with your data</strong></h4><p>To interact with your Teradata database in natural language, you’ll need to use Copilot in Agent Mode, which allows GitHub Copilot Chat to talk directly to the MCP server.</p><p>Here’s how to do it.</p><h4><strong>Open GitHub Copilot Chat in VS Code</strong></h4><ul><li>In your sidebar, click on the GitHub Copilot Chat icon.</li><li>If you don’t see it, open the Command Palette (Ctrl + Shift + P) and search for “GitHub Copilot: Open Chat View”</li></ul><h4><strong>Switch to Agent Mode</strong></h4><ul><li>Go to the Copilot Chat panel in VS Studio</li><li>Ensure it’s set to Agent Mode (not Chat Mode). If you don’t see this, go to command palette to enter Copilot: Toggle Agent Mode.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*M_kRqPZZ9hk2mgvQlMnWBA.png" /><figcaption>Set GitHub Copilot Chat to Agent Mode to allow MCP-based tool and database interactions.</figcaption></figure><p>Once in Agent Mode, Copilot will be ready to route your questions through the MCP server to your Teradata backend and return results right inside your chat window.</p><p><strong>Note on data:</strong> By default, our walkthrough defines demo_user as the default database. This database is empty in a new ClearScape Analytics® environment, you should load sample data first or update the database/table names in the examples to match tables that exist in your environment. This ensures you can explore and query your data.</p><p>Try prompts like:</p><ul><li>“Show me the first 5 rows from demo_user.dim_customers”</li><li>“What tables exist in the demo_user schema?”</li><li>“Give me the total sales by region from demo_user.sales_fact”</li><li>“Which columns are available in the demo_user.orders table?”</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*vOm7VydQFTEoRqisrsJyjg.png" /><figcaption>Querying the Teradata database in natural language via GitHub Copilot Chat, with requests executed through the MCP server.</figcaption></figure><h3><strong>Conclusion</strong></h3><p>By combining Teradata’s MCP Server with GitHub Copilot Chat in Visual Studio Code, we unlock a powerful way to make data feel more conversational, accessible, and integrated into the daily developer workflow.</p><p>If you’re someone who works with Teradata and builds in VS Code, I highly recommend giving this workflow a try. Running the queries and getting results in the chat might boost your productivity.</p><p>Got questions or want to share your own use case? Feel free to reach out or leave a comment.</p><h3>About Joshitha Recharla</h3><p>Joshitha Recharla is a developer advocate intern at Teradata and a graduate student in data science at SUNY Buffalo. She enjoys breaking down complex AI and analytics workflows into hands-on, developer-friendly tutorials and demos. Through her internship, she has explored how tools like MCP can make data more accessible and useful where developers work.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=071673bd3ffe" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/chat-with-your-data-in-visual-studio-code-through-an-mcp-server-071673bd3ffe">Chat with your data in Visual Studio Code through an MCP server</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
    </channel>
</rss>