<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[Teradata - Medium]]></title>
        <description><![CDATA[Technical articles for developers using the Teradata Vantage Platform - Medium]]></description>
        <link>https://medium.com/teradata?source=rss----8adfdb056496---4</link>
        <image>
            <url>https://cdn-images-1.medium.com/proxy/1*TGH72Nnw24QL3iV9IOm4VA.png</url>
            <title>Teradata - Medium</title>
            <link>https://medium.com/teradata?source=rss----8adfdb056496---4</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Thu, 23 Apr 2026 09:42:17 GMT</lastBuildDate>
        <atom:link href="https://medium.com/feed/teradata" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[Carbon Footprint Analytics: a data integration story]]></title>
            <link>https://medium.com/teradata/carbon-footprint-analytics-a-data-integration-story-d9f37e95f2d7?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/d9f37e95f2d7</guid>
            <category><![CDATA[data-analysis]]></category>
            <category><![CDATA[ghg-emissions]]></category>
            <category><![CDATA[esg]]></category>
            <category><![CDATA[advanced-analytics]]></category>
            <category><![CDATA[climate-change]]></category>
            <dc:creator><![CDATA[Gregory Leduc]]></dc:creator>
            <pubDate>Mon, 13 Apr 2026 14:50:47 GMT</pubDate>
            <atom:updated>2026-04-13T14:30:40.598Z</atom:updated>
            <content:encoded><![CDATA[<p>Measuring GHG emissions is a data integration challenge. See how Teradata’s CFA solution accelerator tackles it.</p><figure><img alt="Galina Nelyubova
 Earth Day — a collection of environmental 3D illustrations" src="https://cdn-images-1.medium.com/max/975/1*gXIDV8mS_xphxbw0XDSGYg.png" /></figure><p><em>Galina Nelyubova ; Earth Day — a collection of environmental 3D illustrations</em></p><p>Climate emergency intensifies, and companies must play a significant role in ensuring a sustainable future. As climate change is one of the biggest systemic risks, mitigation of adverse financial effects and adaptation to rapidly changing circumstances is key for companies to ensure long-term going concern and growth. Centerpiece for these activities is proper reporting and reducing of companies’ greenhouse gas (GHG) emissions. And doing so requires much more than good intentions: it demands robust data integration, and the right platform.</p><h4><strong>Why Carbon footprint analytics matters</strong></h4><p>Every company’s operations — from logistics and manufacturing to IT and facilities — generate emissions. These emissions are environmental concerns, and they can also become a financial and reputational risk. Regulatory bodies are moving toward mandatory enterprise-wide GHG reporting, and stakeholders are demanding transparency and action.</p><p>To meet these expectations, companies must answer critical questions:</p><p>· How much GHG emissions are we responsible for?</p><p>· Which activities contribute the most?</p><p>· What’s our roadmap to a sustainable level of GHG emissions? (Ones would say “path to net-zero”, but let’s focus on reduction first before compensating).</p><p>The answers lie within the data companies already possess: financial systems, operational logs, energy consumption records… All these contain the clues needed to build a comprehensive GHG emissions profile. And here is precisely the challenge most companies are facing: transforming existing data into actionable insights.</p><figure><img alt="Figure 1 — Calculating a carbon footprint is easy… until your start digging into the details!" src="https://cdn-images-1.medium.com/max/661/1*gPTBQxdqaN6fy4vYg3v9ew.png" /></figure><p><em>Figure 1 — Calculating a carbon footprint is easy… until your start digging into the details!</em></p><h4>Why it’s a data integration story</h4><p>Carbon footprint analytics is fundamentally a data integration challenge. As the saying goes, “What gets measured, gets managed”. And measuring emissions accurately means harmonizing data from disparate sources, each with its own format, granularity, units, and quality. Trying to report GHG emissions without data integration is like assembling a jigsaw puzzle where pieces are taken from different puzzles: You’ll never see the full picture.</p><p>The Carbon Footprint Analytics (CFA) solution accelerator designed by Teradata tackles this complexity challenge. Based on a flexible data model, it ingests enterprise-specific data using proven algorithms, and outputs detailed GHG emissions across all scopes (1, 2 and 3) defined by the <a href="https://ghgprotocol.org/">GHG Protocol</a>.</p><p>Key capabilities include:</p><p>· <strong>Unit harmonization:</strong> automatically converts measures to configurable preferred units (pounds vs kilograms, kWh vs MWh, etc.)</p><figure><img alt="Figure 2 — Automated management of measure systems heterogeneity" src="https://cdn-images-1.medium.com/max/704/1*_MOnELgq78QslHRBBj064Q.png" /></figure><p><em>Figure 2 — Automated management of measure systems heterogeneity</em></p><p>· <strong>Geospatial awareness:</strong> identifies the right emission factors based on location, which is particularly critical for electricity consumption (electricity mix, thus GHG emissions, significantly differ <a href="https://app.electricitymaps.com/map/all/monthly">between countries and regions</a>)</p><p>· <strong>Temporal modeling:</strong> the CFA design reflects evolving emission factors and activity periods.</p><p>A key point to understand is that CFA is not a reporting tool. It is an engine in which we can connect reporting tools. This engine integrates and enriches data to deliver a harmonized, enterprise-wide view of emissions.</p><h4>Think before you jump: design matters</h4><p>Before diving into carbon analytics, companies must design a robust and flexible data model. A poorly structured model can lead to wasted efforts and inconsistent results.</p><p>The CFA solution accelerator emphasizes thoughtful architecture:</p><p>· A core data model that supports multiple aggregation levels and units</p><p>· Pre-integrated calculation rules for seamless processing and explainable results</p><p>· Advanced analytics capabilities, including machine learning, geospatial, time-series and string similarities.</p><figure><img alt="Figure 3 — Overview of the CFA data model, designed for scope 1, 2 and 3 GHG emissions" src="https://cdn-images-1.medium.com/max/975/1*nT9Z8eWEynEGFAvll72CNw.png" /></figure><p><em>Figure 3 — Overview of the CFA data model, designed for scope 1, 2 and 3 GHG emissions</em></p><p>This foundation ensures that companies can not only report emissions, but also forecast, simulate and optimize their GHG emissions reduction strategy.</p><h4>Why Teradata is the right platform</h4><p>Teradata Vantage is well positioned to support Carbon Footprint Analytics:</p><p>· <strong>Geospatial capabilities</strong>: visualize GHG emissions and apply location-specific factors.</p><figure><img alt="Figure 4 — Geospatial capabilities facilitate segregation of GHG emissions according to location" src="https://cdn-images-1.medium.com/max/700/1*1zHzQzNv2UL5vrF--gvr4A.png" /></figure><p><em>Figure 4 — Geospatial capabilities facilitate segregation of GHG emissions according to location</em></p><p>· <strong>Temporal intelligence</strong>: track changes over time with time-aware tables.</p><p>· <strong>Advanced analytics</strong>: leverage ClearScape Analytics™ features to observe the trends and to simulate the effect of your strategic decisions for multiple scenarios.</p><p>· <strong>Scalable integration</strong>: harmonize data from multiple sources with high performance and reliability.</p><figure><img alt="Figure 5 — Multiple data sources must be efficiently integrated, both internal and external" src="https://cdn-images-1.medium.com/max/668/1*QQfSGE-14XddC6sXGakWgQ.png" /></figure><p><em>Figure 5 — Multiple data sources must be efficiently integrated, both internal and external</em></p><p>The <a href="https://clearscape.teradata.com">ClearScape Analytics Experience</a> demo showcases CFA in action — transforming raw data into <a href="https://nbviewer.org/github/Teradata/jupyter-demos/blob/85d6dd55cfe0e8b509e536bb4e61d9e45a2cb5c9/UseCases/Carbon_Footprint_Analytics/Pop_Carbon_Footprint_Analytics_PY_SQL.ipynb">meaningful carbon insights</a>. So, what are you waiting for? Deploy your own CFA solution accelerator, with Teradata ready to support you every step of the way.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=d9f37e95f2d7" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/carbon-footprint-analytics-a-data-integration-story-d9f37e95f2d7">Carbon Footprint Analytics: a data integration story</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Building Multimodal Agentic Search Over Images and Documents — A DevTalk]]></title>
            <link>https://medium.com/teradata/building-multimodal-agentic-search-over-images-and-documents-a-devtalk-fdf1bf3029e5?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/fdf1bf3029e5</guid>
            <category><![CDATA[retrieval-augmented-gen]]></category>
            <category><![CDATA[unstructured-data]]></category>
            <category><![CDATA[teradata]]></category>
            <category><![CDATA[image-recognition]]></category>
            <category><![CDATA[multimodal-ai]]></category>
            <dc:creator><![CDATA[Daniel Herrera]]></dc:creator>
            <pubDate>Thu, 09 Apr 2026 12:55:14 GMT</pubDate>
            <atom:updated>2026-04-09T12:55:13.358Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="" src="https://cdn-images-1.medium.com/max/480/1*mNnGddOhjpio4Mw-tnHa3w.jpeg" /><figcaption>Teradata DevTalk — Multimodal Agentic Semantic Search</figcaption></figure><p>Recently I had the opportunity to host a <a href="https://www.youtube.com/live/G_KXqo8jXuc">DevTalk</a> that focused on something many teams are actively experimenting with right now: how to reason over unstructured data using multimodal embeddings and agentic retrieval patterns.</p><p>Together with <a href="https://www.linkedin.com/in/ajaykrish/">Ajay Gopalan</a> from <a href="https://unstructured.io/">Unstructured.io</a>, we walked through a live, <a href="https://github.com/Teradata/jupyter-demos/blob/main/VantageCloud_Lake/UseCases/Multimodal_Agentic_Semantic_Search/Agentic_Semantic_Search_Teradata_Unstructured.ipynb">end‑to‑end demo</a> that shows how images, documents, and text can be brought together. The goal start from raw unstructured data and end with an experience where you can select an image, ask questions about it, and receive answers grounded in real enterprise documentation.</p><p>The full session is available on YouTube, and if you are working with vector search, multimodal models, or agentic workflows, it is worth watching the demo in full.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FG_KXqo8jXuc%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DG_KXqo8jXuc&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FG_KXqo8jXuc%2Fhqdefault.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/79e5bd6086933a7020f876e3cdf68e87/href">https://medium.com/media/79e5bd6086933a7020f876e3cdf68e87/href</a></iframe><p>Below is a summary of what we covered and why it matters:</p><h4>Why Multimodal Search Keeps Coming Up</h4><p>Images exist everywhere in enterprise environments. They appear inside PDFs, research documents, medical records, financial reports, and technical documentation. On their own, images rarely provide enough context to answer useful questions. The surrounding text, tables, and metadata usually carry the explanation.</p><p>The challenge is connecting those pieces in a way that scales.</p><p>During the session, we framed the problem around a simple interaction. A user selects an image and asks a question. The system retrieves the relevant documentation that explains that image and uses it to produce a grounded answer. That interaction requires coordinated ingestion, embedding, storage, retrieval, and reasoning across multiple data modalities.</p><h4>Structuring Unstructured Data with Unstructured</h4><p>We started with ingestion, using Unstructured to process a mix of composite PDFs and standalone image files. Ajay walked through how Unstructured workflows break documents into structured elements such as text blocks, tables, and images. These elements are then enriched and prepared for downstream use.</p><p>One point that came up during the demo is how useful the Unstructured UI is during experimentation. You can design and test workflows visually, refine parsing strategies, and then transition to the API when you are ready to automate ingestion at scale. The same pipeline is used in both cases.</p><h4>Storing and Searching Embeddings in Teradata VantageCloud</h4><p>Once the documents and images are processed, embeddings are generated and stored inside Teradata using Enterprise Vector Store capabilities. This keeps vector search close to the rest of the enterprise data, with the same governance, access control, and operational model teams already rely on.</p><p>In the demo, we created a vector store from the generated embeddings and preserved metadata such as file names and record locators. That metadata becomes important later, when retrieved results need to be traced back to their original documents.</p><p>From a data platform perspective, this aligns well with how many teams already think about indexing. The mental model remains focused on accelerating discovery and retrieval, with similarity used as the matching mechanism instead of exact values.</p><h4>Adding Agentic Reasoning on Top of Retrieval</h4><p>Retrieval alone gets you relevant context. Agentic workflows turn that context into an interactive experience.</p><p>Using LangChain integrations with Teradata, we built a small set of tools that the agent can use. One tool performs similarity search over embeddings to locate the most relevant documentation for a selected image. Another tool retrieves and displays the source document associated with that result.</p><p>The agent prompt describes how these tools should be used together. When a user selects an image and asks a question, the agent retrieves the closest matching documentation and generates an answer grounded in that evidence. The reasoning step stays tied to retrieved content instead of relying on general knowledge.</p><p>During the demo, we asked questions about medical images, including tissue health and visible markers. Each response referenced information asdpulled from the most relevant documents in the library.</p><h4>Observations from the Live Discussion</h4><p>The audience questions touched on topics many teams are evaluating right now.</p><p>Cost considerations came up early, particularly around embeddings and multimodal models. As with most architectural decisions, the balance depends on scale and workload. Bringing vector search into the same platform as analytics removes certain integration and duplication costs that often appear with external systems.</p><p>We also discussed entry points. Multimodal search is compelling, though many teams begin with document retrieval, tables, policies, and reports. These use cases already benefit from structured parsing and vector search and provide a solid foundation before adding image‑based reasoning.</p><p>Another area of interest was familiarity. Vector stores can feel unfamiliar at first, though the underlying ideas map closely to indexing strategies DBA teams already use. Teradata’s approach keeps vector operations accessible through both SQL and Python, which helps bridge that gap.</p><h4>Try the Demo Yourself</h4><p>The full demo runs inside <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">Teradata.com/experience</a>, which is available at no cost. It includes the notebooks, ingestion pipeline, vector store setup, and agentic workflow shown during the livestream. Unstructured also provides free credits that make it easy to reproduce and extend the example.</p><p>If you are exploring multimodal embeddings, agentic retrieval patterns, or enterprise‑grade vector search, the livestream walks through the full system working together, we have also prepared a detailed <a href="https://medium.com/teradata/end-to-end-multimodal-rag-and-agents-with-teradata-and-unstructured-c3686b5c6198">blog for step by step guidance.</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=fdf1bf3029e5" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/building-multimodal-agentic-search-over-images-and-documents-a-devtalk-fdf1bf3029e5">Building Multimodal Agentic Search Over Images and Documents — A DevTalk</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[End-to-End Multimodal RAG and Agents With Teradata and Unstructured]]></title>
            <link>https://medium.com/teradata/end-to-end-multimodal-rag-and-agents-with-teradata-and-unstructured-c3686b5c6198?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/c3686b5c6198</guid>
            <category><![CDATA[unstructured-data]]></category>
            <category><![CDATA[retrieval-augmented-gen]]></category>
            <category><![CDATA[teradata]]></category>
            <category><![CDATA[agentic-ai]]></category>
            <category><![CDATA[vector-database]]></category>
            <dc:creator><![CDATA[Daniel Herrera]]></dc:creator>
            <pubDate>Mon, 23 Mar 2026 12:06:15 GMT</pubDate>
            <atom:updated>2026-03-23T12:06:14.172Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="Multimodal RAG and Agents with Teradata &amp; Unstructured" src="https://cdn-images-1.medium.com/max/1024/1*6hQeuUsHuF70O4j6KCXSXg.png" /><figcaption>Multimodal RAG and Agents with Teradata &amp; Unstructured</figcaption></figure><h3><strong>The challenge of grounding the work of agentic systems</strong></h3><p>Agentic systems work across multiple steps to achieve a specific goal. Precision and predictability in achieving that goal require rich context that often goes beyond structured transactional data. Policies, procedures, diagnostics, design documents, and customer conversations are often stored as documents, PDFs, images, and audio files. When that content is transformed into embeddings, making it searchable, agents gain something closer to long-term memory: they can retrieve relevant context, cite sources, and ground their responses in what is true for the organization, making them accurate and traceable.</p><p>Most enterprise data stacks split these two worlds cleanly. Transactional facts live in relational tables. Policies, manuals, and media live somewhere else, in object storage or file systems. Bringing them together requires additional data-processing pipelines and adds governance overhead.</p><p>It is possible, though, to bridge the gap between these two worlds, and it can be done on the same platform, without data duplication, data movement overhead, and extra governance layers.</p><h3><strong>Multimodal agentic development with Teradata and Unstructured</strong></h3><p>Teradata Enterprise Vector Store has unleashed a myriad of use cases for Teradata customers, including:</p><ul><li>Customer 360 enrichment and contact center intelligence</li><li>Autonomous customer insights</li><li>Compliance and policy intelligence</li><li>Insurance claims and contract analysis</li><li>R&amp;D and engineering knowledge retrieval</li><li>Large‑scale enterprise search</li><li>Defense and public‑sector operational intelligence</li></ul><p>These use cases are enabled by performing vector and metadata-aware search across structured and unstructured data in a single operation, directly within the Teradata system as the single source of truth.</p><p>Teradata’s partnership with Unstructured extends these capabilities by introducing a unified multimodal ingestion and enrichment layer capable of parsing <a href="https://docs.unstructured.io/api-reference/supported-file-types">70+ file types</a> and generating embeddings for text and images, with audio support on the roadmap. The Unstructured pipelines handle chunking, metadata extraction, enrichment, adaptive model selection, and embedding generation, and deliver these enriched artifacts directly into Teradata’s enterprise analytics platform. The Teradata platform provides a native security model, object‑level governance, and lineage propagation. Because the Enterprise Vector Store is implemented on top of Teradata’s MPP execution framework, vector indexing, similarity search, and hybrid semantic–lexical retrieval scale transparently with other analytic and operational workloads, inheriting Teradata’s concurrency, data locality, and high‑throughput characteristics.</p><p>For development, the <a href="https://github.com/Teradata/langchain-teradata">langchain-teradata</a> library offers a familiar interface for incorporating Teradata Enterprise Vector Store into your agentic workflows, providing the standard agentic primitives: chat models, tools, and retrievers. The result is a single system where agents can search, and act, without the extra pipelines or governance overhead that come with stitching separate tools together.</p><h3><strong>Experience multimodal agentic development with Teradata and Unstructured</strong></h3><p>In the walkthrough that follows, we will build a life sciences image question-and-answering agent powered by Teradata Enterprise Vector Store, using data ingested and embedded by Unstructured.</p><p>The final product is an agentic system where a user can pick an image from a gallery of medical image samples and ask questions about it. The agent relies on a library of medical image documentation, finds the documentation most like the selected image, and uses it to answer the user’s question.</p><figure><img alt="Image 1 — Image grid gallery" src="https://cdn-images-1.medium.com/max/736/1*UgaSv1ekYAZWxiyI7x3oFQ.png" /><figcaption>Image 1 — Image grid gallery</figcaption></figure><p>The process of building this agent consists of the following general steps:</p><ul><li>Build an ingestion and processing pipeline for composite PDF documents and images using Unstructured, which also generates the embeddings.</li><li>Store the embeddings in Teradata’s enterprise analytics platform and index them with Teradata Enterprise Vector Store.</li><li>Match a user‑selected image to its most similar documents using vector similarity search.</li><li>Wrap the workflow in a LangChain agent that allows users to query the results in natural language.</li></ul><p><strong>Setting up your development environment</strong></p><ul><li>If you haven’t already, start by creating an account at <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">teradata.com/experience.</a></li><li>Once you’re logged in, create <strong>a new environment.</strong> This ensures you’re working with the latest features and capabilities.</li></ul><figure><img alt="Image 2 — Create an environment" src="https://cdn-images-1.medium.com/max/607/1*PMyrgYrB4mpixfnFdYfbZw.png" /><figcaption>Image 2 — Create an environment</figcaption></figure><ul><li>Start the Jupyter Notebook environment by clicking “Run demos.”</li></ul><figure><img alt="Image 3 — Run demos" src="https://cdn-images-1.medium.com/max/624/1*1g_Xy-nEcg24xL4lFUT59g.png" /><figcaption>Image 3 — Run demos</figcaption></figure><p><strong>Exploring the notebook</strong></p><p><strong>Setting up needed libraries and access to Teradata Enterprise Vector Store</strong></p><ul><li>First, we install the packages needed.</li></ul><pre>%%capture<br>!pip install -r &quot;./utils/requirements.txt&quot; --quiet</pre><ul><li>After restarting the kernel, we import the libraries that will power our agentic multimodal system.</li></ul><pre># Required imports<br><br># General imports<br>from teradatagenai import VSManager<br>from langchain_teradata import TeradataVectorStore<br>from teradataml import *<br>import os<br>import json<br>import re<br>import time<br><br># Credentials and configuration management<br>from dotenv import load_dotenv<br>from getpass import getpass<br><br><br>#Langchain Imports<br>from langchain.chat_models import init_chat_model<br>from langchain.agents import create_agent<br>from langchain_core.messages import HumanMessage<br>from langchain.tools import tool<br><br># Suppress warnings<br>import warnings<br>warnings.filterwarnings(&#39;ignore&#39;)<br>display.suppress_vantage_runtime_warnings = True<br><br># Widget display<br>from IPython.display import display, HTML, IFrame<br>import ipywidgets as widgets<br><br># Import utilities<br>from unstructured_utils.teradata_ingest import ingest<br>from utils.image_grid import display_image_grid</pre><ul><li>Our trial experience provides access to a Teradata instance with Enterprise Vector Store capabilities.</li></ul><pre>ues_uri=env_vars.get(&quot;ues_uri&quot;)<br>if ues_uri.endswith(&quot;/open-analytics&quot;):<br>  ues_uri = ues_uri[:-15]  # remove last 5 chars (&quot;/open-analytics&quot;)<br><br>if set_auth_token(base_url=ues_uri,<br>         pat_token=env_vars.get(&quot;access_token&quot;), <br>         pem_file=env_vars.get(&quot;pem_file&quot;)<br>         ):<br>  print(&quot;UES Authentication successful&quot;)<br>else:<br>  print(&quot;UES Authentication failed. Check credentials.&quot;)<br>  sys.exit(1)<br><br>VSManager.health()<br>```</pre><p><strong>Ingestion and processing pipeline with Unstructured</strong></p><ul><li>Unstructured offers a graphical user interface (GUI) to create workflows. The same building blocks used in the GUI are available from <a href="https://docs.unstructured.io/api-reference/overview">Unstructured API</a>. Namely: sources, destinations, workflows, and jobs.</li><li>Workflows orchestrate data processing between a source and a destination with steps such as partitioning, chunking, enriching, embedding, etc.</li><li>Jobs are specific execution of a workflow.</li><li>We have built a function that abstracts our defined workflow. The full implementation is contained in the /unstructured_utils folder and imported into the notebook.</li></ul><p>The workflow has three sequential nodes:</p><ol><li><strong>Partitioning:</strong> Uses Claude as a vision-language model (VLM) to extract structured content from raw documents. Using a VLM rather than a traditional text extractor means it can handle complex layouts, images, and tables by “looking” at the document rather than just parsing text.</li><li><strong>Chunking:</strong> Splits the partitioned content into pieces using a title-aware strategy. It targets 1,500 characters per chunk, caps at 2,048, and adds a 100-character overlap between consecutive chunks so context isn’t lost at boundaries. It also retains the original elements alongside the chunks, which is useful for traceability.</li><li><strong>Embedding:</strong> Converts each chunk into a vector using Voyage AI’s multimodal model. The multimodal part matters here because the upstream VLM partitioner may have preserved image-based content, and a multimodal embedder can represent that meaningfully rather than ignoring it.</li></ol><ul><li>In the notebook, we call this function to execute the ingestion of both composite PDFs and images.</li></ul><pre>ingest(api_key=unstructured_api_key, <br>    td_host=env_vars.get(&quot;host&quot;), <br>    td_user=dbuser, <br>    td_password=dbpwd, <br>    td_database=default_db, <br>    td_table=&#39;composite_pdfdocs_embedded&#39;, <br>    s3_uri=&quot;s3://dev-rel-demos/teradata-unstructured/healthcare-assets/composite-pdfs/&quot;, <br>    s3_anonymous=True)<br><br>ingest(api_key=unstructured_api_key, <br>    td_host=env_vars.get(&quot;host&quot;), <br>    td_user=dbuser, <br>    td_password=dbpwd, <br>    td_database=default_db, <br>    td_table=&#39;image_samples_embedded&#39;, <br>    s3_uri=&quot;s3://dev-rel-demos/teradata-unstructured/healthcare-assets/images/&quot;, <br>    s3_anonymous=True)</pre><p><strong>Teradata Enterprise Vector Store: The engine for processing unstructured data</strong></p><ul><li>The first step to create our Vector Store is to ingest the data loaded by Unstructured into teradataml DataFrames. We can inspect a sample of the records.</li><li>The preview displays the <strong>record_id</strong>, <strong>filename</strong>, and <strong>record_locator</strong> columns. The embedding vectors are omitted here for readability, but they are present in the underlying table and will be used for similarity search in later steps.</li></ul><figure><img alt="Image 1 — Ingested data in Teradata DataFrame" src="https://cdn-images-1.medium.com/max/1024/1*Uglp77liNx_3XVnPumogzg.png" /><figcaption>Image 4 — Ingested data in Teradata DataFrame</figcaption></figure><ul><li>Once the documents are ingested and embedded, we create a Vector Store based on the embeddings through the `TeradataVectorStore` class `from_embeddings` method. This is a class inherited from `langchain-teradata`.</li></ul><pre>vs_emb = TeradataVectorStore.from_embeddings(name = f&quot;{default_db}_mm_embeddings_vs&quot;,<br>                   data = image_documentation_bank,<br>                   data_columns = &quot;embeddings&quot;,<br>                   key_columns = [&quot;id&quot;, &quot;record_id&quot;],<br>                   embedding_data_columns = &quot;text&quot;,<br>                   metadata_columns = [&quot;text&quot;,&quot;date_created&quot;, &quot;date_modified&quot;, &quot;record_locator&quot;, &quot;filename&quot;],)</pre><ul><li>The Vector Store registers and creates a new vector index in VantageCloud over the embedding column, enabling sub-second approximate nearest-neighbor lookups at scale. Key parameters include the data table, the column containing the embedding vectors `data_columns`, the primary key columns for deduplication, and the metadata columns to carry through into search results.</li><li>The `TeradataVectorStore` class includes methods for performing searches, semantical or lexical, asking questions about the results of searches (powered by an included LLM client), and to update, re-index or destroy the Vector Store.</li><li>We can use the Vector Store to perform similarity searches between any image on the image bank and the documentation library.</li></ul><p>In this case, we are retrieving the top record from the DataFrame containing the image data, and we perform a similarity search by vector retrieving the most similar document.</p><pre>response = vs_emb.similarity_search_by_vector(data = raw_images_df.head(1), column=&#39;embeddings&#39;, top_k=1)</pre><figure><img alt="Image 5 — Similarity search results" src="https://cdn-images-1.medium.com/max/872/1*GFwesq9rPgw3E8YE6Fmecg.png" /><figcaption>Image 5 — Similarity search results</figcaption></figure><p>The match is driven by the image contained in the composite document. The retrieval returns the associated text along with metadata included when creating the Vector Store index, such as the `record_locator` and `filename`. In an agentic system, this metadata allows an agent to fetch the original document if needed, while the text provides information to answer questions about the selected image.</p><p>Teradata Enterprise Vector Store includes a built-in LLM integration that enables natural language queries over vector search results.</p><pre>question=&#39;I need to recover The title, description and record id, and locator of the most similar record?&#39;<br>prompt=&#39;Format the response in a conversational way.&#39;<br>response = vs_emb.prepare_response(question=question, similarity_results=response, prompt=prompt)</pre><p>The response object contains the requested data in the requested format as processed by the chat model included in the Teradata Enterprise Vector Store.</p><blockquote>Based on the provided data, here is the information you requested:<br> <br> Title: Image 4: Immune Cell Infiltration in Solid Tumor Tissue<br> Description: Microscopic image showing immune cell infiltration in solid tumor tissue. The image displays a colorful fluorescent staining pattern with bright yellow circular structures scattered throughout a predominantly pink and magenta background with blue areas. Dark black voids are visible throughout the tissue section. The staining reveals cellular structures and immune markers in the tumor microenvironment.<br> <br> This image captures a solid tumor section with visible immune cell infiltration. The bright yellow circular structures likely represent lipid droplets or specific immune cell subtypes such as macrophages or dendritic cells. The dark voids may be necrotic foci or gland-like structures within the tumor. The predominant pink and magenta signals indicate widespread expression of tumor or immune markers throughout the tissue. This kind of staining is commonly used in studies examining the tumor immune microenvironment, particularly in the context of immunotherapy response.<br> <br> Record ID: c724ea12-fa51–554d-b552-cb2d5de3b0a7<br> Record Locator: {“protocol”: “s3”, “remote_file_path”: “s3://dev-rel-demos/teradata-unstructured/healthcare-assets/composite-pdfs/”}</blockquote><p>The capabilities we have explored unlock a lot of possibilities for an agentic system, such as the one we are building.</p><h3><strong>Building a LangChain agent powered by Teradata Enterprise Vector Store</strong></h3><p>Our agent is wrapped in a widget UI that allows users to select an image from the gallery, and a chat widget that enables conversations. These widgets are merely UI abstractions; the components of the agent itself are the same as any LangChain agent. We need a `chat_llm` component, and a set of tools.</p><p>Our trial experience includes a proxy LLM client to power the `chat_llm`, while Teradata Enterprise Vector Store includes all the functionality needed to power the core tool.</p><pre>llm_key = env_vars.get(&quot;litellm_key&quot;)<br>llm_url = env_vars.get(&quot;litellm_base_url&quot;)<br><br>llm = init_chat_model(<br>  model=&quot;openai-gpt-41&quot;,<br>  model_provider=&quot;openai&quot;,<br>  base_url=llm_url,<br>  api_key=llm_key,<br>)</pre><p>The main tool is a thin abstraction over the `similarity_search` and `prepare_response` methods of Teradata Enterprise Vector Store. The image grid widget is generated from the DataFrame containing the image data, and selection buttons that update the `selected_id` parameter to look up the corresponding embedding. From there, the process follows the same steps described above.</p><pre>@tool<br>def search_similar_image_documentation(dummy: str = &quot;&quot;) -&gt; str:<br>  &quot;&quot;&quot;<br>  Runs a similarity search using the currently selected image and displays<br>  the most similar result. Call this whenever the user asks about similar images.<br>  &quot;&quot;&quot;<br>  if result.selected_id is None:<br>    return &quot;No image is selected. Please select an image from the grid first.&quot;<br><br>  response_similarity = vs_emb.similarity_search_by_vector(<br>    data=raw_images_df[raw_images_df[&quot;record_id&quot;] == result.selected_id],<br>    column=&quot;embeddings&quot;,<br>    top_k=1,<br>    return_type=&quot;json&quot;,<br>  )<br>  question = &quot;I need to recover the title, description, filename, record id, and file_locator of the most similar record?&quot;<br>  prompt = &quot;Format the response as JSON object.&quot;<br>  response_chat = vs_emb.prepare_response(<br>    question=question,<br>    similarity_results=response_similarity,<br>    prompt=prompt,<br>  )<br><br>  with chat_output:<br>    display(HTML(<br>      f&quot;&lt;div style=&#39;font-family:sans-serif; padding:12px; background:#f9f9f9;&quot;<br>      f&quot;border-radius:6px; margin:8px 0;&#39;&gt;{response_chat}&lt;/div&gt;&quot;<br>    ))<br><br>  return response_chat</pre><p>The agent itself includes specific instructions to interact with the tools and respond to the users’ intent.</p><blockquote>You are an image analysis assistant.</blockquote><blockquote>Tools available:</blockquote><blockquote>- search_similar_image_documentation: Find images similar to a given input and retrieve associated documentation</blockquote><blockquote>- display_pdf_from_locator: Render a PDF document</blockquote><blockquote>Workflow:</blockquote><blockquote>1. When a user provides an image or asks for similar images, call search_similar_image_documentation</blockquote><blockquote>2. If the user asks to view/open a PDF from the results, extract `filename` and `remote_file_path` from the search response and pass them to display_pdf_from_locator</blockquote><blockquote>3. If the user asks to describe features from the image in questions like: — What are the XYZ features of this image -</blockquote><blockquote>a. First call search_similar_image_documentation</blockquote><blockquote>b. If the documentation contains relevant information, answer based on it</blockquote><blockquote>c. If not, inform the user that the image library documentation does not contain information to answer their question, then display the PDF using display_pdf_from_locator so they can review it directly</blockquote><blockquote>d. Never ask to upload the image as the user is referring to the selected image.</blockquote><blockquote>If a search returns no results or the PDF fields are missing, let the user know.</blockquote><p>The widget UI and the agent enable the fully agentic experience.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/736/1*4oU2PMznbUXAn2u1xYwlGg.png" /><figcaption>Image 6— Agentic medical image question-and-answer experience</figcaption></figure><p><a href="https://www.teradata.com/platform/clearscape-analytics/enterprise-vector-store">Learn more about Teradata multimodal agentic capabilities.</a></p><p>Build the demo yourself at <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">teradata.com/experience</a>.</p><p><strong>Conclusion</strong></p><p>Bringing together Teradata Enterprise Vector Store and Unstructured creates a unified foundation for multimodal agentic systems, one where structured and unstructured data coexist naturally, governance stays simple, and developers can build advanced retrieval-augmented workflows without extra pipelines or infrastructure overhead. High-fidelity ingestion, scalable vector indexing, and agentic tooling combine into a stack that reduces the friction between experimentation and production. The walkthrough here is one example of the pattern, but the same approach applies to customer experience, automation, and context-aware intelligence across industries. As multi-modal data and agentic AI continue to expand, this integrated stack gives organizations a practical path to accurate, grounded, and enterprise-ready agentic systems built for scale.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=c3686b5c6198" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/end-to-end-multimodal-rag-and-agents-with-teradata-and-unstructured-c3686b5c6198">End-to-End Multimodal RAG and Agents With Teradata and Unstructured</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Introducing Teradata Enterprise MCP: Build Your First Data Analyst Agent]]></title>
            <link>https://medium.com/teradata/introducing-teradata-enterprise-mcp-build-your-first-data-analyst-agent-c546dbd2d424?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/c546dbd2d424</guid>
            <category><![CDATA[data-analysis]]></category>
            <category><![CDATA[mcp-server]]></category>
            <category><![CDATA[teradata]]></category>
            <category><![CDATA[generative-ai-tools]]></category>
            <category><![CDATA[agentic-ai]]></category>
            <dc:creator><![CDATA[Daniel Herrera]]></dc:creator>
            <pubDate>Fri, 30 Jan 2026 12:02:23 GMT</pubDate>
            <atom:updated>2026-02-02T10:16:57.661Z</atom:updated>
            <content:encoded><![CDATA[<p>Teradata Enterprise MCP is designed to empower organizations to build advanced data analyst agents through secure, scalable, and agentic AI-powered analytics.</p><figure><img alt="Teradata Enterprise MCP" src="https://cdn-images-1.medium.com/max/975/1*vQUNvGB84MEO_xJNxgPCYw.png" /></figure><p>An <a href="https://www.teradata.com/insights/ai-and-machine-learning/autonomous-ai-knowledge-platform">autonomous enterprise</a> is one that uses its knowledge, applies agentic reasoning, and delivers faster, higher‑quality outcomes for customers. It operates through a modern stack that unifies a knowledge layer, an intelligent agent layer, and an outcomes layer, enabling systems to reason and act with context rather than rely solely on manual workflows.</p><p>The agentic layer combines Large Language Models (LLMs) reasoning capabilities with trusted data, policies, context, and analytics, enabling agents to perform actions such as querying structured and unstructured data, executing advanced analytics, running predictions, retrieving documents, and grounding responses in enterprise reality. This connection transforms AI from experimentation into operational impact.</p><p>Teradata Enterprise MCP acts as a core pillar of this autonomy by providing secure access to data, tool standardization, context management, auditability, and high scalability for agent actions. It ensures that agents can interact with enterprise systems in a trusted, governed, and consistent way while using analytics and knowledge capabilities at production scale.</p><p><strong>The Model Context Protocol –MCP– in brief:</strong></p><p>MCP is an <a href="https://www.anthropic.com/news/model-context-protocol">open-source standard designed to make AI systems talk to the outside world in a predictable, structured way</a>. Instead of every developer inventing a new format for tools, capabilities, or metadata, MCP provides a common contract defining all these aspects, standardizing LLMs interactions with external tools. For a developer, this means spending more time designing the behavior of AI powered agents, not worrying about how to package or describe needed tools or necessary parameters. The protocol takes care of that for you.</p><p>Teradata research teams and a group of community contributors built the community driven <a href="https://github.com/Teradata/teradata-mcp-server">Teradata MCP</a>, this open-source effort has taken off in ways none of us fully expected. Suddenly, people were building agents that could explore data, run analytics, and automate workflows directly inside the Teradata ecosystem. The best part is that the project keeps getting better as contributors continue to refine the server, add new tools, tighten the interface, and push the boundaries of what MCP powered agents can do.</p><figure><img alt="Teradata MCP server architecture" src="https://cdn-images-1.medium.com/max/734/1*lWtDvq-4GS5Pf3AdIgdAUg.png" /><figcaption>Teradata MCP server architecture</figcaption></figure><p><strong>The challenges of deploying and implementing an MCP server</strong></p><p>MCP isn’t without its challenges. While implementing an MCP server for an enterprise knowledge platform, the same questions always come up. Can it scale across teams? Can it connect to sensitive data with the right governance and privacy controls? What happens when teams start chaining servers together and bump into context‑overflow issues?</p><p>In consideration of these pain-points, Teradata has developed Teradata Enterprise MCP. Teradata Enterprise MCP tackles the core operational and governance challenges head‑on. Enabling data professionals to focus on building agents that deliver value, easing the concerns around deployment, scaling, and access controls common to MCP as a protocol.</p><p><strong>How Teradata Enterprise MCP contributes to your Agentic AI projects</strong></p><p>Teradata Enterprise MCP ships with a set of features that tackle common implementation challenges.</p><figure><img alt="Features of Teradata Enterprise MCP" src="https://cdn-images-1.medium.com/max/975/1*vQUNvGB84MEO_xJNxgPCYw.png" /><figcaption>Features of Teradata Enterprise MCP</figcaption></figure><p>Security:</p><ul><li>Every user connects to Teradata Enterprise MCP using their own credentials, whether that’s basic TD2 authentication or JWT and Bearer tokens validated through the organization’s identity provider. This ensures the server only exposes the data that specific users are authorized to see, instead of routing all users through a single service account, unless the latter is the chosen implementation decision of the organization.</li><li>RBAC controls tool visibility. MCP roles, with their corresponding assigned toolset, can be mapped to Teradata database roles so you can define exactly which tools a user can access in the system according to the user’s role.</li><li>SQL injection protection, security scans, and end-to-end logging provide an enterprise grade security posture without adding friction for developers.</li></ul><p>Performance:</p><ul><li>The server is built for multiuser, multi-session workloads with automatic connection pooling and sub second response times. Teams across the company can reliably share one MCP server without creating bottlenecks.</li></ul><p>Tool Context Management:</p><ul><li>A deep catalog of more than a hundred production ready tools; from database queries and DBA functions to ML, vector search, and RAG might be difficult to manage for LLMs, however, because tool exposure is governed per role, each agent stays within a clean and compliant sandbox.</li></ul><p><strong>Experience Teradata Enterprise MCP</strong></p><p>You can experience Teradata Enterprise MCP directly inside <a href="https://www.teradata.com/platform/clearscape-analytics">ClearScape Analytics Experience</a>. We’ve put together a <a href="https://github.com/JH255095/teradata_enterprise_mcp/blob/main/Langchain_Teradata_MCP.ipynb">sample notebook</a> that shows a simple LangChain integration. This is only a starting point that you can extend according to your needs.</p><p>You can easily experiment on your own by following the steps below:</p><p><strong>Setting up your development environment</strong></p><ul><li>If you haven’t used ClearScape Analytics Experience before, start by creating an account at <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">teradata.com/experience.</a></li><li>Once you’re logged in, create <strong>a new environment.</strong> This ensures you’re working with the latest features and capabilities. Take note of your chosen password as you will need it to create the MCP server configuration later.</li></ul><figure><img alt="Create an environment on ClearScape Analytics Experience" src="https://cdn-images-1.medium.com/max/948/1*MdF9wcmXWgHqiXYCb7IGAA.png" /><figcaption>Create an environment on ClearScape Analytics Experience</figcaption></figure><ul><li>Start the Jupyter Notebook environment by clicking “Run demos”.</li></ul><figure><img alt="Run demos on ClearScape Analytics Experience" src="https://cdn-images-1.medium.com/max/975/1*SRlqJmCqOQpQYOxGaZ82Hg.png" /><figcaption>Run demos on ClearScape Analytics Experience</figcaption></figure><ul><li>In the Jupyter Notebook environment, open the “Getting Started” folder look for the folder named Enterprise_MCP_Data_Analyst_Agent, open the notebook inside that folder.</li></ul><p><strong>Exploring the notebook</strong></p><ul><li>First, we install the packages needed for the MCP integration with LangChain.</li></ul><pre>pip install -U langchain langchain-mcp-adapters langchain-openai httpx</pre><ul><li>After restarting the kernel, we import the libraries that will power our agent: asyncio for async operations, the MCP client adapter, and LangChain’s chat model and agent components.</li></ul><pre>import asyncio, sys, os, httpx<br>import time<br>from dotenv import load_dotenv<br>from langchain_mcp_adapters.client import MultiServerMCPClient<br>from langchain.chat_models import init_chat_model<br>from langchain.agents import create_agent<br>from langchain_core.messages import HumanMessage<br>from langchain_core.messages import AIMessage</pre><ul><li>The notebook pulls in your environment settings from a .env file. This keeps your credentials secure and your code clean.</li></ul><pre>environment_path = &quot;/home/jovyan/JupyterLabRoot/VantageCloud_Lake/.config/.env&quot;<br>load_dotenv(dotenv_path=environment_path) <br>llm_key = os.getenv(&#39;litellm_key&#39;)<br>llm_url = os.getenv(&#39;litellm_base_url&#39;)<br>username = os.getenv(&#39;username&#39;)<br>password = os.getenv(&#39;my_variable&#39;)</pre><ul><li>Here we establish the connection to the Enterprise MCP Server using LangChain’s MultiServerMCPClient. The MCP server uses the credentials retrieved from the environment variables to connect on your behalf and returns the available tools.</li></ul><pre>async def get_mcp_tools():<br>    client = MultiServerMCPClient(<br>        {<br>            &quot;teradata&quot;: {<br>                &quot;transport&quot;: &quot;http&quot;,<br>                &quot;url&quot;: &quot;http://host.docker.internal:8001/mcp&quot;,<br>                &quot;auth&quot;: httpx.BasicAuth(username, password),<br>            }<br>        }<br>    )<br>    tools = await client.get_tools()<br>    return client, tools<br><br>client, tools = await get_mcp_tools()<br>print([t.name for t in tools])</pre><ul><li>We initialize a chat model through the LLM proxy available in ClearScape Analytics Experience, then create a LangChain agent by passing in the Teradata MCP server tools. The system prompt tells the agent it’s a data analyst working with Teradata.</li></ul><pre>llm = init_chat_model( <br>    model=&quot;openai-gpt-41&quot;, <br>    model_provider=&quot;openai&quot;, <br>    base_url=&quot;https://llmlite.ci.clearscape.teradata.com&quot;, <br>    api_key=llm_key, ) <br><br>agent = create_agent(<br>        model=llm, <br>        tools=tools, <br>        system_prompt=(<br>        &#39;&#39;&#39; You are a data analyst that responds to user questions regarding data in a teradata system &#39;&#39;&#39; <br>        ), )</pre><ul><li>With everything set up, you can send natural language queries to the agent. It interprets your question, picks the right MCP tools, and executes them against your Teradata environment.</li></ul><pre>user_query = input(&#39;\n What would you like to know about your data? &#39;)<br>result = await agent.ainvoke(<br>    {&quot;messages&quot;: [HumanMessage(content=user_query)]}<br>)</pre><figure><img alt="Query submitted to the agent" src="https://cdn-images-1.medium.com/max/878/1*oPkm_4QXOs3kL2ImAPKZTw.png" /><figcaption>Query submitted to the agent</figcaption></figure><ul><li>The answer of the agent to this specific question is as follows:</li></ul><pre>print(result[&quot;messages&quot;][-1].content)</pre><figure><img alt="" src="https://cdn-images-1.medium.com/max/975/1*b7YgVCl-bsObqlVOvEmdOg.png" /><figcaption>Agent response to the user query</figcaption></figure><p>ClearScape Analytics Experience is a free of cost experience with access to limited resources, for a full demo on a production grade Teradata environment feel free to reach us out at <a href="https://www.teradata.com/platform/agent-stack">Teradata Enterprise AgentStack</a>.</p><p><strong>Conclusion:</strong></p><p>The shift toward autonomous enterprises is no longer theoretical. With agentic AI, unified knowledge layers, and governed execution environments, organizations now have the foundations they need to build systems that reason, act, and deliver value with less manual intervention. The Teradata Enterprise MCP Server plays a significant role in making this real. It removes the operational and security hurdles that traditionally slow down MCP adoption and replaces them with a platform that feels stable, scalable, and ready for enterprise use. By giving teams, a secure, governed, and high‑performance way to connect LLMs and agents to trusted data, the Enterprise MCP Server frees developers and data practitioners to do what they do best: build. Whether you are exploring early prototypes, scaling agentic workloads, or preparing for production‑grade deployments, the Enterprise MCP Server provides the backbone that lets you move faster, stay compliant, and unlock the full potential of agentic AI inside your organization.</p><p><strong>About the Author:</strong></p><p>Daniel Herrera is Principal Developer Advocate at Teradata, specializing in AI/ML and advanced analytics. He helps developers and data practitioners unlock enterprise-scale innovation through ClearScape Analytics and VantageCloud. With deep expertise in data engineering, cloud architecture, and model deployment, Daniel bridges technical complexity with practical solutions. A frequent speaker and community contributor, he focuses on enabling scalable AI use cases across industries.</p><p>You can reach out to Daniel in <a href="https://www.linkedin.com/in/daniel-herrera/">LinkedIn</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=c546dbd2d424" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/introducing-teradata-enterprise-mcp-build-your-first-data-analyst-agent-c546dbd2d424">Introducing Teradata Enterprise MCP: Build Your First Data Analyst Agent</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[A Gateway to Enhanced Agents & AI Workloads: Arrow Flight SQL with QueryGrid®]]></title>
            <link>https://medium.com/teradata/a-gateway-to-enhanced-agents-ai-workloads-arrow-flight-sql-with-querygrid-e6b4c6721c2a?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/e6b4c6721c2a</guid>
            <category><![CDATA[data-fabric]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[ai-agent]]></category>
            <category><![CDATA[apache-arrow]]></category>
            <dc:creator><![CDATA[Sebastian]]></dc:creator>
            <pubDate>Thu, 29 Jan 2026 14:36:49 GMT</pubDate>
            <atom:updated>2026-01-29T14:36:48.428Z</atom:updated>
            <content:encoded><![CDATA[<h4>Unveiling the power of the Arrow Flight SQL connector for QueryGrid®</h4><p>The introduction of the Arrow Flight SQL connector for QueryGrid® marks a significant advancement for organizations seeking to accelerate data-driven &amp; AI-driven innovation. By enabling streamlined, high-performance access to diverse data sources without the overhead of traditional ETL processes, this capability empowers teams to unlock faster insights, enhance collaboration, and optimize their AI workflows. As a result, customers can maximize the value of their existing data infrastructure while future-proofing their operations for emerging AI and agentic workloads.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*GXpNQqtsXD4crfriTOFtyg.png" /><figcaption>QueryGrid(r) connector stack</figcaption></figure><h3>Introduction to Apache Arrow</h3><p><a href="https://arrow.apache.org/">Apache Arrow</a> is a cross-language development platform designed for in-memory data processing. It provides a standardized columnar memory format for flat and hierarchical data, enabling efficient analytics operations. Arrow’s key strength lies in its ability to handle large datasets with minimal overhead, facilitating rapid data access and reduced memory usage. Over the last few years, Arrow has gained traction and interest in the Data Science community, with commits to its codebase added every day.</p><h3>Arrow Flight and Arrow Flight SQL</h3><p><a href="https://arrow.apache.org/docs/format/Flight.html">Arrow Flight</a> is an extension of Apache Arrow that introduces high-performance messaging for data transfer. It leverages gRPC to enable efficient, secure, and highly scalable data transport between systems. Building upon Arrow Flight, <a href="https://arrow.apache.org/docs/format/FlightSql.html">Arrow Flight SQL</a> provides a protocol for executing SQL queries over Arrow Flight, allowing seamless and efficient interaction with SQL-based systems. The performance achieved is over 15 times higher than what a single-threaded protocol like JDBC can deliver (based on lab tests).</p><h3>Meet QueryGrid® Arrow Flight SQL connector</h3><p>QueryGrid® <a href="https://medium.com/teradata/understanding-querygrid-a-gateway-to-optimized-distributed-data-access-7d381db1e2ac">is a sophisticated query fabric</a> that seamlessly connects diverse data sources and processing engines. It enables businesses to execute queries across multiple platforms, integrating disparate data environments into a <a href="https://www.teradata.com/platform/data-fabric">unified analytical ecosystem</a>. The integration of QueryGrid® with Arrow Flight SQL extends this capability to a broader range of engines and systems, and provides a net new capability: now client tools can also connect through the QueryGrid® Arrow Flight SQL connector. All is done leveraging Arrow’s modern and efficient data format and transport protocols, maintaining QueryGrid’s legacy: providing high-throughput parallel data exchange, and minimizing data movement when possible.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Xv8xGSasO3uMLAkG0CZghg.png" /><figcaption>QueryGrid(r) Arrow Flight SQL Connector</figcaption></figure><h4>Client Server Architecture</h4><p>QueryGrid® Arrow Flight SQL connector comes in two flavors: Client &amp; Server. When the client capability is configured, the <a href="https://www.teradata.com/">Teradata Vantage</a> database can read data in Arrow format from other engines on the fly, without the need to copy the data upfront or perform any ETL processes (capability is available today). When the Server is configured, the Teradata Vantage database is exposed as an Arrow Flight SQL endpoint, allowing not only engines, but also tools, to read and write data in parallel, leveraging this highly efficient columnar format and transfer protocol. The server capability will be released soon, but don’t let this block you from planning how to leverage it in your future architecture; you can start the conversation with us today and <a href="https://www.teradata.com/insights/videos/data-federation-for-ai">see this in action</a>.</p><h4>Expanding the Ecosystem</h4><p>Integrating QueryGrid® with Apache Arrow allows enterprises to tap into a wide array of data processing engines supported by Arrow Flight SQL, such as DuckDB, Dremio, and Apache Doris, just to name a few. It also allows any ADBC-compliant driver to read data from Vantage <em>(coming soon with server capability).</em></p><h4>Enhancements in Data Processing</h4><p>QueryGrid® and Apache Arrow facilitate smooth, high-speed parallel data transfer and query execution across multiple data sources. Combined with the intelligent push-down processing that QueryGrid® provides, this enables efficient data exchange between applications, something that is key in a distributed environment.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/997/1*Q-PhpcDU8TMBYylEOCIvmw.png" /><figcaption>QueryGrid MCP Server CE enables Agentic workflows</figcaption></figure><h3>Built for AI Workloads &amp; Agents</h3><p>The integration of QueryGrid® with Apache Arrow is particularly advantageous for AI and Agentic workloads. These workloads often require the processing of massive datasets, real-time data streaming, and the execution of complex queries. Apache Arrow’s columnar format and the high-performance transmission via Arrow Flight and Arrow Flight SQL provide the robustness and speed necessary for these demanding tasks, while also simplifying the interface between systems.</p><h4>Accelerating Training and Inference</h4><p>With the capability to efficiently handle large volumes of data, the integration aids in speeding up both the training and inference phases of machine learning models. Data scientists can seamlessly access and manipulate data from various sources, reducing the time spent on data preparation and transformation.</p><h4>Enhanced Interoperability and Flexibility</h4><p>This integration ensures that models can leverage the best-suited data processing engines, enhancing flexibility and scalability. We know that Agents are greedy and require massive amounts of data, while also the need for consolidating multiple systems. QueryGrid® Arrow Flight SQL connector exposes a simple interface for agents to efficiently grab data from multiple engines and perform complex tasks. Imagine the power you can get if you also connect this with <a href="https://medium.com/teradata/chat-with-your-data-in-visual-studio-code-through-an-mcp-server-071673bd3ffe">Teradata’s</a> and <a href="https://medium.com/teradata/why-youll-want-querygrid-mcp-server-community-edition-in-your-toolkit-cb0872bf9c8a">QueryGrid</a> MCP Servers, or the <a href="https://www.teradata.com/press-releases/2026/teradata-unveils-enterprise-agentstack">Teradata AgentStack</a>.</p><h3>Conclusion</h3><p>The QueryGrid® Arrow Flight SQL connector delivers high performance and seamless integration for AI and Agentic workloads. By utilizing Apache Arrow’s efficient columnar data format and high-speed transport protocols, QueryGrid® enables robust connectivity across numerous engines and systems, empowering organizations to execute advanced analytics and AI/ML tasks with ease. This integration not only streamlines data workflows but also provides agents with an efficient and unified interface to access large volumes of data from diverse sources, unlocking new possibilities in data science and machine learning.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=e6b4c6721c2a" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/a-gateway-to-enhanced-agents-ai-workloads-arrow-flight-sql-with-querygrid-e6b4c6721c2a">A Gateway to Enhanced Agents &amp; AI Workloads: Arrow Flight SQL with QueryGrid®</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Why You’ll Want QueryGrid® MCP Server Community Edition in Your Toolkit]]></title>
            <link>https://medium.com/teradata/why-youll-want-querygrid-mcp-server-community-edition-in-your-toolkit-cb0872bf9c8a?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/cb0872bf9c8a</guid>
            <category><![CDATA[data-fabric]]></category>
            <category><![CDATA[cloud]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[mcps]]></category>
            <category><![CDATA[agents]]></category>
            <dc:creator><![CDATA[Sebastian]]></dc:creator>
            <pubDate>Tue, 06 Jan 2026 12:41:04 GMT</pubDate>
            <atom:updated>2026-01-06T12:40:32.667Z</atom:updated>
            <content:encoded><![CDATA[<h4>Streamline, automate, and secure your QueryGrid management with this new release</h4><p>Let’s face it, managing large data resources can become complicated quickly. If you’re working with <a href="https://www.teradata.com/platform/data-fabric">QueryGrid</a>®, you already know that staying on top of your environment means juggling a lot of moving parts. Enter the <a href="https://github.com/Teradata/teradata-qg-mcp-server">QueryGrid® MCP Server Community Edition</a> (CE), a freshly launched toolkit that’s about to make your life a whole lot easier.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/997/1*O4X21e5sQbPu0XpFvenwQQ.png" /><figcaption><em>QueryGrid(r) MCP Server components</em></figcaption></figure><h3>What Is the QueryGrid MCP Server CE?</h3><p>Think of the <a href="https://www.teradata.com/insights/ai-and-machine-learning/model-context-protocol-mcp">MCP Server CE</a> as your new command center for all things QueryGrid® management. It’s designed to help <a href="https://www.teradata.com/press-releases/2025/mcp-server-agentic-ai-at-scale">AI assistants and other MCP clients</a> talk directly to the QueryGrid Manager (QGM) using a standardized protocol. The focus here isn’t on querying data directly: it’s about making administrative and operational tasks faster, smarter, and more secure. And if you’re already running QueryGrid 3.x, you’re good to go; this release is fully compatible with your environment.</p><h3>Key Features</h3><ul><li><strong>Massive Toolkit</strong>: Over 120 built-in tools support just about every QueryGrid operation you can think of: monitoring resources, adding and removing sources, tweaking configs, creating and dropping links, and tons more.</li><li><strong>Controlled Security</strong>: Basic authentication for QueryGrid Manager means better access control. Plus, logging is totally configurable, so you get the transparency you need.</li><li><strong>RESTful API Integration</strong>: Full support for Create, Read, Update, and Delete (CRUD) operations. All you need for the automation you’ve been looking for.</li></ul><h3>Why Switch to MCP Server CE?</h3><ul><li><strong>Boosted Efficiency</strong>: The toolkit is comprehensive and fully tested, backed by a robust API layer provided by QueryGrid® Manager service. That means even complex management tasks are handled quickly and consistently, so you can accelerate your development and deployment cycles.</li><li><strong>Automation is King</strong>: With the ability to programmatically manage QueryGrid® resources, you can automate all your admin workflows — whether through chatbots, agents, or other tools — and get back to the fun stuff.</li><li><strong>Built-in Security</strong>: Only authorized users or systems can make changes, thanks to secure authentication mechanisms.</li><li><strong>Ready for Scale</strong>: Built for production from the ground up, with configurable logging for transparency and traceability. Start with small projects and climb to mission-critical deployments.</li></ul><h3>Takeways</h3><p>Whether you’re a developer, data engineer, DBA, or someone who wants smoother QueryGrid® management, the QueryGrid® MCP Server Community Edition is a game-changer. It’s secure, flexible, and packed with features that will make your admin processes more efficient and reliable. If you’re ready to embrace modern automation and development practices, this toolkit deserves a spot in your environment. And remember, you can expand the power of this freshly released server with <a href="https://github.com/Teradata/teradata-mcp-server/">Teradata’s MCP Server Community Edition</a>.</p><p>Start your journey now by downloading your copy from <a href="https://github.com/Teradata/teradata-qg-mcp-server">Teradata GitHub</a>.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=cb0872bf9c8a" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/why-youll-want-querygrid-mcp-server-community-edition-in-your-toolkit-cb0872bf9c8a">Why You’ll Want QueryGrid® MCP Server Community Edition in Your Toolkit</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Building Autonomous Customer Intelligence: A Developer’s Guide to Teradata’s Customer Intelligence…]]></title>
            <link>https://medium.com/teradata/building-autonomous-customer-intelligence-a-developers-guide-to-teradata-s-customer-intelligence-a28d8dd77c41?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/a28d8dd77c41</guid>
            <category><![CDATA[mcp-server]]></category>
            <category><![CDATA[ai-agent]]></category>
            <category><![CDATA[llm]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[customer-experience]]></category>
            <dc:creator><![CDATA[Vidhan Bhonsle]]></dc:creator>
            <pubDate>Tue, 06 Jan 2026 12:32:58 GMT</pubDate>
            <atom:updated>2026-01-06T12:32:55.861Z</atom:updated>
            <content:encoded><![CDATA[<h3>Building Autonomous Customer Intelligence: A Developer’s Guide to Teradata’s Customer Intelligence Framework</h3><h3>Introduction</h3><p>The way customer data systems are built is undergoing a remarkable transformation. Across the industry, significant milestones have been reached — unified Customer 360 views, sophisticated machine learning (ML) models, and real-time streaming pipelines. These foundations set up the next leap: systems that don’t just analyze customer behavior, but actively shape experiences in real time.</p><p>The <a href="https://www.teradata.com/solutions/customer-experience"><strong>Teradata Customer Intelligence Framework</strong></a> represents this next generation of customer intelligence architecture. Built on established data engineering and machine learning practices, it enables autonomous, real-time decision-making. The framework equips data scientists, data engineers, ML engineers, and platform architects to create systems that sense needs, understand context, and take immediate action; powered by AI for CX that delivers measurable business impact.</p><h3>The challenge</h3><p>Modern customer data architectures hit the same three walls repeatedly. Data lives in disconnected systems, making a reliable Customer 360 fragile at best. Batch-heavy pipelines introduce latency, so insights arrive after the moment has passed. And unresolved customer tasks often escalate to costly channels, driving up operations while eroding satisfaction.</p><figure><img alt="Pain points of customer experiences" src="https://cdn-images-1.medium.com/max/1024/1*UpWTFRDjVpFES2mS7jtLeQ.png" /><figcaption>Pain points of customer experiences</figcaption></figure><p>The Customer Intelligence Framework addresses these limits with an event-driven approach: unify raw streams into reusable data products, detect signals that carry business meaning in real time, and trigger guard-railed agent decisions that act immediately. In short, move from fragmented data and delayed reports to a closed loop that senses, decides, and acts.</p><h3>Architecture: from data to intelligence</h3><p>The Customer Intelligence Framework transforms raw data into autonomous intelligence through three core components working across five integrated layers:</p><h4><strong>Foundational Elements</strong></h4><figure><img alt="Core components of Customer Intelligence Framework" src="https://cdn-images-1.medium.com/max/1018/1*WEE7_RXJCmdzs_5J1WJsNg.png" /><figcaption>Core components of Customer Intelligence Framework</figcaption></figure><ul><li><strong>Data products</strong> — Reusable, structured assets that organize complex customer data. These multi-dimensional building blocks can be assembled differently for various use cases, with full version control and API accessibility via REST, Python SDKs, or SQL.</li><li><strong>Signals</strong> — Real-time patterns detected from customer behavior, not just transactions. The framework continuously identifies meaningful events like engagement drops, anomalous patterns, or service escalations, enabling immediate response rather than batch processing.</li><li><strong>Agents</strong> — Intelligent entities that autonomously execute decisions within configured guardrails. These agents operate across the end-to-end framework — from data orchestration and feature engineering to signal detection and service activation. They provide natural language interfaces, coordinate with other agents, and maintain full audit trails of their reasoning and actions.</li></ul><h4><strong>Architecture Components</strong></h4><figure><img alt="Customer Intelligence Framework Diagram" src="https://cdn-images-1.medium.com/max/1024/1*gZ40YGrZBBh6ERjXtX3Edw.png" /><figcaption>Customer Intelligence Framework Diagram</figcaption></figure><ul><li><strong>Data ingestion</strong>. Structured, semi-structured, and unstructured sources land through streaming, ETL, object storage, and APIs; catalogued for lineage.</li><li><strong>Feature engineering</strong> (in-database). Transform inputs into robust features and vector representations close to the data to minimize movement.</li><li><strong>Models and rules</strong>. Hybrid decisioning blends ML models, heuristics, and policy rules to balance accuracy, cost, and control.</li><li><strong>Signal processing</strong>. Detect, evaluate, and fuse patterns; a semantic layer maps technical signals (scores, thresholds, sequences) to business concepts (intent, risk, opportunity).</li><li><strong>Service activation</strong> (VCX). Publish decisions and signals via pub/sub so downstream applications subscribe to just-in-time intelligence (NBA/NBO, retention, remediation). Outcomes flow back for continuous learning.</li></ul><p>The architecture flows from data ingestion (supporting structured, semi-structured, and unstructured sources) through features engineering and model deployment using <a href="https://www.teradata.com/platform/clearscape-analytics">ClearScape Analytics®</a> for in-database execution.</p><p>Signal processing interprets these patterns through a semantic layer that maps technical signals to business concepts, while the service activation layer, powered by <a href="https://docs.teradata.com/r/Enterprise_IntelliFlex_VMware/Vantage-Customer-Experience-User-Guide/What-Is-Vantage-Customer-Experience">Teradata VCX</a>, streams real-time intelligence to downstream applications via publish/subscribe patterns.</p><p>Key technical advantages:</p><ul><li>Industry Data Models provide pre-built schemas for rapid deployment</li><li>In-database processing eliminates data movement overhead</li><li>Multi-model orchestration combines rules, ML models, and heuristics</li><li>Every application can subscribe to relevant signals via REST APIs</li></ul><h3>Customer lifetime value multi-agent system demo</h3><p>This demo shows how autonomous customer intelligence turns CLV forecasting and growth into a production workflow. The multi-agent system runs in <a href="https://www.teradata.com/insights/ai-and-machine-learning/agentic-ai-stack-with-teradata-agentbuilder?utm_medium=organic">Teradata AgentBuilder</a>, keeps computation inside Teradata Vantage, and analyzes historical behavior directly in-database. LLMs — GPT-4.1 (routing/response) and Claude Sonnet-4 (analysis/strategy), drive intent understanding, exploration, and decisioning, combining reasoning with real-time signals to answer questions like “Which customer type has the highest CLV and why?” and to recommend next-best actions that lift retention and cross-sell.</p><p>The implementation uses Flowise inside AgentBuilder for low-code composition; the pattern is integration-friendly and supports additional connectors.</p><figure><img alt="Architecture of the Customer Lifetime Value multi-agent system" src="https://cdn-images-1.medium.com/max/1024/1*UeERc7qoAtCQnZTsZ7HVzg.png" /><figcaption>Architecture of the Customer Lifetime Value multi-agent system</figcaption></figure><p>In the multi-agent system architecture, a request enters at “Start”, the conditional agent classifies intent, and exactly one specialist is selected — data exploration, insights generator, or strategic reasoning. The “check if chart required” gate then decides whether a visual adds clarity. If a chart is warranted, the “Visualization Agent” prepares a spec and a custom JavaScript function renders the graphic using <a href="https://docs.teradata.com/r/Enterprise_IntelliFlex_VMware/Teradata-Package-for-Python-User-Guide/Plotting-in-teradataml">Teradata plotting</a>; otherwise, the LLM node composes a concise text response. The result is a governed path from question to decision, with visuals used only when they improve understanding.</p><p>The multi-agent system includes the following nodes:</p><ul><li><strong>Start node</strong>: Initializes the run by capturing the user’s input and shared context (customer/segment, time window, policy flags) for downstream nodes.</li><li><strong>Condition agent nodes</strong>: Route the flow based on intent (exploration/insights/strategy) and presentation needs (whether a chart is required).</li><li><strong>Agent nodes</strong>: Execute scoped tasks — data exploration, insight generation, strategic reasoning, and visualization using <a href="https://github.com/Teradata/teradata-mcp-server">Teradata MCP</a> tools against Vantage with clear guardrails.</li><li><strong>LLM node</strong>: Composes the final, business-ready response by fusing agent outputs, rationale, KPIs, and (if present) a chart caption.</li><li><strong>Custom function node</strong>: Renders artifacts like charts from a validated spec and returns an embeddable image/URI for the response.</li></ul><p>Let’s look at the flow of the whole system!</p><h4><strong>Start node</strong></h4><p>The start node initializes the flow, captures the user’s input (chat or form) and sets shared variables — question, optional customer/segment, time window, table handles, and policy flags, for all downstream nodes.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/589/1*a9nTjKmaVv3icO6aKfw_Kw.png" /><figcaption>Start Node configuration</figcaption></figure><h4><strong>Condition agent node: intent identification agent</strong></h4><p>The Intent Identification Agent is the first of the two condition-agent nodes in the system. It decides which agent to call from the three available specialists, based on the scenario expressed in the user’s query.</p><p>There are three scenarios:</p><ul><li>User is asking for basic details about tables, databases, columns, etc.<br>- <em>Routes to </em><strong><em>Data Exploration Agent</em></strong></li><li>User is asking about banking churn, customer lifetime value, CLV, correlation, sentiments analysis, charts like pie chart, histogram etc.<br>- Routes to <strong>Insights Generator Agent</strong></li><li>User is asking about recommendation to reduce the churn, what are the main factors leading to a churn or low CLV, recommend me method to reduce churn, which are the important factor for churn.<br>- Routes to <strong>Strategic Reasoning Agent</strong></li></ul><figure><img alt="Intent Identification Agent node configuration" src="https://cdn-images-1.medium.com/max/577/1*NQB6iwiVU96AMAnSJjYELg.png" /><figcaption>Intent Identification Agent node configuration</figcaption></figure><p>Based on the prompt instructions, an LLM such as <a href="https://azure.microsoft.com/en-us/blog/announcing-the-gpt-4-1-model-series-for-azure-ai-foundry-developers/">Azure ChatOpenAI (GPT-4.1)</a> interprets the intent, selects the appropriate scenario, and triggers the corresponding agent node. The model can be chosen per task; in this agent node, GPT-4.1 is used and configured with an API key (plus endpoint/deployment where applicable).</p><figure><img alt="LLM configuration of Intent Identification Agent node" src="https://cdn-images-1.medium.com/max/541/1*x0jEDP-vjvZ-eu9sVyUJwQ.png" /><figcaption>LLM configuration of Intent Identification Agent node</figcaption></figure><h4><strong>Agent node: data exploration agent</strong></h4><p>Reports what data is available (e.g., databases/tables, key columns, grain, joins, and freshness) and can return small, PII-safe previews using Teradata MCP tools against Vantage.</p><figure><img alt="Data Exploration Agent node configuration" src="https://cdn-images-1.medium.com/max/571/1*lj42_xbEpnT6lq4JwA427Q.png" /><figcaption>Data Exploration Agent node configuration</figcaption></figure><p>The LLM used here is <a href="https://www.anthropic.com/news/claude-4">Anthropic Claude Sonnet 4</a>; it can be configured the same way as GPT-4.1. All access to LLMs is provided through a cloud service provider. The specific models available, along with credentials and base URLs, depend on the deployment of Teradata and AgentBuilder.</p><figure><img alt="LLM configuration of Data Exploration Agent node" src="https://cdn-images-1.medium.com/max/549/1*Cca7AQQZkh5uvIqim3tdCg.png" /><figcaption>LLM configuration of Data Exploration Agent node</figcaption></figure><p>Access to the Teradata MCP server is configured through the “Tool” panel. As with LLM integration, the connection URL and credentials depend on the specific deployment of Teradata and AgentBuilder.</p><p>Providing the right tools via Teradata MCP is critical for accurate LLM outputs. Select and scope these tools in the “Available Actions” panel so each agent can call exactly what it needs for governed, reliable results.</p><p>Tools can only be selected under Available Actions once MCP server is configured.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/526/1*phUieatgMOndHCMqIJPlJg.png" /><figcaption>MCP configuration of Data Exploration Agent node</figcaption></figure><h4><strong>Agent node: insight generator agent</strong></h4><p>Interprets the question and derives findings only from the governed data, following a reasoning-first workflow: quick data checks (columns, types, missingness), method selection (aggregations, correlations, simple tests), and a brief rationale. It performs the calculation, not just a description, and explains the process before stating results.</p><p>For simple scalar asks, it stays concise; for richer questions, it returns a compact business summary. Output is markdown with “Reasoning” and “Conclusion” sections for clarity and reuse.</p><figure><img alt="Insight Generator Agent node configuration" src="https://cdn-images-1.medium.com/max/564/1*fFdRn3NQzMp3NoVb8oGWEg.png" /><figcaption>Insight Generator Agent node configuration</figcaption></figure><p>The LLM and MCP configurations mirror the Data Exploration Agents configuration, with the addition of <strong>Sentiment Extractor</strong> in Available Actions when needed.</p><figure><img alt="MCP tools configuration of Insight Generator Agent node" src="https://cdn-images-1.medium.com/max/468/1*nTjveqMvf5MDrfDpennfjg.png" /><figcaption>MCP tools configuration of Insight Generator Agent node</figcaption></figure><h4><strong>Agent node: strategic reasoning agent</strong></h4><p>Converts analytical findings into actionable churn-reduction strategies by reviewing feature importance, explaining how key factors drive attrition, and prioritizing where to intervene. It proposes evidence-based methods (e.g., onboarding improvements, complaint triage, targeted cross-sell) and links each recommendation to the contributing drivers with clear rationale and guardrails. Responses are concise and structured in markdown with an intro, numbered reasoning steps, recommended methods to reduce banking churn, and a brief summary of the highest-impact actions.</p><figure><img alt="Strategic Reasoning Agent node configuration" src="https://cdn-images-1.medium.com/max/565/1*LVhcY53VzB0HR3WiCRXfZQ.png" /><figcaption>Strategic Reasoning Agent node configuration</figcaption></figure><p>The MCP and LLM configuration remains the same, with only changes in MCPs Available Actions.</p><figure><img alt="MCP tool configuration of Strategic Reasoning Agent node" src="https://cdn-images-1.medium.com/max/423/1*dnGQcblY4uTp_Bf6q-DBeQ.png" /><figcaption>MCP tool configuration of Strategic Reasoning Agent node</figcaption></figure><h4><strong>Condition agent: check if chart required</strong></h4><p>Decides whether the answer needs a visualization or should remain textual. It inspects the prompt and, only when the user asks to plot, selects a chart type from <strong>Line</strong>, <strong>Polar</strong>, <strong>Pie</strong>, or <strong>Radar</strong>; single-value outputs (e.g., <em>corr: 0.89</em>) skip charts. If a valid chart is warranted, it passes a clean spec to the Visualization Agent and Custom Function to render, otherwise it defaults to a table or plain response and routes directly to the “Generate Response” LLM node.</p><figure><img alt="Check if chart required node configuration" src="https://cdn-images-1.medium.com/max/555/1*2Nba8xE4xrcnS7BElBzpTw.png" /><figcaption>Check if chart required node configuration</figcaption></figure><p>The LLM configuration mirrors the earlier condition agent, this node takes the question as input and chooses between two scenarios — chart-related or general. If chart required is selected, it triggers the Visualization Agent to plot a graph, otherwise, it returns a straightforward textual output.</p><figure><img alt="Input and Scenarios detail of Check if chart required node" src="https://cdn-images-1.medium.com/max/564/1*vkh7GETDIbrnuT1UabtWZA.png" /><figcaption>Input and Scenarios detail of Check if chart required node</figcaption></figure><h4><strong>LLM node: generate response</strong></h4><p>Composes the final, business-ready answer using the context passed from upstream agents, focusing on clear reasoning and actionable guidance (no code). It identifies the domain of the ask (visualization, tables, churn reduction, database design), applies best-practice logic, and balances detail with brevity for simple scalar questions. The output is 2–4 tight paragraphs that explain the “why,” list concrete next steps, note caveats, and tie recommendations to measurable impact.</p><figure><img alt="Generate Response node configuration" src="https://cdn-images-1.medium.com/max/559/1*pdbSI2alUgN0GbkiALHttg.png" /><figcaption>Generate Response node configuration</figcaption></figure><p>For clean presentation, <strong>Return Response As</strong> is set to <em>Assistant Message.</em></p><figure><img alt="Selection of Assistant message in Generate Response node" src="https://cdn-images-1.medium.com/max/547/1*zz8JOI6PTJQPHh1i8D2vsg.png" /><figcaption>Selection of Assistant message in Generate Response node</figcaption></figure><h4><strong>Agent node: visualization agent</strong></h4><p>Emits a strict JSON payload — no prose, notes, or code. If unsure about columns, it first verifies with MCP tools before invoking any chart tools. When a chart is needed, it returns only this structure:</p><pre>{ &quot;type&quot;: &quot;&lt;line|polar|pie|radar&gt;&quot;, &quot;title&quot;: &quot;&lt;chart title&gt;&quot;, &quot;labels&quot;: […], &quot;datasets&quot;: […] }</pre><p>If a chart isn’t appropriate, it returns JSON for a simple table instead, without adding or removing fields or including extra text.</p><figure><img alt="Visualization Agent node configuration" src="https://cdn-images-1.medium.com/max/544/1*IFIh11AxpGEvLGp2XaDv1Q.png" /><figcaption>Visualization Agent node configuration</figcaption></figure><p>Teradata MCP <strong>Available Actions</strong> include Teradata plotting tools that help generate the requested visual.</p><figure><img alt="MCP tool configuration of Visualization Agent node" src="https://cdn-images-1.medium.com/max/409/1*-r-5LnX593SScn01nDCiNg.png" /><figcaption>MCP tool configuration of Visualization Agent node</figcaption></figure><p>It can generate four chart types:</p><ul><li><strong>Line</strong> — trends over time (e.g., weekly churn rate by segment)</li><li><strong>Polar</strong> — circular comparison of magnitudes (e.g., complaints by category)</li><li><strong>Pie</strong> — part-of-whole (e.g., share of customers by risk tier)</li><li><strong>Radar</strong> — multi-metric profile (e.g., cohort scores across engagement, spend, friction)</li></ul><h4><strong>Custom function: draw function</strong></h4><p>The Draw Function turns the Visualization Agent’s JSON into a live chart. It expects a clean payload with type, title, labels, and datasets; if that structure is present, it renders an embedded Chart.js graphic in an iframe. If the input isn’t valid chart data (or the task doesn’t require a chart), it gracefully falls back to a plain message instead of breaking the flow. Supported chart types match the gating node: line, polar, pie, and radar.</p><p>To wire it in the flow, pass the Visualization Agent’s output to the function as “<em>input_data = {{$flow.state.main_vis_op}}</em>”. The function parses that value, validates that labels and datasets are arrays, and then builds the chart with the provided title. If validation fails, the function simply returns the text it received (e.g., “No chart needed”), keeping responses clean and predictable.</p><p>This approach keeps rendering logic outside the agents, avoids code in the final answer, and ensures visuals appear only when they add clarity, everything else stays fast, lightweight, and easy to audit.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/568/1*gjxRblF_w4tm9_FZnY4b1Q.png" /><figcaption>Draw Function node configuration</figcaption></figure><h4><strong>Output</strong></h4><p>The CLV multi-agent system tailors each response to the question — returning plain, business-ready text when a single number suffices (e.g., “What is the average CLV?” → $412.37), and a chart when it adds clarity. Two quick examples from the demo:</p><ul><li><strong>Text only ask</strong>: “What is the average value of CLV”<br>- <strong>Flow</strong>: Start → Intent Identification Agent → Insights Generator Agent → Generate Response</li></ul><p>The assistant computes the mean CLV in-database and returns a concise explanation (no visualization).</p><figure><img alt="Response and process flow details of “What is the average value of CLV” question" src="https://cdn-images-1.medium.com/max/883/1*Imm730hwR6qvZmKzSiT7wA.png" /><figcaption>Response and process flow details of “What is the average value of CLV” question</figcaption></figure><ul><li><strong>Visualization a</strong>sk: “Create a visualization for average value of balance and CLV”<br>- <strong>Flow</strong>: Start → Intent Identification Agent → Insights Generator Agent → Check if chart required → Visualization Agent → Draw Function</li></ul><figure><img alt="Response details of “Create a visualization for average value of balance and CLV” question" src="https://cdn-images-1.medium.com/max/883/1*d7DiESAzc0xfHXnOArvfTQ.png" /><figcaption>Response details of “Create a visualization for average value of balance and CLV” question</figcaption></figure><p>The chart gate detects a plotting request, the Visualization Agent emits a JSON spec, and the Draw Function renders a bar chart.</p><figure><img alt="Process flow details of “Create a visualization for average value of balance and CLV” question" src="https://cdn-images-1.medium.com/max/861/1*9WgMLf2DKkOwOXgDdXTD1w.png" /><figcaption>Process flow details of “Create a visualization for average value of balance and CLV” question</figcaption></figure><p>Every response is traceable, the process flow shows which nodes executed, and the output links back to the data and policies used. Visuals appear only when requested or clearly helpful, otherwise, the answer stays fast, readable, and ready to act on.</p><h3><strong>Conclusion</strong></h3><p>Teradata’s Customer Intelligence Framework closes the loop from <strong>data → signal → activation</strong>. Data remains governed in Vantage; signals are detected and interpreted in real time; and activation happens through guard-railed next-best actions that are explainable and measurable. Built in <a href="https://www.teradata.com/press-releases/2025/teradata-unveils-agentbuilder">AgentBuilder</a>, the CLV demo shows <strong>autonomous customer intelligence</strong> in practice: agents understand intent, assemble evidence, and trigger the right response — text for simple answers, visuals when they add clarity, and recommendations when action is warranted.</p><p>The pattern extends beyond customer lifetime value to use cases such as churn reduction, cross-sell optimization, and service issue deflection, without redesigning the underlying architecture. Performance comes from keeping computation in Teradata Vantage; auditability comes from end-to-end traces of signals, decisions, and actions; and delivery speed comes from reusable prompts, tools, and node templates assembled in AgentBuilder.</p><p><a href="https://www.teradata.com/platform/clearscape-analytics/agent-ai">Ready to turn signals into outcomes?</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=a28d8dd77c41" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/building-autonomous-customer-intelligence-a-developers-guide-to-teradata-s-customer-intelligence-a28d8dd77c41">Building Autonomous Customer Intelligence: A Developer’s Guide to Teradata’s Customer Intelligence…</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Chat with your data in Visual Studio Code through an MCP server]]></title>
            <link>https://medium.com/teradata/chat-with-your-data-in-visual-studio-code-through-an-mcp-server-071673bd3ffe?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/071673bd3ffe</guid>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[teradata]]></category>
            <category><![CDATA[mcp-server]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[ai-agent]]></category>
            <dc:creator><![CDATA[Joshitha Recharla]]></dc:creator>
            <pubDate>Wed, 17 Dec 2025 17:45:17 GMT</pubDate>
            <atom:updated>2025-12-17T17:44:46.853Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="Teradata’s Model Context Protocol (MCP)" src="https://cdn-images-1.medium.com/max/1024/1*ibiQZwku-XO4iMWjaV4Pvg.png" /><figcaption>Teradata’s Model Context Protocol (MCP) with GitHub Copilot</figcaption></figure><p>If you build data pipelines or analytic Python scripts, like dbt models, in Visual Studio Code (VS Code), you know the pain of context-switching just to sanity-check a column list or scan for NULLs. Traditional SQL clients are great, but bouncing between these database clients and VS Code can be distracting and break your flow when your focus is the code you’re writing on VS Code.</p><p>Teradata’s Model Context Protocol (MCP) server helps you by allowing GitHub Copilot Chat to talk to your data directly, providing a handy assistant for drafting analytics scripts.</p><p>Teradata MCP server is an open and extensible backend server that acts as a bridge between large language models (LLMs) and your Teradata environment,<strong> </strong>allowing you to query data, explore metadata, and even validate logic just by asking questions in plain English.</p><p>You stay in VS Code, GitHub Copilot routes data-related requests to the MCP server, and results come back in the chat.</p><p><strong>With this setup, GitHub Copilot can:</strong></p><ul><li>Interpret your natural language questions</li><li>Translate them into SQL queries behind the scenes</li><li>Connect to your Teradata Vantage® instance</li><li>Retrieve structured answers directly in your Copilot chat window</li></ul><p><strong>In this article, I’ll walk you through:</strong></p><ul><li>How to set up your development environment to take advantage of Teradata MCP Server</li><li>How to set up your own MCP server to work with GitHub Copilot Chat in VS Code</li><li>How to query your Teradata database as if<em> </em>you’re having a conversation, simplifying chores like discovering tables, validating filters, and validating data quality in VS Code</li></ul><p>Let’s begin the setup.</p><h3><strong>Requirements</strong></h3><ul><li>A free Teradata VantageCloud environment. <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">Get one in just a few minutes via ClearScape Analytics® Experience.</a></li><li>Visual Studio Code with the Copilot Chat and MCP-Client extensions. The download is available at <a href="https://code.visualstudio.com/">https://code.visualstudio.com/</a>.</li><li>A Teradata MCP server. An open-source community edition is available in Teradata’s GitHub organization; later we’ll cover the details for cloning this repository and setting up the server.</li></ul><h3><strong>Preparing the development environment</strong></h3><h4><strong>Log into ClearScape Analytics® Experience</strong></h4><ul><li><a href="https://www.teradata.com/getting-started/demos/clearscape-reg?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016R000001n3bCQAQ">Sign up for a free account</a> if you haven’t already, or log in.</li><li>In the console, select “create new environment.” Take note of your host and password and the username and database, which will also be needed to set up your server later, default to demo_user in ClearScape Analytics® Experience.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*vFATLUA3xFoU8Qe4FnzFUg.png" /><figcaption>Creating a new ClearScape Analytics environment</figcaption></figure><h4><strong>Install uv</strong></h4><p>We utilize the uv package manager to run the Teradata MCP server in an isolated Python environment to avoid package conflicts and ease the setup.</p><ul><li>Install uv by following the steps at <a href="https://docs.astral.sh/uv/getting-started/installation/">https://docs.astral.sh/uv/getting-started/installation/</a></li><li>Verify if uv is properly installed, open your terminal, and run: ```uv –v```</li></ul><p>This will display the installed version of uv. If you see any error, make sure you have followed the installation instructions and restart the console.</p><h3><strong>Setting up your own Teradata MCP server to work with GitHub Copilot in Visual Studio Code</strong></h3><h4><strong>Clone the Teradata MCP Server</strong></h4><p>git clone <a href="https://github.com/Teradata/teradata-mcp-server.git">https://github.com/Teradata/teradata-mcp-server.git</a></p><h4><strong>Set environment variables</strong></h4><p>Create a .env file with the following structure:</p><pre>DATABASE_URI= teradata://username:password@host:1025/databasename<br>LOGMECH=TD2 #TD2 or LDAP<br>MCP_TRANSPORT=stdio<br>PROFILE = dba</pre><p>We’ll use the stdio transport method in this setup, though other methods are also available. Check the <a href="https://github.com/Teradata/teradata-mcp-server/blob/main/docs/server_guide/CONFIGURATION.md#-transport-modes">relevant documentation</a> in the repository for details.</p><ul><li>Replace username, password, host, and database name with your own Teradata Vantage® credentials.</li><li>The DATABASE_URI tells the server where to point your queries.</li><li>Given the tasks we’re performing, we need a profile that enables the required tools. Specifically, set PROFILE=dba to access the required toolset.</li></ul><h4><strong>Start the MCP Server</strong></h4><p>From your project directory (i.e., the folder where you cloned the teradata-mcp-server), test starting the MCP server using the following command.</p><pre>uv run teradata-mcp-server</pre><p>This boots up the MCP server to verify that everything is working as expected. If everything is in order, you can interrupt this run by pressing Ctrl + C.</p><p><strong>Connecting the MCP server to GitHub Copilot in VS Code</strong></p><p>Now let’s link this server to your GitHub Copilot:</p><ol><li>Open VS Code</li><li>Press Ctrl + Shift + P to open the Command Palette</li><li>Select MCP: Add Server</li><li>Choose <strong>stdio</strong> as the transport type</li><li>Fill in the dialog as follows:</li><li>Command: “uv”</li><li>ID: For example, “teradata-mcp-server”</li><li>Then confirm and save now</li></ol><p>An mcp.json file will be created automatically. Complete the required arguments as shown below:</p><pre>{<br>&quot;mcp&quot;: {<br>&quot;servers&quot;: {<br>&quot;teradatastdio&quot;: {<br>&quot;type&quot;: &quot;stdio&quot;,<br>&quot;command&quot;: &quot;uv&quot;,<br>&quot;args&quot;: [<br>&quot; - directory&quot;,<br>&quot;&lt;Full Path&gt;/teradata-mcp-server&quot;,<br>&quot;run&quot;,<br>&quot;teradata-mcp-server&quot;<br>],<br>}<br>}<br>}<br>}</pre><p>Make sure to replace &lt;full path&gt; with the actual location of the teradata-mcp-server directory if it didn’t load itself.</p><h4><strong>Launch the MCP Server in VS Code</strong></h4><p>Once configured, start the server from within VS Code if it didn’t start itself:</p><ul><li>Friendly controls to start, restart, and stop the server overlay on top of the mcp.json file for convenient operation</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*mAmw1CjrQIrEHgClGBZElA.png" /><figcaption>Configuring the MCP server in mcp.json to run the Teradata MCP server via stdio.</figcaption></figure><h4><strong>Chat with your data</strong></h4><p>To interact with your Teradata database in natural language, you’ll need to use Copilot in Agent Mode, which allows GitHub Copilot Chat to talk directly to the MCP server.</p><p>Here’s how to do it.</p><h4><strong>Open GitHub Copilot Chat in VS Code</strong></h4><ul><li>In your sidebar, click on the GitHub Copilot Chat icon.</li><li>If you don’t see it, open the Command Palette (Ctrl + Shift + P) and search for “GitHub Copilot: Open Chat View”</li></ul><h4><strong>Switch to Agent Mode</strong></h4><ul><li>Go to the Copilot Chat panel in VS Studio</li><li>Ensure it’s set to Agent Mode (not Chat Mode). If you don’t see this, go to command palette to enter Copilot: Toggle Agent Mode.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*M_kRqPZZ9hk2mgvQlMnWBA.png" /><figcaption>Set GitHub Copilot Chat to Agent Mode to allow MCP-based tool and database interactions.</figcaption></figure><p>Once in Agent Mode, Copilot will be ready to route your questions through the MCP server to your Teradata backend and return results right inside your chat window.</p><p><strong>Note on data:</strong> By default, our walkthrough defines demo_user as the default database. This database is empty in a new ClearScape Analytics® environment, you should load sample data first or update the database/table names in the examples to match tables that exist in your environment. This ensures you can explore and query your data.</p><p>Try prompts like:</p><ul><li>“Show me the first 5 rows from demo_user.dim_customers”</li><li>“What tables exist in the demo_user schema?”</li><li>“Give me the total sales by region from demo_user.sales_fact”</li><li>“Which columns are available in the demo_user.orders table?”</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*vOm7VydQFTEoRqisrsJyjg.png" /><figcaption>Querying the Teradata database in natural language via GitHub Copilot Chat, with requests executed through the MCP server.</figcaption></figure><h3><strong>Conclusion</strong></h3><p>By combining Teradata’s MCP Server with GitHub Copilot Chat in Visual Studio Code, we unlock a powerful way to make data feel more conversational, accessible, and integrated into the daily developer workflow.</p><p>If you’re someone who works with Teradata and builds in VS Code, I highly recommend giving this workflow a try. Running the queries and getting results in the chat might boost your productivity.</p><p>Got questions or want to share your own use case? Feel free to reach out or leave a comment.</p><h3>About Joshitha Recharla</h3><p>Joshitha Recharla is a developer advocate intern at Teradata and a graduate student in data science at SUNY Buffalo. She enjoys breaking down complex AI and analytics workflows into hands-on, developer-friendly tutorials and demos. Through her internship, she has explored how tools like MCP can make data more accessible and useful where developers work.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=071673bd3ffe" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/chat-with-your-data-in-visual-studio-code-through-an-mcp-server-071673bd3ffe">Chat with your data in Visual Studio Code through an MCP server</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Teradata Enables AI Factories with NVIDIA: On-premises, secure, and predictable AI development]]></title>
            <link>https://medium.com/teradata/teradata-enables-ai-factories-with-nvidia-on-premises-secure-and-predictable-ai-development-8640ef2e714f?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/8640ef2e714f</guid>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[ai-tools]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[data-scientist]]></category>
            <dc:creator><![CDATA[Vidhan Bhonsle]]></dc:creator>
            <pubDate>Wed, 17 Dec 2025 16:48:20 GMT</pubDate>
            <atom:updated>2025-12-17T16:47:58.162Z</atom:updated>
            <content:encoded><![CDATA[<h3>Introduction</h3><p>Operationalizing AI in a secure, on-premises environment comes with its own set of challenges, especially when the goal is to build production-ready applications that handle sensitive, high-volume, or unstructured data.</p><p>For a data scientist, that might mean testing large language models against internal documents. For a Machine Learning engineer, it could involve deploying and scaling models efficiently across GPU resources. A data engineer may be focused on integrating structured and unstructured data into unified pipelines, while an AI architect is looking for a system that supports end-to-end control, compliance, and performance at scale.</p><p>Teradata and NVIDIA enable AI factories that support all these needs with integrated tools for model development, governance, inference, and vector storage, all working together within your trusted infrastructure.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FQObGG4JWJHY%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DQObGG4JWJHY&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FQObGG4JWJHY%2Fhqdefault.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/85f7163ba46b42ce324975341c119a39/href">https://medium.com/media/85f7163ba46b42ce324975341c119a39/href</a></iframe><p>For a more detailed overview of how Teradata enables AI factories with NVIDIA, <a href="https://medium.com/teradata/teradata-enables-ai-factories-with-nvidia-for-on-premises-ai-ml-697a076dbc81">​​explore this article</a>.</p><p>In this article, we walk through how it works, from accessing ​Teradata ​AI Workbench and deploying a foundation model, to running GPU-accelerated inference and building a RAG pipeline using a PDF document.</p><h3>How Teradata ​Enables ​AI Factories</h3><p>Automating the full claims lifecycle from document extraction to settlement without human intervention. Monitoring conversations in real time to detect compliance risks before they escalate. Enhancing Anti-Money Laundering detection by bringing together structured records and unstructured sources like social media.</p><p>These are just a few of the practical, high-impact solutions developers can build with Teradata.</p><p>With integrated vector storage, GPU-accelerated inference, ClearScape Analytics®, and a full-featured AI Workbench, Teradata turns these complex workloads into secure, scalable, and executable pipelines entirely within your own infrastructure.</p><p>Let’s take a look at how you can leverage Teradata and NVIDIA for your AI factory.</p><h4>Accessing the ​Teradata​ ​​AI Workbench</h4><p>Your journey begins at the ​​Teradata ​​AI Workbench landing page, your central hub for everything AI. From here, you can launch notebooks, utilize ModelOps, manage users, and access support if needed, all from a single interface.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*DZQRNVGSgAIv1o5hlcB9Yg.png" /><figcaption>Teradata AI Workbench landing page</figcaption></figure><h4>Exploring ModelOps for advanced AI management</h4><p>Inside the Workbench, ModelOps provides a complete environment for managing your AI models, including access to <a href="https://developer.nvidia.com/nim?sortBy=developer_learning_library%2Fsort%2Ffeatured_in.nim%3Adesc%2Ctitle%3Aasc&amp;hitsPerPage=12">NVIDIA NIM</a> microservices and a collection of over 80+ pre-trained Large Language Models (LLMs), all ready to be deployed on your GPUs using NIM​.​</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*9qFsDgRBO4coI0WYaAopBw.png" /><figcaption>List of AI Foundation Models in Teradata AI Workbench</figcaption></figure><h4>Selecting a model</h4><p>Selecting a model is as simple as a click. Choose any ​​NIM​​, and a model card will open with details such as model evaluation, benchmark performance, license information, and resource profile.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/999/1*AGM83aVUeTbWZ8Yx5JXzLQ.png" /><figcaption>Model card for Llama-3-Swallow-70B Instruct v0.1</figcaption></figure><h4>One-click model deployment</h4><p>After reviewing a model, you can deploy it with just one click. Once deployed, you receive an endpoint and a code snippet, perfect for integrating the model into your Generative AI applications, including Retrieval-Augmented Generation (RAG) workflows with Teradata Enterprise Vector Store.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*cdSF_bKqdPUoCztZg_H6hQ.png" /><figcaption>Model deployment steps in Teradata AI Workbench</figcaption></figure><h4>Running inferences and building applications</h4><p>With your model deployed and the endpoint ready, you can immediately start running inferences against your data. This enables you to quickly build applications for demonstration, experimentation, or even production use cases.</p><h4>A practical use case: PDF Vector Store with RAG</h4><p>One of the standout capabilities is its ability to transform unstructured documents into actionable insights using RAG.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*qnzFyrEmZqiOeYAdNkvSKw.png" /><figcaption>Teradata implementation for AI Factories</figcaption></figure><p>Teradata AI Microservices with NVIDIA enables access to GPU-accelerated tools for processing PDFs, extracting text, images, tables, and charts using OCR, object detection, and table parsing. These documents are then enriched using NVIDIA NIM to generate high-dimensional vector embeddings via contextual LLMs.</p><p>Finally, the embeddings are indexed in Teradata’s Enterprise Vector Store, enabling fast, intelligent search across the data. This pipeline makes it easy to build RAG applications that can analyze, summarize, and surface relevant insights from unstructured documents — quickly and securely.</p><h4>Code example: building a Vector Store with an NVIDIA model</h4><p>Let’s walk through how to create a vector store and ask contextual questions using a PDF document. In this example, we’re using a sample file named <em>data_access_policy.pdf</em> to demonstrate the steps. You can replace it with your own document as needed.</p><ol><li><strong>Install and import required modules</strong></li></ol><p>Install the <a href="https://docs.teradata.com/r/Lake-Analyze-Your-Data-with-ClearScape-AnalyticsTM/Teradata-Package-for-Generative-AI">teradatagenai</a> module, a Teradata package for Generative AI and import the required classes.</p><pre>!pip install teradatagenai<br>from teradatagenai import VSManager, VectorStore, VSPattern, VSApi<br>import os</pre><p><strong>2. Initialize the Vector Store</strong></p><p>Create an instance of the vector store named “vs_chatbot”</p><pre>chatbot = VectorStore(&#39;vs_chatbot&#39;)</pre><p>You can also check the system health to know if your system is running fine or not with “VSManager.health()”.</p><p><strong>3. Load the document and define the embedding model</strong></p><p>Load a sample PDF and specify the NVIDIA embedding model (nv-embedqa-mistral-7b-v2).</p><pre>files = [os.path.join(os.getcwd(), &quot;data_access_policy.pdf&quot;)]<br>embeddings_model = &#39;nvidia/nv-embedqa-mistral-7b-v2&#39;</pre><p><a href="https://build.nvidia.com/nvidia/nv-embedqa-mistral-7b-v2">nv-embedqa-mistral-7b-v2</a> is a multilingual text question-answering retrieval model from NVIDIA.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*NXRjSBwD0pgQC8OHslkm3w.png" /><figcaption>data_access_policy.pdf content</figcaption></figure><p><strong>4. Create the Vector Store</strong></p><p>Generate the vector embeddings for the document and store them using Teradata’s Enterprise Vector Store.</p><pre>chatbot.create(embeddings_model=embeddings_model,<br>search_algorithm= &#39;VECTORDISTANCE&#39;,<br>top_k= 10,<br>object_names= [&#39;data_access_policy&#39;],<br>data_columns= [&#39;chunks&#39;],<br>vector_column= &#39;VectorIndex&#39;,<br>chunk_size= 500,<br>optimized_chunking=False,<br>document_files=files)</pre><p><strong>5. Check Vector Store Status</strong></p><p>Confirm if the vector store is ready for querying.</p><pre>chatbot.status()</pre><p><strong>6. Ask a contextual question</strong></p><p>You can query the vector store with a business-specific question using ask method. The vector store will use its similarity search to find the most relevant information and provide a detailed response.</p><pre>question = &#39;What are the levels of access to patient data allowed in our organization?&#39;<br>prompt = &#39;List the levels of access and briefly describe them.&#39;<br>response = chatbot.ask(question=question, prompt=prompt)<br>print(response)</pre><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*67zcxoXMwhE34EKzkkJoPw.png" /><figcaption>Output of ask method</figcaption></figure><p>This response was generated directly from the uploaded PDF using the ask() function in Teradata AI Factory. It shows how unstructured documents can be converted into actionable insights accurately, securely, and instantly.</p><h3>Conclusion</h3><p>Teradata and NVIDIA provide everything needed to build, deploy, and manage AI workloads entirely on-premises by combining model development, GPU-powered inference, vector management, and workflow automation into a single integrated environment.</p><p>AI factories leveraging Teradata and NVIDIA equip data scientists, machine learning engineers, data engineers, and AI architects with the tools to experiment, scale, and operationalize AI, all while maintaining full control over performance, compliance, and data security.</p><p>To learn more about AI factories leveraging Teradata and NVIDIA, explore <a href="https://www.teradata.com/platform/teradata-ai-factory">here</a>.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=8640ef2e714f" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/teradata-enables-ai-factories-with-nvidia-on-premises-secure-and-predictable-ai-development-8640ef2e714f">Teradata Enables AI Factories with NVIDIA: On-premises, secure, and predictable AI development</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Agentic AI Stack with Teradata AgentBuilder]]></title>
            <link>https://medium.com/teradata/agentic-ai-stack-with-teradata-agentbuilder-aea2b3e3c718?source=rss----8adfdb056496---4</link>
            <guid isPermaLink="false">https://medium.com/p/aea2b3e3c718</guid>
            <category><![CDATA[flowise]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[agentic-ai]]></category>
            <category><![CDATA[data-analysis]]></category>
            <category><![CDATA[teradata]]></category>
            <dc:creator><![CDATA[Daniel Herrera]]></dc:creator>
            <pubDate>Wed, 26 Nov 2025 16:41:55 GMT</pubDate>
            <atom:updated>2025-11-26T16:41:53.939Z</atom:updated>
            <content:encoded><![CDATA[<p>Accelerate enterprise AI with integrated agentic tools, trusted data, and seamless orchestration using Teradata AgentBuilder.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/541/1*Q2rydByxql36zbJx3QfMMA.png" /><figcaption>Teradata AgentBuilder</figcaption></figure><p><a href="https://www.teradata.com/insights/ai-and-machine-learning/ai-agents-in-action?utm_medium=organic">Agentic AI</a> refers to artificial intelligence systems that can operate autonomously, make decisions, and take actions to achieve specific goals with minimal human intervention. These systems exhibit characteristics like reasoning, planning, adaptability, and goal-directed behavior.</p><p>Building agentic AI systems in the enterprise often means manually integrating components and frameworks for connecting to large language model providers, managing prompts, executing tools, handling memory and context, and ensuring agents have access to relevant data and business domain knowledge. For AI engineering, data science, and data engineering teams, this creates considerable overhead in infrastructure setup before they can even begin experimenting or building agentic solutions for their customers and organizations.</p><p>Teradata AgentBuilder simplifies this process by offering a unified developer experience that connects autonomous agents to trusted data, domain expertise, and scalable tools. Everything is integrated and deployed alongside Teradata. Teradata removes the infrastructure burden and provides an enterprise-grade agentic data stack out of the box.</p><p>Teradata AgentBuilder is a set of capabilities that enables enterprises to build and manage autonomous AI agents with trusted data, domain expertise, and hybrid deployment flexibility.</p><p>These capabilities include:</p><ul><li>UI: A UI interface to build agents and develop multi-agent workflows. Current choices include open-source tools such as <a href="https://flowiseai.com/">Flowise</a>, <a href="https://www.crewai.com/">CrewAI</a>, <a href="https://www.langchain.com/langgraph">LangGraph</a> as well as partner tools from CSPs, and NVIDIA.</li><li>Context intelligence combines Teradata IP such as expert prompts, contextual understanding, and the right tools with open-and-connected LLM APIs.</li><li>Teradata MCP Server and tools, which enables agents to leverage tools for descriptive, predictive, and prescriptive analytics. The toolset will comprise Basic tools, DBA Tools, EFS (Enterprise Feature Store), Quality Tools, vector store, RAG, metadata, SQL tools, and a few select ClearScape Analytics® tools.</li><li>Pre-built Agents — Teradata developed agents: Teradata SQL agent, Teradata data science agent and Teradata monitoring agent.</li></ul><p>All these components are designed to work together, giving teams a robust foundation for building intelligent multi-agent systems.</p><p>In this post, we’ll walk through a credit risk profiling use-case powered by a multi-agent system, built on top of Teradata AgentBuilder. While the demo focuses on core MCP tools and agent workflows, Teradata also provides <a href="https://www.teradata.com/resources/datasheets/industry-data-models-idm">industry data models</a> that can be embedded to enrich agent context. These models offer a powerful way to infuse agents with business logic and domain expertise, making them more effective in solving real-world problems.</p><h3>A credit risk profiling multi-agent system</h3><p>We built a multi-agent AI system to analyze credit risk data using Teradata AgentBuilder. The data is stored and processed in Teradata Vantage. Based on historical records, we deployed a credit risk assessment model using ClearScape Analytics. This agentic system can query and interpret data from the database, act as a data analyst, score applicants against the model to assess their risk profile and explain the logic behind the model.</p><p>This system was implemented using Flowise integration within AgentBuilder, and while this demo highlights Flowise, additional integrations are already supported, and more are on the way.</p><p>The architecture of the system is the following:</p><figure><img alt="Query routing in AgentBuilder" src="https://cdn-images-1.medium.com/max/624/0*3ochMjP3khqNOZox" /><figcaption>Query router in AgentBuilder</figcaption></figure><p>With the following nodes:</p><ul><li>Start node:<br>Defines how the workflow is initiated, in this case through a chat interface (form inputs are also available).</li><li>Condition agent node:<br>Routes the flow to purpose specific agents according to the user’s intent as inferred from the user’s prompt.</li><li>Purpose specific agent nodes:<br>Execute the actual task needed to respond to the user’s prompt. This can be either providing information regarding what data is available in the relevant database, performing summaries and aggregations of data, acting as a data analyst, scoring data against the deployed model or explaining the model itself.</li></ul><p>The risk assessment model is a Generalized Linear Model created and deployed through Teradata ClearScape Analytics. The specific steps followed to create the model can be explored in this <a href="https://github.com/Teradata/jupyter-demos/blob/main/UseCases/Credit_Risk_Assessment/Credit_Risk_Assessment_Python.ipynb">notebook</a> (also available ready to execute in <a href="https://www.teradata.com/getting-started/demos/clearscape-analytics?utm_campaign=gbl-clearscape-analytics-devrel&amp;utm_content=demo&amp;utm_id=7016r000001n3bcqaq&amp;utm_medium=organic">ClearScape Analytics Experience</a>); they are not covered in detail in this blog.</p><h3>The start node</h3><p>This node initializes the flow by capturing the user input, as a chat prompt or as a form submission.</p><p>The flow state, or variables used by the flow are defined in the start node.</p><figure><img alt="Start node configuration" src="https://cdn-images-1.medium.com/max/584/0*O586R9y8dowRWfv5" /><figcaption>Start node configuration</figcaption></figure><h3>Agent nodes</h3><p>The flow consists of five agent nodes, one of these agent nodes is a router node, which routes the flow to the specific purpose agents according to the user intent. The other four are purpose-specific agents that respond to the user with a clearly defined scope.</p><h4>Common configurations to all agent nodes</h4><p>Each node is individually configured through a low-code user interface. <br>All agent nodes require access to large language models (LLMs), which can be integrated via cloud service providers or on-premise deployments. Additionally, the four purpose-specific agents also require access to <a href="https://github.com/Teradata/teradata-mcp-server">Teradata MCP server</a>.</p><p>Teradata AgentBuilder seamlessly incorporates these components, making them straightforward to configure within each corresponding node.</p><h4>LLM configuration</h4><p>In this case, access to LLMs is provided through a cloud service provider. The specific models available, along with the required credentials and base URL, depend on your particular deployment of Teradata and AgentBuilder.</p><figure><img alt="LLM configuration" src="https://cdn-images-1.medium.com/max/536/0*DO1NyzU7H_V9jutq" /><figcaption>LLM Configuration</figcaption></figure><h4>MCP configuration</h4><p>Access to the Teradata MCP server is configured through the tools panel. As with LLM integration, the connection URL and credentials depend on your specific deployment of Teradata and AgentBuilder.</p><figure><img alt="MCP configuration" src="https://cdn-images-1.medium.com/max/540/0*k5mmUzDwwwWojJdw" /><figcaption>MCP Configuration</figcaption></figure><p>After configuring the MCP server, all related tools are available for selection.</p><figure><img alt="MCP Toolset" src="https://cdn-images-1.medium.com/max/454/0*rVi1kO3SgBDtDE6B" /><figcaption>MCP Toolset</figcaption></figure><h4>Question evaluation agent</h4><p>A condition agent node routes the flow according to the user intent, the most important configurations in this node are the general instructions which detail the logical flow and rationale of each of the individual routes, and the scenario list which creates each of the separate branches of the flow (four in total, one shown in the graph below). The input, in this example, is the {{question}} variable, this is a global variable that contains the initial prompt of the user to the chatbot agent and that can be retrieved from any of the nodes in the flow.</p><figure><img alt="Query evaluation" src="https://cdn-images-1.medium.com/max/499/0*UYv-h3HGf80rIUxD" /><figcaption>Query evaluation agent</figcaption></figure><p>This agent evaluates the user’s query against the following four scenarios:</p><ul><li>User is asking about tables, columns or types of columns → Route to the data exploration agent</li><li>User is asking about summarizing, aggregating or otherwise calculating on the data of a given table → Route to the data aggregation agent</li><li>User is asking for assessing the risk profile of a specific record → Route to the credit profiling agent</li><li>User is asking for an analysis of the assessment model → Route to the model analysis agent</li></ul><p>The agent is instructed to:</p><ul><li>Analyze the input: Determine the user’s intent based on phrasing and keywords.</li><li>Match to a scenario: Compare the input against the list of scenarios.</li><li>Route accordingly: Return the selected scenario in JSON format, which triggers the corresponding agent node.</li></ul><figure><img alt="Routing of user query" src="https://cdn-images-1.medium.com/max/624/0*bheFue4JGK-hvLEQ" /><figcaption>Routing of user query</figcaption></figure><h3>Purpose-specific agents</h3><p>The primary configuration for each agent is its corresponding system prompt. It is important that every purpose-specific agent has a clearly defined scope of action. This allows designing system prompts that deliver precise and relevant instructions tailored to the agent’s task. One agent handles general data exploration, another focuses on summarization and aggregation, a third performs data scoring, and a fourth interprets the credit risk model.</p><h4>Data exploration and data summarization agents</h4><p>These are the simplest agents, they answer questions regarding the data present in our risk credit profiling database, and perform summarizations, acting like a data analyst for this specific dataset.</p><figure><img alt="System prompt data summarization agent" src="https://cdn-images-1.medium.com/max/526/0*9jSst9YJR4kff01p" /><figcaption>System prompt data summarization agent</figcaption></figure><p>These agents respond to queries like the one shown below:</p><figure><img alt="Data analysis agent in action" src="https://cdn-images-1.medium.com/max/400/0*u5QxWFkL_JEFGuzA" /><figcaption>Data analysis agent in action</figcaption></figure><h4>Credit profiling agent</h4><p>The credit profiling agent is the core of the project. It retrieves the data from a given credit application and based on that data, assesses the credit profile of the applicant. This assessment is guided by the scoring logic of the ClearScape Analytics credit profiling model, which is deployed directly within the Teradata Vantage database.</p><figure><img alt="Credit profiling agent configuration" src="https://cdn-images-1.medium.com/max/532/0*bs9iSEvhjxlwsRu2" /><figcaption>Credit profiling agent configuration</figcaption></figure><p>The details of this process are detailed in the corresponding system prompt for the agent:</p><figure><img alt="System prompt credit profiling agent" src="https://cdn-images-1.medium.com/max/624/0*kPlRidYhINLUiVoK" /><figcaption>System prompt credit profiling agent</figcaption></figure><p>This prompt utilizes ClearScape Analytics experience scoring capabilities to score the requested data upon a user’s request.</p><figure><img alt="System scoring a specific customer" src="https://cdn-images-1.medium.com/max/371/0*xrgb_67TLfRwSKb4" /><figcaption>System scoring a specific customer</figcaption></figure><p>What determines a strong or weak credit profile within the context of the currently deployed model? The final agent in the system analyzes the model’s parameters and provides responses to potential queries from business users regarding the workings of the model.</p><h4>Model analysis agent</h4><p>This agent retrieves the parameters stored in the table where the credit profiling model resides and provides insights into how specific features influence the resulting credit profile.</p><figure><img alt="Model analysis agent" src="https://cdn-images-1.medium.com/max/380/0*wJJSVtLXgvwzWNGf" /><figcaption>Model analysis agent</figcaption></figure><h3>Conclusion</h3><p>With Teradata AgentBuilder, teams can move beyond infrastructure setup and start building intelligent multi-agent systems that are deeply integrated with enterprise data and domain knowledge. By combining open-source agentic frameworks, scalable tooling, and trusted data access through Vantage, AgentBuilder provides a streamlined path from experimentation to production. Whether you’re modeling credit risk, optimizing supply chains, or building customer-facing assistants, your agentic data stack starts here, with the flexibility, performance, and reliability that enterprise AI demands.</p><p><a href="http://teradata.com/agentai">Ready to start building?</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=aea2b3e3c718" width="1" height="1" alt=""><hr><p><a href="https://medium.com/teradata/agentic-ai-stack-with-teradata-agentbuilder-aea2b3e3c718">Agentic AI Stack with Teradata AgentBuilder</a> was originally published in <a href="https://medium.com/teradata">Teradata</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
    </channel>
</rss>