Mcp Server Redshift Setup
Mcp Server Redshift Setup
Written by The Data Workers Team — 14 autonomous agents shipping production data infrastructure since 2026.
Technically reviewed by the Data Workers engineering team.
Last updated .
Running MCP against Redshift means provisioning a dedicated workload management (WLM) queue for agents, connecting through the Data API rather than JDBC, and enforcing a query timeout per statement. That setup keeps analyst dashboards fast while giving agents the access they need to ship answers.
Redshift is the most common AWS warehouse, and teams running it want agent access without risking the cluster. This guide covers authentication, WLM queues, Data API vs JDBC, query policies, and observability, so an MCP server can query Redshift the same way a BI tool would.
Use the Data API, Not JDBC
Redshift's Data API runs over IAM-authenticated HTTPS and returns results asynchronously. That is a much better fit for MCP than classic JDBC connections, which hold long-lived TCP sockets and require VPC networking. The Data API gives you IAM auth for free, eliminates connection pooling, and works from any environment the agent runs in.
If you already run JDBC for legacy BI tools, keep those connections and add the Data API on top for MCP. The two protocols do not conflict. For Redshift Serverless, the Data API is effectively the only sensible option.
Dedicated WLM Queue
Redshift's Workload Management lets you carve out a queue for agent traffic so it cannot starve the analyst queue. Create a queue called agents with a concurrency of 5, a memory share of 20%, and a query timeout of 120 seconds. Route MCP queries to that queue via a query group label, and the agent cannot block an analyst dashboard no matter how badly it misbehaves.
- •Dedicated queue — isolates agent traffic
- •Query group label — routes agent SQL to the right queue
- •Concurrency cap — 5 slots maximum
- •Memory share — 20% of cluster
- •Statement timeout — 120 seconds hard cap
IAM Roles and Least Privilege
The Data API authenticates via IAM, so the MCP server runs as a named IAM role (mcp-redshift-agent) with permissions to call redshift-data:ExecuteStatement, GetStatementResult, and DescribeStatement on a specific cluster. Inside Redshift, that role maps to a database user via IAM:GetClusterCredentials, scoped to a read-only schema.
| Layer | Control | Purpose |
|---|---|---|
| IAM | redshift-data:* on one cluster | API-level access |
| Redshift user | SELECT on analytics_v | Schema-level access |
| WLM queue | agents, 5 slots | Resource isolation |
| Statement timeout | 120s | Kills runaway SQL |
| Query group | agent | Routes to correct queue |
| Audit | STL_QUERY + CloudTrail | Full record of activity |
Column-Level Security
Redshift supports column-level grants, so you can grant the agent SELECT on specific columns while excluding PII. A safer pattern is to expose a curated view that already redacts sensitive columns and grant the agent SELECT on the view. Either approach keeps the raw table invisible to the MCP server.
Observability with STL and SVV Tables
Redshift exposes query history via STL_QUERY (last 2-5 days) and SYS_QUERY_HISTORY (serverless). Join those against the MCP audit log using the query group label and you have a complete record of what the agent ran, how long it took, and how much spillage it caused. Ship the joined stream to CloudWatch for dashboards.
Data Workers on Redshift
Data Workers' Redshift connector uses the Data API, routes queries to the agent WLM queue, and tags every statement with the agent ID. The catalog agent discovers tables via SVV_COLUMNS, the cost agent watches for spillage, and the pipeline agent can trigger downstream transforms. See AI for data infrastructure for the full stack or compare with MCP server BigQuery production setup.
To see a Redshift MCP server running in production with WLM isolation and Data API auth, book a demo. We will walk through cluster setup, queue configuration, and query attribution.
Redshift Serverless deserves a special mention because it changes the MCP story. With classic Redshift you pay for cluster hours whether the agent is active or not, and WLM queues are the main isolation mechanism. Serverless charges per query based on RPU-seconds consumed, which makes agent workloads much cheaper — you pay only when the agent runs a query. For MCP that do not need round-the-clock availability, Serverless is often the right default.
The Data API also solves a long-standing pain point for Redshift: connection limits. Classic Redshift caps the number of concurrent connections per cluster, and a misbehaving MCP client can exhaust that pool. The Data API is stateless — every call is a short-lived HTTP request — so it sidesteps the connection limit entirely. Agents that poll frequently, or multiple agents sharing one cluster, see none of the connection pressure they would hit with JDBC.
One often-missed optimization is result set caching. Redshift caches query results automatically for 24 hours when the source tables have not changed, and cache hits are free. Agents that ask similar questions repeatedly get massive speedups from this cache, and the MCP server does not need to do anything special — Redshift handles it transparently. Make sure the agent's query tag does not inadvertently poison the cache by varying on every call.
Redshift MCP works once you give agents their own WLM queue, route through the Data API, and enforce a statement timeout. The rest is standard warehouse hygiene — column grants, curated views, and query history observability.
Further Reading
Sources
See Data Workers in action
15 autonomous AI agents working across your entire data stack. MCP-native, open-source, deployed in minutes.
Book a DemoRelated Resources
- Mcp Server Snowflake Production Setup — Mcp Server Snowflake Production Setup
- Mcp Server Bigquery Production Setup — Mcp Server Bigquery Production Setup
- Why AI Agents Need MCP Servers for Data Engineering — MCP servers give AI agents structured access to your data tools — Snowflake, BigQuery, dbt, Airflow, and more. Here is why MCP is the int…
- MCP Server Analytics: Understanding How Your AI Tools Are Actually Used — Your team uses dozens of MCP tools every day. MCP analytics tracks adoption, measures ROI, identifies unused tools, and provides the usag…
- How to Build an MCP Server for Your Data Warehouse (Tutorial) — MCP servers give AI agents structured access to your data warehouse. This tutorial walks through building one from scratch — TypeScript,…
- MCP Server Security: Authentication, Authorization, and Audit Trails — MCP servers expose powerful capabilities to AI agents. Securing them requires OAuth 2.1 authentication, scoped authorization, least-privi…
- MCP Server for Snowflake: Connect AI Agents to Your Data Warehouse — Snowflake's MCP server exposes Cortex Analyst, Cortex Search, and schema metadata to AI agents. Here's how to set it up and how Data Work…
- MCP Server for BigQuery: Give AI Agents Access to Your Analytics — BigQuery's MCP server gives AI agents access to schemas, query execution, and cost estimation. Here's how to connect it and use Data Work…
- MCP Server Tutorial: Build a Data Warehouse Integration in 30 Minutes (Python) — Build an MCP server for your data warehouse in 30 minutes with Python. Step-by-step tutorial covering schema exposure, query execution, a…
- MCP Server for Databases: Connect AI Agents to Postgres, BigQuery, and Snowflake — Connect AI agents to Postgres, BigQuery, and Snowflake via MCP servers. Database-specific patterns, schema exposure, and query execution.
- Remote MCP Servers: Deploy AI Tool Integrations to Production — Remote MCP servers move AI tool integrations from local development to production — with OAuth authentication, mTLS security, Kubernetes…
- MCP Server for Postgres: Connect AI Agents to Your Relational Database — Connect AI agents to PostgreSQL via MCP. Covers core query tools, advanced features (pgvector, TimescaleDB, PostGIS), and security best p…
Explore Topic Clusters
- Data Governance: The Complete Guide — Policies, access controls, PII, and compliance at scale.
- Data Catalog: The Complete Guide — Discovery, metadata, lineage, and the modern catalog stack.
- Data Lineage: The Complete Guide — Column-level lineage, impact analysis, and observability.
- Data Quality: The Complete Guide — Tests, SLAs, anomaly detection, and data reliability engineering.
- AI Data Engineering: The Complete Guide — LLMs, agents, and autonomous workflows across the data stack.
- MCP for Data: The Complete Guide — Model Context Protocol servers, tools, and agent integration.
- Data Mesh & Data Fabric: The Complete Guide — Federated ownership, domain-oriented architecture, and interop.
- Open-Source Data Stack: The Complete Guide — dbt, Airflow, Iceberg, DuckDB, and the modern OSS toolkit.
- AI for Data Infra — The complete category for AI agents built specifically for data engineering, data governance, and data infrastructure work.