AI and Analytics for Data Teams: What It Is, What’s Available, and How to Get Started

AI and Analytics for Data Teams: What It Is, What’s Available, and How to Get Started – SQLYARD

AI and Analytics for Data Teams: What It Is, What’s Available, and How to Get Started


Leadership is asking for AI-powered reports and analytics. Business analysts are being asked to deliver insights faster. Data warehouse teams and data engineers are evaluating new platforms. DBAs are fielding questions about where the database fits in an AI strategy. Everyone is hearing the same words — vectors, embeddings, semantic search, RAG, Copilot — and not everyone has had the time to understand what they actually mean or what options are available.

This article is for all of those people. It starts with plain-language explanations that anyone on the team can read and understand. It covers the full landscape of options — Microsoft, Snowflake, Google, AWS, Databricks, and standalone tools — honestly and without assuming you have already picked a platform. And for the technical members of the team, it goes all the way through to working code examples. Every section is labeled so you can read what is relevant to your role and skip what is not.

Who this article is for: Leadership Business Analysts Report Developers DBAs Data Engineers Data Warehouse Teams. Sections are labeled by audience and skill level so you can navigate to what matters for your role.

1 Why Leadership Is Asking About AI and Analytics Right Now Everyone

For most organizations, data has always been abundant and insight has been scarce. You had the transactions, the orders, the customer records, the support tickets — gigabytes or terabytes of it sitting in databases and data warehouses. But turning that data into answers took time. An analyst would write a report, a developer would build a dashboard, and by the time the insight reached the decision-maker it was days or weeks old and answered a specific pre-defined question.

AI changes that relationship in two fundamental ways.

First, AI makes it possible to ask questions that were not pre-defined. Instead of a fixed report that shows sales by region for last quarter, an executive can ask “Why did the Northeast underperform in Q1 compared to the same period last year, and which product categories drove the gap?” and get an answer from the data without a developer building a new report first. This is what leadership means when they ask for AI-powered analytics.

Second, AI makes it possible to find things in your data that keyword search cannot. A customer types “I can’t get online” into your support system. A keyword search might miss that ticket unless the database contains exactly those words. Semantic AI search understands the meaning — it connects “can’t get online” to “internet connectivity issue,” “network not working,” and “WiFi not connecting” and surfaces all related tickets regardless of exact wording. This enables a quality of analysis that was simply not possible before.

A real-world example A retail company has five years of customer support tickets. Leadership wants to know what the most common unresolved frustrations are — not the ones that were escalated and labeled, but the underlying themes in free-text customer comments that never got formally categorized. Traditional analytics cannot answer this — there is no column to group by. AI analytics can: generate embedding vectors for every ticket, cluster similar ones together, and surface the dominant themes with representative examples. This analysis takes hours with AI-enabled tools. It was previously impractical.

2 What AI in Analytics Actually Means — Plain Language Everyone

When people say “AI analytics” they are usually describing one or more of these four things. Understanding which one you are dealing with matters because each has different implications for your data platform, your team, and your budget.

1. Natural Language to Data (Text-to-SQL)

A non-technical person types a question in plain English — “Show me the top ten customers by revenue this year” — and the system generates the correct query, runs it against your database, and returns the result as a chart or table. No SQL knowledge required. This is what Copilot in Power BI, Snowflake Cortex Analyst, and similar tools do. The business value is immediate: analysts and executives can explore data without waiting for a developer to write a report.

2. Semantic Search

Instead of finding records that match exact keywords, semantic search finds records that match the meaning of a query. Search for “affordable accommodation near the city center” and get results that include “budget hotel downtown,” “cheap apartment in the CBD,” and “low-cost hostel central location” — none of which contain your exact words. This is used in product catalogs, support knowledge bases, document retrieval, and any scenario where the data is text and exact matching is too rigid.

3. AI-Powered Classification and Enrichment

Run an AI model against your existing data to add new information automatically. Label customer support tickets as positive, negative, or neutral. Identify product categories from free-text descriptions. Extract entities from unstructured documents. Flag potentially fraudulent transactions. The AI reads each row, applies its understanding, and writes the result back — turning unstructured data into structured, queryable insights.

4. Retrieval Augmented Generation (RAG)

Connect a large language model to your organization’s private data so it can answer questions with accurate, current, specific information — not just general knowledge. Ask a chatbot “What is our return policy for electronics purchased online?” and it retrieves the relevant section of your actual policy document and uses it to generate an accurate answer. This is the foundation of enterprise AI assistants built on your own data.

3 The Core Concepts Explained Simply Everyone

These terms appear constantly in AI analytics conversations. Here they are in plain language — no mathematics required.

Vectors and Embeddings

An AI model can convert any piece of text, image, or data into a list of numbers that represents its meaning. That list of numbers is called an embedding (or a vector). Two pieces of text with similar meaning produce similar lists of numbers. Two pieces of text with unrelated meaning produce very different lists. The database stores these lists of numbers and can find similar ones very quickly — which is how semantic search works.

Think of it like a map. Every piece of content in your database gets a location on a very large map based on its meaning. When you search for something, you find your own location on that map and then find everything nearby — everything with a similar meaning — regardless of the exact words used.

Large Language Models (LLMs)

LLMs are AI systems trained on enormous amounts of text — billions of documents — that can understand and generate human language at a high level. GPT-4, Claude, Llama, and Gemini are examples. They can write, summarize, classify, explain, and answer questions. They do not have access to your private data unless you explicitly provide it to them as part of a query — which is what RAG does.

RAG — Retrieval Augmented Generation

RAG is the pattern that connects your organization’s data to an LLM. When someone asks a question, the system first searches your database for relevant information (retrieval), then provides that information to the LLM along with the question (augmentation), and the LLM uses it to generate an accurate answer (generation). The LLM does not need to have memorized your data — it just needs the relevant section retrieved and handed to it at the moment of the question.

The SQL Connection

All of the major data platforms are now building these AI capabilities directly into their database engines. This means your existing data — the data already in your SQL Server, Snowflake, BigQuery, or Redshift — can become the foundation of AI analytics without copying it to a separate AI system first. Your security policies, access controls, and governance rules all apply automatically because the AI is operating inside the database, not outside it.

4 How AI Changes What Reports and Analytics Can Do Everyone

Traditional analytics answers questions you knew to ask. You defined the dimensions, the metrics, and the filters ahead of time. An analyst or developer built the report. Leadership read the output. The gap between question and answer was measured in days or weeks.

AI-enabled analytics changes this in three practical ways that affect every role in the data team:

CapabilityTraditional AnalyticsAI-Enabled Analytics
Who can ask questions Anyone who can use a pre-built dashboard or knows SQL Anyone — natural language interfaces allow non-technical users to explore data directly
What data can be searched Structured data with defined columns and exact values Structured and unstructured — text, documents, support tickets, product descriptions, emails
How questions are matched to data Exact keyword or filter match Meaning-based match — finds conceptually related content even with different words
Speed from question to insight Hours to days for new questions not covered by existing reports Seconds for natural language queries against pre-built semantic layers
Unstructured data analysis Requires manual tagging, pre-defined categories, or export to specialist tools Classify, cluster, and summarize unstructured text directly in the database

What this means for the report team: AI does not replace analysts and report developers — it changes what they spend time on. Less time building one-off SQL queries for ad-hoc questions. More time building the semantic layers, AI models, and data pipelines that enable self-service analytics. The skill set expands; the volume of manual report-writing decreases.

5 The Major Platforms and What Each One Offers Leadership / Architects

Every major data platform has launched or is actively building AI analytics capabilities. The right choice for your organization depends on where your data already lives, what your team already knows, your cloud strategy, and your budget. There is no universally correct answer — but there are clear patterns that make one platform a better fit than another for specific situations.

Microsoft: SQL Server 2025, Azure SQL, Fabric

Best for organizations already on the Microsoft stack. Native AI in T-SQL, Power BI integration, zero-ETL Fabric Mirroring, Copilot in SSMS and Power BI.

Snowflake Cortex AI

Best for organizations already on Snowflake. AI functions in SQL, Cortex Analyst for natural language queries, fully managed — no infrastructure to run.

Databricks Mosaic AI

Best for data science-heavy teams, open-source preference, or organizations with complex ML pipelines. Unity Catalog governs AI models alongside data.

Google BigQuery ML and Vertex AI

Best for Google Cloud organizations. Create and run ML models in SQL, integrated with Vertex AI for managed model deployment and Gemini integration.

AWS Redshift ML and Bedrock

Best for AWS-primary organizations. Create ML models with SQL using Amazon SageMaker under the hood, Amazon Bedrock for LLM access.

Standalone Vector DBs: Pinecone, Weaviate, Milvus

Best when building custom AI applications that need purpose-built vector search. Higher operational overhead but maximum flexibility and performance.

6 Microsoft: SQL Server 2025, Azure SQL, and Fabric Everyone

Microsoft’s AI analytics story in 2026 is built around three connected products. SQL Server 2025 brings AI natively into the database engine for on-premises and cloud VM environments. Azure SQL Database is the fully managed cloud version with the same AI capabilities, always up to date. Microsoft Fabric is the unified analytics platform that sits above both — connecting operational data to analytics, AI, and Power BI in one environment.

What Is New in SQL Server 2025

SQL Server 2025 is the most significant SQL Server release in a decade. The headline addition is native AI built directly into T-SQL — not a bolt-on, not an external service, but SQL functions that generate embeddings, perform semantic search, and call AI models using the same language DBAs and developers already know. Key additions include a native VECTOR data type for storing AI-generated embeddings, VECTOR_DISTANCE for semantic similarity queries, CREATE EXTERNAL MODEL to connect SQL Server to Azure OpenAI or any compatible AI model, AI_GENERATE_EMBEDDINGS to generate embeddings directly in T-SQL, and Change Event Streaming to feed real-time data changes into analytics pipelines.

Microsoft Fabric

Microsoft Fabric is a single platform that brings together data engineering, data warehousing, real-time analytics, Power BI, and AI. It stores everything in OneLake — a shared data lake — so data does not have to be copied between services. Fabric Mirroring continuously replicates your SQL Server databases into Fabric within seconds, enabling near-real-time Power BI reports and AI analytics on operational data without ETL batch jobs. As of March 2026, Fabric serves more than 31,000 customers and is the fastest-growing data platform in Microsoft’s history.

Copilot Across the Microsoft Stack

Copilot is Microsoft’s AI assistant integration. In Power BI it generates reports and visuals from natural language descriptions. In SSMS 21 it generates T-SQL from plain English questions about your data. In the Azure Portal it helps manage and query Azure SQL databases. In Fabric it assists across all workloads — notebooks, pipelines, reports, and more.

Best fit for: Organizations already running SQL Server, using Azure, or invested in the Microsoft 365 and Power BI ecosystem. The integration between SQL Server data and Power BI analytics through Fabric is a genuine competitive advantage — no competitor offers the same depth of integration between the operational database and the analytics and reporting layer.

7 Snowflake Cortex AI Everyone

Snowflake Cortex AI is Snowflake’s fully managed AI and machine learning capability built directly into the Snowflake platform. It requires no infrastructure management, no model deployment, and no data movement — AI functions run against the data already in your Snowflake tables using standard SQL.

What Cortex AI Includes

Cortex LLM Functions — a set of SQL functions that call large language models directly from Snowflake queries. SNOWFLAKE.CORTEX.COMPLETE calls an LLM for text generation. SNOWFLAKE.CORTEX.SENTIMENT classifies text as positive, negative, or neutral. SNOWFLAKE.CORTEX.SUMMARIZE generates a concise summary of a long text field. SNOWFLAKE.CORTEX.TRANSLATE translates text between languages. SNOWFLAKE.CORTEX.EXTRACT_ANSWER answers a question based on a provided document. All of these run as SQL functions — no Python, no external API calls, no configuration.

Cortex Search — hybrid search that combines keyword and semantic search over your Snowflake data. Point it at a table, define which columns to search, and Cortex Search handles the embedding generation, indexing, and query execution automatically. Accessible via SQL or a REST API for application developers.

Cortex Analyst — a natural language interface for business analysts. Connect it to your Snowflake data and semantic model (a YAML file describing your business metrics and relationships), and users can ask questions in plain English through a chat interface. Cortex Analyst generates the SQL, runs it, and returns results. No SQL knowledge required from the end user.

Arctic and Partner Models — Snowflake hosts its own open-source LLM called Arctic, optimized for enterprise SQL generation. It also provides access to models from Mistral, Meta Llama, Google, and others through the Cortex functions, giving you model choice without managing infrastructure.

Best fit for: Organizations already on Snowflake, or those evaluating a cloud data warehouse without an existing Microsoft commitment. Cortex AI’s fully managed, zero-infrastructure approach makes it one of the fastest paths to production AI analytics. The natural language to SQL capability via Cortex Analyst is mature and well-suited for business analyst self-service.

8 Databricks with Unity Catalog and Mosaic AI Intermediate

Databricks is a data intelligence platform built on Apache Spark with strong open-source foundations. Its AI capabilities are grouped under the Mosaic AI brand, and its governance layer is Unity Catalog — a unified catalog that manages data, models, and AI assets in one place.

What Databricks Mosaic AI Includes

Model Serving — deploy and serve AI models at production scale, with auto-scaling and monitoring. Supports open-source models (Llama, Mistral) as well as external models like Azure OpenAI and Anthropic.

Vector Search — a managed vector database built into Databricks. Create vector indexes on Delta Lake tables, run similarity queries, and build RAG pipelines — all within the same platform as your data engineering and data science workloads.

AI Playground and Prompt Engineering — experiment with models and prompts directly in the Databricks UI without writing application code. Useful for rapid prototyping before committing to a production implementation.

Unity Catalog for AI Governance — models, feature tables, and AI assets are governed by the same Unity Catalog that manages data access, lineage, and auditing. This is Databricks’ key differentiator — unified governance across data and AI.

Best fit for: Data science-heavy organizations, teams with existing Spark expertise, or companies that want open-source model flexibility and strong ML lifecycle management. Databricks is the most powerful option for complex custom AI model development but has a steeper learning curve than the managed SQL-first approaches from Microsoft or Snowflake.

9 Google BigQuery ML and Vertex AI Intermediate

Google’s AI analytics strategy combines BigQuery ML — machine learning directly in SQL — with Vertex AI, Google’s managed machine learning platform, and Gemini, Google’s large language model family.

BigQuery ML

BigQuery ML allows you to create and run ML models using standard SQL CREATE MODEL statements directly in BigQuery. Supported model types include linear and logistic regression, k-means clustering, time series forecasting with ARIMA_PLUS, deep neural networks, and imported TensorFlow or scikit-learn models. You can also call Vertex AI hosted models and Google’s embedding APIs directly from BigQuery SQL using ML.GENERATE_TEXT and ML.GENERATE_EMBEDDING functions — similar in concept to SQL Server 2025’s AI_GENERATE_EMBEDDINGS.

Gemini in BigQuery and Looker

Google’s Gemini model is integrated across the analytics stack. In BigQuery it assists with query writing and explanation. In Looker (Google’s BI platform) it enables natural language questions against your semantic model — equivalent to Microsoft Copilot in Power BI or Snowflake Cortex Analyst for business users.

Best fit for: Organizations running on Google Cloud, particularly those with strong data science teams comfortable with Python and TensorFlow, or those that want tight integration between their analytics and Google’s AI research capabilities.

10 AWS: Redshift ML and Amazon Bedrock Intermediate

AWS provides AI analytics capabilities through Redshift ML for in-database machine learning and Amazon Bedrock for LLM access, with Amazon SageMaker as the underlying ML platform.

Redshift ML

Redshift ML uses CREATE MODEL SQL statements to train machine learning models on data in Amazon Redshift, using Amazon SageMaker Autopilot to handle the training and optimization automatically. Once trained, models are exposed as SQL functions you can call in queries. This follows the same pattern as BigQuery ML — ML in SQL — but is more tightly integrated with the broader AWS SageMaker ecosystem for teams that already use it.

Amazon Bedrock

Bedrock is AWS’s managed service for accessing large language models from multiple providers — Anthropic Claude, Meta Llama, Mistral, Amazon Titan, and others — through a single API. Data engineers and developers can call Bedrock models from applications, Lambda functions, or SageMaker notebooks to build RAG patterns on top of Redshift or RDS data.

Best fit for: Organizations with significant AWS infrastructure investment, particularly those already using SageMaker for ML workflows. The Redshift ML and Bedrock combination is powerful but requires more integration work compared to Microsoft’s or Snowflake’s more tightly integrated approaches.

11 Standalone Vector Databases — Pinecone, Weaviate, Milvus, pgvector Intermediate

Before the major database platforms added native vector capabilities, organizations building AI applications had to use purpose-built vector databases. These tools still have a role, though the case for using them standalone has narrowed significantly now that SQL Server 2025, Snowflake, BigQuery, and Databricks all support vectors natively.

The Main Options

Pinecone — fully managed cloud vector database. Extremely fast approximate nearest-neighbor search, no infrastructure to manage, simple API. Best for applications that need dedicated high-performance vector search and are willing to pay for a separate managed service. Does not store your relational data — you still need your primary database alongside it.

Weaviate — open-source vector database that can be self-hosted or used as a managed service. Supports hybrid search combining vector and keyword results. Includes built-in modules for automatic vectorization using popular embedding models.

Milvus — open-source vector database designed for very large scale — billions of vectors. More complex to operate than Pinecone but no per-query costs. Used by organizations that need extreme scale with full control over infrastructure.

pgvector — an open-source extension for PostgreSQL that adds vector storage and similarity search. If your organization runs PostgreSQL, pgvector adds vector search without a new database system. Performance at very large scale is more limited than dedicated vector databases but entirely adequate for most enterprise use cases.

The key question before choosing a standalone vector database: Does your primary data platform already support vectors natively? If you are on SQL Server 2025, Azure SQL, Snowflake, or BigQuery, a standalone vector database adds infrastructure, synchronization overhead, and a separate security boundary for what those platforms can now do internally. Standalone vector databases make the most sense for custom AI applications with extreme performance requirements or for organizations not yet on a platform with native vector support.

12 How to Choose — A Framework for Your Organization Leadership / Architects

Platform selection for AI analytics is not primarily a technical decision — it is an organizational one. The best platform is almost always the one your team already knows and your data already lives in. Migration cost, retraining cost, and integration complexity almost always outweigh the differences between mature platforms in the same tier.

If your organization…Start hereWhy
Runs SQL Server on-premises and uses Power BISQL Server 2025 + Microsoft FabricZero data movement, native T-SQL AI, Power BI Copilot — all in the stack you already own
Is already on Azure, Microsoft 365, or TeamsAzure SQL + Microsoft FabricCopilot everywhere, Entra ID integration, Fabric Mirroring from existing Azure SQL — fastest path
Is already on SnowflakeSnowflake Cortex AICortex LLM functions, Cortex Analyst, and Cortex Search require no new platform — run on your existing data
Has a strong data science team, uses SparkDatabricks Mosaic AIBest open-source model flexibility, Unity Catalog for governance, MLflow for experiment tracking
Is on Google Cloud with BigQueryBigQuery ML + Vertex AI + GeminiNative ML in SQL, Gemini integration, Looker for natural language analytics
Is AWS-primary with RedshiftRedshift ML + Amazon BedrockNative ML in SQL, Bedrock for LLM access, SageMaker for model development
Needs a custom AI application with extreme vector search scalePinecone or Milvus alongside primary DBPurpose-built vector performance where native DB vector support is not sufficient

Questions to Ask Before Committing

  • Where does the data already live? Moving data to a new platform to access AI features is rarely worth the cost if your existing platform offers comparable capabilities.
  • What does your team already know? A technically superior platform that your team cannot operate is not actually superior in practice.
  • What does leadership need from AI analytics — self-service reporting, semantic search, customer-facing AI features, or internal knowledge management? The answer changes the optimal tool.
  • What are the governance requirements? Regulated industries need to know where data goes when an AI model processes it and who has access to what. All the major platforms have answered this — but the answers differ.
  • What is the total cost? LLM calls are metered per token. Embedding generation has a cost. Vector index storage has a cost. Model the usage before committing to a cost estimate.

13 What Changes for Report Teams and Business Analysts Analysts / Report Teams

The most immediate impact of AI analytics for report teams and business analysts is the shift from building reports to enabling questions. Instead of a business user submitting a request and waiting for a developer to build a new query or dashboard, an AI-enabled analytics layer allows them to ask questions directly and get answers from the data in real time.

This does not eliminate the report developer’s role — it changes it. Someone still needs to build and maintain the semantic layer that defines what “revenue” means, what the right date hierarchies are, which tables relate to which, and what the business rules are. That semantic layer is what enables the natural language query tools to generate correct SQL rather than plausible-sounding but wrong SQL. Building and governing the semantic layer becomes one of the most valuable things a report team can do.

What Business Analysts Gain

  • Ability to ask new ad-hoc questions without waiting for a developer to build a query
  • Semantic search over unstructured data — product descriptions, customer comments, support tickets — that keyword search could not previously reach
  • Automatic classification and tagging of unstructured content, reducing manual categorization work
  • Summarization of large text datasets — “What are the five most common themes in this year’s customer feedback?” answered in seconds

What Business Analysts Need to Understand

  • Natural language query tools generate SQL — they do not always generate correct SQL. Results should be validated, especially for financial or compliance reporting.
  • AI-generated answers are only as good as the data and the semantic model behind them. Garbage in, garbage out still applies.
  • Semantic search returns relevant results, not exact results. For compliance or audit scenarios where you need complete and exact records, traditional structured queries are still appropriate.

14 Copilot in Power BI and Fabric — Natural Language to Reports Analysts / Report Teams

Microsoft Copilot is integrated across Power BI and Microsoft Fabric, providing AI assistance at every step of the analytics workflow. For business analysts and report teams, the most immediately useful capabilities are natural language report generation and the ability to ask questions about data without writing DAX or SQL.

What Copilot in Power BI Can Do

  • Generate report pages from a description. Type “Create a sales performance report showing revenue by region, product category, and month with a comparison to last year” and Copilot builds the page with appropriate visuals, selecting measures and dimensions from your semantic model.
  • Answer questions about your data in natural language. Ask “Which product category had the highest growth rate in Q1 2026?” and Copilot queries your model and returns an answer with the supporting visual.
  • Summarize report insights. Generate a plain-language executive summary of what the data in a report page is showing — useful for stakeholder communications where not everyone will read the full dashboard.
  • Explain visuals. Ask “Why is this bar chart showing a dip in February?” and Copilot analyzes the data to provide a narrative explanation.

Copilot in Power BI works best with a well-built semantic model. It uses your semantic model’s field names, measure definitions, and relationships to generate queries. Poorly named fields, missing descriptions, and ambiguous relationships all reduce the quality of Copilot’s output. Investing in a clean, well-documented semantic model is the single most impactful thing a report team can do to make Copilot useful.

15 Snowflake Cortex Analyst — Asking Questions in Plain English Analysts / Report Teams

Snowflake Cortex Analyst is Snowflake’s natural language to SQL interface for business users. You provide a semantic model — a YAML file that describes your business metrics, dimensions, and the relationships between tables — and Cortex Analyst uses it to convert plain English questions into accurate SQL queries against your Snowflake data.

The semantic model definition is the critical piece. A good semantic model defines revenue with the exact calculation your business uses, specifies that “customer” and “client” mean the same thing, and explains that orders are tied to customers through the order_customer_id foreign key. With that context, Cortex Analyst can answer “What was our revenue last quarter broken down by customer segment?” with a query that will return the right answer.

# Example Snowflake Cortex Analyst semantic model (simplified YAML)
# This is what a report developer creates to enable natural language queries

name: sales_analytics
description: Sales performance metrics for business analysis

tables:
  - name: fact_orders
    description: All customer orders
    base_table: SALES_DW.PUBLIC.FACT_ORDERS
    measures:
      - name: total_revenue
        description: Total order revenue after discounts
        expr: SUM(order_amount - discount_amount)
        data_type: number
      - name: order_count
        description: Number of orders placed
        expr: COUNT(DISTINCT order_id)
        data_type: number
    dimensions:
      - name: order_date
        description: Date the order was placed
        expr: order_date
        data_type: date
      - name: customer_segment
        description: Customer tier - Enterprise, Mid-Market, or SMB
        expr: c.segment
        data_type: text

# Once this model is deployed, business users can ask:
# "What was our total revenue last quarter by customer segment?"
# "Which product had the highest order count in January?"
# "Show me month-over-month revenue growth for the past year"
# Cortex Analyst generates the SQL, runs it, returns results

16 Microsoft Fabric Mirroring and Real-Time Analytics DBAs / Data Engineers

Microsoft Fabric Mirroring solves one of the longest-standing problems in analytics — keeping your reporting and analytics environment current with operational data without complex ETL pipelines and batch windows. It continuously replicates changes from your SQL Server 2025 databases into Fabric’s OneLake within seconds using the transaction log change feed.

Once data is in OneLake, every Fabric workload can use it: Power BI Direct Lake reports that reflect near-real-time operational state, Spark notebooks for data engineering and ML, the Fabric Data Warehouse for SQL analytics, and AI workloads including vector search and RAG patterns. All without copying data between services — OneLake is the single store that all workloads share.

Change Event Streaming for Real-Time AI Pipelines

Change Event Streaming (CES) is a newer capability that goes further than Mirroring for event-driven architectures. CES streams individual row-level changes from the SQL Server transaction log to Azure Event Hubs or Fabric Eventstream as they happen. This enables architectures where a product catalog change triggers an immediate re-embedding of the new description, updating the semantic search index in real time — no batch job, no overnight delay.

-- Confirm SQL Server 2025 is Arc-enabled (required for Fabric Mirroring on-premises)
-- Run on the SQL Server instance
SELECT SERVERPROPERTY('IsArcEnabled') AS [Arc Enabled];

-- Check recovery model (required: FULL)
SELECT name, recovery_model_desc
FROM sys.databases
WHERE name = 'YourDatabase';

-- Fabric Mirroring is configured in the Fabric portal, not in SQL Server T-SQL
-- Once mirrored, connect SSMS to the Fabric SQL Analytics Endpoint:
-- Server: yourworkspace-xxxx.sql.fabric.microsoft.com
-- Database: YourDatabase (the mirrored copy in Fabric)
-- Then query as normal T-SQL for analytics workloads

17 SQL Server 2025: Vector Search in T-SQL — Step by Step Advanced / DBAs

This is the complete implementation for semantic search using SQL Server 2025’s native vector capabilities. The pattern works for product catalogs, knowledge bases, support articles, or any text data where meaning-based search is more useful than keyword matching.

-- ============================================================
-- End-to-end semantic search in SQL Server 2025
-- Prerequisites: SQL Server 2025, Azure OpenAI resource
-- ============================================================

-- Step 1: Enable preview features (for DiskANN vector index)
ALTER DATABASE SCOPED CONFIGURATION SET PREVIEW_FEATURES = ON;
GO

-- Step 2: Create credential and external model
CREATE MASTER KEY ENCRYPTION BY PASSWORD = 'YourStrongPassword!';
GO

CREATE DATABASE SCOPED CREDENTIAL [AzureOpenAICred]
WITH IDENTITY = 'HTTPEndpointHeaders',
SECRET = '{"api-key": "your-azure-openai-api-key"}';
GO

CREATE EXTERNAL MODEL [EmbeddingModel]
WITH (
    LOCATION   = 'https://your-resource.openai.azure.com/openai/deployments/text-embedding-ada-002',
    API_FORMAT = 'Azure_OpenAI',
    MODEL_TYPE = EMBEDDINGS,
    CREDENTIAL = [AzureOpenAICred]
);
GO

-- Step 3: Create table with VECTOR column
-- Dimension 1536 matches text-embedding-ada-002 output
CREATE TABLE dbo.KnowledgeBase (
    ArticleId   INT IDENTITY PRIMARY KEY,
    Title       NVARCHAR(200) NOT NULL,
    Content     NVARCHAR(MAX),
    Category    NVARCHAR(50),
    IsPublished BIT DEFAULT 1,
    ContentVec  VECTOR(1536)
);
GO

-- Step 4: Insert content (embeddings generated in next step)
INSERT INTO dbo.KnowledgeBase (Title, Content, Category)
VALUES
('How to reset your password', 'To reset your password click Forgot Password on the login page...', 'Account'),
('Wifi connection troubleshooting', 'If you cannot connect to wifi check that your router is powered on...', 'Network'),
('Return and refund policy', 'Items may be returned within 30 days of purchase with original receipt...', 'Policy'),
('Setting up two-factor authentication', 'Two-factor authentication adds an extra layer of security to your account...', 'Security');
GO

-- Step 5: Generate embeddings for all articles
UPDATE dbo.KnowledgeBase
SET ContentVec = AI_GENERATE_EMBEDDINGS(Content USE MODEL EmbeddingModel)
WHERE ContentVec IS NULL;
GO

-- Step 6: Create DiskANN vector index for fast approximate search
CREATE VECTOR INDEX IX_KB_ContentVec
ON dbo.KnowledgeBase (ContentVec)
WITH (METRIC = 'cosine', TYPE = 'diskann');
GO

-- Step 7: Semantic search -- find articles relevant to a user query
-- The query "I forgot how to log in" should surface password reset article
-- even though it does not contain those exact words
DECLARE @UserQuery VECTOR(1536) =
    AI_GENERATE_EMBEDDINGS(N'I forgot how to log in' USE MODEL EmbeddingModel);

SELECT TOP(3)
    ArticleId, Title, Category,
    VECTOR_DISTANCE('cosine', @UserQuery, ContentVec) AS Relevance
FROM dbo.KnowledgeBase
WHERE IsPublished = 1
ORDER BY Relevance ASC;
-- Expected: password reset article ranks #1 despite no keyword overlap

-- Step 8: Hybrid search -- semantic + relational filters
DECLARE @NetQuery VECTOR(1536) =
    AI_GENERATE_EMBEDDINGS(N'cannot get online wireless' USE MODEL EmbeddingModel);

SELECT TOP(3)
    ArticleId, Title, Category,
    VECTOR_DISTANCE('cosine', @NetQuery, ContentVec) AS Relevance
FROM dbo.KnowledgeBase
WHERE Category = 'Network'           -- relational filter
AND IsPublished = 1
ORDER BY Relevance ASC;

18 Snowflake Cortex AI in SQL — Step by Step Advanced / Data Engineers

Snowflake Cortex AI functions run as standard SQL — no infrastructure to configure, no API keys to manage. The functions call Snowflake-hosted models with the costs billed to your Snowflake account per token used.

-- ============================================================
-- Snowflake Cortex AI functions -- run in Snowflake SQL
-- No setup required beyond standard Snowflake access
-- ============================================================

-- Sentiment analysis on customer reviews
SELECT
    review_id,
    customer_name,
    review_text,
    SNOWFLAKE.CORTEX.SENTIMENT(review_text)       AS sentiment_score,
    -- Returns -1 (very negative) to 1 (very positive)
    CASE
        WHEN SNOWFLAKE.CORTEX.SENTIMENT(review_text) > 0.3  THEN 'Positive'
        WHEN SNOWFLAKE.CORTEX.SENTIMENT(review_text) < -0.3 THEN 'Negative'
        ELSE 'Neutral'
    END                                           AS sentiment_label
FROM customer_reviews
WHERE review_date >= DATEADD('day', -30, CURRENT_DATE());

-- Summarize long customer support tickets
SELECT
    ticket_id,
    SNOWFLAKE.CORTEX.SUMMARIZE(ticket_description) AS summary
FROM support_tickets
WHERE LENGTH(ticket_description) > 500
AND created_date >= DATEADD('day', -7, CURRENT_DATE());

-- Extract answers from documents using LLM
SELECT
    doc_id,
    SNOWFLAKE.CORTEX.EXTRACT_ANSWER(
        document_text,
        'What is the warranty period for this product?'
    ) AS warranty_answer
FROM product_documents
WHERE product_category = 'Electronics';

-- Generate embeddings for semantic search (Cortex Search handles this automatically
-- but you can also generate manually for custom use cases)
SELECT
    article_id,
    SNOWFLAKE.CORTEX.EMBED_TEXT_1024(
        'snowflake-arctic-embed-m',   -- Snowflake's own embedding model
        article_content
    ) AS content_embedding
FROM knowledge_base_articles;

-- Translate content for multilingual analytics
SELECT
    feedback_id,
    original_language,
    feedback_text,
    SNOWFLAKE.CORTEX.TRANSLATE(
        feedback_text,
        'auto',   -- detect source language automatically
        'en'      -- translate to English
    ) AS feedback_english
FROM customer_feedback
WHERE original_language != 'en';

-- Use COMPLETE for custom prompts and classification
SELECT
    product_id,
    description,
    SNOWFLAKE.CORTEX.COMPLETE(
        'mistral-7b',    -- model choice
        CONCAT(
            'Classify this product description into one of these categories: ',
            'Electronics, Clothing, Home & Garden, Sports, Food & Beverage. ',
            'Return only the category name. Description: ',
            description
        )
    ) AS predicted_category
FROM unclassified_products
LIMIT 100;

19 Building a RAG Pattern — Grounding AI in Your Own Data Advanced / Data Engineers

RAG is the architecture that makes enterprise AI assistants actually useful. Without it, an LLM answers questions based only on its training data — it knows nothing about your organization’s policies, products, or procedures. With RAG, every answer is grounded in your actual data, retrieved at query time from your database.

The pattern is the same regardless of platform: embed your knowledge base, retrieve the most relevant context for each question using vector search, provide that context to the LLM alongside the question, get a grounded answer back. The implementation differs by platform.

-- ============================================================
-- RAG implementation in SQL Server 2025 T-SQL
-- A stored procedure that answers questions using your knowledge base
-- ============================================================
CREATE OR ALTER PROCEDURE dbo.AnswerFromKnowledgeBase
    @Question NVARCHAR(2000),
    @Answer   NVARCHAR(MAX) OUTPUT
AS
BEGIN
    SET NOCOUNT ON;

    -- Step 1: Embed the user's question
    DECLARE @QuestionVec VECTOR(1536) =
        AI_GENERATE_EMBEDDINGS(@Question USE MODEL EmbeddingModel);

    -- Step 2: Retrieve top 5 most relevant articles
    DECLARE @Context NVARCHAR(MAX) = '';
    SELECT @Context = @Context + CHAR(13) + Title + ': ' + Content + CHAR(13)
    FROM (
        SELECT TOP(5) Title, Content,
               VECTOR_DISTANCE('cosine', @QuestionVec, ContentVec) AS Dist
        FROM dbo.KnowledgeBase
        WHERE IsPublished = 1
        ORDER BY Dist ASC
    ) AS ctx;

    -- Step 3: Call GPT-4 with the retrieved context
    DECLARE @Payload NVARCHAR(MAX) = CONCAT(
        N'{"messages":[',
        N'{"role":"system","content":"Answer using only the provided context. If the answer is not in the context say so."},',
        N'{"role":"user","content":"Context:\n', REPLACE(@Context, '"', '\"'),
        N'\n\nQuestion: ', REPLACE(@Question, '"', '\"'), N'"}',
        N'],"max_tokens":400}'
    );

    DECLARE @Response NVARCHAR(MAX);
    DECLARE @Status   INT;

    EXEC sp_invoke_external_rest_endpoint
        @url         = N'https://your-resource.openai.azure.com/openai/deployments/gpt-4/chat/completions?api-version=2024-02-01',
        @method      = 'POST',
        @headers     = N'{"api-key":"your-key","Content-Type":"application/json"}',
        @payload     = @Payload,
        @response    = @Response OUTPUT,
        @status_code = @Status OUTPUT;

    SET @Answer = JSON_VALUE(@Response, '$.choices[0].message.content');
END;
GO

-- Test the RAG procedure
DECLARE @Result NVARCHAR(MAX);
EXEC dbo.AnswerFromKnowledgeBase
    @Question = 'How do I get a refund for a product I bought last week?',
    @Answer   = @Result OUTPUT;
SELECT @Result AS [Answer];
-- ============================================================
-- RAG pattern in Snowflake using Cortex Search + COMPLETE
-- Simpler implementation -- Cortex Search handles embedding automatically
-- ============================================================

-- First: create a Cortex Search service on your knowledge base table
-- (done once, in Snowflake UI or SQL)
CREATE OR REPLACE CORTEX SEARCH SERVICE knowledge_search
    ON content                          -- column to search
    ATTRIBUTES title, category          -- filterable metadata
    WAREHOUSE = compute_wh
    TARGET_LAG = '1 minute'            -- how often to refresh the index
AS (
    SELECT title, content, category
    FROM knowledge_base_articles
    WHERE is_published = TRUE
);

-- Then: build a RAG query using Cortex Search for retrieval
-- and COMPLETE for generation
SELECT
    SNOWFLAKE.CORTEX.COMPLETE(
        'mistral-large',
        CONCAT(
            'Answer this question using ONLY the context below. ',
            'If the answer is not in the context say you do not know.',
            CHR(10), 'Context:', CHR(10),
            -- Cortex Search returns JSON with matching results
            -- In a real implementation this is called via the Search API
            -- and the results are passed here as context
            retrieved_context,
            CHR(10), 'Question: ', @user_question
        )
    ) AS ai_answer
FROM retrieved_context_table;

20 What Is Coming Next Across All Platforms Everyone

The pace of change in AI analytics is faster than any other area of the data platform landscape. Here is what is confirmed or strongly indicated across the major platforms heading into the rest of 2026:

Microsoft

  • VECTOR_SEARCH and DiskANN vector index moving from preview to GA in a near-term SQL Server 2025 cumulative update
  • SSMS 22 replacing Azure OpenAI endpoint requirement with GitHub Copilot account for Copilot in SSMS
  • Fabric migration assistant for SQL Server — AI-assisted migration from on-premises SQL Server into Microsoft Fabric
  • dbt Fusion support in Fabric expected Q2 2026
  • Fabric IQ — natural language analytics across all Fabric workloads powered by semantic models
  • GPU acceleration for AI inference workloads in SQL Server 2025 on supported hardware

Snowflake

  • Cortex Analyst expanding model choice and semantic model capabilities
  • Snowflake Arctic model family updates for improved SQL generation accuracy
  • Deeper integration between Cortex Search and Cortex Analyst for end-to-end RAG pipelines in SQL

Industry-Wide

  • Every major platform is investing heavily in natural language to SQL — the quality of AI-generated queries will continue to improve as semantic models become richer and models become more capable
  • Agentic AI — AI systems that take multi-step actions against databases rather than answering a single question — is the next frontier. Microsoft, Snowflake, and Databricks are all building agentic capabilities into their platforms
  • Multimodal data — queries that combine text, images, and structured data in a single search — is moving from research into early production availability
  • Open-source model quality continues to close the gap with commercial models. The ability to run capable AI models on-premises using Ollama or similar tools matters significantly for regulated industries and organizations with strict data residency requirements

The practical takeaway for every role: AI analytics is not a future investment — it is available today on every major platform, at varying levels of maturity. The organizations that start building with it now — learning what works, building the semantic layers and data pipelines that enable it, and developing the governance practices that make it trustworthy — will have a substantial advantage over those waiting for the technology to mature further. The technology is mature enough. The organizational readiness is the variable to work on.

References


Discover more from SQLYARD

Subscribe to get the latest posts sent to your email.

Leave a Reply

Discover more from SQLYARD

Subscribe now to keep reading and get access to the full archive.

Continue reading