This post is a part of the DP-900: Microsoft Azure Data Fundamentals Exam Prep Hub.
This topic falls under these sections:
Describe core data concepts (25–30%)
--> Describe common data workloads
--> Describe features of analytical workloads
Note that there are 10 practice questions (with answers and explanations) for each section to help you solidify your knowledge of the material. Also, there are 2 practice tests with 60 questions each available on the hub below the exam topics section.
Analytical workloads are essential for deriving insights from data. Unlike transactional workloads — which support day-to-day operations — analytical workloads focus on querying, aggregating, summarizing, and analyzing large volumes of data to help with reporting, decision making, and trends.
What Is an Analytical Workload?
An analytical workload refers to data processing that is oriented toward analysis, rather than operational updates. It is optimized for:
- Complex queries
- Aggregations across large datasets
- Historical analysis and reporting
- Business intelligence (BI)
Analytical workloads are often associated with OLAP (Online Analytical Processing) systems.
Key Features of Analytical Workloads
1. Large Volumes of Data
Analytical systems often operate on datasets that are:
- Much larger than transactional tables
- Historical — spanning months or years of records
- Combined from multiple sources (e.g., transactional systems, logs, external data)
These datasets can be stored in data warehouses, data lakes, or big data systems.
2. Complex, Read-Heavy Queries
Analytical workloads are dominated by complex SELECT queries, often involving:
- Aggregations (SUM, AVG, COUNT)
- Grouping by categories
- Filtering on multiple dimensions
- Joining large tables
These queries can be computationally intensive and are often used for reporting and dashboards.
3. Denormalized or Columnar Storage
Unlike transactional systems that use normalized schemas, analytical workloads often use:
- Denormalized schemas (e.g., star or snowflake schemas)
- Columnar storage formats (e.g., Parquet, ORC)
These formats improve query performance by minimizing I/O and enabling efficient aggregation.
4. Longer Query Response Times (But High Throughput)
Queries in analytical systems are not always expected to return results in milliseconds, as they:
- Scan large amounts of data
- Compute aggregates and summaries
- May be optimized for throughput rather than low latency
This contrasts with transactional systems where fast, small transactions are critical.
5. Batch or Bulk Processing
Analytical workloads often rely on:
- Batch ingestion of data (e.g., nightly ETL jobs)
- Data transformation pipelines (cleaning, aggregating, enriching)
- Tools like Azure Data Factory, Databricks, or Synapse pipelines
These pipelines prepare data for analytics and reporting.
6. Support for BI and Reporting Tools
Analytical workloads integrate with business intelligence tools, such as:
- Power BI
- Excel
- Azure Synapse Analytics Studio
These tools connect directly to analytical stores to produce dashboards, charts, and insights.
Analytical vs Transactional Workloads — Quick Comparison
| Feature | Transactional | Analytical |
|---|---|---|
| Primary Purpose | Operational processing (OLTP) | Decision support & reporting (OLAP) |
| Data Size | Small to moderate | Large or very large |
| Workload Type | Frequent inserts/updates/deletes | Complex queries/aggregations |
| Schema | Normalized | Often denormalized |
| Query Focus | Single record operations | Scanning many records |
| Typical Tools | Relational OLTP databases | Data warehouses, big data systems |
Where Analytical Workloads Run in Azure
Azure offers several services optimized for analytical workloads:
Azure Synapse Analytics
A unified analytics service that enables:
- Data warehousing
- Big data processing
- Integration with Spark and SQL
- High-performance analytics
It is ideal for large-scale reporting and BI scenarios.
Azure Data Lake Storage + Analytics
Azure Data Lake Storage Gen2 works with:
- Apache Spark
- Azure Databricks
- Synapse Analytics
This combination supports big data analytics, machine learning, and data science workloads.
Azure SQL Data Warehouse (Synapse Dedicated SQL Pools)
This is the former SQL DW offering (now part of Synapse) optimized for:
- Massive parallel processing
- Distributed query execution
- High-volume analytical queries
Why Analytical Workloads Matter for DP-900
For DP-900, you should be able to:
- Define analytical workloads and distinguish them from transactional workloads
- Recognize use cases where analytical workloads are appropriate
- Identify Azure services designed for analytical processing
- Understand schema design and storage options that support analytics
Being able to describe these features shows your understanding of how modern data ecosystems support business intelligence and analytics.
Summary — Exam-Relevant Takeaways
✔ Analytical workloads focus on complex queries and analysis across large datasets
✔ They use denormalized schemas and columnar storage to boost performance
✔ They are optimized for throughput and summarization, not real-time transactions
✔ They typically support reports, dashboards, and insights
✔ Azure services like Azure Synapse Analytics, Azure Data Lake, and Databricks support these workloads
Go to the Practice Exam Questions for this topic.
Go to the DP-900 Exam Prep Hub main page.
