Category: Microsoft Certification

Practice Questions: Identify appropriate visualizations for data (DP-900 Exam Prep)

Practice Questions


Question 1

Which visualization is BEST for showing sales trends over the past 12 months?

A. Pie chart
B. Scatter chart
C. Line chart
D. Gauge chart

Answer: C

Explanation:
Line charts are ideal for displaying trends over time.


Question 2

Which visualization is MOST appropriate for comparing revenue across different product categories?

A. Map
B. Bar chart
C. Card
D. Gauge chart

Answer: B

Explanation:
Bar charts are commonly used for comparing values across categories.


Question 3

Which visualization is BEST suited for displaying a single metric such as total profit?

A. Matrix
B. Pie chart
C. Card
D. Scatter chart

Answer: C

Explanation:
Cards are used to display a single important value or KPI.


Question 4

Which visualization would be MOST useful for displaying sales performance by country?

A. Map
B. Gauge chart
C. Pie chart
D. Card

Answer: A

Explanation:
Maps are designed for visualizing geographic or location-based data.


Question 5

Which visualization is BEST for showing the relationship between advertising spend and sales revenue?

A. Line chart
B. Scatter chart
C. Pie chart
D. Matrix

Answer: B

Explanation:
Scatter charts help identify relationships and correlations between numeric variables.


Question 6

A report needs to show how each region contributes to total company sales. Which visualization is MOST appropriate?

A. Gauge chart
B. Pie chart
C. Table
D. Scatter chart

Answer: B

Explanation:
Pie charts are commonly used to show proportions or percentages of a whole.


Question 7

Which Power BI visualization is MOST appropriate for displaying detailed transactional records?

A. Table
B. Card
C. Gauge chart
D. Pie chart

Answer: A

Explanation:
Tables display detailed row-level data effectively.


Question 8

Which visualization is BEST for monitoring progress toward a sales target?

A. Scatter chart
B. Matrix
C. Gauge chart
D. Pie chart

Answer: C

Explanation:
Gauge charts show progress toward a goal or threshold.


Question 9

Which Power BI feature allows users to interactively filter report visuals?

A. Relationships
B. Measures
C. Slicers
D. Cards

Answer: C

Explanation:
Slicers provide interactive filtering capabilities in reports.


Question 10

Which visualization is MOST appropriate for summarized cross-tab reporting with grouped data?

A. Matrix
B. Card
C. Pie chart
D. Gauge chart

Answer: A

Explanation:
Matrices are useful for grouped summaries and pivot-style analysis.


✅ Quick Exam Takeaways

✔ Visualization selection matters

✔ Common visualizations:

  • Line chart → trends over time
  • Bar/Column chart → comparisons
  • Pie/Donut chart → proportions
  • Scatter chart → relationships/correlation
  • Map → geographic data
  • Card → single metric
  • Gauge → progress toward target
  • Table → detailed records
  • Matrix → grouped summaries

✔ Interactive feature:

  • Slicers → filtering

✔ Exam tips:
👉 Trend over time = line chart
👉 Category comparison = bar chart
👉 Relationship between values = scatter chart
👉 Single KPI = card
👉 Geographic data = map


Go to the DP-900 Exam Prep Hub main page.

Describe features of data models in Power BI (DP-900 Exam Prep)

This post is a part of the DP-900: Microsoft Azure Data Fundamentals Exam Prep Hub. 
This topic falls under these sections:
Describe an analytics workload (25–30%)
--> Describe data visualization in Microsoft Power BI
--> Describe features of data models in Power BI


Note that there are 10 practice questions (with answers and explanations) for each section to help you solidify your knowledge of the material. Also, there are 2 practice tests with 60 questions each available on the hub below the exam topics section.

A data model is the foundation of any effective report in Microsoft Power BI. It defines how data is structured, related, and calculated, enabling efficient analysis and meaningful visualizations.

For the DP-900 exam, you should understand how data models work, their key components, and best practices.


What Is a Data Model in Power BI?

A data model is a logical representation of data that includes:

  • Tables
  • Relationships
  • Calculations

It allows Power BI to:

  • Combine data from multiple sources
  • Enable filtering and aggregation
  • Support interactive reporting

Key Features of Power BI Data Models


1. Tables

Data models consist of one or more tables, which can come from:

  • Databases
  • Files (Excel, CSV)
  • Cloud sources

✔ Tables contain rows (records) and columns (fields)


2. Relationships

Relationships define how tables are connected.

Types of Relationships

  • One-to-many (1:*) → Most common
  • Many-to-one (*:1)
  • Many-to-many (:)

Key Concepts

  • Primary key → Unique identifier in one table
  • Foreign key → Reference in another table

✔ Relationships enable filtering across tables


3. Schema Design (Star Schema)

Power BI models commonly follow a star schema:

  • Fact tables → Contain measurable data (e.g., sales)
  • Dimension tables → Contain descriptive data (e.g., customer, product)

✔ This structure improves performance and usability


4. Measures and Calculated Columns

Power BI uses DAX (Data Analysis Expressions) for calculations.

Measures

  • Calculated at query time
  • Used in aggregations (e.g., SUM, AVERAGE)

Calculated Columns

  • Computed during data load
  • Stored in the model

✔ Measures are preferred for performance


5. Data Types

Each column has a defined data type:

  • Text
  • Number
  • Date/Time
  • Boolean

✔ Correct data types ensure accurate calculations and visuals


6. Hierarchies

Hierarchies allow users to drill down into data.

Example

  • Year → Quarter → Month → Day

✔ Used for interactive reporting and exploration


7. Filtering and Cross-Filtering

Relationships enable:

  • Filter propagation between tables
  • Cross-filtering in visuals

✔ Example:
Selecting a product filters related sales data


8. Data Granularity

Granularity refers to the level of detail in data.

  • Fine-grained → detailed (e.g., individual transactions)
  • Coarse-grained → summarized (e.g., monthly totals)

✔ Consistent granularity is important for accurate analysis


9. Model Optimization

Well-designed models:

  • Use fewer tables when possible
  • Avoid unnecessary columns
  • Use measures instead of calculated columns
  • Follow star schema design

✔ Improves performance and usability


10. Relationships Direction (Filter Direction)

Relationships can filter:

  • Single direction (default, recommended)
  • Both directions (used cautiously)

✔ Incorrect settings can lead to ambiguous results


Typical Data Modeling Workflow in Power BI

  1. Load data into Power BI
  2. Clean and transform data (Power Query)
  3. Define relationships
  4. Create measures and calculations
  5. Build reports and visuals

Why This Matters for DP-900

On the exam, you may be asked to:

  • Identify components of a data model
  • Understand relationships and keys
  • Differentiate between measures and calculated columns
  • Recognize star schema design
  • Understand filtering behavior

Summary — Exam-Relevant Takeaways

✔ A data model includes:

  • Tables
  • Relationships
  • Calculations

✔ Key features:

  • Relationships (1:*, :)
  • Star schema (fact + dimension tables)
  • Measures vs calculated columns
  • Hierarchies and filtering

✔ Best practices:

  • Use star schema
  • Prefer measures over calculated columns
  • Maintain consistent granularity

✔ Exam tips:
👉 Fact table = metrics (numbers)
👉 Dimension table = descriptive attributes
👉 Measure = dynamic calculation
👉 Calculated column = stored value


Go to the Practice Exam Questions for this topic.

Go to the DP-900 Exam Prep Hub main page.

Practice Questions: Describe features of data models in Power BI (DP-900 Exam Prep)

Practice Questions


Question 1

What is the primary purpose of a data model in Power BI?

A. To store raw files
B. To define relationships and enable data analysis
C. To manage network connections
D. To create dashboards only

Answer: B

Explanation:
A data model organizes data and defines relationships and calculations for analysis.


Question 2

Which component connects tables together in a Power BI data model?

A. Measures
B. Relationships
C. Dashboards
D. Queries

Answer: B

Explanation:
Relationships define how tables interact and allow filtering across them.


Question 3

Which type of relationship is MOST common in Power BI models?

A. Many-to-many
B. One-to-many
C. One-to-one
D. No relationship

Answer: B

Explanation:
The one-to-many (1:*) relationship is the most common in analytical models.


Question 4

In a star schema, which table typically contains numeric values used for analysis?

A. Dimension table
B. Lookup table
C. Fact table
D. Bridge table

Answer: C

Explanation:
Fact tables store measurable data (e.g., sales, revenue).


Question 5

What is the role of a dimension table in a data model?

A. Store raw transaction data
B. Store aggregated values only
C. Provide descriptive attributes for filtering and grouping
D. Execute calculations

Answer: C

Explanation:
Dimension tables contain descriptive data like customer or product details.


Question 6

Which type of calculation is evaluated at query time in Power BI?

A. Calculated column
B. Measure
C. Table relationship
D. Data type

Answer: B

Explanation:
Measures are calculated dynamically during query execution.


Question 7

Which language is used to create measures and calculated columns in Power BI?

A. SQL
B. Python
C. DAX
D. Java

Answer: C

Explanation:
DAX (Data Analysis Expressions) is used for calculations in Power BI.


Question 8

What is the benefit of using a star schema in Power BI?

A. Increased data duplication
B. Simplified relationships and improved performance
C. Elimination of fact tables
D. Reduced data types

Answer: B

Explanation:
Star schema improves performance and usability by simplifying relationships.


Question 9

What happens when you create a relationship between two tables?

A. Data is duplicated
B. Tables are merged into one
C. Filters can propagate between tables
D. Data types are changed

Answer: C

Explanation:
Relationships allow filtering across related tables.


Question 10

Which feature allows users to drill down through levels such as Year → Month → Day?

A. Measures
B. Hierarchies
C. Relationships
D. Dashboards

Answer: B

Explanation:
Hierarchies enable drill-down analysis in reports.


✅ Quick Exam Takeaways

✔ Data model components:

  • Tables
  • Relationships
  • Measures & calculated columns

✔ Key concepts:

  • Fact table → numeric data
  • Dimension table → descriptive data
  • Relationships → connect tables

✔ Calculations:

  • Measures → dynamic
  • Calculated columns → stored

✔ Design best practice:

  • Use star schema

✔ Exam tip:
👉 Measure = calculated at query time
👉 Calculated column = stored in table
👉 Fact = numbers, Dimension = descriptions


Go to the DP-900 Exam Prep Hub main page.

Identify capabilities of Power BI (DP-900 Exam Prep)

This post is a part of the DP-900: Microsoft Azure Data Fundamentals Exam Prep Hub. 
This topic falls under these sections:
Describe an analytics workload (25–30%)
--> Describe data visualization in Microsoft Power BI
--> Identify capabilities of Power BI


Note that there are 10 practice questions (with answers and explanations) for each section to help you solidify your knowledge of the material. Also, there are 2 practice tests with 60 questions each available on the hub below the exam topics section.

Microsoft Power BI is Microsoft’s business intelligence (BI) and data visualization platform. It enables users to connect to data, transform it, and create interactive reports and dashboards for data-driven decision-making.

For the DP-900 exam, you should understand what Power BI can do, its core components, and its role in an analytics solution.


What Is Power BI?

Power BI is a self-service and enterprise BI tool that allows users to:

  • Connect to multiple data sources
  • Transform and model data
  • Create visualizations and reports
  • Share insights across an organization

Core Capabilities of Power BI


1. Data Connectivity

Power BI can connect to a wide range of data sources:

  • Cloud services (Azure, SaaS apps)
  • Databases (SQL Server, Azure SQL)
  • Files (Excel, CSV)
  • Streaming data sources

✔ Supports both import and direct query modes


2. Data Transformation (Power Query)

Power BI includes Power Query, a tool for:

  • Cleaning data
  • Shaping and transforming data
  • Merging and filtering datasets

✔ Uses a visual interface (no coding required, though M language is available)


3. Data Modeling

Power BI enables users to create data models by:

  • Defining relationships between tables
  • Creating calculated columns and measures
  • Using DAX (Data Analysis Expressions)

✔ Supports star schema design (common in analytics)


4. Data Visualization

Power BI provides a rich set of visualizations:

  • Charts (bar, line, pie, etc.)
  • Tables and matrices
  • Maps and geographic visuals
  • KPIs and gauges

✔ Visuals are interactive and dynamic


5. Reports

A report in Power BI:

  • Is a collection of visualizations
  • Typically spans multiple pages
  • Allows filtering, slicing, and drill-down

✔ Built in Power BI Desktop and published to the cloud


6. Dashboards

A dashboard:

  • Is a single-page view of key metrics
  • Displays pinned visuals from reports
  • Provides a high-level overview

✔ Used for quick insights and monitoring


7. Data Refresh

Power BI supports:

  • Scheduled refresh (periodic updates)
  • Real-time/streaming data updates

✔ Ensures reports reflect current data


8. Sharing and Collaboration

Power BI enables users to:

  • Publish reports to the Power BI Service
  • Share dashboards with others
  • Collaborate across teams

✔ Integrates with Microsoft 365 (Teams, SharePoint)


9. Security

Power BI provides:

  • Row-Level Security (RLS)
  • Data access controls
  • Integration with Azure Active Directory

✔ Ensures users only see authorized data


10. Integration with Azure and Microsoft Ecosystem

Power BI integrates with:

  • Azure Synapse Analytics
  • Azure Data Lake Storage
  • Microsoft Fabric
  • Excel and other Microsoft tools

✔ Plays a key role in end-to-end analytics solutions


Power BI Components


Power BI Desktop

  • Authoring tool for reports
  • Installed on a local machine

Power BI Service

  • Cloud-based platform
  • Used for sharing and collaboration

Power BI Mobile

  • View dashboards and reports on mobile devices

Typical Analytics Workflow with Power BI

  1. Connect to data sources
  2. Transform data (Power Query)
  3. Model data (relationships, DAX)
  4. Create visualizations
  5. Publish reports
  6. Share dashboards

Why This Matters for DP-900

On the exam, you may be asked to:

  • Identify Power BI capabilities
  • Differentiate between reports and dashboards
  • Understand data connectivity and refresh options
  • Recognize Power BI’s role in analytics solutions

Summary — Exam-Relevant Takeaways

✔ Power BI is used for:

  • Data visualization
  • Reporting
  • Business intelligence

✔ Key capabilities:

  • Data connectivity
  • Data transformation (Power Query)
  • Data modeling (relationships, DAX)
  • Interactive visualizations
  • Sharing and collaboration

✔ Key components:

  • Power BI Desktop → report creation
  • Power BI Service → sharing
  • Dashboards → single-page overview
  • Reports → multi-page detailed analysis

✔ Exam tips:
👉 Report = multi-page, detailed
👉 Dashboard = single-page, summary
👉 Power Query = data transformation
👉 DAX = calculations and measures


Go to the Practice Exam Questions for this topic.

Go to the DP-900 Exam Prep Hub main page.

Practice Questions: Identify capabilities of Power BI (DP-900 Exam Prep)

Practice Questions


Question 1

What is the primary purpose of Microsoft Power BI?

A. Managing databases
B. Running virtual machines
C. Creating reports and visualizations from data
D. Developing applications

Answer: C

Explanation:
Power BI is a business intelligence tool used to create reports, dashboards, and visualizations.


Question 2

Which Power BI component is used to create reports?

A. Power BI Service
B. Power BI Mobile
C. Power BI Desktop
D. Azure Portal

Answer: C

Explanation:
Power BI Desktop is the primary tool for building reports and data models.


Question 3

What is the main difference between a report and a dashboard in Power BI?

A. Reports are single-page, dashboards are multi-page
B. Reports are multi-page, dashboards are single-page
C. Reports are only for developers
D. Dashboards cannot contain visuals

Answer: B

Explanation:
Reports are multi-page and detailed, while dashboards are single-page summaries.


Question 4

Which feature in Power BI is used to clean and transform data?

A. DAX
B. Power Query
C. Power Pivot
D. Azure Data Factory

Answer: B

Explanation:
Power Query is used for data transformation and preparation.


Question 5

Which language is used in Power BI for creating calculations and measures?

A. SQL
B. Python
C. DAX
D. Java

Answer: C

Explanation:
DAX (Data Analysis Expressions) is used for calculations and measures.


Question 6

Which Power BI feature allows users to restrict data access to specific rows?

A. Data refresh
B. Row-Level Security (RLS)
C. Power Query
D. Dashboards

Answer: B

Explanation:
Row-Level Security (RLS) ensures users only see data they are authorized to access.


Question 7

Which of the following is a key capability of Power BI?

A. Running operating systems
B. Hosting web applications
C. Connecting to multiple data sources
D. Managing network traffic

Answer: C

Explanation:
Power BI can connect to many different data sources, including databases, files, and cloud services.


Question 8

Where are Power BI reports typically published for sharing and collaboration?

A. Power BI Desktop
B. Power BI Service
C. Azure Virtual Machines
D. SQL Server

Answer: B

Explanation:
Reports are published to the Power BI Service for sharing and collaboration.


Question 9

Which capability allows Power BI to display near real-time data?

A. Scheduled refresh only
B. Streaming datasets
C. Static reports
D. Data export

Answer: B

Explanation:
Streaming datasets enable real-time or near real-time updates.


Question 10

What is the purpose of a Power BI dashboard?

A. To store raw data
B. To create data pipelines
C. To provide a single-page view of key metrics
D. To manage user accounts

Answer: C

Explanation:
Dashboards provide a high-level, single-page summary of important data.


✅ Quick Exam Takeaways

✔ Power BI is used for:

  • Data visualization
  • Reporting
  • Business intelligence

✔ Key features:

  • Power Query → data transformation
  • DAX → calculations
  • Reports → multi-page
  • Dashboards → single-page

✔ Components:

  • Power BI Desktop → build reports
  • Power BI Service → share and collaborate

✔ Security:

  • Row-Level Security (RLS)

✔ Exam tip:
👉 Transform data → Power Query
👉 Create calculations → DAX
👉 Share reports → Power BI Service


Go to the DP-900 Exam Prep Hub main page.

Identify Microsoft Cloud Services for real-time analytics (DP-900 Exam Prep)

This post is a part of the DP-900: Microsoft Azure Data Fundamentals Exam Prep Hub. 
This topic falls under these sections:
Describe an analytics workload (25–30%)
--> Describe considerations for real-time data analytics
--> Identify Microsoft Cloud Services for real-time analytics


Note that there are 10 practice questions (with answers and explanations) for each section to help you solidify your knowledge of the material. Also, there are 2 practice tests with 60 questions each available on the hub below the exam topics section.

Real-time analytics enables organizations to ingest, process, and analyze data as it is generated, allowing for immediate insights and actions. Microsoft Azure provides several services specifically designed to support real-time analytics workloads.

For the DP-900 exam, you should understand which services are used, their roles, and how they work together in a streaming architecture.


What Is Real-Time Analytics?

Real-time analytics refers to:

  • Processing data as it arrives (streaming data)
  • Producing insights with low latency (seconds or milliseconds)
  • Supporting immediate decision-making

Key Components of a Real-Time Analytics Solution

A typical real-time pipeline includes:

  1. Ingestion → Capture streaming data
  2. Processing → Analyze and transform data
  3. Storage → Persist results
  4. Visualization → Display insights

Core Azure Services for Real-Time Analytics


1. Event Ingestion Services


Azure Event Hubs

Purpose

  • High-throughput event ingestion service

Key Features

  • Handles millions of events per second
  • Scalable and distributed
  • Supports real-time data pipelines

Use Cases

  • IoT telemetry ingestion
  • Application logs
  • Streaming data pipelines

Think: “Entry point for streaming data”


Azure IoT Hub

Purpose

  • Specialized ingestion for IoT devices

Key Features

  • Device-to-cloud communication
  • Secure device management

Use Cases

  • Sensor data
  • Connected devices

Think: “Event Hubs for IoT scenarios”


2. Stream Processing Services


Azure Stream Analytics

Purpose

  • Real-time data processing using SQL-like queries

Key Features

  • Low-latency processing
  • Easy-to-use query language
  • Built-in integrations with Azure services

Use Cases

  • Real-time dashboards
  • Fraud detection
  • Alerting systems

Think: “Real-time analytics with SQL”


Azure Databricks

Purpose

  • Advanced stream and batch processing using Apache Spark

Key Features

  • Supports structured streaming
  • Handles large-scale data processing
  • Integrates with machine learning workflows

Use Cases

  • Complex event processing
  • Advanced analytics
  • Machine learning pipelines

Think: “Powerful, flexible streaming + big data processing”


3. Real-Time Analytics & Query Services


Azure Synapse Analytics

Purpose

  • Analyze streaming and batch data

Key Features

  • Integrates with streaming pipelines
  • Supports near real-time analytics

✔ Often used as part of a larger analytics architecture


Microsoft Fabric

Purpose

  • End-to-end analytics including real-time capabilities

Key Features

  • Real-Time Analytics workloads
  • Integrated with OneLake and Power BI
  • Unified platform for ingestion, processing, and visualization

Think: “All-in-one analytics platform (including real-time)”


How These Services Work Together

Typical Real-Time Pipeline

  1. Ingestion
    • Azure Event Hubs / Azure IoT Hub
  2. Processing
    • Azure Stream Analytics / Azure Databricks
  3. Storage
    • Data Lake / Synapse / Fabric OneLake
  4. Visualization
    • Power BI / Fabric dashboards

Service Selection Guidance


Use Azure Event Hubs when:

  • You need high-throughput event ingestion
  • Handling streaming data at scale

Use Azure IoT Hub when:

  • You are working with connected devices (IoT)

Use Azure Stream Analytics when:

  • You want simple, SQL-based real-time processing
  • Need quick setup and low complexity

Use Azure Databricks when:

  • You need advanced processing or machine learning
  • Working with complex or large-scale streaming data

Use Microsoft Fabric when:

  • You want a unified platform with real-time analytics built in
  • Need end-to-end analytics (data + BI)

Why This Matters for DP-900

On the exam, you may be asked to:

  • Identify which service handles streaming ingestion vs processing
  • Choose the correct service for real-time scenarios
  • Understand how services work together in a pipeline

Summary — Exam-Relevant Takeaways

✔ Real-time analytics = low-latency insights from streaming data

✔ Core services:

  • Ingestion
    • Azure Event Hubs
    • Azure IoT Hub
  • Processing
    • Azure Stream Analytics
    • Azure Databricks
  • Analytics / Platform
    • Azure Synapse Analytics
    • Microsoft Fabric

✔ Key distinctions:

  • Event Hubs → ingestion
  • Stream Analytics → real-time processing
  • Databricks → advanced processing
  • Fabric → unified analytics platform

✔ Exam tip:
👉 Streaming ingestion → Event Hubs
👉 Real-time processing → Stream Analytics
👉 Advanced analytics → Databricks
👉 Unified solution → Fabric


Go to the Practice Exam Questions for this topic.

Go to the DP-900 Exam Prep Hub main page.

Practice Questions: Identify Microsoft Cloud Services for real-time analytics (DP-900 Exam Prep)

Practice Questions


Question 1

Which Azure service is primarily used for ingesting large volumes of streaming data?

A. Azure Data Factory
B. Azure Event Hubs
C. Azure SQL Database
D. Azure Files

Answer: B

Explanation:
Azure Event Hubs is designed for high-throughput event ingestion in real time.


Question 2

Which Azure service is specifically designed for ingesting data from IoT devices?

A. Azure Blob Storage
B. Azure IoT Hub
C. Azure Synapse Analytics
D. Azure Table Storage

Answer: B

Explanation:
Azure IoT Hub enables secure communication with IoT devices and ingests telemetry data.


Question 3

Which Azure service allows real-time data processing using a SQL-like query language?

A. Azure Databricks
B. Azure Data Factory
C. Azure Stream Analytics
D. Azure Virtual Machines

Answer: C

Explanation:
Azure Stream Analytics processes streaming data using SQL-like queries.


Question 4

Which service is BEST suited for advanced real-time analytics and machine learning on streaming data?

A. Azure Files
B. Azure Databricks
C. Azure Table Storage
D. Azure DNS

Answer: B

Explanation:
Azure Databricks supports advanced analytics, Spark processing, and ML workflows.


Question 5

Which service provides a unified analytics platform that includes real-time analytics capabilities?

A. Azure Virtual Machines
B. Azure Blob Storage
C. Microsoft Fabric
D. Azure Files

Answer: C

Explanation:
Microsoft Fabric integrates real-time analytics, data engineering, and BI into one platform.


Question 6

Which component of a real-time analytics solution is responsible for capturing incoming data?

A. Processing
B. Storage
C. Visualization
D. Ingestion

Answer: D

Explanation:
The ingestion layer is responsible for capturing streaming data.


Question 7

You need to process streaming data with minimal setup using SQL-like queries. Which service should you choose?

A. Azure Databricks
B. Azure Synapse Analytics
C. Azure Stream Analytics
D. Azure Data Factory

Answer: C

Explanation:
Stream Analytics is ideal for simple, real-time processing with SQL syntax.


Question 8

Which service is MOST appropriate for handling millions of streaming events per second?

A. Azure SQL Database
B. Azure Files
C. Azure Event Hubs
D. Azure Table Storage

Answer: C

Explanation:
Event Hubs is built for high-throughput event ingestion at scale.


Question 9

Which of the following describes a typical real-time analytics pipeline?

A. Storage → Visualization → Ingestion → Processing
B. Processing → Ingestion → Storage → Visualization
C. Ingestion → Processing → Storage → Visualization
D. Visualization → Storage → Processing → Ingestion

Answer: C

Explanation:
The standard flow is:
Ingestion → Processing → Storage → Visualization


Question 10

Which scenario BEST demonstrates a real-time analytics use case?

A. Generating a yearly financial report
B. Archiving historical data
C. Monitoring live sensor data and triggering alerts
D. Migrating legacy databases

Answer: C

Explanation:
Real-time analytics is used for immediate insights and actions, such as alerts from live data.


✅ Quick Exam Takeaways

✔ Real-time analytics = low-latency insights from streaming data

✔ Core services:

  • Ingestion
    • Azure Event Hubs
    • Azure IoT Hub
  • Processing
    • Azure Stream Analytics
    • Azure Databricks
  • Platform
    • Microsoft Fabric

✔ Key roles:

  • Event Hubs → ingestion
  • Stream Analytics → real-time processing
  • Databricks → advanced analytics
  • Fabric → unified analytics platform

✔ Exam tip:
👉 Ingest streaming data → Event Hubs
👉 Process with SQL → Stream Analytics
👉 Advanced analytics → Databricks
👉 End-to-end solution → Fabric


Go to the DP-900 Exam Prep Hub main page.

Practice Questions: Describe the difference between Batch and Streaming data (DP-900 Exam Prep)

Practice Questions


Question 1

What is the primary characteristic of batch data processing?

A. Continuous data flow
B. Real-time processing
C. Processing data in scheduled chunks
D. Immediate event handling

Answer: C

Explanation:
Batch processing handles data in groups at scheduled intervals, not continuously.


Question 2

Which type of processing is BEST suited for real-time analytics?

A. Batch processing
B. Stream processing
C. Periodic processing
D. Manual processing

Answer: B

Explanation:
Stream processing enables real-time or near real-time insights.


Question 3

Which Azure service is commonly used for streaming data ingestion?

A. Azure Data Factory
B. Azure Event Hubs
C. Azure Synapse Analytics
D. Azure SQL Database

Answer: B

Explanation:
Azure Event Hubs is designed for high-throughput, real-time data ingestion.


Question 4

Which scenario is BEST suited for batch processing?

A. Monitoring live stock prices
B. Detecting fraud in real time
C. Generating a monthly financial report
D. Tracking website clicks instantly

Answer: C

Explanation:
Batch processing is ideal for scheduled, periodic workloads like reports.


Question 5

What is the typical latency for streaming data processing?

A. Hours
B. Days
C. Seconds or milliseconds
D. Weeks

Answer: C

Explanation:
Streaming processing provides low-latency, near real-time results.


Question 6

Which Azure service is used to process streaming data in real time?

A. Azure Blob Storage
B. Azure Stream Analytics
C. Azure Files
D. Azure Virtual Machines

Answer: B

Explanation:
Azure Stream Analytics processes streaming data in real time.


Question 7

Which statement about batch processing is TRUE?

A. It processes data continuously
B. It always requires real-time data sources
C. It is typically more cost-effective than streaming
D. It has lower latency than streaming

Answer: C

Explanation:
Batch processing is generally more cost-efficient than continuous streaming.


Question 8

Which scenario requires streaming processing?

A. Archiving old data
B. Processing annual tax records
C. Monitoring IoT sensor data in real time
D. Generating quarterly reports

Answer: C

Explanation:
Streaming is needed for continuous, real-time data flows like IoT.


Question 9

What is a key difference between batch and streaming processing?

A. Batch uses structured data, streaming does not
B. Streaming has higher latency than batch
C. Batch processes data in chunks, streaming processes data continuously
D. Streaming is always cheaper than batch

Answer: C

Explanation:
Batch = periodic chunks, Streaming = continuous flow.


Question 10

Which approach would you choose if immediate action is required based on incoming data?

A. Batch processing
B. Stream processing
C. Scheduled processing
D. Offline processing

Answer: B

Explanation:
Streaming is required when real-time decisions are needed.


✅ Quick Exam Takeaways

Batch processing

  • Scheduled
  • High latency
  • Cost-effective
  • Best for historical analysis

Streaming processing

  • Continuous
  • Low latency
  • Real-time insights
  • More complex

✔ Azure services:

  • Batch → Azure Data Factory, Azure Synapse Analytics
  • Streaming → Azure Event Hubs, Azure Stream Analytics

✔ Exam tip:
👉 Real-time = Streaming
👉 Scheduled/historical = Batch


Go to the DP-900 Exam Prep Hub main page.

Describe options for analytical data stores (DP-900 Exam Prep)

This post is a part of the DP-900: Microsoft Azure Data Fundamentals Exam Prep Hub. 
This topic falls under these sections:
Describe an analytics workload (25–30%)
--> Describe common elements of large-scale analytics
--> Describe options for analytical data stores


Note that there are 10 practice questions (with answers and explanations) for each section to help you solidify your knowledge of the material. Also, there are 2 practice tests with 60 questions each available on the hub below the exam topics section.

Analytical data stores are designed to support reporting, business intelligence, and large-scale data analysis. For the DP-900 exam, you should understand the different types of analytical stores, their characteristics, and when to use each.


What Is an Analytical Data Store?

An analytical data store is optimized for:

  • Querying large volumes of data
  • Aggregations and reporting
  • Historical analysis

✔ Unlike transactional systems, analytical stores focus on read-heavy workloads rather than frequent updates.


Key Characteristics

  • Optimized for complex queries and aggregations
  • Stores historical data
  • Handles large datasets (TBs to PBs)
  • Typically uses denormalized schemas
  • Designed for high-performance reads

Main Types of Analytical Data Stores


1. Data Warehouse

Definition

A structured repository designed for relational analytical queries.

Key Features

  • Uses structured data
  • Schema-based (often star or snowflake schema)
  • Supports SQL queries

Azure Example

Azure Synapse Analytics

Use Cases

  • Business intelligence reporting
  • Financial analysis
  • Enterprise dashboards

Best for: Structured data and SQL-based analytics


2. Data Lake

Definition

A storage repository for raw data in its native format.

Key Features

  • Supports structured, semi-structured, and unstructured data
  • Schema-on-read (schema applied when querying)
  • Highly scalable and cost-effective

Azure Example

Azure Data Lake Storage

Use Cases

  • Big data analytics
  • Machine learning
  • Storing raw ingestion data

Best for: Flexible, large-scale data storage


3. Data Lakehouse (Conceptual)

Definition

A hybrid approach combining features of data lakes and data warehouses.

Key Features

  • Stores raw data like a data lake
  • Supports structured queries like a warehouse
  • Often uses open formats (e.g., Parquet, Delta)

Azure Context

  • Often implemented using:
    • Azure Data Lake Storage
    • Azure Synapse Analytics

Best for: Unified analytics platform


4. Analytical Databases / Big Data Processing Systems

Definition

Systems designed for distributed processing of large datasets.

Azure Example

Azure Synapse Analytics

Key Features

  • Parallel processing
  • Handles massive datasets
  • Supports batch and interactive queries

Best for: Large-scale analytics workloads


Comparison of Analytical Data Stores

FeatureData WarehouseData LakeLakehouse
Data TypeStructuredAll typesAll types
SchemaSchema-on-writeSchema-on-readHybrid
CostHigherLowerModerate
FlexibilityLowHighHigh
Query PerformanceHighVariableHigh

Key Design Considerations


1. Data Structure

  • Structured → Data warehouse
  • Mixed or raw → Data lake

2. Query Requirements

  • Complex SQL queries → Data warehouse
  • Exploratory analytics → Data lake

3. Cost

  • Data lakes are generally more cost-effective
  • Warehouses provide optimized performance at higher cost

4. Scalability

  • All Azure analytical stores scale
  • Data lakes excel in massive data storage

5. Performance Needs

  • Warehouses → optimized for speed
  • Lakes → optimized for storage and flexibility

Typical Analytics Architecture

  1. Data Ingestion
    • Batch or streaming
  2. Storage
    • Data lake or data warehouse
  3. Processing
    • Transformations and aggregations
  4. Visualization
    • BI tools (e.g., Power BI)

Why This Matters for DP-900

On the exam, you may be asked to:

  • Identify the correct analytical store for a scenario
  • Compare data lakes vs data warehouses
  • Understand schema-on-read vs schema-on-write
  • Recognize Azure services used for analytics

Summary — Exam-Relevant Takeaways

✔ Analytical data stores are used for:

  • Reporting
  • Analytics
  • Historical data analysis

✔ Main types:

  • Data Warehouse → structured, high-performance queries
  • Data Lake → raw, flexible storage
  • Lakehouse → hybrid approach

✔ Key concepts:

  • Schema-on-write (warehouse)
  • Schema-on-read (lake)

✔ Azure services to know:

  • Azure Synapse Analytics → data warehouse & analytics
  • Azure Data Lake Storage → scalable data lake

✔ Exam tip:
👉 Structured + SQL analytics → Data Warehouse
👉 Raw + flexible + big data → Data Lake


Go to the Practice Exam Questions for this topic.

Go to the DP-900 Exam Prep Hub main page.

Practice Questions: Describe options for analytical data stores (DP-900 Exam Prep)

Practice Questions


Question 1

What is the primary purpose of an analytical data store?

A. To process high-volume transactions
B. To store temporary application data
C. To support reporting and data analysis
D. To manage user authentication

Answer: C

Explanation:
Analytical data stores are optimized for reporting, querying, and analysis, not transactions.


Question 2

Which type of data store is BEST suited for structured data and complex SQL queries?

A. Data lake
B. Data warehouse
C. File storage
D. Key-value store

Answer: B

Explanation:
Data warehouses are designed for structured data and high-performance SQL queries.


Question 3

Which Azure service is commonly used as a data warehouse?

A. Azure Data Lake Storage
B. Azure Synapse Analytics
C. Azure Files
D. Azure Table Storage

Answer: B

Explanation:
Azure Synapse Analytics provides data warehousing and large-scale analytics capabilities.


Question 4

What is a key characteristic of a data lake?

A. Requires predefined schema before loading data
B. Stores only structured data
C. Stores data in its raw format
D. Optimized for transactional workloads

Answer: C

Explanation:
Data lakes store raw data in native formats, supporting schema-on-read.


Question 5

Which concept describes applying schema when data is read rather than when it is written?

A. Schema-on-write
B. Schema-on-read
C. Data normalization
D. Data partitioning

Answer: B

Explanation:
Schema-on-read is used in data lakes, allowing flexible analysis.


Question 6

Which scenario is BEST suited for a data lake?

A. Financial reporting with strict schema
B. Running complex SQL joins on structured data
C. Storing raw IoT and log data for later analysis
D. Processing online transactions

Answer: C

Explanation:
Data lakes are ideal for large volumes of raw, diverse data.


Question 7

Which analytical data store typically uses schema-on-write?

A. Data lake
B. Data warehouse
C. Object storage
D. Key-value store

Answer: B

Explanation:
Data warehouses require a defined schema before data is loaded.


Question 8

Which of the following best describes a data lakehouse?

A. A transactional database system
B. A file storage system only
C. A hybrid of data lake and data warehouse
D. A key-value storage solution

Answer: C

Explanation:
A lakehouse combines flexibility of data lakes with performance of warehouses.


Question 9

Which factor is MOST important when choosing between a data lake and a data warehouse?

A. Screen resolution
B. Data structure and query requirements
C. Programming language
D. User interface design

Answer: B

Explanation:
The choice depends on data type (structured vs raw) and query needs.


Question 10

Which Azure service is BEST suited for storing large volumes of raw, unstructured data?

A. Azure SQL Database
B. Azure Data Lake Storage
C. Azure Synapse Analytics
D. Azure Table Storage

Answer: B

Explanation:
Azure Data Lake Storage is optimized for large-scale raw data storage.


✅ Quick Exam Takeaways

✔ Analytical data stores support:

  • Reporting
  • Business intelligence
  • Large-scale analytics

✔ Main types:

  • Data Warehouse → structured, SQL, high performance
  • Data Lake → raw, flexible, scalable
  • Lakehouse → hybrid approach

✔ Key concepts:

  • Schema-on-write → warehouse
  • Schema-on-read → lake

✔ Azure services:

  • Azure Synapse Analytics → data warehouse / analytics
  • Azure Data Lake Storage → data lake

✔ Exam tip:
👉 Structured + SQL → Data Warehouse
👉 Raw + flexible → Data Lake


Go to the DP-900 Exam Prep Hub main page.