Category: Data Quality Assurance

Select, Filter, and Aggregate Data Using DAX

This post is a part of the DP-600: Implementing Analytics Solutions Using Microsoft Fabric Exam Prep Hub; and this topic falls under these sections: 
Prepare data
--> Query and analyze data
--> Select, Filter, and Aggregate Data Using DAX

Data Analysis Expressions (DAX) is a formula language used to create dynamic calculations in Power BI semantic models. Unlike SQL or KQL, DAX works within the analytical model and is designed for filter context–aware calculations, interactive reporting, and business logic. For DP-600, you should understand how to use DAX to select, filter, and aggregate data within a semantic model for analytics and reporting.


What Is DAX?

DAX is similar to Excel formulas but optimized for relational, in-memory analytics. It is used in:

  • Measures (dynamic calculations)
  • Calculated columns (row-level derived values)
  • Calculated tables (additional, reusable query results)

In a semantic model, DAX queries run in response to visuals and can produce results based on current filters and slicers.


Selecting Data in DAX

DAX itself doesn’t use a traditional SELECT statement like SQL. Instead:

  • Data is selected implicitly by filter context
  • DAX measures operate over table columns referenced in expressions

Example of a simple DAX measure selecting and displaying sales:

Total Sales = SUM(Sales[SalesAmount])

Here:

  • Sales[SalesAmount] references the column in the Sales table
  • The measure returns the sum of all values in that column

Filtering Data in DAX

Filtering in DAX is context-driven and can be applied in multiple ways:

1. Implicit Filters

Visual-level filters and slicers automatically apply filters to DAX measures.

Example:
A card visual showing Total Sales will reflect only the filtered subset by product or date.

2. FILTER Function

Used within measures or calculated tables to narrow down rows:

HighValueSales = CALCULATE(
    SUM(Sales[SalesAmount]),
    FILTER(Sales, Sales[SalesAmount] > 1000)
)

Here:

  • FILTER returns a table with rows meeting the condition
  • CALCULATE modifies the filter context

3. CALCULATE as Filter Modifier

CALCULATE changes the context under which a measure evaluates:

SalesLastYear = CALCULATE(
    [Total Sales],
    SAMEPERIODLASTYEAR(Date[Date])
)

This measure selects data for the previous year based on current filters.


Aggregating Data in DAX

Aggregation in DAX is done using built-in functions and is influenced by filter context.

Common Aggregation Functions

  • SUM() — totals a numeric column
  • AVERAGE() — computes the mean
  • COUNT() / COUNTA() — row counts
  • MAX() / MIN() — extreme values
  • SUMX() — row-by-row iteration and sum

Example of row-by-row aggregation:

Total Profit = SUMX(
    Sales,
    Sales[SalesAmount] - Sales[Cost]
)

This computes the difference per row and then sums it.


Filter Context and Row Context

Understanding how DAX handles filter context and row context is essential:

  • Filter context: Set by the report (slicers, column filters) or modified by CALCULATE
  • Row context: Used in calculated columns and iteration functions (SUMX, FILTER)

DAX measures always respect the current filter context unless explicitly modified.


Grouping and Summarization

While DAX doesn’t use GROUP BY in the same way SQL does, measures inherently aggregate over groups determined by filter context or visual grouping.

Example:
In a table visual grouped by Product Category, the measure Total Sales returns aggregated values per category automatically.


Time Intelligence Functions

DAX includes built-in functions for time-based aggregation:

  • TOTALYTD(), TOTALQTD(), TOTALMTD() — year-to-date, quarter-to-date, month-to-date
  • SAMEPERIODLASTYEAR() — compare values year-over-year
  • DATESINPERIOD() — custom period

Example:

SalesYTD = TOTALYTD(
    [Total Sales],
    Date[Date]
)


Best Practices

  • Use measures, not calculated columns, for dynamic, filter-sensitive aggregations.
  • Let visuals control filter context via slicers, rows, and columns.
  • Avoid unnecessary row-by-row calculations when simple aggregation functions suffice.
  • Explicitly use CALCULATE to modify filter context for advanced scenarios.

When to Use DAX vs SQL/KQL

ScenarioBest Tool
Static relational queryingSQL
Streaming/event analyticsKQL
Report-level dynamic calculationsDAX
Interactive dashboards with slicersDAX

Example Use Cases

1. Total Sales Measure

Total Sales = SUM(Sales[SalesAmount])

2. Filtered Sales for Big Orders

Big Orders Sales = CALCULATE(
    [Total Sales],
    Sales[SalesAmount] > 1000
)

3. Year-over-Year Sales

Sales YOY = CALCULATE(
    [Total Sales],
    SAMEPERIODLASTYEAR(Date[Date])
)


Key Takeaways for the Exam

  • DAX operates based on filter context and evaluates measures dynamically.
  • There is no explicit SELECT statement — rather, measures compute values based on current context.
  • Use CALCULATE to change filter context.
  • Aggregation functions (e.g., SUM, COUNT, AVERAGE) are fundamental to summarizing data.
  • Filtering functions like FILTER and time intelligence functions enhance analytical flexibility.

Final Exam Tips

  • If a question mentions interactive reports, dynamic filters, slicers, or time-based comparisons, DAX is likely the right language to use for the solution.
  • Measures + CALCULATE + filter context appear frequently.
  • If the question mentions slicers, visuals, or dynamic results, think DAX measure.
  • Time intelligence functions are high-value topics.

Practice Questions:

Here are 10 questions to test and help solidify your learning and knowledge. As you review these and other questions in your preparation, make sure to …

  • Identifying and understand why an option is correct (or incorrect) — not just which one
  • Look for and understand the usage scenario of keywords in exam questions to guide you
  • Expect scenario-based questions rather than direct definitions

1. Which DAX function is primarily used to modify the filter context of a calculation?

A. FILTER
B. SUMX
C. CALCULATE
D. ALL

Correct answer: ✅ C
Explanation: CALCULATE changes the filter context under which an expression is evaluated.


2. A Power BI report contains slicers for Year and Product. A measure returns different results as slicers change. What concept explains this behavior?

A. Row context
B. Filter context
C. Evaluation context
D. Query context

Correct answer: ✅ B
Explanation: Filter context is affected by slicers, filters, and visual interactions.


3. Which DAX function iterates row by row over a table to perform a calculation?

A. SUM
B. COUNT
C. AVERAGE
D. SUMX

Correct answer: ✅ D
Explanation: SUMX evaluates an expression for each row and then aggregates the results.


4. You want to calculate total sales only for transactions greater than $1,000. Which approach is correct?

A.

SUM(Sales[SalesAmount] > 1000)

B.

FILTER(Sales, Sales[SalesAmount] > 1000)

C.

CALCULATE(
    SUM(Sales[SalesAmount]),
    Sales[SalesAmount] > 1000
)

D.

SUMX(Sales, Sales[SalesAmount] > 1000)

Correct answer: ✅ C
Explanation: CALCULATE applies a filter condition while aggregating.


5. Which DAX object is evaluated dynamically based on report filters and slicers?

A. Calculated column
B. Calculated table
C. Measure
D. Relationship

Correct answer: ✅ C
Explanation: Measures respond dynamically to filter context; calculated columns do not.


6. Which function is commonly used to calculate year-to-date (YTD) values in DAX?

A. DATESINPERIOD
B. SAMEPERIODLASTYEAR
C. TOTALYTD
D. CALCULATE

Correct answer: ✅ C
Explanation: TOTALYTD is designed for year-to-date aggregations.


7. A DAX measure returns different totals when placed in a table visual grouped by Category. Why does this happen?

A. The measure contains row context
B. The table visual creates filter context
C. The measure is recalculated per row
D. Relationships are ignored

Correct answer: ✅ B
Explanation: Visual grouping applies filter context automatically.


8. Which DAX function returns a table instead of a scalar value?

A. SUM
B. AVERAGE
C. FILTER
D. COUNT

Correct answer: ✅ C
Explanation: FILTER returns a table that can be consumed by other functions like CALCULATE.


9. Which scenario is the best use case for DAX instead of SQL or KQL?

A. Cleaning raw data before ingestion
B. Transforming streaming event data
C. Creating interactive report-level calculations
D. Querying flat files in a lakehouse

Correct answer: ✅ C
Explanation: DAX excels at dynamic, interactive calculations in semantic models.


10. What is the primary purpose of the SAMEPERIODLASTYEAR function?

A. Aggregate values by fiscal year
B. Remove filters from a date column
C. Compare values to the previous year
D. Calculate rolling averages

Correct answer: ✅ C
Explanation: It shifts the date context back one year for year-over-year analysis.


Select, Filter, and Aggregate Data Using SQL

This post is a part of the DP-600: Implementing Analytics Solutions Using Microsoft Fabric Exam Prep Hub; and this topic falls under these sections: 
Prepare data
--> Query and analyze data
--> Select, Filter, and Aggregate Data Using SQL

Working with SQL to select, filter, and aggregate data is a core skill for analytics engineers using Microsoft Fabric. Whether querying data in a warehouse, lakehouse SQL analytics endpoint, or semantic model via DirectQuery, SQL enables precise data retrieval and summarization for reporting, dashboards, and analytics solutions.

For DP-600, you should understand how to construct SQL queries that perform:

  • Selecting specific data columns
  • Filtering rows based on conditions
  • Aggregating values with grouping and summary functions

SQL Data Selection

Selecting data refers to using the SELECT clause to choose which columns or expressions to return.

Example:

SELECT
    CustomerID,
    OrderDate,
    SalesAmount
FROM Sales;

  • Use * to return all columns:
    SELECT * FROM Sales;
  • Use expressions to compute derived values: SELECT OrderDate, SalesAmount, SalesAmount * 1.1 AS AdjustedRevenue FROM Sales;

Exam Tip: Be purposeful in selecting only needed columns to improve performance.


SQL Data Filtering

Filtering data determines which rows are returned based on conditions using the WHERE clause.

Basic Filtering:

SELECT *
FROM Sales
WHERE OrderDate >= '2025-01-01';

Combined Conditions:

  • AND: WHERE Country = 'USA' AND SalesAmount > 1000
  • OR: WHERE Region = 'East' OR Region = 'West'

Null and Missing Value Filters:

WHERE SalesAmount IS NOT NULL

Exam Tip: Understand how WHERE filters reduce dataset size before aggregation.


SQL Aggregation

Aggregation summarizes grouped rows using functions like SUM, COUNT, AVG, MIN, and MAX.

Basic Aggregation:

SELECT
    SUM(SalesAmount) AS TotalSales
FROM Sales;

Grouped Aggregation:

SELECT
    Country,
    SUM(SalesAmount) AS TotalSales,
    COUNT(*) AS OrderCount
FROM Sales
GROUP BY Country;

Filtering After Aggregation:

Use HAVING instead of WHERE to filter aggregated results:

SELECT
    Country,
    SUM(SalesAmount) AS TotalSales
FROM Sales
GROUP BY Country
HAVING SUM(SalesAmount) > 100000;

Exam Tip:

  • Use WHERE for row-level filters before grouping.
  • Use HAVING to filter group-level aggregates.

Combining Select, Filter, and Aggregate

A complete SQL query often blends all three:

SELECT
    ProductCategory,
    COUNT(*) AS Orders,
    SUM(SalesAmount) AS TotalSales,
    AVG(SalesAmount) AS AvgSale
FROM Sales
WHERE OrderDate BETWEEN '2025-01-01' AND '2025-12-31'
GROUP BY ProductCategory
ORDER BY TotalSales DESC;

This example:

  • Selects specific columns and expressions
  • Filters by date range
  • Aggregates by product category
  • Orders results by summary metric

SQL in Different Fabric Workloads

WorkloadSQL Usage
WarehouseStandard T-SQL for BI queries
Lakehouse SQL AnalyticsSQL against Delta tables
Semantic Models via DirectQuerySQL pushed to source where supported
Dataflows/Power QuerySQL-like operations through M (not direct SQL)

Performance and Pushdown

When using SQL in Fabric:

  • Engines push filters and aggregations down to the data source for performance.
  • Select only needed columns early to limit data movement.
  • Avoid SELECT * in production queries unless necessary.

Key SQL Concepts for the Exam

ConceptWhy It Matters
SELECTDefines what data to retrieve
WHEREFilters data before aggregation
GROUP BYOrganizes rows into groups
HAVINGFilters after aggregation
Aggregate functionsSummarize numeric data

Understanding how these work together is essential for creating analytics-ready datasets.


Common Exam Scenarios

You may be asked to:

  • Write SQL to filter data based on conditions
  • Summarize data across groups
  • Decide whether to use WHERE or HAVING
  • Identify the correct SQL pattern for a reporting requirement

Example exam prompt:

“Which SQL query correctly returns the total sales per region, only for regions with more than 1,000 orders?”

Understanding aggregate filters (HAVING) and groupings will be key.


Final Exam Tips

If a question mentions:

  • “Return summary metrics”
  • “Only include rows that meet conditions”
  • “Group results by category”

…you’re looking at combining SELECT, WHERE, and GROUP BY in SQL.

  • WHERE filters rows before aggregation
  • HAVING filters after aggregation
  • GROUP BY is required for per-group metrics
  • Use aggregate functions intentionally
  • Performance matters — avoid unnecessary columns

Practice Questions:

Here are 10 questions to test and help solidify your learning and knowledge. As you review these and other questions in your preparation, make sure to …

  • Identifying and understand why an option is correct (or incorrect) — not just which one
  • Look for and understand the usage scenario of keywords in exam questions to guide you
  • Expect scenario-based questions rather than direct definitions

1. Which SQL clause is used to filter rows before aggregation occurs?

A. HAVING
B. GROUP BY
C. WHERE
D. ORDER BY

Correct Answer: C

Explanation:
The WHERE clause filters individual rows before any aggregation or grouping takes place. HAVING filters results after aggregation.


2. You need to calculate total sales per product category. Which clause is required?

A. WHERE
B. GROUP BY
C. ORDER BY
D. HAVING

Correct Answer: B

Explanation:
GROUP BY groups rows so aggregate functions (such as SUM) can be calculated per category.


3. Which function returns the number of rows in each group?

A. SUM()
B. COUNT()
C. AVG()
D. MAX()

Correct Answer: B

Explanation:
COUNT() counts the number of rows in a group. It is commonly used to count records or transactions.


4. Which query correctly filters aggregated results?

A.

WHERE SUM(SalesAmount) > 10000

B.

HAVING SUM(SalesAmount) > 10000

C.

GROUP BY SUM(SalesAmount) > 10000

D.

ORDER BY SUM(SalesAmount) > 10000

Correct Answer: B

Explanation:
HAVING is used to filter aggregated values. WHERE cannot reference aggregate functions.


5. Which SQL statement returns the total number of orders?

A.

SELECT COUNT(*) FROM Orders;

B.

SELECT SUM(*) FROM Orders;

C.

SELECT TOTAL(Orders) FROM Orders;

D.

SELECT COUNT(Orders) FROM Orders;

Correct Answer: A

Explanation:
COUNT(*) counts all rows in a table, making it the correct way to return total order count.


6. Which clause is used to sort aggregated query results?

A. GROUP BY
B. WHERE
C. ORDER BY
D. HAVING

Correct Answer: C

Explanation:
ORDER BY sorts the final result set, including aggregated columns.


7. What happens if a column in the SELECT statement is not included in the GROUP BY clause or an aggregate function?

A. The query runs but returns incorrect results
B. SQL automatically groups it
C. The query fails
D. The column is ignored

Correct Answer: C

Explanation:
In SQL, any column in SELECT must either be aggregated or included in GROUP BY.


8. Which query returns average sales amount per country?

A.

SELECT Country, AVG(SalesAmount)
FROM Sales;

B.

SELECT Country, AVG(SalesAmount)
FROM Sales
GROUP BY Country;

C.

SELECT Country, SUM(SalesAmount)
GROUP BY Country;

D.

SELECT AVG(SalesAmount)
FROM Sales
GROUP BY Country;

Correct Answer: B

Explanation:
Grouping by Country allows AVG(SalesAmount) to be calculated per country.


9. Which filter removes rows with NULL values in a column?

A.

WHERE SalesAmount = NULL

B.

WHERE SalesAmount <> NULL

C.

WHERE SalesAmount IS NOT NULL

D.

WHERE NOT NULL SalesAmount

Correct Answer: C

Explanation:
SQL uses IS NULL and IS NOT NULL to check for null values.


10. Which SQL pattern is most efficient for analytics queries in Microsoft Fabric?

A. Selecting all columns and filtering later
B. Using SELECT * for simplicity
C. Filtering early and selecting only needed columns
D. Aggregating without grouping

Correct Answer: C

Explanation:
Filtering early and selecting only required columns improves performance by reducing data movement—an important Fabric best practice.


Select, Filter, and Aggregate Data by Using the Visual Query Editor

This post is a part of the DP-600: Implementing Analytics Solutions Using Microsoft Fabric Exam Prep Hub; and this topic falls under these sections: 
Prepare data
--> Query and analyze data
--> Select, Filter, and Aggregate Data by Using the Visual Query Editor

In Microsoft Fabric, the Visual Query Editor (VQE) provides a low-code, graphical experience for querying data across lakehouses, warehouses, and semantic models. It allows analytics engineers to explore, shape, and summarize data without writing SQL or KQL, while still generating optimized queries behind the scenes.

For the DP-600 exam, you should understand what the Visual Query Editor is, where it’s used, and how to perform common data analysis tasks such as selecting columns, filtering rows, and aggregating data.


What Is the Visual Query Editor?

The Visual Query Editor is a graphical query-building interface available in multiple Fabric experiences, including:

  • Lakehouse SQL analytics endpoint
  • Warehouse
  • Power BI (Direct Lake and DirectQuery scenarios)
  • Data exploration within Fabric items

Instead of writing queries manually, you interact with:

  • Tables and columns
  • Drag-and-drop operations
  • Menus for filters, grouping, and aggregations

Fabric then translates these actions into optimized SQL or engine-specific queries.


Selecting Data

Selecting data in the Visual Query Editor focuses on choosing the right columns and datasets for analysis.

Key Capabilities

  • Select or deselect columns from one or more tables
  • Rename columns for readability
  • Reorder columns for analysis or reporting
  • Combine columns from related tables (via existing relationships)

Exam Tips

  • Selecting fewer columns improves performance and reduces data transfer.
  • Column renaming in VQE affects the query result, not the underlying table schema.
  • The Visual Query Editor respects relationships defined in semantic models and warehouses.

Filtering Data

Filtering allows you to limit rows based on conditions, ensuring only relevant data is included.

Common Filter Types

  • Equality filters (e.g., Status = "Active")
  • Range filters (e.g., dates, numeric thresholds)
  • Text filters (contains, starts with, ends with)
  • Null / non-null filters
  • Relative date filters (last 7 days, current month)

Where Filtering Is Applied

  • At the query level, not permanently in the data source
  • Before aggregation (important for correct results)

Exam Tips

  • Filters applied in the Visual Query Editor are executed at the data source when possible (query folding).
  • Filtering early improves performance and reduces memory usage.
  • Be aware of how filters interact with aggregations.

Aggregating Data

Aggregation summarizes data by grouping rows and applying calculations.

Common Aggregations

  • Sum
  • Count / Count Distinct
  • Average
  • Min / Max

Grouping Data

  • Select one or more columns as group-by fields
  • Apply aggregations to numeric or date columns
  • Results return one row per group

Examples

  • Total sales by product category
  • Count of orders per customer
  • Average response time by day

Exam Tips

  • Aggregations in the Visual Query Editor are conceptually similar to GROUP BY in SQL.
  • Aggregated queries reduce dataset size and improve performance.
  • Understand the difference between row-level data and aggregated results.

Behind the Scenes: Generated Queries

Although the Visual Query Editor is low-code, Fabric generates:

  • SQL queries for warehouses and lakehouse SQL endpoints
  • Optimized engine-specific queries for semantic models

This ensures:

  • Efficient execution
  • Compatibility with Direct Lake and DirectQuery
  • Consistent results across Fabric experiences

Exam Tip

You are not required to read or write the generated SQL, but you should understand that the Visual Query Editor does not bypass query optimization.


When to Use the Visual Query Editor

Use the Visual Query Editor when:

  • Quickly exploring unfamiliar datasets
  • Building queries without writing code
  • Creating reusable query logic for reports
  • Teaching or collaborating with less SQL-focused users

Avoid it when:

  • Complex transformations are required (use SQL, Spark, or Dataflows)
  • Highly customized logic is needed beyond supported operations

Key Exam Takeaways

For the DP-600 exam, remember:

  • The Visual Query Editor is a graphical query-building tool in Microsoft Fabric.
  • It supports selecting columns, filtering rows, and aggregating data.
  • Operations are translated into optimized queries executed at the data source.
  • Filtering occurs before aggregation, affecting results and performance.
  • It is commonly used with lakehouses, warehouses, and semantic models.

Practice Questions:

Here are 10 questions to test and help solidify your learning and knowledge. As you review these and other questions in your preparation, make sure to …

  • Identifying and understand why an option is correct (or incorrect) — not just which one
  • Look for and understand the usage scenario of keywords in exam questions to guide you
  • Expect scenario-based questions rather than direct definitions
  • Know the purpose and scope of the Visual Query Editor
  • Know how to selecting, filtering, and aggregating data
  • Understand execution order and performance implications
  • Know when to use (and not use) the Visual Query Editor

Question 1

What is the primary purpose of the Visual Query Editor in Microsoft Fabric?

A. To permanently modify table schemas
B. To build queries visually without writing SQL
C. To replace semantic models
D. To manage workspace permissions

Correct Answer: B

Explanation:
The Visual Query Editor provides a low-code, graphical interface for building queries. It does not modify schemas, replace models, or manage security.


Question 2

When you deselect a column in the Visual Query Editor, what happens?

A. The column is deleted from the source table
B. The column is hidden permanently for all users
C. The column is excluded only from the query results
D. The column data type is changed

Correct Answer: C

Explanation:
Column selection affects only the query output, not the underlying data or schema.


Question 3

Why is it considered a best practice to select only required columns in a query?

A. It enforces data security
B. It reduces query complexity and improves performance
C. It enables Direct Lake mode
D. It prevents duplicate rows

Correct Answer: B

Explanation:
Selecting fewer columns reduces data movement and memory usage, leading to better performance.


Question 4

Which type of filter is commonly used to restrict data to a recent time period?

A. Equality filter
B. Text filter
C. Relative date filter
D. Aggregate filter

Correct Answer: C

Explanation:
Relative date filters (e.g., “Last 30 days”) dynamically adjust based on the current date and are commonly used in analytics.


Question 5

At what stage of query execution are filters applied in the Visual Query Editor?

A. After aggregation
B. After the query result is returned
C. Before aggregation
D. Only in the Power BI report layer

Correct Answer: C

Explanation:
Filters are applied before aggregation, ensuring accurate summary results and better performance.


Question 6

Which aggregation requires grouping to produce meaningful results?

A. SUM
B. COUNT
C. GROUP BY
D. MIN

Correct Answer: C

Explanation:
Grouping defines how rows are summarized. Aggregations like SUM or COUNT rely on GROUP BY logic to produce per-group results.


Question 7

You want to see total sales by product category. Which Visual Query Editor actions are required?

A. Filter Product Category and sort by Sales
B. Group by Product Category and apply SUM to Sales
C. Count Product Category and filter Sales
D. Rename Product Category and aggregate rows

Correct Answer: B

Explanation:
This scenario requires grouping on Product Category and applying a SUM aggregation to the Sales column.


Question 8

What happens behind the scenes when you build a query using the Visual Query Editor?

A. Fabric stores a cached dataset only
B. Fabric generates optimized SQL or engine-specific queries
C. Fabric converts the query into DAX
D. Fabric disables query folding

Correct Answer: B

Explanation:
The Visual Query Editor translates visual actions into optimized queries (such as SQL) that execute at the data source.


Question 9

Which Fabric items commonly support querying through the Visual Query Editor?

A. Pipelines and notebooks only
B. Dashboards only
C. Lakehouses, warehouses, and semantic models
D. Eventhouses only

Correct Answer: C

Explanation:
The Visual Query Editor is widely used across lakehouses, warehouses, and semantic models in Fabric.


Question 10

When should you avoid using the Visual Query Editor?

A. When exploring new datasets
B. When building quick aggregations
C. When complex transformation logic is required
D. When filtering data

Correct Answer: C

Explanation:
For advanced or complex transformations, SQL, Spark, or Dataflows are more appropriate than the Visual Query Editor.


Filter Data

This post is a part of the DP-600: Implementing Analytics Solutions Using Microsoft Fabric Exam Prep Hub; and this topic falls under these sections: 
Prepare data
--> Transform data
--> Filter data

Filtering data is one of the most fundamental transformation operations used when preparing analytics data. It ensures that only relevant, valid, and accurate records are included in curated tables or models. Filtering improves performance, reduces unnecessary processing overhead, and helps enforce business logic early in the analytics pipeline.

In Microsoft Fabric, filtering occurs at multiple transformation layers — from ingestion tools to interactive modeling. For the DP-600 exam, you should understand where, why, and how to filter data effectively using various tools and technologies within Fabric.


Why Filter Data?

Filtering data serves several key purposes in analytics:

1. Improve Query and Report Performance

  • Reduces the amount of data scanned and processed
  • Enables faster refresh and retrieval

2. Enforce Business Logic

  • Excludes irrelevant segments (e.g., test data, canceled transactions)
  • Supports clean analytical results

3. Prepare Analytics-Ready Data

  • Limits datasets to required time periods or categories
  • Produces smaller, focused outputs for reporting

4. Reduce Cost

  • Smaller processing needs reduce compute and storage overhead

Where Filtering Happens in Microsoft Fabric

Filtering can be implemented at multiple stages:

LayerHow You Filter
Power Query (Dataflows Gen2 / Lakehouse)UI filters or M code
SQL (Warehouse & Lakehouse SQL analytics)WHERE clauses
Spark (Lakehouse Notebooks)DataFrame filter() / where()
Pipelines (Data Movement)Source filters or query-based extraction
Semantic Models (Power BI / DAX)Query filters, slicers, and row-level security

Filtering early, as close to the data source as possible, ensures better performance downstream.


Tools and Techniques

1. Power Query (Low-Code)

Power Query provides a user-friendly interface to filter rows:

  • Text filters: Equals, Begins With, Contains, etc.
  • Number filters: Greater than, Between, Top N, etc.
  • Date filters: Before, After, This Month, Last 12 Months, etc.
  • Remove blank or null values

These filters are recorded as transformation steps and can be reused or versioned.


2. SQL (Warehouses & Lakehouses)

SQL filtering uses the WHERE clause:

SELECT *
FROM Sales
WHERE OrderDate >= '2025-01-01'
  AND Country = 'USA';

SQL filtering is efficient and pushed down to the engine, reducing row counts early.


3. Spark (Notebooks)

Filtering in Spark (PySpark example):

filtered_df = df.filter(df["SalesAmount"] > 1000)

Or with SQL in Spark:

SELECT *
FROM sales
WHERE SalesAmount > 1000;

Spark filtering is optimized for distributed processing across big datasets.


4. Pipelines (Data Movement)

During ingestion or ETL, you can apply filters in:

  • Copy activity query filters
  • Source queries
  • Pre-processing steps

This ensures only needed rows land in the target store.


5. Semantic Model Filters

In Power BI and semantic models, filtering can happen as:

  • Report filters
  • Slicers and visuals
  • Row-Level Security (RLS) — security-driven filtering

These filters control what users see rather than what data is stored.


Business and Data Quality Scenarios

Filtering is often tied to business needs such as:

  • Excluding invalid, test, or archived records
  • Restricting to active customers only
  • Selecting a specific date range (e.g., last fiscal year)
  • Filtering data for regional or product segments

Filtering vs Security

It’s important to distinguish filtering for transformation from security filters:

FilteringSecurity
Removes unwanted rows during transformationControls what users are allowed to see
Improves performanceEnforces access control
Happens before modelingHappens during query evaluation

Best Practices

When filtering data in Microsoft Fabric:

  • Filter early in the pipeline to reduce volume
  • Use pushdown filters in SQL when querying large sources
  • Document filtering logic for audit and governance
  • Combine filters logically (AND/OR) to match business rules
  • Avoid filtering in the semantic model when it can be done upstream

Common Exam Scenarios

You may be asked to:

  • Choose the correct tool and stage for filtering
  • Translate business rules into filter logic
  • Recognize when filtering improves performance
  • Identify risks of filtering too late or in the wrong layer

Example exam prompt:
A dataset should exclude test transactions and include only the last 12 months of sales. Which transformation step should be applied and where?
The correct answer will involve filtering early with SQL or Power Query before modeling.


Key Takeaways

  • Filtering data is a core part of preparing analytics-ready datasets.
  • Multiple Fabric components support filtering (Power Query, SQL, Spark, pipelines).
  • Filtering early improves performance and reduces unnecessary workload.
  • Understand filtering in context — transformation vs. security.

Final Exam Tips

  • When a question asks about reducing dataset size, improving performance, or enforcing business logic before loading into a model, filtering is often the correct action — and it usually belongs upstream.
  • Filter early and upstream whenever possible
  • Use SQL or Power Query for transformation-level filtering
  • Avoid relying solely on report-level filters for large datasets
  • Distinguish filtering for performance from security filtering

Practice Questions:

Here are 10 questions to test and help solidify your learning and knowledge. As you review these and other questions in your preparation, make sure to …

  • Identifying and understand why an option is correct (or incorrect) — not just which one
  • Look for and understand the usage scenario of keywords in exam questions to guide you
  • Expect scenario-based questions rather than direct definitions

Question 1

What is the primary purpose of filtering data during the transformation phase?

A. To enforce user-level security
B. To reduce data volume and improve performance
C. To encrypt sensitive columns
D. To normalize data structures

Correct Answer: B

Explanation:
Filtering removes unnecessary rows early in the pipeline, reducing data volume, improving performance, and lowering compute costs. Security and normalization are separate concerns.


Question 2

Which Fabric component allows low-code, UI-driven row filtering during data preparation?

A. Spark notebooks
B. SQL warehouse
C. Power Query (Dataflows Gen2)
D. Semantic models

Correct Answer: C

Explanation:
Power Query provides a graphical interface for filtering rows using text, numeric, and date-based filters, making it ideal for low-code transformations.


Question 3

Which SQL clause is used to filter rows in a lakehouse or warehouse?

A. GROUP BY
B. HAVING
C. WHERE
D. ORDER BY

Correct Answer: C

Explanation:
The WHERE clause filters rows before aggregation or sorting, making it the primary SQL mechanism for data filtering.


Question 4

Which filtering approach is most efficient for very large datasets?

A. Filtering in Power BI visuals
B. Filtering after loading data into a semantic model
C. Filtering at the source using SQL or ingestion queries
D. Filtering using calculated columns

Correct Answer: C

Explanation:
Filtering as close to the source as possible minimizes data movement and processing, making it the most efficient approach for large datasets.


Question 5

In a Spark notebook, which method is commonly used to filter a DataFrame?

A. select()
B. filter() or where()
C. join()
D. distinct()

Correct Answer: B

Explanation:
Spark DataFrames use filter() or where() to remove rows based on conditions.


Question 6

Which scenario is an example of business-rule filtering?

A. Removing duplicate rows
B. Converting text to numeric data types
C. Excluding canceled orders from sales analysis
D. Creating a star schema

Correct Answer: C

Explanation:
Business-rule filtering enforces organizational logic, such as excluding canceled or test transactions from analytics.


Question 7

What is the key difference between data filtering and row-level security (RLS)?

A. Filtering improves query speed; RLS does not
B. Filtering removes data; RLS restricts visibility
C. Filtering is applied only in SQL; RLS is applied only in Power BI
D. Filtering is mandatory; RLS is optional

Correct Answer: B

Explanation:
Filtering removes rows from the dataset, while RLS controls which rows users can see without removing the data itself.


Question 8

Which filtering method is typically applied after data has already been loaded?

A. Source query filters
B. Pipeline copy activity filters
C. Semantic model report filters
D. Power Query transformations

Correct Answer: C

Explanation:
Report and visual filters in semantic models are applied at query time and do not reduce stored data volume.


Question 9

Why is filtering data early in the pipeline considered a best practice?

A. It increases data redundancy
B. It simplifies semantic model design
C. It reduces processing and storage costs
D. It improves data encryption

Correct Answer: C

Explanation:
Early filtering minimizes unnecessary data processing and storage, improving efficiency across the entire analytics solution.


Question 10

A dataset should include only the last 12 months of data. Where should this filter ideally be applied?

A. In Power BI slicers
B. In the semantic model
C. During data ingestion or transformation
D. In calculated measures

Correct Answer: C

Explanation:
Applying time-based filters during ingestion or transformation ensures only relevant data is processed and stored, improving performance and consistency.


Convert Column Data Types

This post is a part of the DP-600: Implementing Analytics Solutions Using Microsoft Fabric Exam Prep Hub; and this topic falls under these sections: 
Prepare data
--> Transform data
--> Convert column data types

Converting data types is a fundamental transformation task in data preparation. It helps ensure data consistency, accurate calculations, filter behavior, sorting, joins, and overall query correctness. In Microsoft Fabric, data type conversion can happen in Power Query, SQL, or Spark depending on the workload and where you are in your data pipeline.

This article explains why, where, and how you convert data types in Fabric, with an emphasis on real-world scenarios and exam relevance.

Why Convert Data Types?

Data type mismatches can lead to:

  • Erroneous joins (e.g., joining text to numeric)
  • Incorrect aggregations (e.g., sums treating numbers as text)
  • Filtering issues (e.g., date strings not filtering as dates)
  • Unexpected sort order (e.g., text sorts differently from numbers)

In analytics, getting data types right is critical for both the correctness of results and query performance.

Common Data Types in Analytics

Here are some common data types you’ll work with:

CategoryExamples
NumericINT, BIGINT, DECIMAL, FLOAT
TextSTRING, VARCHAR
Date/TimeDATE, TIME, DATETIME, TIMESTAMP
BooleanTRUE / FALSE

Where Data Type Conversion Occurs in Fabric

Depending on workload and tool, you may convert data types in:

Power Query (Dataflows Gen2 & Lakehouses)

  • Visual change type steps (Menu → Transform → Data Type)
  • Applied steps stored in the query
  • Useful for low-code transformation

SQL (Warehouse & Lakehouse SQL Analytics)

  • CAST, CONVERT, or TRY_CAST in SQL
  • Applies at query time or when persisting transformed data

Spark (Lakehouse Notebooks)

  • Explicit schema definitions
  • Transformation commands like withColumn() with type conversion functions

Each environment has trade-offs. For example, Power Query is user-friendly but may not scale like SQL or Spark for very large datasets.

How to Convert Data Types

In Power Query

  1. Select the column
  2. Go to Transform → Data Type
  3. Choose the correct type (e.g., Whole Number, Decimal Number, Date)

Power Query generates a Change Type step that applies at refresh.

In SQL

SELECT

    CAST(order_amount AS DECIMAL(18,2)) AS order_amount,

    CONVERT(DATE, order_date) AS order_date

FROM Sales;

  • CAST() and CONVERT() are standard.
  • Some engines support TRY_CAST() to avoid errors on incompatible values.

In Spark (PySpark or SQL)

PySpark example:

df = df.withColumn(“order_date”, df[“order_date”].cast(“date”))

SQL example in Spark:

SELECT CAST(order_amount AS DOUBLE) AS order_amount

FROM sales;

When to Convert Data Types

You should convert data types:

  • Before joins (to ensure matching keys)
  • Before aggregations (to ensure correct math operations)
  • Before loading into semantic models
    (to ensure correct behavior in Power BI)
  • When cleaning source data
    (e.g., text fields that actually represent numbers or dates)

Common Conversion Scenarios

1. Text to Numeric

Often needed when source systems export numbers as text:

SourceTarget
“1000”1000 (INT/DECIMAL)

2. Text to Date/Time

Date fields often arrive as text:

SourceTarget
“2025-08-01”2025-08-01 (DATE)

3. Numeric to Text

Sometimes required when composing keys:

CONCAT(customer_id, order_id)

4. Boolean Conversion

Often used in logical flags:

SourceTarget
“Yes”/”No”TRUE/FALSE

Handling Conversion Errors

Not all values convert cleanly. Options include:

  • TRY_CAST / TRY_CONVERT
    • Returns NULL instead of error
  • Error handling in Power Query
    • Replacing errors or invalid values
  • Filtering out problematic rows
    • Before casting

Example:

SELECT TRY_CAST(order_amount AS DECIMAL(18,2)) AS order_amount

FROM sales;

Performance and Governance Considerations

  • Convert as early as possible to support accurate joins/filters
  • Document transformations for transparency
  • Use consistent type conventions across the organization
  • Apply sensitivity labels appropriately — type conversion doesn’t affect security labels

Impact on Semantic Models

When creating semantic models (Power BI datasets):

  • Data types determine field behavior (e.g., date hierarchies)
  • Incorrect types can cause:
    • Incorrect aggregations
    • Misleading visuals
    • DAX errors

Always validate types before importing data into the model.

Best Practices

  • Always validate data values before conversion
  • Use schema enforcement where possible (e.g., Spark schema)
  • Avoid implicit type conversions during joins
  • Keep logs or steps of transformations for reproducibility

Key Takeaways for the DP-600 Exam

  • Know why data type conversion matters for analytics
  • Be able to choose the right tool (Power Query / SQL / Spark) for the context
  • Understand common conversions (text→numeric, text→date, boolean conversion)
  • Recognize when conversion must occur in the pipeline for correctness and performance

Practice Questions:

Here are 10 questions to test and help solidify your learning and knowledge. As you review these and other questions in your preparation, make sure to …

  • Expect scenario-based questions rather than direct definitions
  • Identifying and understand why an option is correct (or incorrect) — not just which one
  • Look for and understand the usage scenario of keywords in exam questions to guide you
  • Keep in mind that if a question mentions unexpected calculations, broken joins, or filtering issues, always consider data type mismatches as a possible root cause.

Question 1

Why is converting column data types important in an analytics solution?

A. It reduces storage costs
B. It ensures accurate calculations, joins, and filtering
C. It improves report visuals automatically
D. It encrypts sensitive data

Correct Answer: B

Explanation:
Correct data types ensure accurate aggregations, proper join behavior, correct filtering, and predictable sorting.

Question 2

Which Fabric tool provides a visual, low-code interface for changing column data types?

A. SQL Analytics endpoint
B. Spark notebooks
C. Power Query
D. Eventhouse

Correct Answer: C

Explanation:
Power Query allows users to change data types through a graphical interface and automatically records the steps.

Question 3

What is a common risk when converting text values to numeric data types?

A. Increased storage usage
B. Duplicate rows
C. Conversion errors or null values
D. Slower report rendering

Correct Answer: C

Explanation:
Text values that are not valid numbers may cause conversion failures or be converted to nulls, depending on the method used.

Question 4

Which SQL function safely attempts to convert a value and returns NULL if conversion fails?

A. CAST
B. CONVERT
C. TRY_CAST
D. FORMAT

Correct Answer: C

Explanation:
TRY_CAST avoids query failures by returning NULL when a value cannot be converted.

Question 5

When should data types ideally be converted in a Fabric analytics pipeline?

A. At report query time
B. After publishing reports
C. Early in the transformation process
D. Only in the semantic model

Correct Answer: C

Explanation:
Converting data types early prevents downstream issues in joins, aggregations, and semantic models.

Question 6

Which data type is most appropriate for calendar-based filtering and time intelligence?

A. Text
B. Integer
C. Date or DateTime
D. Boolean

Correct Answer: C

Explanation:
Date and DateTime types enable proper time-based filtering, hierarchies, and time intelligence calculations.

Question 7

Which Spark operation converts a column’s data type?

A. changeType()
B. convert()
C. cast()
D. toType()

Correct Answer: C

Explanation:
The cast() method is used in Spark to convert a column’s data type.

Question 8

Why can implicit data type conversion during joins be problematic?

A. It improves performance
B. It hides data lineage
C. It may cause incorrect matches or slow performance
D. It automatically removes duplicates

Correct Answer: C

Explanation:
Implicit conversions can prevent index usage and lead to incorrect or inefficient joins.

Question 9

A numeric column is stored as text and sorts incorrectly (e.g., 1, 10, 2). What is the cause?

A. Incorrect aggregation
B. Missing values
C. Wrong data type
D. Duplicate rows

Correct Answer: C

Explanation:
Text sorting is lexicographical, not numeric, leading to incorrect ordering.

Question 10

What is the impact of incorrect data types in a Power BI semantic model?

A. Only visuals are affected
B. Aggregations, filters, and DAX behavior may be incorrect
C. Reports fail to load
D. Sensitivity labels are removed

Correct Answer: B

Explanation:
Data types influence how fields behave in calculations, visuals, and DAX expressions.

Identify and Resolve Duplicate Data, Missing Data, or Null Values

This post is a part of the DP-600: Implementing Analytics Solutions Using Microsoft Fabric Exam Prep Hub; and this topic falls under these sections: 
Prepare data
--> Transform data
--> Identify and resolve duplicate data, missing data, or null values

Ensuring data quality is foundational for reliable analytics. Duplicate records, missing values, and nulls can lead to inaccurate aggregations, misleading insights, and broken joins. Microsoft Fabric provides multiple tools and techniques to identify, investigate, and resolve these issues during data preparation.

Why Data Quality Matters

Poor data quality can cause:

  • Incorrect business metrics (e.g., inflated counts)
  • Failed joins or mismatches
  • Incorrect aggregates or KPIs
  • Discrepancies across reports

The DP-600 exam expects you to know how to detect and fix these issues using Fabric’s transformation tools — without degrading performance or losing important data.

Key Data Quality Issues

1. Duplicate Data

Duplicates occur when the same record appears multiple times.
Common causes:

  • Repeated ingestion jobs
  • Incorrect joins
  • Source system errors

Impact of duplicates:

  • Inflated metrics
  • Misleading counts
  • Distorted analytics

2. Missing Data

Missing data refers to complete absence of expected rows for certain categories or time periods.

Examples:

  • No sales records for a specific store in a date range
  • Missing customer segments

Impact:

  • Bias in analysis
  • Understated performance

3. Null Values

Nulls represent unknown or undefined values in a dataset.

Common cases:

  • Missing customer name
  • Missing numeric values
  • Unpopulated fields in incomplete records

Consequences:

  • SQL functions may ignore nulls
  • Aggregations may be skewed
  • Joins may fail or produce incorrect results

Tools and Techniques in Microsoft Fabric

1. Power Query (Dataflows Gen2 / Lakehouse)

Power Query provides a visual and programmatic interface to clean data:

  • Remove duplicates:
    Home → Remove Rows → Remove Duplicates
  • Replace or fill nulls:
    Transform → Replace Values
    Or use Fill Up / Fill Down
  • Filter nulls:
    Filter rows where column is null or not null

Benefits:

  • No-code/low-code
  • Reusable transformation steps
  • Easy preview and validation

2. SQL (Warehouses / Lakehouse SQL Analytics)

Using SQL, you can identify and fix issues:

Detect duplicates:

SELECT Col1, Col2, COUNT(*) AS Cnt
FROM table
GROUP BY Col1, Col2
HAVING COUNT(*) > 1;

Remove duplicates (example pattern):

WITH RankedRows AS (
  SELECT *, ROW_NUMBER() OVER (PARTITION BY keycol ORDER BY keycol) AS rn
  FROM table
)
SELECT * FROM RankedRows WHERE rn = 1;

Replace nulls:

SELECT COALESCE(column, 0) AS column_fixed
FROM table;

3. Spark (Lakehouses via Notebooks)

Identify nulls:

df.filter(df["column"].isNull()).show()

Drop duplicates:

df.dropDuplicates(["keycol"])

Fill nulls:

df.na.fill({"column": "Unknown"})

Best Practices for Resolution

Addressing Duplicates

  • Use business keys (unique identifiers) to define duplicates
  • Validate whether duplicates are true duplicates or legitimate repeats
  • Document deduplication logic

Handling Nulls

  • Use domain knowledge to decide substitute values
    • Zero for numeric
    • “Unknown” or “Not Provided” for text
  • Preserve nulls when they carry meaning (e.g., missing responses)

Handling Missing Data

  • Understand the business meaning
    • Is absence valid?
    • Should data be imputed?
    • Or should missing rows be generated via reference tables?

Data Profiling

  • Use profiling to understand distributions and quality:
    • Column completeness
    • Unique value distribution
    • Null frequency

Data profiling helps you decide which cleaning steps are required.

When to Clean Data in Fabric

Data quality transformations should be performed:

  • Early in the pipeline (at the ingestion or transformation layer)
  • Before building semantic models
  • Before aggregations or joins
  • Before publishing curated datasets

Early cleaning prevents issues from propagating into semantic models and reports.

Exam Scenarios

In DP-600 exam questions, you might see scenarios like:

  • Metrics appear inflated due to duplicate records
  • Reports show missing date ranges
  • Joins fail due to null key values
  • Aggregations ignore null values

Your job is to choose the correct transformation action — e.g., filtering nulls, deduplicating, replacing values, or imputing missing data — and the best tool (Power Query vs SQL vs Spark).

Key Takeaways

  • Duplicate rows inflate counts and distort analytics.
  • Missing rows can bias time-series or segment analysis.
  • Null values can break joins and cause incorrect aggregation results.
  • Effective resolution relies on understanding business context and using the right Fabric tools.
  • Clean data early for better downstream performance and governance.

Final Exam Tip
If a metric doesn’t look right, think data quality first — missing or null values and duplicates are one of the most common real-world issues covered in DP-600 scenarios.

Practice Questions:

Here are 10 questions to test and help solidify your learning and knowledge. As you review these and other questions in your preparation, make sure to …

  • Identifying and understand why an option is correct (or incorrect) — not just which one
  • Look for and understand the usage scenario of keywords in exam questions to guide you
  • Expect scenario-based questions rather than direct definitions

General Exam Tips for this section
If something looks wrong in a report:

  • Too high? → Check for duplicates
  • Blank or missing? → Check for nulls or missing rows
  • Not joining correctly? → Check nulls and key integrity

Question 1

Which issue is most likely to cause inflated totals in aggregated metrics?

A. Null values in numeric columns
B. Missing rows for a time period
C. Duplicate records
D. Incorrect column data types

Correct Answer: C

Explanation:
Duplicate records result in the same data being counted more than once, which inflates sums, counts, and averages.

Question 2

In Power Query, which action is used to remove duplicate rows?

A. Filter Rows
B. Group By
C. Remove Duplicates
D. Replace Values

Correct Answer: C

Explanation:
The Remove Duplicates operation removes repeated rows based on selected columns.

Question 3

Which SQL function is commonly used to replace null values with a default value?

A. NULLIF
B. ISNULL or COALESCE
C. COUNT
D. CAST

Correct Answer: B

Explanation:
ISNULL() and COALESCE() return a specified value when a column contains NULL.

Question 4

Why can null values cause problems in joins?

A. Nulls increase query runtime
B. Nulls are treated as zero
C. Nulls never match other values
D. Nulls are automatically filtered

Correct Answer: C

Explanation:
NULL values do not match any value (including other NULLs), which can cause rows to be excluded from join results.

Question 5

Which scenario best justifies keeping null values rather than replacing them?

A. The column is used in joins
B. The null indicates “unknown” or “not applicable”
C. The column is numeric
D. The column has duplicates

Correct Answer: B

Explanation:
Nulls may carry important business meaning and should be preserved when they accurately represent missing or unknown information.

Question 6

Which Fabric tool is most appropriate for visual data profiling to identify missing and null values?

A. Power BI visuals
B. Power Query
C. Semantic models
D. Eventhouse

Correct Answer: B

Explanation:
Power Query provides built-in data profiling features such as column distribution, column quality, and column profile.

Question 7

What is the purpose of using an anti join when checking data quality?

A. To merge tables
B. To append data
C. To identify unmatched records
D. To replace null values

Correct Answer: C

Explanation:
Anti joins return rows that do not have a match in another table, making them ideal for identifying missing or orphaned records.

Question 8

Which approach is considered a best practice for handling data quality issues?

A. Fix issues only in reports
B. Clean data as late as possible
C. Resolve issues early in the pipeline
D. Ignore null values

Correct Answer: C

Explanation:
Resolving data quality issues early prevents them from propagating into semantic models and reports.

Question 9

Which Spark operation removes duplicate rows from a DataFrame?

A. filter()
B. groupBy()
C. dropDuplicates()
D. distinctColumns()

Correct Answer: C

Explanation:
dropDuplicates() removes duplicate rows based on one or more columns.

Question 10

A report is missing values for several dates. What is the most likely cause?

A. Duplicate rows
B. Incorrect aggregation logic
C. Missing source data
D. Incorrect data type conversion

Correct Answer: C

Explanation:
Missing dates usually indicate that source records are absent rather than null or duplicated.

Aggregate Data

This post is a part of the DP-600: Implementing Analytics Solutions Using Microsoft Fabric Exam Prep Hub; and this topic falls under these sections: 
Prepare data
--> Transform data
--> Aggregate data

Aggregating data is a foundational data transformation technique used to compute summaries and roll-ups, such as totals, averages, counts, and other statistical measures. In analytics solutions—even ones built in Microsoft Fabric—aggregation enables faster performance, simplified reporting, and clearer insights.

In the context of DP-600, you should understand why and when to aggregate data, how aggregation affects downstream analytics, and where it is implemented in Fabric workloads.

What Is Data Aggregation?

Aggregation refers to the process of summarizing detailed records into higher-level metrics. Common aggregation operations include:

  • SUM – total of a numeric field
  • COUNT / COUNT DISTINCT – number of records or unique values
  • AVG – average
  • MIN / MAX – lowest or highest value
  • GROUP BY – group records before applying aggregate functions

Aggregation turns row-level data into summary tables that are ideal for dashboards, KPIs, and trend analysis.

Why Aggregate Data?

Performance

Large detailed tables can be slow to query. Pre-aggregated data:

  • Reduces data scanned at query time
  • Improves report responsiveness

Simplicity

Aggregated data simplifies reporting logic for end users by providing ready-to-use summary metrics.

Consistency

When aggregations are standardized at the data layer, multiple reports can reuse the same durable summaries, ensuring consistent results.

When to Aggregate

Consider aggregating when:

  • Working with large detail tables (e.g., web logs, transaction history)
  • Reports require summary metrics (e.g., monthly totals, regional averages)
  • Users frequently query the same roll-ups
  • You want to offload compute from the semantic model or report layer

Where to Aggregate in Microsoft Fabric

Lakehouse

  • Use Spark SQL or SQL analytics endpoints
  • Good for large-scale transformations on big data
  • Ideal for creating summarized tables

Warehouse

  • Use T-SQL for aggregations
  • Supports highly optimized analytical queries
  • Can store aggregated tables for BI performance

Dataflows Gen2

  • Use Power Query transformations to aggregate and produce curated tables
  • Fits well in ETL/ELT pipelines

Notebooks

  • Use Spark (PySpark or SQL) for advanced or complex aggregations

Semantic Models (DAX)

  • Create aggregated measures
  • Useful for scenarios when aggregation logic must be defined at analysis time

Common Aggregation Patterns

Rollups by Time

Aggregating by day, week, month, quarter, or year:

SELECT
  YEAR(OrderDate) AS Year,
  MONTH(OrderDate) AS Month,
  SUM(SalesAmount) AS TotalSales
FROM Sales
GROUP BY
  YEAR(OrderDate),
  MONTH(OrderDate);

Aggregations with Dimensions

Combining filters and groupings:

SELECT
  Region,
  ProductCategory,
  SUM(SalesAmount) AS TotalSales,
  COUNT(*) AS OrderCount
FROM Sales
GROUP BY
  Region,
  ProductCategory;

Aggregations vs. Detailed Tables

AspectDetailed TableAggregated Table
Query flexibilityHighLower (fixed aggregates)
PerformanceLowerHigher
StorageModerateLower
BI simplicityModerateHigh

Best practice: store both detail and aggregated tables when storage and refresh times permit.

Aggregation and Semantic Models

Semantic models often benefit from pre-aggregated tables:

  • Improves report performance
  • Reduces row scans on large datasets
  • Can support composite models that combine aggregated tables with detail tables

Within semantic models:

  • Calculated measures define aggregation rules
  • Aggregated physical tables can be imported for performance

Governance and Refresh Considerations

  • Aggregated tables must be refreshed on a schedule that matches business needs.
  • Use pipelines or automation to update aggregated data regularly.
  • Ensure consistency between fact detail and aggregated summaries.
  • Document and version aggregation logic for maintainability.

Example Use Cases

Sales KPI Dashboard

  • Monthly total sales
  • Year-to-date sales
  • Average order value

Operational Reporting

  • Daily website visits by category
  • Hourly orders processed per store

Executive Scorecards

  • Quarter-to-date profits
  • Customer acquisition counts by region

Best Practices for DP-600

  • Aggregate as close to the data source as practical to improve performance
  • Use Dataflows Gen2, Lakehouse SQL, or Warehouse SQL for durable aggregated tables
  • Avoid over-aggregation that removes necessary detail for other reports
  • Use semantic model measures for dynamic aggregation needs

Key Takeaway
In DP-600 scenarios, aggregating data is about preparing analytics-ready datasets that improve performance and simplify reporting. Understand how to choose the right place and method for aggregation—whether in a lakehouse, warehouse, dataflow, or semantic model—and how that choice impacts downstream analytics.

Practice Questions:

Here are 10 questions to test and help solidify your learning and knowledge. As you review these and other questions in your preparation, make sure to …

  • Identifying and understand why an option is correct (or incorrect) — not just which one
  • Look for and understand the usage scenario of keywords in exam questions to guide you
  • Expect scenario-based questions rather than direct definitions

Question 1

What is the primary purpose of aggregating data in analytics solutions?

A. To increase data granularity
B. To reduce data quality issues
C. To summarize detailed data into meaningful metrics
D. To enforce security rules

Correct Answer: C

Explanation:
Aggregation summarizes detailed records (for example, transactions) into higher-level metrics such as totals, averages, or counts, making data easier to analyze and faster to query.

Question 2

Which SQL clause is required when using aggregate functions like SUM() or COUNT() with non-aggregated columns?

A. ORDER BY
B. GROUP BY
C. WHERE
D. HAVING

Correct Answer: B

Explanation:
GROUP BY defines how rows are grouped before aggregate functions are applied. Any non-aggregated column in the SELECT clause must appear in the GROUP BY clause.

Question 3

Which scenario is the best candidate for creating a pre-aggregated table in Microsoft Fabric?

A. Ad-hoc exploratory analysis
B. Frequently queried KPIs used across multiple reports
C. Data with unpredictable schema changes
D. Small lookup tables

Correct Answer: B

Explanation:
Pre-aggregated tables are ideal for commonly used KPIs because they improve performance and ensure consistent results across reports.

Question 4

Where can durable aggregated tables be created in Microsoft Fabric?

A. Only in semantic models
B. Only in notebooks
C. Lakehouses and warehouses
D. Power BI reports

Correct Answer: C

Explanation:
Both Lakehouses (via Spark SQL or SQL analytics endpoints) and Warehouses (via T-SQL) support persistent aggregated tables.

Question 5

Which aggregation function returns the number of unique values in a column?

A. COUNT
B. SUM
C. AVG
D. COUNT DISTINCT

Correct Answer: D

Explanation:
COUNT DISTINCT counts only unique values, which is commonly used for metrics like unique customers or unique orders.

Question 6

What is a key benefit of aggregating data before loading it into a semantic model?

A. Increased storage usage
B. Improved query performance
C. More complex DAX expressions
D. Higher data latency

Correct Answer: B

Explanation:
Pre-aggregated data reduces the number of rows scanned at query time, resulting in faster report and dashboard performance.

Question 7

Which Fabric component is best suited for performing aggregation as part of an ETL or ELT process using Power Query?

A. Notebooks
B. Dataflows Gen2
C. Eventhouses
D. Semantic models

Correct Answer: B

Explanation:
Dataflows Gen2 use Power Query and are designed for repeatable data transformations, including grouping and aggregating data.

Question 8

What is a common tradeoff when using aggregated tables instead of detailed fact tables?

A. Higher storage costs
B. Reduced data security
C. Loss of granular detail
D. Slower refresh times

Correct Answer: C

Explanation:
Aggregated tables improve performance but reduce flexibility because detailed, row-level data is no longer available.

Question 9

Which aggregation pattern is commonly used for time-based analysis?

A. GROUP BY product category
B. GROUP BY customer ID
C. GROUP BY date, month, or year
D. GROUP BY transaction ID

Correct Answer: C

Explanation:
Time-based aggregations (daily, monthly, yearly) are fundamental for trend analysis and KPI reporting.

Question 10

Which approach is considered a best practice when designing aggregated datasets for analytics?

A. Aggregate all data at the highest level only
B. Store only aggregated tables and discard detail data
C. Maintain both detailed and aggregated tables when possible
D. Avoid aggregations until the reporting layer

Correct Answer: C

Explanation:
Keeping both detail-level and aggregated tables provides flexibility while still achieving strong performance for common analytical queries.

Denormalize Data

This post is a part of the DP-600: Implementing Analytics Solutions Using Microsoft Fabric Exam Prep Hub; and this topic falls under these sections: 
Prepare data
--> Transform data
--> Denormalize Data

Data denormalization is a transformation strategy that restructures data to improve query performance and simplify analytics—especially in analytical workloads such as reporting, dashboarding, and BI. In Microsoft Fabric, denormalization plays a key role when preparing data for efficient consumption in lakehouses, warehouses, and semantic models.

This article explains what denormalization means, why it’s important for analytics, how to implement it in Fabric, and when to use it versus normalized structures.

What Is Denormalization?

Denormalization is the process of combining data from multiple tables or sources into a single, flattened structure. The goal is to reduce the number of joins and simplify querying at the expense of some redundancy.

In contrast:

  • Normalized data avoids redundancy by splitting data into many related tables.
  • Denormalized data often duplicates data intentionally to speed up analytical queries.

Why Denormalize Data for Analytics?

Denormalization is widely used in analytics because it:

  • Improves query performance: Fewer joins mean faster queries—especially for BI tools like Power BI.
  • Simplifies report logic: Flattened tables make it easier for report authors to understand and use data.
  • Reduces semantic model complexity: Fewer relationships and tables can improve both model performance and maintainability.
  • Optimizes storage access: Pre-computed joins and aggregated structures reduce run-time computation.

Beneficial Scenarios for Denormalization

Denormalization is especially helpful when:

  • Building star schemas or analytical data marts.
  • Preparing data for semantic models that are consumed by BI tools.
  • Performance is critical for dashboards and reports.
  • Data rarely changes (or changes can be managed with refresh logic).
  • Users require self-service analytics with minimal SQL complexity.

Where to Denormalize in Microsoft Fabric

Denormalization can be implemented in different Fabric components depending on workload and transformation needs:

1. Dataflows Gen2

  • Use Power Query to merge tables and create flattened structures
  • Ideal for low-code scenarios targeting OneLake
  • Great for building reusable tables

2. Lakehouses

  • Use Spark SQL or T-SQL to perform joins and build denormalized tables
  • Useful for large-scale ELT transformations

3. Warehouse

  • Use SQL to create flattened analytic tables optimized for BI
  • Supports indexing and performance tuning

4. Notebooks

  • Use PySpark or Spark SQL for complex or iterative denormalization logic

How to Denormalize Data

Typical Techniques

  • Merge or Join tables: Combine fact and dimension tables into a single analytic table
  • Pre-compute derived values: Compute metrics or concatenated fields ahead of time
  • Flatten hierarchies: Add attributes from parent tables directly into child records
  • Pivot or unpivot: Adjust layout to match analytics needs

Example (Conceptual Join)

Instead of querying these tables:

SELECT

    s.SalesID,

    d.CustomerName,

    p.ProductName

FROM FactSales s

JOIN DimCustomer d ON s.CustomerID = d.CustomerID

JOIN DimProduct p ON s.ProductID = p.ProductID;

Create a denormalized “SalesAnalytics” table:

SELECT

    s.SalesID,

    s.SalesDate,

    d.CustomerName,

    p.ProductName,

    s.SalesAmount

INTO DenormSalesAnalytics

FROM FactSales s

JOIN DimCustomer d ON s.CustomerID = d.CustomerID

JOIN DimProduct p ON s.ProductID = p.ProductID;

This single table can then be queried directly by BI tools without joins.

Trade-Offs of Denormalization

While denormalization improves performance and simplicity, it also introduces trade-offs:

Pros

  • Faster, simpler queries
  • Better analytics experience
  • Easier semantic model design

Cons

  • Data redundancy
  • Larger storage footprint
  • More complex refresh and update logic
  • Higher maintenance if source schemas change

Integrating Denormalization with Semantic Models

Denormalized tables are often used as sources for Power BI semantic models to:

  • Reduce row-level relationships
  • Improve report refresh times
  • Simplify model structure
  • Support consistent business metrics

Because semantic models work best with wide tables and straightforward relationships, denormalized sources are ideal.

Best Practices for Denormalization

  • Denormalize only where it delivers clear performance or usability benefits
  • Document transformation logic for future maintainability
  • Use pipelines or Dataflows Gen2 for repeatable and auditable ELT flows
  • Monitor refresh performance and adjust partitions or indexes

When Not to Denormalize

Avoid denormalization when:

  • Data integrity rules are strict and must avoid redundancy
  • Source systems change frequently
  • You are performing OLTP-style operations (transactional systems)
  • Storage and refresh cost outweigh performance gains

What to Know for the DP-600 Exam

You should be comfortable with:

  • The definition and purpose of denormalization
  • Recognizing when it’s appropriate in analytics workloads
  • How to implement denormalization in Fabric components
  • The trade-offs involved in denormalizing data
  • How denormalized structures optimize semantic models and BI

Final Exam Tip
If a question emphasizes reducing joins, improving query performance, and simplifying reporting, you’re likely dealing with denormalization.
If it emphasizes transactional integrity and normalized structures, that’s not the scenario for denormalization.

Practice Questions:

Here are 10 questions to test and help solidify your learning and knowledge. As you review these and other questions in your preparation, make sure to …

  • Identifying and understand why an option is correct (or incorrect) — not just which one
  • Look for and understand the usage scenario of keywords in exam questions to guide you
  • Expect scenario-based questions rather than direct definitions

1. What is the primary purpose of denormalizing data for analytics workloads?

A. Reduce data duplication
B. Improve transactional integrity
C. Improve query performance and simplify analytics
D. Enforce strict normalization rules

Correct Answer: C

Explanation:
Denormalization intentionally introduces redundancy to reduce joins, simplify queries, and improve performance—key requirements for analytics and BI workloads.

2. Which type of workload benefits most from denormalized data?

A. OLTP transaction processing
B. Real-time device telemetry ingestion
C. BI reporting and dashboarding
D. Application logging

Correct Answer: C

Explanation:
BI reporting and analytics benefit from flattened, denormalized structures because they reduce query complexity and improve performance.

3. What is a common technique used to denormalize data?

A. Normalizing dimension tables
B. Splitting wide tables into smaller ones
C. Merging multiple related tables into one
D. Removing foreign keys

Correct Answer: C

Explanation:
Denormalization commonly involves merging fact and dimension data into a single table to reduce the need for joins during querying.

4. Which Microsoft Fabric component is best suited for low-code denormalization?

A. Notebooks
B. SQL analytics endpoint
C. Dataflows Gen2
D. Eventhouse

Correct Answer: C

Explanation:
Dataflows Gen2 use Power Query to perform low-code transformations such as merging tables and creating flattened datasets.

5. What is a key trade-off introduced by denormalization?

A. Reduced query performance
B. Increased data redundancy
C. Reduced storage reliability
D. Loss of query flexibility

Correct Answer: B

Explanation:
Denormalization duplicates data across rows or tables, which increases redundancy and can complicate updates and refresh processes.

6. Why is denormalized data often used as a source for Power BI semantic models?

A. Power BI cannot handle relationships
B. Denormalized tables simplify models and improve performance
C. Semantic models require flattened data only
D. Denormalized data reduces licensing costs

Correct Answer: B

Explanation:
Flattened tables reduce the number of relationships and joins, improving performance and making semantic models easier to design and maintain.

7. In which scenario should denormalization generally be avoided?

A. Preparing a reporting data mart
B. Building a self-service analytics dataset
C. Supporting frequent transactional updates
D. Optimizing dashboard query speed

Correct Answer: C

Explanation:
Denormalization is not ideal for transactional systems where frequent updates and strict data integrity are required.

8. Where is denormalization commonly implemented in Microsoft Fabric?

A. User interface settings
B. Workspace-level permissions
C. Lakehouses, warehouses, and Dataflows Gen2
D. Real-Time hub only

Correct Answer: C

Explanation:
Denormalization is a data transformation task typically implemented in Fabric lakehouses, warehouses, notebooks, or Dataflows Gen2.

9. What is a common benefit of denormalizing data earlier in the data pipeline?

A. Reduced need for data validation
B. Improved consistency across analytics assets
C. Automatic enforcement of row-level security
D. Lower data ingestion costs

Correct Answer: B

Explanation:
Denormalizing upstream ensures that all downstream analytics assets consume the same enriched and flattened datasets, improving consistency.

10. Which phrase best indicates that denormalization is an appropriate solution?

A. “Strict transactional consistency is required”
B. “Data must be updated in real time per record”
C. “Queries require many joins and are slow”
D. “Source systems change frequently”

Correct Answer: C

Explanation:
Denormalization is commonly applied when complex joins cause performance issues and simplified querying is required.

Create Views, Functions, and Stored Procedures

This post is a part of the DP-600: Implementing Analytics Solutions Using Microsoft Fabric Exam Prep Hub; and this topic falls under these sections: 
Prepare data
--> Transform data
--> Create views, functions, and stored procedures

Creating views, functions, and stored procedures is a core data transformation and modeling skill for analytics engineers working in Microsoft Fabric. These objects help abstract complexity, improve reusability, enforce business logic, and optimize downstream analytics and reporting.

This section of the DP-600 exam focuses on when, where, and how to use these objects effectively across Fabric components such as Lakehouses, Warehouses, and SQL analytics endpoints.

Views

What are Views?

A view is a virtual table defined by a SQL query. It does not store data itself but presents data dynamically from underlying tables.

Where Views Are Used in Fabric

  • Fabric Data Warehouse
  • Lakehouse SQL analytics endpoint
  • Exposed to Power BI semantic models and other consumers

Common Use Cases

  • Simplify complex joins and transformations
  • Present curated, analytics-ready datasets
  • Enforce column-level or row-level filtering logic
  • Provide a stable schema over evolving raw data

Key Characteristics

  • Always reflect the latest data
  • Can be used like tables in SELECT statements
  • Improve maintainability and readability
  • Can support security patterns when combined with permissions

Exam Tip

Know that views are ideal for logical transformations, not heavy compute or data persistence.

Functions

What are Functions?

Functions encapsulate reusable logic and return a value or a table. They help standardize calculations and transformations across queries.

Types of Functions (SQL)

  • Scalar functions: Return a single value (e.g., formatted date, calculated metric)
  • Table-valued functions (TVFs): Return a result set that behaves like a table

Where Functions Are Used in Fabric

  • Fabric Warehouses
  • SQL analytics endpoints for Lakehouses

Common Use Cases

  • Standardized business calculations
  • Reusable transformation logic
  • Parameterized filtering or calculations
  • Cleaner and more modular SQL code

Key Characteristics

  • Improve consistency across queries
  • Can be referenced in views and stored procedures
  • May impact performance if overused in large queries

Exam Tip

Functions promote reuse and consistency, but should be used thoughtfully to avoid performance overhead.

Stored Procedures

What are Stored Procedures?

Stored procedures are precompiled SQL code blocks that can accept parameters and perform multiple operations.

Where Stored Procedures Are Used in Fabric

  • Fabric Data Warehouses
  • SQL endpoints that support procedural logic

Common Use Cases

  • Complex transformation workflows
  • Batch processing logic
  • Conditional logic and control-of-flow (IF/ELSE, loops)
  • Data loading, validation, and orchestration steps

Key Characteristics

  • Can perform multiple SQL statements
  • Can accept input and output parameters
  • Improve performance by reducing repeated compilation
  • Support automation and operational workflows

Exam Tip

Stored procedures are best for procedural logic and orchestration, not ad-hoc analytics queries.

Choosing Between Views, Functions, and Stored Procedures

ObjectBest Used For
ViewsSimplifying data access and shaping datasets
FunctionsReusable calculations and logic
Stored ProceduresComplex, parameter-driven workflows

Understanding why you would choose one over another is frequently tested on the DP-600 exam.

Integration with Power BI and Analytics

  • Views are commonly consumed by Power BI semantic models
  • Functions help ensure consistent calculations across reports
  • Stored procedures are typically part of data preparation or orchestration, not directly consumed by reports

Governance and Best Practices

  • Use clear naming conventions (e.g., vw_, fn_, sp_)
  • Document business logic embedded in SQL objects
  • Minimize logic duplication across objects
  • Apply permissions carefully to control access
  • Balance reusability with performance considerations

What to Know for the DP-600 Exam

You should be comfortable with:

  • When to use views vs. functions vs. stored procedures
  • How these objects support data transformation
  • Their role in analytics-ready data preparation
  • How they integrate with Lakehouses, Warehouses, and Power BI
  • Performance and governance implications

Practice Questions:

Here are 10 questions to test and help solidify your learning and knowledge. As you review these and other questions in your preparation, make sure to …

  • Identifying and understand why an option is correct (or incorrect) — not just which one
  • Look for and understand the usage scenario of keywords in exam questions to guide you
  • Expect scenario-based questions rather than direct definitions

1. What is the primary purpose of creating a view in a Fabric lakehouse or warehouse?

A. To permanently store transformed data
B. To execute procedural logic with parameters
C. To provide a virtual, query-based representation of data
D. To orchestrate batch data loads

Correct Answer: C

Explanation:
A view is a virtual table defined by a SQL query. It does not store data but dynamically presents data from underlying tables, making it ideal for simplifying access and shaping analytics-ready datasets.

2. Which Fabric component commonly exposes views directly to Power BI semantic models?

A. Eventhouse
B. SQL analytics endpoint
C. Dataflow Gen2
D. Real-Time hub

Correct Answer: B

Explanation:
The SQL analytics endpoint (for lakehouses and warehouses) exposes tables and views that Power BI semantic models can consume using SQL-based connectivity.

3. When should you use a scalar function instead of a view?

A. When you need to return a dataset with multiple rows
B. When you need to encapsulate reusable calculation logic
C. When you need to perform batch updates
D. When you want to persist transformed data

Correct Answer: B

Explanation:
Scalar functions are designed to return a single value and are ideal for reusable calculations such as formatting, conditional logic, or standardized metrics.

4. Which object type can return a result set that behaves like a table?

A. Scalar function
B. Stored procedure
C. Table-valued function
D. View index

Correct Answer: C

Explanation:
A table-valued function (TVF) returns a table and can be used in FROM clauses, similar to a view but with parameterization support.

5. Which scenario is the best use case for a stored procedure?

A. Creating a simplified reporting dataset
B. Applying row-level filters for security
C. Running conditional logic with multiple SQL steps
D. Exposing data to Power BI reports

Correct Answer: C

Explanation:
Stored procedures are best suited for procedural logic, including conditional branching, looping, and executing multiple SQL statements as part of a workflow.

6. Why are views commonly preferred over duplicating transformation logic in reports?

A. Views improve report rendering speed automatically
B. Views centralize and standardize transformation logic
C. Views permanently store transformed data
D. Views replace semantic models

Correct Answer: B

Explanation:
Views allow transformation logic to be defined once and reused consistently across multiple reports and consumers, improving maintainability and governance.

7. What is a potential downside of overusing functions in large SQL queries?

A. Increased storage costs
B. Reduced data freshness
C. Potential performance degradation
D. Loss of security enforcement

Correct Answer: C

Explanation:
Functions, especially scalar functions, can negatively impact query performance when used extensively on large datasets due to repeated execution per row.

8. Which object is most appropriate for parameter-driven data preparation steps in a warehouse?

A. View
B. Scalar function
C. Table
D. Stored procedure

Correct Answer: D

Explanation:
Stored procedures support parameters, control-of-flow logic, and multiple statements, making them ideal for complex, repeatable data preparation tasks.

9. How do views support governance and security in Microsoft Fabric?

A. By encrypting data at rest
B. By defining workspace-level permissions
C. By exposing only selected columns or filtered rows
D. By controlling OneLake storage access

Correct Answer: C

Explanation:
Views can limit the columns and rows exposed to users, helping implement logical data access patterns when combined with permissions and security models.

10. Which statement best describes how these objects fit into Fabric’s analytics lifecycle?

A. They replace Power BI semantic models
B. They are primarily used for real-time streaming
C. They prepare and standardize data for downstream analytics
D. They manage infrastructure-level security

Correct Answer: C

Explanation:
Views, functions, and stored procedures play a key role in transforming, standardizing, and preparing data for consumption by semantic models, reports, and analytics tools.

Ingest or Access Data as Needed

This post is a part of the DP-600: Implementing Analytics Solutions Using Microsoft Fabric Exam Prep Hub; and this topic falls under these sections: 
Prepare data
--> Get data
--> Ingest or access data as needed

A core responsibility of a Microsoft Fabric Analytics Engineer is deciding how data should be brought into Fabric—or whether it should be brought in at all. For the DP-600 exam, this topic focuses on selecting the right ingestion or access pattern based on performance, freshness, cost, and governance requirements.

Ingest vs. Access: Key Concept

Before choosing a tool or method, understand the distinction:

  • Ingest data: Physically copy data into Fabric-managed storage (OneLake)
  • Access data: Query or reference data where it already lives, without copying

The exam frequently tests your ability to choose the most appropriate option—not just a working one.

Common Data Ingestion Methods in Microsoft Fabric

1. Dataflows Gen2

Best for:

  • Low-code ingestion and transformation
  • Reusable ingestion logic
  • Business-friendly data preparation

Key characteristics:

  • Uses Power Query Online
  • Supports scheduled refresh
  • Stores results in OneLake (Lakehouse or Warehouse)
  • Ideal for centralized, governed ingestion

Exam tip:
Use Dataflows Gen2 when reuse, transformation, and governance are priorities.

2. Data Pipelines (Copy Activity)

Best for:

  • High-volume or frequent ingestion
  • Orchestration across multiple sources
  • ELT-style workflows

Key characteristics:

  • Supports many source and sink types
  • Enables scheduling, dependencies, and retries
  • Minimal transformation (primarily copy)

Exam tip:
Choose pipelines when performance and orchestration matter more than transformation.

3. Notebooks (Spark)

Best for:

  • Complex transformations
  • Data science or advanced engineering
  • Custom ingestion logic

Key characteristics:

  • Full control using Spark (PySpark, Scala, SQL)
  • Suitable for large-scale processing
  • Writes directly to OneLake

Exam tip:
Notebooks are powerful but require engineering skills—don’t choose them for simple ingestion scenarios.

Accessing Data Without Ingesting

1. OneLake Shortcuts

Best for:

  • Avoiding data duplication
  • Reusing data across workspaces
  • Accessing external storage

Key characteristics:

  • Logical reference only (no copy)
  • Supports ADLS Gen2 and Amazon S3
  • Appears native in Lakehouse tables or files

Exam tip:
Shortcuts are often the best answer when the question mentions avoiding duplication or reducing storage cost.

2. DirectQuery

Best for:

  • Near-real-time data access
  • Large datasets that cannot be imported
  • Centralized source-of-truth systems

Key characteristics:

  • Queries run against the source system
  • Performance depends on source
  • Limited modeling flexibility compared to Import

Exam tip:
Expect trade-off questions involving DirectQuery vs. Import.

3. Real-Time Access (Eventstreams / KQL)

Best for:

  • Streaming and telemetry data
  • Operational and real-time analytics

Key characteristics:

  • Event-driven ingestion
  • Supports near-real-time dashboards
  • Often discovered via Real-Time hub

Exam tip:
Use real-time ingestion when freshness is measured in seconds, not hours.

Choosing the Right Approach (Exam-Critical)

You should be able to decide based on these factors:

RequirementBest Option
Reusable ingestion logicDataflows Gen2
High-volume copyData pipelines
Complex transformationsNotebooks
Avoid duplicationOneLake shortcuts
Near real-time reportingDirectQuery / Eventstreams
Governance and trustIngestion + endorsement

Governance and Security Considerations

  • Ingested data can inherit sensitivity labels
  • Access-based methods rely on source permissions
  • Workspace roles determine who can ingest or access data
  • Endorsed datasets should be preferred for reuse

DP-600 often frames ingestion questions within a governance context.

Common Exam Scenarios

You may be asked to:

  • Choose between ingesting data or accessing it directly
  • Identify when shortcuts are preferable to ingestion
  • Select the right tool for a specific ingestion pattern
  • Balance data freshness vs. performance
  • Reduce duplication across workspaces

Best Practices to Remember

  • Ingest when performance and modeling flexibility are required
  • Access when freshness, cost, or duplication is a concern
  • Centralize ingestion logic for reuse
  • Prefer Fabric-native patterns over external tools
  • Let business requirements drive architectural decisions

Key Takeaway
For the DP-600 exam, “Ingest or access data as needed” is about making intentional, informed choices. Microsoft Fabric provides multiple ways to bring data into analytics solutions, and the correct approach depends on scale, freshness, reuse, governance, and cost. Understanding why one method is better than another is far more important than memorizing features.

Practice Questions:

Here are 10 questions to test and help solidify your learning and knowledge. As you review these and other questions in your preparation, make sure to …

  • Identifying and understand why an option is correct (or incorrect) — not just which one
  • Look for and understand the usage scenario of keywords in exam questions (for example, low code/no code, large dataset, high-volume data, reuse, complex transformations)
  • Expect scenario-based questions rather than direct definitions

Also, keep in mind that …

  • DP-600 questions often include multiple valid options, but only one that best aligns with the scenario’s constraints. Always identify and consider factors such as:
    • Data volume
    • Freshness requirements
    • Reuse and duplication concerns
    • Transformation complexity

1. What is the primary difference between ingesting data and accessing data in Microsoft Fabric?

A. Ingested data cannot be secured
B. Accessed data is always slower
C. Ingesting copies data into OneLake, while accessing queries data in place
D. Accessed data requires a gateway

Correct Answer: C

Explanation:
Ingestion physically copies data into Fabric-managed storage (OneLake), while access-based approaches query or reference data where it already exists.

2. Which option is BEST when the goal is to avoid duplicating large datasets across multiple workspaces?

A. Import mode
B. Dataflows Gen2
C. OneLake shortcuts
D. Notebooks

Correct Answer: C

Explanation:
OneLake shortcuts allow data to be referenced without copying it, making them ideal for reuse and cost control.

3. A team needs reusable, low-code ingestion logic with scheduled refresh. Which Fabric feature should they use?

A. Spark notebooks
B. Data pipelines
C. Dataflows Gen2
D. DirectQuery

Correct Answer: C

Explanation:
Dataflows Gen2 provide Power Query–based ingestion with refresh scheduling and reuse across Fabric items.

4. Which ingestion method is MOST appropriate for complex transformations requiring custom logic?

A. Dataflows Gen2
B. Copy activity in pipelines
C. OneLake shortcuts
D. Spark notebooks

Correct Answer: D

Explanation:
Spark notebooks offer full control over transformation logic and are suited for complex, large-scale processing.

5. When should DirectQuery be preferred over Import mode?

A. When the dataset is small
B. When data freshness is critical
C. When transformations are complex
D. When performance must be maximized

Correct Answer: B

Explanation:
DirectQuery is preferred when near-real-time access to data is required, even though performance depends on the source system.

6. Which Fabric component is BEST suited for orchestrating high-volume data ingestion with dependencies and retries?

A. Dataflows Gen2
B. Data pipelines
C. Semantic models
D. Power BI Desktop

Correct Answer: B

Explanation:
Data pipelines are designed for orchestration, handling large volumes of data, scheduling, and dependency management.

7. A dataset is queried infrequently but must support advanced modeling features. Which approach is most appropriate?

A. DirectQuery
B. Access via shortcut
C. Import into OneLake
D. Eventstream ingestion

Correct Answer: C

Explanation:
Import mode supports full modeling capabilities and high query performance, making it suitable even for infrequently accessed data.

8. Which scenario best fits the use of real-time ingestion methods such as Eventstreams or KQL databases?

A. Monthly financial reporting
B. Static reference data
C. IoT telemetry and operational monitoring
D. Slowly changing dimensions

Correct Answer: C

Explanation:
Real-time ingestion is designed for continuous, event-driven data such as IoT telemetry and operational metrics.

9. Why might ingesting data be preferred over accessing it directly?

A. It always reduces storage costs
B. It eliminates the need for security
C. It improves performance and modeling flexibility
D. It avoids data refresh

Correct Answer: C

Explanation:
Ingesting data into OneLake enables faster query performance and full support for modeling features.

10. Which factor is MOST important when deciding between ingesting data and accessing it?

A. The color of the dashboard
B. The number of reports
C. Business requirements such as freshness, scale, and governance
D. The Fabric region

Correct Answer: C

Explanation:
The decision to ingest or access data should be driven by business needs, including performance, freshness, cost, and governance—not technical convenience alone.