Exam Prep Hubs available on The Data Community

Below are the free Exam Prep Hubs currently available on The Data Community.
Bookmark the hubs you are interested in and use them to ensure you are fully prepared for the respective exam.

Each hub contains:

  1. The topic-by-topic (from the official study guide) coverage of the material, making it easy for you to ensure you are covering all aspects of the exam material.
  2. Practice exam questions for each section.
  3. Bonus material to help you prepare
  4. Two (2) Practice Exams with 60 questions each, along with answer keys.
  5. Links to useful resources, such as Microsoft Learn content, YouTube video series, and more.




Exam Prep Hub for AI-900: Microsoft Azure AI Fundamentals

Welcome to the one-stop hub with information for preparing for the AI-900: Microsoft Azure AI Fundamentals certification exam. The content for this exam helps you to “Demonstrate fundamental AI concepts related to the development of software and services of Microsoft Azure to create AI solutions”. Upon successful completion of the exam, you earn the Microsoft Certified: Azure AI Fundamentals certification.

This hub provides information directly here (topic-by-topic as outlined in the official study guide), links to a number of external resources, tips for preparing for the exam, practice tests, and section questions to help you prepare. Bookmark this page and use it as a guide to ensure that you are fully covering all relevant topics for the AI-900 exam and making use of as many of the resources available as possible.


Audience profile (from Microsoft’s site)

This exam is an opportunity for you to demonstrate knowledge of machine learning and AI concepts and related Microsoft Azure services. As a candidate for this exam, you should have familiarity with Exam AI-900’s self-paced or instructor-led learning material.
This exam is intended for you if you have both technical and non-technical backgrounds. Data science and software engineering experience are not required. However, you would benefit from having awareness of:
- Basic cloud concepts
- Client-server applications
You can use Azure AI Fundamentals to prepare for other Azure role-based certifications like Azure Data Scientist Associate or Azure AI Engineer Associate, but it’s not a prerequisite for any of them.

Skills measured at a glance (as specified in the official study guide)

  • Describe Artificial Intelligence workloads and considerations (15–20%)
  • Describe fundamental principles of machine learning on Azure (15–20%)
  • Describe features of computer vision workloads on Azure (15–20%)
  • Describe features of Natural Language Processing (NLP) workloads on Azure (15–20%)
  • Describe features of generative AI workloads on Azure (20–25%)
Click on each hyperlinked topic below to go to the preparation content and practice questions for that topic. Also, there are 2 practice exams provided below.

Describe Artificial Intelligence workloads and considerations (15–20%)

Identify features of common AI workloads

Identify guiding principles for responsible AI

Describe fundamental principles of machine learning on Azure (15-20%)

Identify common machine learning techniques

Describe core machine learning concepts

Describe Azure Machine Learning capabilities

Describe features of computer vision workloads on Azure (15–20%)

Identify common types of computer vision solution

Identify Azure tools and services for computer vision tasks

Describe features of Natural Language Processing (NLP) workloads on Azure (15–20%)

Identify features of common NLP Workload Scenarios

Identify Azure tools and services for NLP workloads

Describe features of generative AI workloads on Azure (20–25%)

Identify features of generative AI solutions

Identify generative AI services and capabilities in Microsoft Azure


AI-900 Practice Exams

We have provided 2 practice exams (with answer keys) to help you prepare:

AI-900 Practice Exam 1 (60 questions with answers)

AI-900 Practice Exam 2 (60 questions with answers)


Important AI-900 Resources


To Do’s:

  • Schedule time to learn, study, perform labs, and do practice exams and questions
  • Schedule the exam based on when you think you will be ready; scheduling the exam gives you a target and drives you to keep working on it; but keep in mind that it can be rescheduled based on the rules of the provider.
  • Use the various resources above to learn and prepare.
  • Take the free Microsoft Learn practice test, any other available practice tests, and do the practice questions in each section and the two practice tests available on this exam prep hub.

Good luck to you passing the AI-900: Microsoft Azure AI Fundamentals certification exam and earning the Microsoft Certified: Azure AI Fundamentals certification!

Exam Prep Hub for PL-300: Microsoft Power BI Data Analyst

Welcome to the one-stop hub with information for preparing for the PL-300: Microsoft Power BI Data Analyst certification exam. Upon successful completion of the exam, you earn the Microsoft Certified: Power BI Data Analyst Associate certification.

This hub provides information directly here (topic-by-topic), links to a number of external resources, tips for preparing for the exam, practice tests, and section questions to help you prepare. Bookmark this page and use it as a guide to ensure that you are fully covering all relevant topics for the PL-300 exam and making use of as many of the resources available as possible.


Skills tested at a glance (as specified in the official study guide)

  • Prepare the data (25–30%)
  • Model the data (25–30%)
  • Visualize and analyze the data (25–30%)
  • Manage and secure Power BI (15–20%)
Click on each hyperlinked topic below to go to the preparation content and practice questions for that topic. And there are also 2 practice exams provided below.

Prepare the data (25–30%)

Get or connect to data

Profile and clean the data

Transform and load the data

Model the data (25–30%)

Design and implement a data model

Create model calculations by using DAX

Optimize model performance

Visualize and analyze the data (25–30%)

Create reports

Enhance reports for usability and storytelling

Identify patterns and trends

Manage and secure Power BI (15–20%)

Create and manage workspaces and assets

Secure and govern Power BI items


Practice Exams

We have provided 2 practice exams (with answer keys) to help you prepare:


Important PL-300 Resources

To Do’s:

  • Schedule time to learn, study, perform labs, and do practice exams and questions
  • Schedule the exam based on when you think you will be ready; scheduling the exam gives you a target and drives you to keep working on it; but keep in mind that it can be rescheduled based on the rules of the provider.
  • Use the various resources above and below to learn
  • Take the free Microsoft Learn practice test, any other available practice tests, and do the practice questions in each section and the two practice tests available on this hub.

Good luck to you passing the PL-300: Microsoft Power BI Data Analyst certification exam and earning the Microsoft Certified: Power BI Data Analyst Associate certification!

Exam Prep Hub for DP-600: Implementing Analytics Solutions Using Microsoft Fabric

This is your one-stop hub with information for preparing for the DP-600: Implementing Analytics Solutions Using Microsoft Fabric certification exam. Upon successful completion of the exam, you earn the Fabric Analytics Engineer Associate certification.

This hub provides information directly here, links to a number of external resources, tips for preparing for the exam, practice tests, and section questions to help you prepare. Bookmark this page and use it as a guide to ensure that you are fully covering all relevant topics for the exam and using as many of the resources available as possible. We hope you find it convenient and helpful.

Why do the DP-600: Implementing Analytics Solutions Using Microsoft Fabric exam to gain the Fabric Analytics Engineer Associate certification?

Most likely, you already know why you want to earn this certification, but in case you are seeking information on its benefits, here are a few:
(1) there is a possibility for career advancement because Microsoft Fabric is a leading data platform used by companies of all sizes, all over the world, and is likely to become even more popular
(2) greater job opportunities due to the edge provided by the certification
(3) higher earnings potential,
(4) you will expand your knowledge about the Fabric platform by going beyond what you would normally do on the job and
(5) it will provide immediate credibility about your knowledge, and
(6) it may, and it should, provide you with greater confidence about your knowledge and skills.


Important DP-600 resources:


DP-600: Skills measured as of October 31, 2025:

Here you can learn in a structured manner by going through the topics of the exam one-by-one to ensure full coverage; click on each hyperlinked topic below to go to more information about it:

Skills at a glance

  • Maintain a data analytics solution (25%-30%)
  • Prepare data (45%-50%)
  • Implement and manage semantic models (25%-30%)

Maintain a data analytics solution (25%-30%)

Implement security and governance

Maintain the analytics development lifecycle

Prepare data (45%-50%)

Get Data

Transform Data

Query and analyze data

Implement and manage semantic models (25%-30%)

Design and build semantic models

Optimize enterprise-scale semantic models


Practice Exams:

We have provided 2 practice exams with answers to help you prepare.

DP-600 Practice Exam 1 (60 questions with answer key)

DP-600 Practice Exam 2 (60 questions with answer key)


Good luck to you passing the DP-600: Implementing Analytics Solutions Using Microsoft Fabric certification exam and earning the Fabric Analytics Engineer Associate certification!

What Exactly Does a Data Architect Do?

A Data Architect is responsible for designing the overall structure of an organization’s data ecosystem. While Data Engineers build pipelines and Analytics Engineers shape analytics-ready data, Data Architects define how all data systems fit together, both today and in the future.

Their work ensures that data platforms are scalable, secure, consistent, and aligned with long-term business goals.


The Core Purpose of a Data Architect

At its core, the role of a Data Architect is to:

  • Design end-to-end data architectures
  • Define standards, patterns, and best practices
  • Ensure data platforms support business and analytics needs
  • Balance scalability, performance, cost, and governance

Data Architects think in systems, not individual pipelines or reports.


Typical Responsibilities of a Data Architect

While responsibilities vary by organization, Data Architects typically work across the following areas.


Designing the Data Architecture

Data Architects define:

  • How data flows from source systems to consumption
  • The structure of data lakes, warehouses, and lakehouses
  • Integration patterns for batch, streaming, and real-time data
  • How analytics, AI, and operational systems access data

They create architectural blueprints that guide implementation.


Selecting Technologies and Platforms

Data Architects evaluate and recommend:

  • Data storage technologies
  • Integration and processing tools
  • Analytics and AI platforms
  • Metadata, governance, and security tooling

They ensure tools work together and align with strategic goals.


Establishing Standards and Patterns

Consistency is critical at scale. Data Architects define:

  • Data modeling standards
  • Naming conventions
  • Integration and transformation patterns
  • Security and access control frameworks

These standards reduce complexity and technical debt over time.


Ensuring Security, Privacy, and Compliance

Data Architects work closely with security and governance teams to:

  • Design access control models
  • Support regulatory requirements
  • Protect sensitive and regulated data
  • Enable auditing and lineage

Security and compliance are designed into the architecture—not added later.


Supporting Analytics, AI, and Self-Service

A well-designed architecture enables:

  • Reliable analytics and reporting
  • Scalable AI and machine learning workloads
  • Consistent metrics and semantic layers
  • Self-service analytics without chaos

Data Architects ensure the platform supports current and future use cases.


Common Tools Used by Data Architects

While Data Architects are less tool-focused than engineers, they commonly work with:

  • Cloud Data Platforms
  • Data Warehouses, Lakes, and Lakehouses
  • Integration and Streaming Technologies
  • Metadata, Catalog, and Lineage Tools
  • Security and Identity Systems
  • Architecture and Modeling Tools

The focus is on fit and integration, not day-to-day development.


What a Data Architect Is Not

Clarifying this role helps prevent confusion.

A Data Architect is typically not:

  • A data engineer writing daily pipeline code
  • A BI developer building dashboards
  • A data scientist training models
  • A purely theoretical designer disconnected from implementation

They work closely with implementation teams but operate at a higher level.


What the Role Looks Like Day-to-Day

A typical day for a Data Architect may include:

  • Reviewing or designing architectural diagrams
  • Evaluating new technologies or platforms
  • Aligning with stakeholders on future needs
  • Defining standards or reference architectures
  • Advising teams on design decisions
  • Reviewing implementations for architectural alignment

The role balances strategy and execution.


How the Role Evolves Over Time

As organizations mature, the Data Architect role evolves:

  • From point solutions → cohesive platforms
  • From reactive design → proactive strategy
  • From tool selection → ecosystem orchestration
  • From technical focus → business alignment

Senior Data Architects often shape enterprise data strategy.


Why Data Architects Are So Important

Data Architects add value by:

  • Preventing fragmented and brittle data ecosystems
  • Reducing long-term cost and complexity
  • Enabling scalability and innovation
  • Ensuring data platforms can evolve with the business

They help organizations avoid rebuilding their data foundations every few years.


Final Thoughts

A Data Architect’s job is not to choose tools—it is to design a data ecosystem that can grow, adapt, and endure.

When Data Architects do their work well, data teams move faster, platforms remain stable, and organizations can confidently build analytics and AI capabilities on top of a solid foundation.

What Exactly Does a BI Developer Do?

A BI (Business Intelligence) Developer focuses on designing, building, and optimizing dashboards, reports, and semantic models that deliver insights to business users. While Data Analysts focus on analysis and interpretation, BI Developers focus on how insights are packaged, delivered, and consumed at scale.

BI Developers ensure that data is not only accurate—but also usable, intuitive, and performant for decision-makers.


The Core Purpose of a BI Developer

At its core, the role of a BI Developer is to:

  • Turn data into clear, usable dashboards and reports
  • Design semantic models that support consistent metrics
  • Optimize performance and usability
  • Enable data consumption across the organization

BI Developers focus on the last mile of analytics.


Typical Responsibilities of a BI Developer

While responsibilities vary by organization, BI Developers typically work across the following areas.


Designing Dashboards and Reports

BI Developers:

  • Translate business requirements into visual designs
  • Choose appropriate charts and layouts
  • Focus on clarity, usability, and storytelling
  • Design for different audiences (executives, managers, operators)

Good BI design reduces cognitive load and increases insight adoption.


Building and Maintaining Semantic Models

BI Developers often:

  • Define relationships, measures, and calculations
  • Implement business logic in semantic layers
  • Optimize models for performance and reuse
  • Ensure metric consistency across reports

This layer is critical for trusted analytics.


Optimizing Performance and Scalability

BI Developers:

  • Improve query performance
  • Reduce unnecessary complexity in reports
  • Manage aggregations and caching strategies
  • Balance flexibility with performance

Slow or unreliable dashboards quickly lose trust.


Enabling Self-Service Analytics

By building reusable models and templates, BI Developers:

  • Empower users to build their own reports
  • Reduce duplication and rework
  • Provide guardrails for self-service
  • Support governance without limiting agility

They play a key role in self-service success.


Collaborating Across Data Teams

BI Developers work closely with:

  • Data Analysts on requirements and insights
  • Analytics Engineers on data models
  • Data Engineers on performance and data availability
  • Data Architects on standards and platform alignment

They often act as a bridge between technical teams and business users.


Common Tools Used by BI Developers

BI Developers typically work with:

  • BI & Data Visualization Tools
  • Semantic Modeling and Metrics Layers
  • SQL for validation and analysis
  • DAX or Similar Expression Languages
  • Performance Tuning and Monitoring Tools
  • Collaboration and Sharing Platforms

The focus is on usability, performance, and trust.


What a BI Developer Is Not

Clarifying boundaries helps avoid role confusion.

A BI Developer is typically not:

  • A data engineer building ingestion pipelines
  • A data scientist creating predictive models
  • A purely business-facing analyst
  • A graphic designer focused only on aesthetics

They combine technical skill with analytical and design thinking.


What the Role Looks Like Day-to-Day

A typical day for a BI Developer may include:

  • Designing or refining dashboards
  • Validating metrics and calculations
  • Optimizing report performance
  • Responding to user feedback
  • Supporting self-service users
  • Troubleshooting data or visualization issues

Much of the work is iterative and user-driven.


How the Role Evolves Over Time

As organizations mature, the BI Developer role evolves:

  • From static reports → interactive analytics
  • From individual dashboards → standardized platforms
  • From report builders → analytics product owners
  • From reactive fixes → proactive design and governance

Senior BI Developers often lead analytics UX and standards.


Why BI Developers Are So Important

BI Developers add value by:

  • Making insights accessible and actionable
  • Improving adoption of analytics
  • Ensuring consistency and trust
  • Scaling analytics across diverse audiences

They turn data into something people actually use.


Final Thoughts

A BI Developer’s job is not just to build dashboards—it is to design experiences that help people understand and act on data.

When BI Developers do their job well, analytics becomes intuitive, trusted, and embedded into everyday decision-making.

What Exactly Does a Machine Learning Engineer Do?

A Machine Learning (ML) Engineer is responsible for turning machine learning models into reliable, scalable, production-grade systems. While Data Scientists focus on model development and experimentation, ML Engineers focus on deployment, automation, performance, and lifecycle management.

Their work ensures that models deliver real business value beyond notebooks and prototypes.


The Core Purpose of a Machine Learning Engineer

At its core, the role of a Machine Learning Engineer is to:

  • Productionize machine learning models
  • Build scalable and reliable ML systems
  • Automate training, deployment, and monitoring
  • Ensure models perform well in real-world conditions

ML Engineers sit at the intersection of software engineering, data engineering, and machine learning.


Typical Responsibilities of a Machine Learning Engineer

While responsibilities vary by organization, ML Engineers typically work across the following areas.


Deploying and Serving Machine Learning Models

ML Engineers:

  • Package models for production
  • Deploy models as APIs or batch jobs
  • Manage model versions and rollouts
  • Ensure low latency and high availability

This is where ML becomes usable by applications and users.


Building ML Pipelines and Automation

ML Engineers design and maintain:

  • Automated training pipelines
  • Feature generation and validation workflows
  • Continuous integration and deployment (CI/CD) for ML
  • Scheduled retraining processes

Automation is critical for scaling ML across use cases.


Monitoring and Maintaining Models in Production

Once deployed, ML Engineers:

  • Monitor model performance and drift
  • Track data quality and feature distributions
  • Detect bias, degradation, or failures
  • Trigger retraining or rollback when needed

Models are living systems, not one-time deployments.


Optimizing Performance and Reliability

ML Engineers focus on:

  • Model inference speed and scalability
  • Resource usage and cost optimization
  • Fault tolerance and resiliency
  • Security and access control

Production ML must meet engineering standards.


Collaborating Across Teams

ML Engineers work closely with:

  • Data Scientists on model design and validation
  • Data Engineers on data pipelines and feature stores
  • AI Engineers on broader AI systems
  • Software Engineers on application integration
  • Data Architects on platform design

They translate research into production systems.


Common Tools Used by Machine Learning Engineers

ML Engineers commonly work with:

  • Machine Learning Frameworks
  • Model Serving and API Frameworks
  • ML Platforms and Pipelines
  • Feature Stores
  • Monitoring and Observability Tools
  • Cloud Infrastructure and Containers

Tool choice is driven by scalability, reliability, and maintainability.


What a Machine Learning Engineer Is Not

Clarifying this role helps avoid confusion.

A Machine Learning Engineer is typically not:

  • A data analyst creating reports
  • A data scientist focused only on experimentation
  • A general software engineer with no ML context
  • A research scientist working on novel algorithms

Their focus is operational ML.


What the Role Looks Like Day-to-Day

A typical day for a Machine Learning Engineer may include:

  • Deploying or updating models
  • Reviewing training or inference pipelines
  • Monitoring production performance
  • Investigating model or data issues
  • Improving automation and reliability
  • Collaborating on new ML use cases

Much of the work happens after the model is built.


How the Role Evolves Over Time

As organizations mature, the ML Engineer role evolves:

  • From manual deployments → automated MLOps
  • From isolated models → shared ML platforms
  • From single use cases → enterprise ML systems
  • From reactive fixes → proactive optimization

Senior ML Engineers often lead ML platform and MLOps strategy.


Why Machine Learning Engineers Are So Important

ML Engineers add value by:

  • Bridging the gap between research and production
  • Making ML reliable and scalable
  • Reducing operational risk
  • Enabling faster delivery of AI-powered features

Without ML Engineers, many ML initiatives fail to reach production.


Final Thoughts

A Machine Learning Engineer’s job is not to invent new models—it is to make machine learning work reliably in the real world.

When ML Engineers do their job well, organizations can confidently deploy, scale, and trust machine learning systems as part of everyday operations.

Practice Questions: Identify Document Processing Workloads (AI-900 Exam Prep)

Practice Questions


Question 1

A finance team wants to automatically extract the invoice number, vendor name, and total amount from scanned PDF invoices.

Which AI workload is required?

A. Natural language processing
B. Computer vision
C. Document processing
D. Speech recognition

Correct Answer: C

Explanation: Document processing is designed to extract structured fields and data from documents such as invoices and PDFs.


Question 2

An organization wants to digitize thousands of paper forms by converting printed text into machine-readable text.

Which capability is required first?

A. Sentiment analysis
B. Optical Character Recognition (OCR)
C. Text classification
D. Language translation

Correct Answer: B

Explanation: OCR extracts printed or handwritten text from scanned documents and images, enabling further processing.


Question 3

A company processes expense receipts and needs to extract dates, merchant names, totals, and line items.

Which Azure AI service is most appropriate?

A. Azure AI Vision
B. Azure AI Language
C. Azure AI Document Intelligence
D. Azure AI Bot Service

Correct Answer: C

Explanation: Azure AI Document Intelligence (formerly Form Recognizer) is designed for receipt, invoice, and form processing.


Question 4

A business wants to extract rows and columns from tables embedded in scanned reports.

Which document processing capability is required?

A. Image classification
B. Table extraction
C. Sentiment analysis
D. Language detection

Correct Answer: B

Explanation: Table extraction identifies and extracts structured tabular data from documents.


Question 5

A healthcare provider wants to process standardized patient intake forms and store field values in a database.

Which workload best fits this scenario?

A. Computer vision only
B. Natural language processing
C. Document processing with form extraction
D. Speech AI

Correct Answer: C

Explanation: Form extraction is a document processing workload that captures structured key-value pairs from standardized forms.


Question 6

Which scenario most clearly represents a document processing workload?

A. Detecting objects in security camera footage
B. Translating chat messages between languages
C. Extracting contract terms from scanned agreements
D. Converting speech recordings to text

Correct Answer: C

Explanation: Extracting structured information from scanned contracts is a classic document processing use case.


Question 7

A system extracts handwritten notes from scanned documents.

Which capability enables this?

A. Language detection
B. Handwritten text recognition
C. Image tagging
D. Sentiment analysis

Correct Answer: B

Explanation: Handwritten text recognition is part of document processing and OCR capabilities.


Question 8

Which clue in a scenario most strongly indicates a document processing workload?

A. Audio recordings are analyzed
B. Photos are classified into categories
C. Structured data is extracted from PDFs or forms
D. Customer reviews are summarized

Correct Answer: C

Explanation: Document processing focuses on extracting structured information from documents such as PDFs, forms, and invoices.


Question 9

A developer only needs to read plain text from an image without extracting structured fields.

Which Azure AI service is sufficient?

A. Azure AI Document Intelligence
B. Azure AI Language
C. Azure AI Vision
D. Azure AI Bot Service

Correct Answer: C

Explanation: Azure AI Vision provides basic OCR capabilities suitable for simple text extraction from images.


Question 10

An organization wants to ensure responsible use of AI when processing documents that contain personal data.

Which consideration is most relevant?

A. Image resolution
B. Bounding box accuracy
C. Data privacy and access control
D. Model training speed

Correct Answer: C

Explanation: Document processing often involves sensitive information, making privacy and data protection critical considerations.


Final Exam Tip

If a scenario involves forms, invoices, receipts, contracts, PDFs, or extracting structured data from documents, the correct choice is almost always a document processing workload, commonly using Azure AI Document Intelligence.


Go to the PL-300 Exam Prep Hub main page.

Practice Questions: Identify Natural Language Processing Workloads (AI-900 Exam Prep)

Practice Questions


Question 1

A company wants to automatically determine whether customer reviews are positive, negative, or neutral.

Which AI workload is required?

A. Text classification
B. Sentiment analysis
C. Language translation
D. Speech recognition

Correct Answer: B

Explanation: Sentiment analysis evaluates the emotional tone of text, such as opinions expressed in customer reviews.


Question 2

An organization needs to route incoming support emails to the correct department based on their content.

Which NLP capability best fits this scenario?

A. Key phrase extraction
B. Text summarization
C. Text classification
D. Language detection

Correct Answer: C

Explanation: Text classification assigns predefined labels or categories to text, making it ideal for routing emails by topic.


Question 3

A legal team wants to quickly identify names of people, organizations, and locations within long contracts.

Which NLP capability should be used?

A. Sentiment analysis
B. Named entity recognition
C. Text translation
D. Optical character recognition

Correct Answer: B

Explanation: Named entity recognition (NER) extracts structured entities such as people, organizations, and locations from unstructured text.


Question 4

A global company wants to translate product descriptions from English into multiple languages while preserving meaning.

Which AI workload is most appropriate?

A. Language detection
B. Text summarization
C. Language translation
D. Speech synthesis

Correct Answer: C

Explanation: Language translation converts text from one language to another while maintaining its original intent and meaning.


Question 5

An application needs to identify the main topics discussed in thousands of customer feedback messages.

Which NLP capability should be used?

A. Sentiment analysis
B. Key phrase extraction
C. Text classification
D. Question answering

Correct Answer: B

Explanation: Key phrase extraction highlights the most important concepts and terms within text.


Question 6

A chatbot answers common customer questions using a natural conversational interface.

Which AI workload does this represent?

A. Computer vision
B. Conversational AI / NLP
C. Speech AI only
D. Anomaly detection

Correct Answer: B

Explanation: Conversational AI uses NLP to understand user intent and generate natural language responses.


Question 7

A system must determine the language of incoming customer messages before processing them further.

Which NLP capability is required?

A. Text classification
B. Language detection
C. Named entity recognition
D. Text summarization

Correct Answer: B

Explanation: Language detection identifies the language used in a text sample.


Question 8

Which input type most strongly indicates a natural language processing workload?

A. Video streams
B. Audio recordings
C. Images and photos
D. Text documents

Correct Answer: D

Explanation: NLP workloads are centered on understanding and generating text-based data.


Question 9

A manager wants a short summary of long meeting transcripts to quickly understand key points.

Which NLP capability should be used?

A. Text summarization
B. Sentiment analysis
C. Language detection
D. Text classification

Correct Answer: A

Explanation: Text summarization condenses long text into a shorter, meaningful summary.


Question 10

An organization wants to ensure responsible use of AI when analyzing employee emails.

Which consideration is most relevant for NLP workloads?

A. Image resolution
B. Model latency
C. Data privacy and bias
D. Bounding box accuracy

Correct Answer: C

Explanation: NLP systems can introduce bias and raise privacy concerns when processing personal or sensitive text data.


Final Exam Tip

If a scenario focuses on understanding, classifying, translating, summarizing, or responding to text, it is almost always a natural language processing workload.


Go to the PL-300 Exam Prep Hub main page.

Practice Questions: Identify Computer Vision Workloads (AI-900 Exam Prep)

Practice Questions


Question 1

A retail company wants to automatically assign categories such as shirt, shoes, or hat to product photos uploaded by sellers.

Which type of AI workload is this?

A. Natural language processing
B. Image classification
C. Object detection
D. Anomaly detection

Correct Answer: B

Explanation: Image classification assigns one or more labels to an entire image. In this scenario, each product photo is classified into a category.


Question 2

A city uses traffic cameras to identify vehicles and pedestrians and draw boxes around them in each image.

Which computer vision capability is being used?

A. Image tagging
B. Image classification
C. Object detection
D. OCR

Correct Answer: C

Explanation: Object detection identifies multiple objects within an image and locates them using bounding boxes.


Question 3

A company wants to extract text from scanned invoices and store the text in a database for searching.

Which computer vision workload is required?

A. Image description
B. Optical Character Recognition (OCR)
C. Face detection
D. Language translation

Correct Answer: B

Explanation: OCR is used to extract printed or handwritten text from images or scanned documents.


Question 4

An application analyzes photos and generates captions such as “A group of people standing on a beach.”

Which computer vision capability is this?

A. Image classification
B. Image tagging and description
C. Object detection
D. Video analysis

Correct Answer: B

Explanation: Image tagging and description focuses on understanding the overall content of an image and generating descriptive text.


Question 5

A security system needs to determine whether a human face is present in images captured at building entrances.

Which workload is most appropriate?

A. Facial recognition
B. Face detection
C. Image classification
D. Speech recognition

Correct Answer: B

Explanation: Face detection determines whether a face exists in an image. Identity verification (facial recognition) is not the focus of AI-900.


Question 6

A media company wants to analyze recorded videos to identify scenes, objects, and motion over time.

Which Azure AI workload does this represent?

A. Image classification
B. Video analysis
C. OCR
D. Text analytics

Correct Answer: B

Explanation: Video analysis processes visual data across multiple frames, enabling object detection, motion tracking, and scene analysis.


Question 7

A manufacturing company wants to detect defective products by locating scratches or dents in photos taken on an assembly line.

Which computer vision workload should be used?

A. Image classification
B. Object detection
C. Anomaly detection
D. Natural language processing

Correct Answer: B

Explanation: Object detection can be used to locate defects within an image by identifying specific problem areas.


Question 8

A developer needs to train a model using their own labeled images because prebuilt vision models are not sufficient.

Which Azure AI service is most appropriate?

A. Azure AI Vision
B. Azure AI Video Indexer
C. Azure AI Custom Vision
D. Azure AI Language

Correct Answer: C

Explanation: Azure AI Custom Vision allows users to train custom image classification and object detection models using their own data.


Question 9

Which clue in a scenario most strongly indicates a computer vision workload?

A. Audio recordings are analyzed
B. Large amounts of numerical data are processed
C. Images or videos are the primary input
D. Text documents are translated

Correct Answer: C

Explanation: Computer vision workloads always involve visual input such as images or video.


Question 10

An organization wants to ensure responsible use of AI when analyzing images of people.

Which consideration is most relevant for computer vision workloads?

A. Query performance tuning
B. Data normalization
C. Privacy and consent
D. Indexing strategies

Correct Answer: C

Explanation: Privacy, consent, and bias are key responsible AI considerations when working with images and facial data.


Final Exam Tip

If a question mentions photos, images, scanned documents, cameras, or video, think computer vision first, then determine the specific capability (classification, detection, OCR, or description).


Go to the PL-300 Exam Prep Hub main page.

Identify Natural Language Processing Workloads (AI-900 Exam Prep)

Overview

Natural Language Processing (NLP) is a branch of Artificial Intelligence (AI) that enables machines to understand, interpret, and generate human language. For the AI-900: Microsoft Azure AI Fundamentals exam, the goal is not to build language models, but to recognize NLP workloads, understand what problems they solve, and identify when NLP is the correct AI approach.

This topic appears under:

  • Describe Artificial Intelligence workloads and considerations (15–20%)
    • Identify features of common AI workloads

Most exam questions will be scenario-based, asking you to choose the correct AI workload based on how text is used.


What Is a Natural Language Processing Workload?

A natural language processing workload involves analyzing or generating language in written or spoken form (after speech has been converted to text).

NLP workloads typically:

  • Process unstructured text
  • Extract meaning, sentiment, or intent
  • Translate between languages
  • Generate human-like text responses

Common inputs:

  • Emails, chat messages, documents
  • Social media posts
  • Customer reviews
  • Transcribed speech

Common outputs:

  • Sentiment scores
  • Extracted keywords or entities
  • Translated text
  • Generated responses or summaries

Common Natural Language Processing Use Cases

On the AI-900 exam, NLP workloads are presented through everyday business scenarios. The following are the most important ones to recognize.

Text Classification

What it does: Categorizes text into predefined labels.

Example scenarios:

  • Classifying emails as spam or not spam
  • Routing support tickets by topic
  • Detecting abusive or inappropriate content

Key idea: The system assigns one or more labels to a piece of text.


Sentiment Analysis

What it does: Determines the emotional tone of text.

Example scenarios:

  • Analyzing customer reviews to see if feedback is positive or negative
  • Monitoring social media reactions to a product launch

Key idea: Sentiment analysis focuses on opinion and emotion, not topic.


Key Phrase Extraction

What it does: Identifies the main concepts discussed in a document.

Example scenarios:

  • Summarizing customer feedback
  • Highlighting important terms in legal or technical documents

Key idea: Key phrases help quickly understand what a document is about.


Named Entity Recognition (NER)

What it does: Identifies and categorizes entities in text.

Common entity types:

  • People
  • Organizations
  • Locations
  • Dates and numbers

Example scenarios:

  • Extracting company names from contracts
  • Identifying people and places in news articles

Language Detection

What it does: Identifies the language used in a text sample.

Example scenarios:

  • Detecting the language of customer messages before translation
  • Routing requests to region-specific support teams

Language Translation

What it does: Converts text from one language to another.

Example scenarios:

  • Translating product descriptions for global audiences
  • Providing multilingual customer support

Key idea: This workload focuses on preserving meaning, not word-for-word translation.


Question Answering and Conversational AI

What it does: Understands user questions and generates relevant responses.

Example scenarios:

  • Customer support chatbots
  • FAQ systems
  • Virtual assistants

Key idea: The system interprets intent and responds in natural language.


Text Summarization

What it does: Condenses long documents into shorter summaries.

Example scenarios:

  • Summarizing reports or meeting notes
  • Highlighting key points from articles

Azure Services Commonly Associated with NLP

For AI-900, you should recognize these services at a conceptual level.

Azure AI Language

Supports:

  • Sentiment analysis
  • Text classification
  • Key phrase extraction
  • Named entity recognition
  • Language detection
  • Summarization

This is the primary service referenced for NLP workloads on the exam.


Azure AI Translator

Supports:

  • Text translation between languages

Used specifically when scenarios mention multilingual translation.


Azure AI Bot Service

Supports:

  • Conversational AI solutions

Often appears alongside NLP services when building chatbots.


How NLP Differs from Other AI Workloads

Distinguishing NLP from other workloads is a common exam requirement.

AI Workload TypePrimary Input
Natural Language ProcessingText
Speech AIAudio
Computer VisionImages and video
Anomaly DetectionNumerical or time-series data

Exam tip: If the data is text-based and the goal is to understand meaning, sentiment, or intent, it is an NLP workload.


Responsible AI Considerations

NLP systems can introduce risks if not used responsibly.

Key considerations include:

  • Bias in language models
  • Offensive or harmful content generation
  • Data privacy when analyzing personal communications

AI-900 tests awareness, not mitigation techniques.


Exam Tips for Identifying NLP Workloads

  • Look for keywords like text, email, message, document, review, chat
  • Identify the goal: classify, analyze sentiment, extract meaning, translate, or respond
  • Ignore implementation details—focus on what problem is being solved
  • Choose the simplest AI workload that meets the scenario

Summary

For the AI-900 exam, you should be able to:

  • Recognize when a scenario represents a natural language processing workload
  • Identify common NLP use cases and capabilities
  • Associate NLP scenarios with Azure AI Language and related services
  • Distinguish NLP from speech, vision, and other AI workloads

A solid understanding of NLP workloads will significantly improve your confidence across multiple exam questions.


Go to the Practice Exam Questions for this topic.

Go to the PL-300 Exam Prep Hub main page.

Identify Computer Vision Workloads (AI-900 Exam Prep)

Overview

Computer vision is a branch of Artificial Intelligence (AI) that enables machines to interpret, analyze, and understand visual information such as images and videos. In the context of the AI-900: Microsoft Azure AI Fundamentals exam, you are not expected to build complex models or write code. Instead, the focus is on recognizing computer vision workloads, understanding what problems they solve, and knowing which Azure AI services are appropriate for each scenario.

This topic falls under:

  • Describe Artificial Intelligence workloads and considerations (15–20%)
    • Identify features of common AI workloads

A strong conceptual understanding here will help you confidently answer many scenario-based exam questions.


What Is a Computer Vision Workload?

A computer vision workload involves extracting meaningful insights from visual data. These workloads allow systems to:

  • Identify objects, people, or text in images
  • Analyze facial features or emotions
  • Understand the content of photos or videos
  • Detect changes, anomalies, or motion

Common inputs include:

  • Images (JPEG, PNG, etc.)
  • Video streams (live or recorded)

Common outputs include:

  • Labels or tags
  • Bounding boxes around detected objects
  • Extracted text
  • Descriptions of image content

Common Computer Vision Use Cases

On the AI-900 exam, computer vision workloads are usually presented as real-world scenarios. Below are the most common ones you should recognize.

Image Classification

What it does: Assigns a category or label to an image.

Example scenarios:

  • Determining whether an image contains a cat, dog, or bird
  • Classifying products in an online store
  • Identifying whether a photo shows food, people, or scenery

Key idea: The entire image is classified as one or more categories.


Object Detection

What it does: Detects and locates multiple objects within an image.

Example scenarios:

  • Detecting cars, pedestrians, and traffic signs in street images
  • Counting people in a room
  • Identifying damaged items in a warehouse

Key idea: Unlike classification, object detection identifies where objects appear using bounding boxes.


Face Detection and Facial Analysis

What it does: Detects human faces and analyzes facial attributes.

Example scenarios:

  • Detecting whether a face is present in an image
  • Estimating age or emotion
  • Identifying facial landmarks (eyes, nose, mouth)

Important exam note:

  • AI-900 focuses on face detection and analysis, not facial recognition for identity verification.
  • Be aware of ethical and privacy considerations when working with facial data.

Optical Character Recognition (OCR)

What it does: Extracts printed or handwritten text from images and documents.

Example scenarios:

  • Reading text from scanned documents
  • Extracting information from receipts or invoices
  • Recognizing license plate numbers

Key idea: OCR turns unstructured visual text into machine-readable text.


Image Description and Tagging

What it does: Generates descriptive text or tags that summarize image content.

Example scenarios:

  • Automatically tagging photos in a digital library
  • Creating alt text for accessibility
  • Generating captions for images

Key idea: This workload focuses on understanding the overall context of an image rather than specific objects.


Video Analysis

What it does: Analyzes video content frame by frame.

Example scenarios:

  • Detecting motion or anomalies in security footage
  • Tracking objects over time
  • Summarizing video content

Key idea: Video analysis extends image analysis across time, not just a single frame.


Azure Services Commonly Associated with Computer Vision

For the AI-900 exam, you should recognize which Azure AI services support computer vision workloads at a high level.

Azure AI Vision

Supports:

  • Image analysis
  • Object detection
  • OCR
  • Face detection
  • Image tagging and description

This is the most commonly referenced service for computer vision scenarios on the exam.


Azure AI Custom Vision

Supports:

  • Custom image classification
  • Custom object detection

Used when prebuilt models are not sufficient and you need to train a model using your own images.


Azure AI Video Indexer

Supports:

  • Video analysis
  • Object, face, and scene detection in videos

Typically appears in scenarios involving video content.


How Computer Vision Differs from Other AI Workloads

Understanding what is not computer vision is just as important on the exam.

AI Workload TypeFocus Area
Computer VisionImages and videos
Natural Language ProcessingText and speech
Speech AIAudio and voice
Anomaly DetectionPatterns in numerical or time-series data

Exam tip: If the input data is visual (images or video), you are almost certainly dealing with a computer vision workload.


Responsible AI Considerations

Microsoft emphasizes responsible AI, and AI-900 includes high-level awareness of these principles.

For computer vision workloads, key considerations include:

  • Privacy and consent when capturing images or video
  • Avoiding bias in facial analysis
  • Transparency in how visual data is collected and used

You will not be tested on implementation details, but you may see conceptual questions about ethical use.


Exam Tips for Identifying Computer Vision Workloads

  • Focus on keywords like image, photo, video, camera, scanned document
  • Look for actions such as detect, recognize, classify, extract text
  • Match the scenario to the simplest appropriate workload
  • Remember: AI-900 tests understanding, not coding

Summary

To succeed on the AI-900 exam, you should be able to:

  • Recognize when a problem is a computer vision workload
  • Identify common use cases such as image classification, object detection, and OCR
  • Understand which Azure AI services are commonly used
  • Distinguish computer vision from other AI workloads

Mastering this topic will give you a strong foundation for many questions in the Describe Artificial Intelligence workloads and considerations domain.


Go to the Practice Exam Questions for this topic.

Go to the PL-300 Exam Prep Hub main page.