Free Practice Questions for Microsoft Certified: Azure Enterprise Data Analyst Associate (DP-750) Certification

    🔄 Last checked for updates May 3rd, 2026

    Study with 346 exam-style practice questions designed to help you prepare for the Microsoft Certified: Azure Enterprise Data Analyst Associate (DP-750). All questions are aligned with the latest exam guide and include detailed explanations to help you master the material.

    Start Practicing

    Random Questions

    Practice with randomly mixed questions from all topics

    Question MixAll Topics
    FormatRandom Order

    Domain Mode

    Practice questions from a specific topic area

    Quiz History

    Exam Details

    Key information about Microsoft Certified: Azure Enterprise Data Analyst Associate (DP-750)

    Official study guide

    View

    Question formats CertSafari offers
    • Multiple choice
    • Ordering
    • True/False
    • Fill in the blank
    level:

    Associate

    prerequisites:

    Knowledge of ingesting and transforming data using Structured Query Language (SQL) and Python. Experience with software development lifecycle (SDLC) practices, including Git. Familiarity with Microsoft Entra, Azure Data Factory, and Azure Monitor.

    target audience:

    Candidates with subject matter expertise in integrating and modeling data, building and deploying optimized pipelines, and troubleshooting and maintaining workloads in Azure Databricks. They should also have experience applying data quality and data governance best practices in Unity Catalog.

    Exam Topics & Skills Assessed

    Skills measured (from the official study guide)

    Domain 1: Set up and configure an Azure Databricks environment

    Subdomain 1.1: Select and configure compute in a workspace

    - Choose an appropriate compute type, including job compute, serverless, warehouse, classic compute, and shared compute - Configure compute performance settings, including CPU, node count, autoscaling, termination, node type, cluster size, and pooling - Configure compute feature settings, including Photon acceleration, Azure Databricks runtime/Spark version, and machine learning - Install libraries for a compute resource - Configure access permissions to a compute resource

    Subdomain 1.2: Create and organize objects in Unity Catalog

    - Apply naming conventions based on requirements, including isolation, development environment, and external sharing - Create a catalog based on requirements - Create a schema based on requirements - Create volumes based on requirements - Create tables, views, and materialized views - Implement a foreign catalog by configuring connections - Implement data definition language (DDL) operations on managed and external tables - Configure AI/BI Genie instructions for data discovery

    Domain 2: Secure and govern Unity Catalog objects

    Subdomain 2.1: Secure Unity Catalog objects

    - Grant privileges to a principal (user, service principal, or group) for securable objects in Unity Catalog - Implement table- and column-level access control and row-level security - Access Azure Key Vault secrets from within Azure Databricks - Authenticate data access by using service principals - Authenticate resource access by using managed identities

    Subdomain 2.2: Govern Unity Catalog objects

    - Create, implement, and preserve table and column definitions and descriptions for data discovery - Configure attribute-based access control (ABAC) by using tags and policies - Configure row filters and column masks - Apply data retention policies - Set up and manage data lineage tracking by using Catalog Explorer, including owner, history, dependencies, and lineage - Configure audit logging - Design and implement a secure strategy for Delta Sharing

    Domain 3: Prepare and process data

    Subdomain 3.1: Design and implement data modeling in Unity Catalog

    - Design logic for data ingestion and data source configuration, including extraction type and file type - Choose an appropriate data ingestion tool, including Lakeflow Connect, notebooks, and Azure Data Factory - Choose a data loading method, including batch and streaming - Choose a data table format, such as Parquet, Delta, CSV, JSON, or Iceberg - Design and implement a data partitioning scheme - Choose a slowly changing dimension (SCD) type - Choose granularity on a column or table based on requirements - Design and implement a temporal (history) table to record changes over time - Design and implement a clustering strategy, including liquid clustering, Z-ordering, and deletion vectors - Choose between managed and unmanaged tables

    Subdomain 3.2: Ingest data into Unity Catalog

    - Ingest data by using Lakeflow Connect, including batch and streaming - Ingest data by using notebooks, including batch and streaming - Ingest data by using SQL methods, including CREATE TABLE … AS (CTAS), CREATE OR REPLACE TABLE, and COPY INTO - Ingest data by using a change data capture (CDC) feed - Ingest data by using Spark Structured Streaming - Ingest streaming data from Azure Event Hubs - Ingest data by using Lakeflow Spark Declarative Pipelines, including Auto Loader

    Subdomain 3.3: Cleanse, transform, and load data into Unity Catalog

    - Profile data to generate summary statistics and assess data distributions - Choose appropriate column data types - Identify and resolve duplicate, missing, and null values - Transform data, including filtering, grouping, and aggregating data - Transform data by using join, union, intersect, and except operators - Transform data by denormalizing, pivoting, and unpivoting data - Load data by using merge, insert, and append operations

    Subdomain 3.4: Implement and manage data quality constraints in Unity Catalog

    - Implement validation checks, including nullability, data cardinality, and range checking - Implement data type checks - Implement schema enforcement and manage schema drift - Manage data quality with pipeline expectations in Lakeflow Spark Declarative Pipelines

    Domain 4: Deploy and maintain data pipelines and workloads

    Subdomain 4.1: Design and implement data pipelines

    - Design order of operations for a data pipeline - Choose between notebook and Lakeflow Spark Declarative Pipelines - Design task logic for Lakeflow Jobs - Design and implement error handling in data pipelines, notebooks, and jobs - Create a data pipeline by using a notebook, including precedence constraints - Create a data pipeline by using Lakeflow Spark Declarative Pipelines

    Subdomain 4.2: Implement Lakeflow Jobs

    - Create a job, including setup and configuration - Configure job triggers - Schedule a job - Configure alerts for a job - Configure automatic restarts for a job or a data pipeline

    Subdomain 4.3: Implement development lifecycle processes in Azure Databricks

    - Apply version control best practices using Git - Manage branching, pull requests, and conflict resolution - Implement a testing strategy, including unit tests, integration tests, end-to-end tests, and user acceptance testing (UAT) - Configure and package Databricks Asset Bundles - Deploy a bundle by using the Azure Databricks command-line interface (CLI) - Deploy a bundle by using REST APIs

    Subdomain 4.4: Monitor, troubleshoot, and optimize workloads in Azure Databricks

    - Monitor and manage cluster consumption to optimize performance and cost - Troubleshoot and repair issues in Lakeflow Jobs, including repair, restart, stop, and run functions - Troubleshoot and repair issues in Apache Spark jobs and notebooks, including performance tuning, resolving resource bottlenecks, and cluster restart - Investigate and resolve caching, skewing, spilling, and shuffle issues by using a Directed Acyclic Graph (DAG), the Spark UI, and query profile - Optimize Delta tables for performance and cost, including OPTIMIZE and VACUUM commands - Implement log streaming by using Log Analytics in Azure Monitor - Configure alerts by using Azure Monitor

    Techniques & products

    Azure Databricks
    Unity Catalog
    Lakeflow Connect
    Lakeflow Spark Declarative Pipelines
    Lakeflow Jobs
    Apache Spark
    Delta Lake
    Parquet
    CSV
    JSON
    Iceberg
    SQL
    Python
    Git
    Microsoft Entra
    Azure Data Factory
    Azure Monitor
    Azure Key Vault
    Managed Identities
    Service Principals
    Databricks Asset Bundles
    Azure Databricks CLI
    REST APIs
    Catalog Explorer
    Log Analytics
    Photon acceleration
    Z-ordering
    Liquid clustering
    Change Data Capture (CDC)
    Spark Structured Streaming
    Auto Loader
    Data Definition Language (DDL)
    Attribute-Based Access Control (ABAC)
    Delta Sharing
    Slowly Changing Dimension (SCD)
    User Acceptance Testing (UAT)
    Directed Acyclic Graph (DAG)
    Spark UI
    Query profile
    OPTIMIZE command
    VACUUM command

    CertSafari is not affiliated with, endorsed by, or officially connected to Microsoft Corporation. Full disclaimer