Free Practice Questions for Databricks Certified Generative AI Engineer Associate Certification

    🔄 Last checked for updates February 16th, 2026

    Study with 354 exam-style practice questions designed to help you prepare for the Databricks Certified Generative AI Engineer Associate. All questions are aligned with the latest exam guide and include detailed explanations to help you master the material.

    Start Practicing

    Random Questions

    Practice with randomly mixed questions from all topics

    Question MixAll Topics
    FormatRandom Order

    Domain Mode

    Practice questions from a specific topic area

    Exam Information

    Exam Details

    Key information about Databricks Certified Generative AI Engineer Associate

    Official study guide:

    View

    level:

    associate (intermediate)

    renewal:

    Recertification required every two years by taking the full, currently live exam

    prerequisites:

    None required; related course attendance and six months of hands-on experience are highly recommended

    delivery method:

    Online Proctored

    registration fee:

    $200

    time limit minutes:

    90 minutes

    number of questions:

    45 multiple-choice or multiple-selection items

    certification validity:

    2 years

    Exam Topics & Skills Assessed

    Skills measured (from the official study guide)

    Domain 1: Design Applications

    Subdomain 1.1: Design a prompt that elicits a specifically formatted response

    Design a prompt that elicits a specifically formatted response

    Subdomain 1.2: Select model tasks to accomplish a given business requirement

    Select model tasks to accomplish a given business requirement

    Subdomain 1.3: Select chain components for a desired model input and output

    Select chain components for a desired model input and output

    Subdomain 1.4: Translate business use case goals into a description of the desired inputs and outputs for the AI pipeline

    Translate business use case goals into a description of the desired inputs and outputs for the AI pipeline

    Subdomain 1.5: Define and order tools that gather knowledge or take actions for multi-stage reasoning

    Define and order tools that gather knowledge or take actions for multi-stage reasoning

    Domain 2: Data Preparation

    Subdomain 2.1: Apply a chunking strategy for a given document structure and model constraints

    Apply a chunking strategy for a given document structure and model constraints

    Subdomain 2.2: Filter extraneous content in source documents that degrades quality of a RAG application

    Filter extraneous content in source documents that degrades quality of a RAG application

    Subdomain 2.3: Choose the appropriate Python package to extract document content from provided source data and format.

    Choose the appropriate Python package to extract document content from provided source data and format.

    Subdomain 2.4: Define operations and sequence to write given chunked text into Delta Lake tables in Unity Catalog

    Define operations and sequence to write given chunked text into Delta Lake tables in Unity Catalog

    Subdomain 2.5: Identify needed source documents that provide necessary knowledge and quality for a given RAG application

    Identify needed source documents that provide necessary knowledge and quality for a given RAG application

    Subdomain 2.6: Identify prompt/response pairs that align with a given model task

    Identify prompt/response pairs that align with a given model task

    Subdomain 2.7: Use tools and metrics to evaluate retrieval performance

    Use tools and metrics to evaluate retrieval performance

    Subdomain 2.8: Design retrieval systems using advanced chunking strategies.

    Design retrieval systems using advanced chunking strategies.

    Subdomain 2.9: Explain the role of re-ranking in the information retrieval process.

    Explain the role of re-ranking in the information retrieval process.

    Subdomain 2.10: Apply chunking strategy for a given document structure

    Apply chunking strategy for a given document structure

    Domain 3: Application Development

    Subdomain 3.1: Create tools needed to extract data for a given data retrieval need

    Create tools needed to extract data for a given data retrieval need

    Subdomain 3.2: Select Langchain/similar tools for use in a Generative AI application.

    Select Langchain/similar tools for use in a Generative AI application.

    Subdomain 3.3: Identify how prompt formats can change model outputs and results

    Identify how prompt formats can change model outputs and results

    Subdomain 3.4: Qualitatively assess responses to identify common issues such as quality and safety

    Qualitatively assess responses to identify common issues such as quality and safety

    Subdomain 3.5: Select chunking strategy based on model & retrieval evaluation

    Select chunking strategy based on model & retrieval evaluation

    Subdomain 3.6: Augment a prompt with additional context from a user's input based on key fields, terms, and intents

    Augment a prompt with additional context from a user's input based on key fields, terms, and intents

    Subdomain 3.7: Create a prompt that adjusts an LLM's response from a baseline to a desired output

    Create a prompt that adjusts an LLM's response from a baseline to a desired output

    Subdomain 3.8: Implement LLM guardrails to prevent negative outcomes

    Implement LLM guardrails to prevent negative outcomes

    Subdomain 3.9: Write metaprompts that minimize hallucinations or leaking private data

    Write metaprompts that minimize hallucinations or leaking private data

    Subdomain 3.10: Build agent prompt templates exposing available functions

    Build agent prompt templates exposing available functions

    Subdomain 3.11: Select the best LLM based on the attributes of the application to be developed

    Select the best LLM based on the attributes of the application to be developed

    Subdomain 3.12: Select an embedding model context length based on source documents, expected queries, and optimization strategy

    Select an embedding model context length based on source documents, expected queries, and optimization strategy

    Subdomain 3.13: Select a model for from a model hub or marketplace for a task based on model metadata/model cards

    Select a model for from a model hub or marketplace for a task based on model metadata/model cards

    Subdomain 3.14: Select the best model for a given task based on common metrics generated in experiments

    Select the best model for a given task based on common metrics generated in experiments

    Subdomain 3.15: Utilize Agent Framework for developing agentic systems

    Utilize Agent Framework for developing agentic systems

    Domain 4: Assembling and Deploying Applications

    Subdomain 4.1: Code a chain using a pyfunc model with pre- and post-processing

    Code a chain using a pyfunc model with pre- and post-processing

    Subdomain 4.2: Control access to resources from model serving endpoints

    Control access to resources from model serving endpoints

    Subdomain 4.3: Code a simple chain according to requirements

    Code a simple chain according to requirements

    Subdomain 4.4: Code a simple chain using langchain

    Code a simple chain using langchain

    Subdomain 4.5: Choose the basic elements needed to create a RAG application: model flavor, embedding model, retriever, dependencies, input examples, model signature

    Choose the basic elements needed to create a RAG application: model flavor, embedding model, retriever, dependencies, input examples, model signature

    Subdomain 4.6: Register the model to Unity Catalog using MLflow

    Register the model to Unity Catalog using MLflow

    Subdomain 4.7: Sequence the steps needed to deploy an endpoint for a basic RAG application

    Sequence the steps needed to deploy an endpoint for a basic RAG application

    Subdomain 4.8: Create and query a Vector Search index

    Create and query a Vector Search index

    Subdomain 4.9: Identify how to serve an LLM application that leverages Foundation Model APIs

    Identify how to serve an LLM application that leverages Foundation Model APIs

    Subdomain 4.10: Identify resources needed to serve features for a RAG application

    Identify resources needed to serve features for a RAG application

    Subdomain 4.11: Explain the key concepts and components of Mosaic AI Vector Search

    Explain the key concepts and components of Mosaic AI Vector Search

    Subdomain 4.12: Identify batch inference workloads and apply ai_query() appropriately

    Identify batch inference workloads and apply ai_query() appropriately

    Domain 5: Governance

    Subdomain 5.1: Use masking techniques as guard rails to meet a performance objective

    Use masking techniques as guard rails to meet a performance objective

    Subdomain 5.2: Select guardrail techniques to protect against malicious user inputs to a Gen AI application

    Select guardrail techniques to protect against malicious user inputs to a Gen AI application

    Subdomain 5.3: Use legal/licensing requirements for data sources to avoid legal risk

    Use legal/licensing requirements for data sources to avoid legal risk

    Subdomain 5.4: Recommend an alternative for problematic text mitigation in a data source feeding a GenAI application

    Recommend an alternative for problematic text mitigation in a data source feeding a GenAI application

    Domain 6: Evaluation and Monitoring

    Subdomain 6.1: Select an LLM choice (size and architecture) based on a set of quantitative evaluation metrics

    Select an LLM choice (size and architecture) based on a set of quantitative evaluation metrics

    Subdomain 6.2: Select key metrics to monitor for a specific LLM deployment scenario

    Select key metrics to monitor for a specific LLM deployment scenario

    Subdomain 6.3: Evaluate model performance in a RAG application using MLflow

    Evaluate model performance in a RAG application using MLflow

    Subdomain 6.4: Use inference logging to assess deployed RAG application performance

    Use inference logging to assess deployed RAG application performance

    Subdomain 6.5: Use Databricks features to control LLM costs for RAG applications

    Use Databricks features to control LLM costs for RAG applications

    Subdomain 6.6: Use inference tables and Agent Monitoring to track a live LLM endpoint

    Use inference tables and Agent Monitoring to track a live LLM endpoint

    Subdomain 6.7: Identify evaluation judges that require ground truth

    Identify evaluation judges that require ground truth

    Subdomain 6.8: Compare the evaluation and monitoring phases of the Gen AI application life cycle

    Compare the evaluation and monitoring phases of the Gen AI application life cycle

    Techniques & products

    Databricks
    Generative AI
    LLM
    RAG applications
    Agentic systems
    Prompt engineering
    Vector Search
    Model Serving
    MLflow
    Unity Catalog
    Delta Lake
    LangChain
    Hugging Face Transformers
    Python
    Foundation Model APIs
    Mosaic AI Vector Search
    ai_query()
    Inference logging
    Agent Monitoring
    Metaprompts
    LLM guardrails
    Chunking strategies

    CertSafari is not affiliated with, endorsed by, or officially connected to Databricks Inc.. Full disclaimer