Free Practice Questions for Databricks Certified Generative AI Engineer Associate Certification
Study with 354 exam-style practice questions designed to help you prepare for the Databricks Certified Generative AI Engineer Associate. All questions are aligned with the latest exam guide and include detailed explanations to help you master the material.
Start Practicing
Random Questions
Practice with randomly mixed questions from all topics
Domain Mode
Practice questions from a specific topic area
Exam Information
Exam Details
Key information about Databricks Certified Generative AI Engineer Associate
associate (intermediate)
Recertification required every two years by taking the full, currently live exam
None required; related course attendance and six months of hands-on experience are highly recommended
Online Proctored
$200
90 minutes
45 multiple-choice or multiple-selection items
2 years
Exam Topics & Skills Assessed
Skills measured (from the official study guide)
Domain 1: Design Applications
Subdomain 1.1: Design a prompt that elicits a specifically formatted response
Design a prompt that elicits a specifically formatted response
Subdomain 1.2: Select model tasks to accomplish a given business requirement
Select model tasks to accomplish a given business requirement
Subdomain 1.3: Select chain components for a desired model input and output
Select chain components for a desired model input and output
Subdomain 1.4: Translate business use case goals into a description of the desired inputs and outputs for the AI pipeline
Translate business use case goals into a description of the desired inputs and outputs for the AI pipeline
Subdomain 1.5: Define and order tools that gather knowledge or take actions for multi-stage reasoning
Define and order tools that gather knowledge or take actions for multi-stage reasoning
Domain 2: Data Preparation
Subdomain 2.1: Apply a chunking strategy for a given document structure and model constraints
Apply a chunking strategy for a given document structure and model constraints
Subdomain 2.2: Filter extraneous content in source documents that degrades quality of a RAG application
Filter extraneous content in source documents that degrades quality of a RAG application
Subdomain 2.3: Choose the appropriate Python package to extract document content from provided source data and format.
Choose the appropriate Python package to extract document content from provided source data and format.
Subdomain 2.4: Define operations and sequence to write given chunked text into Delta Lake tables in Unity Catalog
Define operations and sequence to write given chunked text into Delta Lake tables in Unity Catalog
Subdomain 2.5: Identify needed source documents that provide necessary knowledge and quality for a given RAG application
Identify needed source documents that provide necessary knowledge and quality for a given RAG application
Subdomain 2.6: Identify prompt/response pairs that align with a given model task
Identify prompt/response pairs that align with a given model task
Subdomain 2.7: Use tools and metrics to evaluate retrieval performance
Use tools and metrics to evaluate retrieval performance
Subdomain 2.8: Design retrieval systems using advanced chunking strategies.
Design retrieval systems using advanced chunking strategies.
Subdomain 2.9: Explain the role of re-ranking in the information retrieval process.
Explain the role of re-ranking in the information retrieval process.
Subdomain 2.10: Apply chunking strategy for a given document structure
Apply chunking strategy for a given document structure
Domain 3: Application Development
Subdomain 3.1: Create tools needed to extract data for a given data retrieval need
Create tools needed to extract data for a given data retrieval need
Subdomain 3.2: Select Langchain/similar tools for use in a Generative AI application.
Select Langchain/similar tools for use in a Generative AI application.
Subdomain 3.3: Identify how prompt formats can change model outputs and results
Identify how prompt formats can change model outputs and results
Subdomain 3.4: Qualitatively assess responses to identify common issues such as quality and safety
Qualitatively assess responses to identify common issues such as quality and safety
Subdomain 3.5: Select chunking strategy based on model & retrieval evaluation
Select chunking strategy based on model & retrieval evaluation
Subdomain 3.6: Augment a prompt with additional context from a user's input based on key fields, terms, and intents
Augment a prompt with additional context from a user's input based on key fields, terms, and intents
Subdomain 3.7: Create a prompt that adjusts an LLM's response from a baseline to a desired output
Create a prompt that adjusts an LLM's response from a baseline to a desired output
Subdomain 3.8: Implement LLM guardrails to prevent negative outcomes
Implement LLM guardrails to prevent negative outcomes
Subdomain 3.9: Write metaprompts that minimize hallucinations or leaking private data
Write metaprompts that minimize hallucinations or leaking private data
Subdomain 3.10: Build agent prompt templates exposing available functions
Build agent prompt templates exposing available functions
Subdomain 3.11: Select the best LLM based on the attributes of the application to be developed
Select the best LLM based on the attributes of the application to be developed
Subdomain 3.12: Select an embedding model context length based on source documents, expected queries, and optimization strategy
Select an embedding model context length based on source documents, expected queries, and optimization strategy
Subdomain 3.13: Select a model for from a model hub or marketplace for a task based on model metadata/model cards
Select a model for from a model hub or marketplace for a task based on model metadata/model cards
Subdomain 3.14: Select the best model for a given task based on common metrics generated in experiments
Select the best model for a given task based on common metrics generated in experiments
Subdomain 3.15: Utilize Agent Framework for developing agentic systems
Utilize Agent Framework for developing agentic systems
Domain 4: Assembling and Deploying Applications
Subdomain 4.1: Code a chain using a pyfunc model with pre- and post-processing
Code a chain using a pyfunc model with pre- and post-processing
Subdomain 4.2: Control access to resources from model serving endpoints
Control access to resources from model serving endpoints
Subdomain 4.3: Code a simple chain according to requirements
Code a simple chain according to requirements
Subdomain 4.4: Code a simple chain using langchain
Code a simple chain using langchain
Subdomain 4.5: Choose the basic elements needed to create a RAG application: model flavor, embedding model, retriever, dependencies, input examples, model signature
Choose the basic elements needed to create a RAG application: model flavor, embedding model, retriever, dependencies, input examples, model signature
Subdomain 4.6: Register the model to Unity Catalog using MLflow
Register the model to Unity Catalog using MLflow
Subdomain 4.7: Sequence the steps needed to deploy an endpoint for a basic RAG application
Sequence the steps needed to deploy an endpoint for a basic RAG application
Subdomain 4.8: Create and query a Vector Search index
Create and query a Vector Search index
Subdomain 4.9: Identify how to serve an LLM application that leverages Foundation Model APIs
Identify how to serve an LLM application that leverages Foundation Model APIs
Subdomain 4.10: Identify resources needed to serve features for a RAG application
Identify resources needed to serve features for a RAG application
Subdomain 4.11: Explain the key concepts and components of Mosaic AI Vector Search
Explain the key concepts and components of Mosaic AI Vector Search
Subdomain 4.12: Identify batch inference workloads and apply ai_query() appropriately
Identify batch inference workloads and apply ai_query() appropriately
Domain 5: Governance
Subdomain 5.1: Use masking techniques as guard rails to meet a performance objective
Use masking techniques as guard rails to meet a performance objective
Subdomain 5.2: Select guardrail techniques to protect against malicious user inputs to a Gen AI application
Select guardrail techniques to protect against malicious user inputs to a Gen AI application
Subdomain 5.3: Use legal/licensing requirements for data sources to avoid legal risk
Use legal/licensing requirements for data sources to avoid legal risk
Subdomain 5.4: Recommend an alternative for problematic text mitigation in a data source feeding a GenAI application
Recommend an alternative for problematic text mitigation in a data source feeding a GenAI application
Domain 6: Evaluation and Monitoring
Subdomain 6.1: Select an LLM choice (size and architecture) based on a set of quantitative evaluation metrics
Select an LLM choice (size and architecture) based on a set of quantitative evaluation metrics
Subdomain 6.2: Select key metrics to monitor for a specific LLM deployment scenario
Select key metrics to monitor for a specific LLM deployment scenario
Subdomain 6.3: Evaluate model performance in a RAG application using MLflow
Evaluate model performance in a RAG application using MLflow
Subdomain 6.4: Use inference logging to assess deployed RAG application performance
Use inference logging to assess deployed RAG application performance
Subdomain 6.5: Use Databricks features to control LLM costs for RAG applications
Use Databricks features to control LLM costs for RAG applications
Subdomain 6.6: Use inference tables and Agent Monitoring to track a live LLM endpoint
Use inference tables and Agent Monitoring to track a live LLM endpoint
Subdomain 6.7: Identify evaluation judges that require ground truth
Identify evaluation judges that require ground truth
Subdomain 6.8: Compare the evaluation and monitoring phases of the Gen AI application life cycle
Compare the evaluation and monitoring phases of the Gen AI application life cycle
Techniques & products