Free Practice Questions for Google Professional Machine Learning Engineer Certification

    🔄 Last checked for updates March 4th, 2026

    Study with 352 exam-style practice questions designed to help you prepare for the Google Professional Machine Learning Engineer.

    Start Practicing

    Random Questions

    Practice with randomly mixed questions from all topics

    Question MixAll Topics
    FormatRandom Order

    Domain Mode

    Practice questions from a specific topic area

    Exam Information

    Exam Details

    Key information about Google Professional Machine Learning Engineer

    Official study guide:

    View

    level:

    Professional

    target audience:

    A Professional Machine Learning Engineer who builds, evaluates, productionizes, and optimizes AI solutions using Google Cloud capabilities and conventional ML approaches.

    Exam Topics & Skills Assessed

    Skills measured (from the official study guide)

    Domain 1: Architecting low-code AI solutions

    Subdomain 1.1: Developing ML models by using BigQuery ML

    Considerations include:

    - Building the appropriate BigQuery ML model (e.g., linear and binary classification, regression, time-series, matrix factorization, boosted trees, autoencoders) based on the business problem - Feature engineering or selection by using BigQuery ML - Generating predictions by using BigQuery ML

    Subdomain 1.2: Building AI solutions by using ML APIs or foundation models

    Considerations include:

    - Building applications by using ML APIs from Model Garden - Building applications by using industry-specific APIs (e.g., Document AI API, Retail API) - Implementing retrieval augmented generation (RAG) applications by using Vertex AI Agent Builder

    Subdomain 1.3: Training models by using AutoML

    Considerations include:

    - Preparing data for AutoML (e.g., feature selection, data labeling, Tabular Workflows on AutoML) - Using available data (e.g., tabular, text, speech, images, videos) to train custom models - Using AutoML for tabular data - Creating forecasting models by using AutoML - Configuring and debugging trained models

    Domain 2: Collaborating within and across teams to manage data and models

    Subdomain 2.1: Exploring and preprocessing organization-wide data (e.g., Cloud Storage, BigQuery, Spanner, Cloud SQL, Apache Spark, Apache Hadoop)

    Considerations include:

    - Organizing different types of data (e.g., tabular, text, speech, images, videos) for efficient training - Managing datasets in Vertex AI - Data preprocessing (e.g., Dataflow, TensorFlow Extended [TFX], BigQuery) - Creating and consolidating features in Vertex AI Feature Store - Privacy implications of data usage and/or collection (e.g., handling sensitive data such as personally identifiable information [PII] and protected health information [PHI]) - Ingesting different data sources (e.g., text documents) into Vertex AI for inference

    Subdomain 2.2: Model prototyping using Jupyter notebooks

    Considerations include:

    - Choosing the appropriate Jupyter backend on Google Cloud (e.g., Vertex AI Workbench, Colab Enterprise, notebooks on Dataproc) - Applying security best practices in Vertex AI Workbench - Using Spark kernels - Integrating code source repositories - Developing models in Vertex AI Workbench by using common frameworks (e.g., TensorFlow, PyTorch, sklearn, Spark, JAX) - Leveraging a variety of foundation and open-source models in Model Garden

    Subdomain 2.3: Tracking and running ML experiments

    Considerations include:

    - Choosing the appropriate Google Cloud environment for development and experimentation (e.g., Vertex AI Experiments, Kubeflow Pipelines, Vertex AI TensorBoard with TensorFlow and PyTorch) given the framework - Evaluating generative AI solutions

    Domain 3: Scaling prototypes into ML models

    Subdomain 3.1: Building models

    Considerations include:

    - Choosing ML framework and model architecture - Modeling techniques given interpretability requirements

    Subdomain 3.2: Training models

    Considerations include:

    - Organizing training data (e.g., tabular, text, speech, images, videos) on Google Cloud (e.g., Cloud Storage, BigQuery) - Ingestion of various file types (e.g., CSV, JSON, images, Hadoop, databases) into training - Training using different SDKs (e.g., Vertex AI custom training, Kubeflow on Google Kubernetes Engine, AutoML, tabular workflows) - Using distributed training to organize reliable pipelines - Hyperparameter tuning - Troubleshooting ML model training failures - Fine-tuning foundation models (e.g., Vertex AI, Model Garden)

    Subdomain 3.3: Choosing appropriate hardware for training

    Considerations include:

    - Evaluation of compute and accelerator options (e.g., CPU, GPU, TPU, edge devices) - Distributed training with TPUs and GPUs (e.g., Reduction Server on Vertex AI, Horovod)

    Domain 4: Serving and scaling models

    Subdomain 4.1: Serving models

    Considerations include:

    - Batch and online inference (e.g., Vertex AI, Dataflow, BigQuery ML, Dataproc) - Using different frameworks (e.g., PyTorch, XGBoost) to serve models - Organizing a model registry - A/B testing different versions of a model

    Subdomain 4.2: Scaling online model serving

    Considerations include:

    - Vertex AI Feature Store - Vertex AI public and private endpoints - Choosing appropriate hardware (e.g., CPU, GPU, TPU, edge) - Scaling the serving backend based on the throughput (e.g., Vertex AI Prediction, containerized serving) - Tuning ML models for training and serving in production (e.g., simplification techniques, optimizing the ML solution for increased performance, latency, memory, throughput)

    Domain 5: Automating and orchestrating ML pipelines

    Subdomain 5.1: Developing end-to-end ML pipelines

    Considerations include:

    - Data and model validation - Ensuring consistent data pre-processing between training and serving - Hosting third-party pipelines on Google Cloud (e.g., MLFlow) - Identifying components, parameters, triggers, and compute needs (e.g., Cloud Build, Cloud Run) - Orchestration framework (e.g., Kubeflow Pipelines, Vertex AI Pipelines, Cloud Composer) - Hybrid or multicloud strategies - System design with TFX components or Kubeflow DSL (e.g., Dataflow)

    Subdomain 5.2: Automating model retraining

    Considerations include:

    - Determining an appropriate retraining policy - Continuous integration and continuous delivery (CI/CD) model deployment (e.g., Cloud Build, Jenkins)

    Subdomain 5.3: Tracking and auditing metadata

    Considerations include:

    - Tracking and comparing model artifacts and versions (e.g., Vertex AI Experiments, Vertex ML Metadata) - Hooking into model and dataset versioning - Model and data lineage

    Domain 6: Monitoring AI solutions

    Subdomain 6.1: Identifying risks to AI solutions

    Considerations include:

    - Building secure AI systems by protecting against unintentional exploitation of data or models (e.g., hacking) - Aligning with Google’s Responsible AI practices (e.g., monitoring for bias) - Assessing AI solution readiness (e.g., fairness, bias) - Model explainability on Vertex AI (e.g., Vertex AI Prediction)

    Subdomain 6.2: Monitoring, testing, and troubleshooting AI solutions

    Considerations include:

    - Establishing continuous evaluation metrics (e.g., Vertex AI Model Monitoring, Explainable AI) - Monitoring for training-serving skew - Monitoring for feature attribution drift - Monitoring model performance against baselines, simpler models, and across the time dimension - Monitoring for common training and serving errors

    Techniques & products

    BigQuery ML
    Linear classification
    Binary classification
    Regression
    Time-series models
    Matrix factorization
    Boosted trees
    Autoencoders
    Feature engineering
    Feature selection
    ML APIs
    Model Garden
    Document AI API
    Retail API
    Retrieval Augmented Generation (RAG)
    Vertex AI Agent Builder
    AutoML
    Data labeling
    Tabular Workflows on AutoML
    Custom models
    Forecasting models
    Cloud Storage
    Spanner
    Cloud SQL
    Apache Spark
    Apache Hadoop
    Vertex AI
    Dataflow
    TensorFlow Extended (TFX)
    Vertex AI Feature Store
    Personally Identifiable Information (PII)
    Protected Health Information (PHI)
    Jupyter notebooks
    Vertex AI Workbench
    Colab Enterprise
    Dataproc
    Spark kernels
    Code source repositories
    TensorFlow
    PyTorch
    sklearn
    JAX
    Foundation models
    Open-source models
    Vertex AI Experiments
    Kubeflow Pipelines
    Vertex AI TensorBoard
    ML framework
    Model architecture
    Interpretability requirements
    Distributed training
    Hyperparameter tuning
    ML model training failures
    Fine-tuning foundation models
    CPU
    GPU
    TPU
    Edge devices
    Reduction Server on Vertex AI
    Horovod
    Batch inference
    Online inference
    Model registry
    A/B testing
    Vertex AI public endpoints
    Vertex AI private endpoints
    Vertex AI Prediction
    Containerized serving
    MLOps
    ML pipelines
    Data validation
    Model validation
    Cloud Build
    Cloud Run
    Cloud Composer
    MLFlow
    Kubeflow DSL
    Continuous Integration (CI)
    Continuous Delivery (CD)
    Jenkins
    Vertex ML Metadata
    Model versioning
    Dataset versioning
    Model lineage
    Data lineage
    Secure AI systems
    Google’s Responsible AI practices
    Bias monitoring
    Fairness assessment
    Model explainability
    Vertex AI Model Monitoring
    Explainable AI
    Training-serving skew
    Feature attribution drift

    CertSafari is not affiliated with, endorsed by, or officially connected to Google LLC. Full disclaimer