Databricks Certified Machine Learning Professional Exam Syllabus

Machine Learning Professional PDF, Machine Learning Professional Dumps, Machine Learning Professional VCE, Databricks Certified Machine Learning Professional Questions PDF, Databricks Certified Machine Learning Professional VCE, Databricks Lakehouse Machine Learning Professional Dumps, Databricks Lakehouse Machine Learning Professional PDFUse this quick start guide to collect all the information about Databricks Machine Learning Professional Certification exam. This study guide provides a list of objectives and resources that will help you prepare for items on the Databricks Certified Machine Learning Professional (Machine Learning Professional) exam. The Sample Questions will help you identify the type and difficulty level of the questions and the Practice Exams will make you familiar with the format and environment of an exam. You should refer this guide carefully before attempting your actual Databricks Certified Machine Learning Professional certification exam.

The Databricks Machine Learning Professional certification is mainly targeted to those candidates who want to build their career in ML Data Scientist domain. The Databricks Certified Machine Learning Professional exam verifies that the candidate possesses the fundamental knowledge and proven skills in the area of Databricks Lakehouse Machine Learning Professional.

Databricks Machine Learning Professional Exam Summary:

Exam Name Databricks Certified Machine Learning Professional
Exam Code Machine Learning Professional
Exam Price $200 (USD)
Duration 120 mins
Number of Questions 60
Passing Score 70%
Books / Training Machine Learning in Production
Schedule Exam Kryterion Webassesor
Sample Questions Databricks Machine Learning Professional Sample Questions
Practice Exam Databricks Machine Learning Professional Certification Practice Exam

Databricks Lakehouse Machine Learning Professional Exam Syllabus Topics:

Topic Details Weights
Experimentation - Data Management
  • Read and write a Delta table
  • View Delta table history and load a previous version of a Delta table
  • Create, overwrite, merge, and read Feature Store tables in machine learning workflows

- Experiment Tracking

  • Manually log parameters, models, and evaluation metrics using MLflow
  • Programmatically access and use data, metadata, and models from MLflow experiments

- Advanced Experiment Tracking

  • Perform MLflow experiment tracking workflows using model signatures and input examples
  • Identify the requirements for tracking nested runs
  • Describe the process of enabling autologging, including with the use of Hyperopt
  • Log and view artifacts like SHAP plots, custom visualizations, feature data, images, and metadata
30%
Model Lifecycle Management - Preprocessing Logic
  • Describe an MLflow flavor and the benefits of using MLflow flavors
  • Describe the advantages of using the pyfunc MLflow flavor
  • Describe the process and benefits of including preprocessing logic and context in custom model classes and objects

- Model Management

  • Describe the basic purpose and user interactions with Model Registry
  • Programmatically register a new model or new model version.
  • Add metadata to a registered model and a registered model version
  • Identify, compare, and contrast the available model stages
  • Transition, archive, and delete model versions

- Model Lifecycle Automation

  • Identify the role of automated testing in ML CI/CD pipelines
  • Describe how to automate the model lifecycle using Model Registry Webhooks and Databricks Jobs
  • Identify advantages of using Job clusters over all-purpose clusters
  • Describe how to create a Job that triggers when a model transitions between stages, given a scenario
  • Describe how to connect a Webhook with a Job
  • Identify which code block will trigger a shown webhook
  • Identify a use case for HTTP webhooks and where the Webhook URL needs to come.
  • Describe how to list all webhooks and how to delete a webhook
30%
Model Deployment - Batch
  • Describe batch deployment as the appropriate use case for the vast majority of deployment use cases
  • Identify how batch deployment computes predictions and saves them somewhere for later use
  • Identify live serving benefits of querying precomputed batch predictions
  • Identify less performant data storage as a solution for other use cases
  • Load registered models with load_model
  • Deploy a single-node model in parallel using spark_udf
  • Identify z-ordering as a solution for reducing the amount of time to read predictions from a table
  • Identify partitioning on a common column to speed up querying
  • Describe the practical benefits of using the score_batch operation

- Streaming

  • Describe Structured Streaming as a common processing tool for ETL pipelines
  • Identify structured streaming as a continuous inference solution on incoming data
  • Describe why complex business logic must be handled in streaming deployments
  • Identify that data can arrive out-of-order with structured streaming
  • Identify continuous predictions in time-based prediction store as a scenario for streaming deployments
  • Convert a batch deployment pipeline inference to a streaming deployment pipeline
  • Convert a batch deployment pipeline writing to a streaming deployment pipeline

- Real-time

  • Describe the benefits of using real-time inference for a small number of records or when fast prediction computations are needed
  • Identify JIT feature values as a need for real-time deployment
  • Query a Model Serving enabled model in the Production stage and Staging stage
  • Identify how cloud-provided RESTful services in containers is the best solution for production-grade real-time deployments
25%
Solution and Data Monitoring - Drift Types
  • Compare and contrast label drift and feature drift
  • Identify scenarios in which feature drift and/or label drift are likely to occur
  • Describe concept drift and its impact on model efficacy

- Drift Tests and Monitoring

  • Describe summary statistic monitoring as a simple solution for numeric feature drift
  • Describe mode, unique values, and missing values as simple solutions for categorical feature drift
  • Describe tests as more robust monitoring solutions for numeric feature drift than simple summary statistics
  • Describe tests as more robust monitoring solutions for categorical feature drift than simple summary statistics
  • Compare and contrast Jenson-Shannon divergence and Kolmogorov-Smirnov tests for numerical drift detection Identify a scenario in which a chi-square test would be useful

- Comprehensive Drift Solutions

  • Describe a common workflow for measuring concept drift and feature drift
  • Identify when retraining and deploying an updated model is a probable solution to drift
  • Test whether the updated model performs better on the more recent data
15%

To ensure success in Databricks Lakehouse Machine Learning Professional certification exam, we recommend authorized training course, practice test and hands-on experience to prepare for Databricks Certified Machine Learning Professional (Machine Learning Professional) exam.

Rating: 5 / 5 (75 votes)