Winter Special - Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: top65certs

Databricks-Machine-Learning-Professional Premium Exam Questions

Databricks Certified Machine Learning Professional Questions and Answers

Question 9

Which of the following tools can assist in real-time deployments by packaging software with its own application, tools, and libraries?

Options:

A.

Cloud-based compute

B.

None of these tools

C.

REST APIs

D.

Containers

E.

Autoscaling clusters

Question 10

A data scientist has written a function to track the runs of their random forest model. The data scientist is changing the number of trees in the forest across each run.

Which of the following MLflow operations is designed to log single values like the number of trees in a random forest?

Options:

A.

mlflow.log_artifact

B.

mlflow.log_model

C.

mlflow.log_metric

D.

mlflow.log_param

E.

There is no way to store values like this.

Question 11

A machine learning engineer and data scientist are working together to convert a batch deployment to an always-on streaming deployment. The machine learning engineer has expressed that rigorous data tests must be put in place as a part of their conversion to account for potential changes in data formats.

Which of the following describes why these types of data type tests and checks are particularly important for streaming deployments?

Options:

A.

Because the streaming deployment is always on, all types of data must be handled without producing an error

B.

All of these statements

C.

Because the streaming deployment is always on, there is no practitioner to debug poor model performance

D.

Because the streamingdeployment is always on, there is a need to confirm that the deployment can autoscale

E.

None of these statements

Question 12

A machine learning engineer has developed a model and registered it using the FeatureStoreClient fs. The model has model URI model_uri. The engineer now needs to perform batch inference on customer-level Spark DataFrame spark_df, but it is missing a few of the static features that were used when training the model. The customer_id column is the primary key of spark_df and the training set used when training and logging the model.

Which of the following code blocks can be used to compute predictions for spark_df when the missing feature values can be found in the Feature Store by searching for features by customer_id?

Options:

A.

df = fs.get_missing_features(spark_df, model_uri)

fs.score_model(model_uri, df)

B.

fs.score_model(model_uri, spark_df)

C.

df = fs.get_missing_features(spark_df, model_uri)

fs.score_batch(model_uri, df)

df = fs.get_missing_features(spark_df)

D.

fs.score_batch(model_uri, df)

E.

fs.score_batch(model_uri, spark_df)