Spring Sale Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 65pass65

Professional-Machine-Learning-Engineer Google Professional Machine Learning Engineer is now Stable and With Pass Result | Test Your Knowledge for Free

Exams4sure Dumps

Professional-Machine-Learning-Engineer Practice Questions

Google Professional Machine Learning Engineer

Last Update 3 days ago
Total Questions : 296

Dive into our fully updated and stable Professional-Machine-Learning-Engineer practice test platform, featuring all the latest Machine Learning Engineer exam questions added this week. Our preparation tool is more than just a Google study aid; it's a strategic advantage.

Our free Machine Learning Engineer practice questions crafted to reflect the domains and difficulty of the actual exam. The detailed rationales explain the 'why' behind each answer, reinforcing key concepts about Professional-Machine-Learning-Engineer. Use this test to pinpoint which areas you need to focus your study on.

Professional-Machine-Learning-Engineer PDF

Professional-Machine-Learning-Engineer PDF (Printable)
$43.75
$124.99

Professional-Machine-Learning-Engineer Testing Engine

Professional-Machine-Learning-Engineer PDF (Printable)
$50.75
$144.99

Professional-Machine-Learning-Engineer PDF + Testing Engine

Professional-Machine-Learning-Engineer PDF (Printable)
$63.7
$181.99
Question # 71

You are tasked with building an MLOps pipeline to retrain tree-based models in production. The pipeline will include components related to data ingestion, data processing, model training, model evaluation, and model deployment. Your organization primarily uses PySpark-based workloads for data preprocessing. You want to minimize infrastructure management effort. How should you set up the pipeline?

Options:

A.  

Set up a TensorFlow Extended (TFX) pipeline on Vertex Al Pipelines to orchestrate the MLOps pipeline. Write a custom component for the PySpark-based workloads on Dataproc.

B.  

Set up a Vertex Al Pipelines to orchestrate the MLOps pipeline. Use the predefined Dataproc component for the PySpark-based workloads.

C.  

Set up Cloud Composer to orchestrate the MLOps pipeline. Use Dataproc workflow templates for the PySpark-based workloads in Cloud Composer.

D.  

Set up Kubeflow Pipelines on Google Kubernetes Engine to orchestrate the MLOps pipeline. Write a custom component for the PySpark-based workloads on Dataproc.

Discussion 0
Question # 72

You are creating a retraining policy for a customer churn prediction model deployed in Vertex AI. New training data is added weekly. You want to implement a model retraining process that minimizes cost and effort. What should you do?

Options:

A.  

Retrain the model when the model ' s latency increases by 10% due to increased traffic.

B.  

Retrain the model when the model accuracy drops by 10% on the new training dataset.

C.  

Retrain the model every week when new training data is available.

D.  

Retrain the model when a significant shift in the distribution of customer attributes is detected in the production data compared to the training data.

Discussion 0
Question # 73

You recently joined a machine learning team that will soon release a new project. As a lead on the project, you are asked to determine the production readiness of the ML components. The team has already tested features and data, model development, and infrastructure. Which additional readiness check should you recommend to the team?

Options:

A.  

Ensure that training is reproducible

B.  

Ensure that all hyperparameters are tuned

C.  

Ensure that model performance is monitored

D.  

Ensure that feature expectations are captured in the schema

Discussion 0
Question # 74

You are developing an ML model using a dataset with categorical input variables. You have randomly split half of the data into training and test sets. After applying one-hot encoding on the categorical variables in the training set, you discover that one categorical variable is missing from the test set. What should you do?

Options:

A.  

Randomly redistribute the data, with 70% for the training set and 30% for the test set

B.  

Use sparse representation in the test set

C.  

Apply one-hot encoding on the categorical variables in the test data.

D.  

Collect more data representing all categories

Discussion 0
Question # 75

You are working on a classification problem with time series data and achieved an area under the receiver operating characteristic curve (AUC ROC) value of 99% for training data after just a few experiments. You haven’t explored using any sophisticated algorithms or spent any time on hyperparameter tuning. What should your next step be to identify and fix the problem?

Options:

A.  

Address the model overfitting by using a less complex algorithm.

B.  

Address data leakage by applying nested cross-validation during model training.

C.  

Address data leakage by removing features highly correlated with the target value.

D.  

Address the model overfitting by tuning the hyperparameters to reduce the AUC ROC value.

Discussion 0
Question # 76

You were asked to investigate failures of a production line component based on sensor readings. After receiving the dataset, you discover that less than 1% of the readings are positive examples representing failure incidents. You have tried to train several classification models, but none of them converge. How should you resolve the class imbalance problem?

Options:

A.  

Use the class distribution to generate 10% positive examples

B.  

Use a convolutional neural network with max pooling and softmax activation

C.  

Downsample the data with upweighting to create a sample with 10% positive examples

D.  

Remove negative examples until the numbers of positive and negative examples are equal

Discussion 0
Question # 77

You are building a MLOps platform to automate your company ' s ML experiments and model retraining. You need to organize the artifacts for dozens of pipelines How should you store the pipelines ' artifacts ' ?

Options:

A.  

Store parameters in Cloud SQL and store the models ' source code and binaries in GitHub

B.  

Store parameters in Cloud SQL store the models ' source code in GitHub, and store the models ' binaries in Cloud Storage.

C.  

Store parameters in Vertex ML Metadata store the models ' source code in GitHub and store the models ' binaries in Cloud Storage.

D.  

Store parameters in Vertex ML Metadata and store the models source code and binaries in GitHub.

Discussion 0
Question # 78

You have trained a model on a dataset that required computationally expensive preprocessing operations. You need to execute the same preprocessing at prediction time. You deployed the model on Al Platform for high-throughput online prediction. Which architecture should you use?

Options:

A.  

• Validate the accuracy of the model that you trained on preprocessed data

• Create a new model that uses the raw data and is available in real time

• Deploy the new model onto Al Platform for online prediction

B.  

• Send incoming prediction requests to a Pub/Sub topic

• Transform the incoming data using a Dataflow job

• Submit a prediction request to Al Platform using the transformed data

• Write the predictions to an outbound Pub/Sub queue

C.  

• Stream incoming prediction request data into Cloud Spanner

• Create a view to abstract your preprocessing logic.

• Query the view every second for new records

• Submit a prediction request to Al Platform using the transformed data

• Write the predictions to an outbound Pub/Sub queue.

D.  

• Send incoming prediction requests to a Pub/Sub topic

• Set up a Cloud Function that is triggered when messages are published to the Pub/Sub topic.

• Implement your preprocessing logic in the Cloud Function

• Submit a prediction request to Al Platform using the transformed data

• Write the predictions to an outbound Pub/Sub queue

Discussion 0
Question # 79

You recently used XGBoost to train a model in Python that will be used for online serving Your model prediction service will be called by a backend service implemented in Golang running on a Google Kubemetes Engine (GKE) cluster Your model requires pre and postprocessing steps You need to implement the processing steps so that they run at serving time You want to minimize code changes and infrastructure maintenance and deploy your model into production as quickly as possible. What should you do?

Options:

A.  

Use FastAPI to implement an HTTP server Create a Docker image that runs your HTTP server and deploy it on your organization ' s GKE cluster.

B.  

Use FastAPI to implement an HTTP server Create a Docker image that runs your HTTP server Upload the image to Vertex Al Model Registry and deploy it to a Vertex Al endpoint.

C.  

Use the Predictor interface to implement a custom prediction routine Build the custom contain upload the container to Vertex Al Model Registry, and deploy it to a Vertex Al endpoint.

D.  

Use the XGBoost prebuilt serving container when importing the trained model into Vertex Al Deploy the model to a Vertex Al endpoint Work with the backend engineers to implement the pre- and postprocessing steps in the Golang backend service.

Discussion 0
Question # 80

You are analyzing customer data for a healthcare organization that is stored in Cloud Storage. The data contains personally identifiable information (PII) You need to perform data exploration and preprocessing while ensuring the security and privacy of sensitive fields What should you do?

Options:

A.  

Use the Cloud Data Loss Prevention (DLP) API to de-identify the PI! before performing data exploration and preprocessing.

B.  

Use customer-managed encryption keys (CMEK) to encrypt the Pll data at rest and decrypt the Pll data during data exploration and preprocessing.

C.  

Use a VM inside a VPC Service Controls security perimeter to perform data exploration and preprocessing.

D.  

Use Google-managed encryption keys to encrypt the Pll data at rest, and decrypt the Pll data during data exploration and preprocessing.

Discussion 0
Get Professional-Machine-Learning-Engineer dumps and pass your exam in 24 hours!

Free Exams Sample Questions