Spring Sale Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 65pass65

Associate-Cloud-Engineer Google Cloud Certified - Associate Cloud Engineer is now Stable and With Pass Result | Test Your Knowledge for Free

Associate-Cloud-Engineer Practice Questions

Google Cloud Certified - Associate Cloud Engineer

Last Update 4 days ago
Total Questions : 332

Dive into our fully updated and stable Associate-Cloud-Engineer practice test platform, featuring all the latest Google Cloud Certified exam questions added this week. Our preparation tool is more than just a Google study aid; it's a strategic advantage.

Our free Google Cloud Certified practice questions crafted to reflect the domains and difficulty of the actual exam. The detailed rationales explain the 'why' behind each answer, reinforcing key concepts about Associate-Cloud-Engineer. Use this test to pinpoint which areas you need to focus your study on.

Associate-Cloud-Engineer PDF

Associate-Cloud-Engineer PDF (Printable)
$43.75
$124.99

Associate-Cloud-Engineer Testing Engine

Associate-Cloud-Engineer PDF (Printable)
$50.75
$144.99

Associate-Cloud-Engineer PDF + Testing Engine

Associate-Cloud-Engineer PDF (Printable)
$63.7
$181.99
Question # 91

Your team is running an on-premises ecommerce application. The application contains a complex set of microservices written in Python, and each microservice is running on Docker containers. Configurations are injected by using environment variables. You need to deploy your current application to a serverless Google Cloud cloud solution. What should you do?

Options:

A.  

Use your existing CI/CD pipeline Use the generated Docker images and deploy them to Cloud Run. Update the configurations and the required endpoints.

B.  

Use your existing continuous integration and delivery (CI/CD) pipeline. Use the generated Docker images and deploy them to Cloud Function. Use the same configuration as on-premises.

C.  

Use the existing codebase and deploy each service as a separate Cloud Function Update the configurations and the required endpoints.

D.  

Use your existing codebase and deploy each service as a separate Cloud Run Use the same configurations as on-premises.

Discussion 0
Question # 92

You recently discovered that your developers are using many service account keys during their development process. While you work on a long term improvement, you need to quickly implement a process to enforce short-lived service account credentials in your company. You have the following requirements:

• All service accounts that require a key should be created in a centralized project called pj-sa.

• Service account keys should only be valid for one day.

You need a Google-recommended solution that minimizes cost. What should you do?

Options:

A.  

Implement a Cloud Run job to rotate all service account keys periodically in pj-sa. Enforce an org policy to deny service account key creation with an exception to pj-sa.

B.  

Implement a Kubernetes Cronjob to rotate all service account keys periodically. Disable attachment ofservice accounts to resources in all projects with an exception to pj-sa.

C.  

Enforce an org policy constraint allowing the lifetime of service account keys to be 24 hours. Enforce an org policy constraint denying service account key creation with an exception on pj-sa.

D.  

Enforce a DENY org policy constraint over the lifetime of service account keys for 24 hours. Disable attachment of service accounts to resources in all projects with an exception to pj-sa.

Discussion 0
Question # 93

Your company uses a large number of Google Cloud services centralized in a single project. All teams have specific projects for testing and development. The DevOps team needs access to all of theproduction services in order to perform their job. You want to prevent Google Cloud product changes from broadening their permissions in the future. You want to follow Google-recommended practices. What should you do?

Options:

A.  

Grant all members of the DevOps team the role of Project Editor on the organization level.

B.  

Grant all members of the DevOps team the role of Project Editor on the production project.

C.  

Create a custom role that combines the required permissions. Grant the DevOps team the custom role on the production project.

D.  

Create a custom role that combines the required permissions. Grant the DevOps team the custom role on the organization level.

Discussion 0
Question # 94

You need to set up a policy so that videos stored in a specific Cloud Storage Regional bucket are moved to Coldline after 90 days, and then deleted after one year from their creation. How should you set up the policy?

Options:

A.  

Use Cloud Storage Object Lifecycle Management using Age conditions with SetStorageClass and Delete actions. Set the SetStorageClass action to 90 days and the Delete action to 275 days (365 – 90)

B.  

Use Cloud Storage Object Lifecycle Management using Age conditions with SetStorageClass and Delete actions. Set the SetStorageClass action to 90 days and the Delete action to 365 days.

C.  

Use gsutil rewrite and set the Delete action to 275 days (365-90).

D.  

Use gsutil rewrite and set the Delete action to 365 days.

Discussion 0
Question # 95

A team of data scientists infrequently needs to use a Google Kubernetes Engine (GKE) cluster that you manage. They require GPUs for some long-running, non-restartable jobs. You want to minimize cost. What should you do?

Options:

A.  

Enable node auto-provisioning on the GKE cluster.

B.  

Create a VerticalPodAutscaler for those workloads.

C.  

Create a node pool with preemptible VMs and GPUs attached to those VMs.

D.  

Create a node pool of instances with GPUs, and enable autoscaling on this node pool with a minimum size of 1.

Discussion 0
Question # 96

The DevOps group in your organization needs full control of Compute Engine resources in your development project. However, they should not have permission to create or update any other resources in the project. You want to follow Google's recommendations for setting permissions for the DevOps group. What should you do?

Options:

A.  

Grant the basic role roles/viewer and the predefined role roles/compute.admin to the DevOps group.

B.  

Create an IAM policy and grant all compute. instanceAdmln." permissions to the policy Attach the policy to the DevOps group.

C.  

Create a custom role at the folder level and grant all compute. instanceAdmln. * permissions to the role Grant the custom role to the DevOps group.

D.  

Grant the basic role roles/editor to the DevOps group.

Discussion 0
Question # 97

You have 32 GB of data in a single file that you need to upload to a Nearline Storage bucket. The WAN connection you are using is rated at 1 Gbps, and you are the only one on the connection. You want to use as much of the rated 1 Gbps as possible to transfer the file rapidly. How should you upload the file?

Options:

A.  

Use the GCP Console to transfer the file instead of gsutil.

B.  

Enable parallel composite uploads using gsutil on the file transfer.

C.  

Decrease the TCP window size on the machine initiating the transfer.

D.  

Change the storage class of the bucket from Nearline to Multi-Regional.

Discussion 0
Question # 98

Your company has developed a new application that consists of multiple microservices. You want to deploy the application to Google Kubernetes Engine (GKE), and you want to ensure that the cluster can scale as more applications are deployed in the future. You want to avoid manual intervention when each new application is deployed. What should you do?

Options:

A.  

Deploy the application on GKE, and add a HorizontalPodAutoscaler to the deployment.

B.  

Deploy the application on GKE, and add a VerticalPodAutoscaler to the deployment.

C.  

Create a GKE cluster with autoscaling enabled on the node pool. Set a minimum and maximum for the size of the node pool.

D.  

Create a separate node pool for each application, and deploy each application to its dedicated node pool.

Discussion 0
Question # 99

Your company's security vulnerability management policy wonts 3 member of the security team to have visibility into vulnerabilities and other OS metadata for a specific Compute Engine instance This Compute Engine instance hosts a critical application in your Goggle Cloud project. You need to implement your company's security vulnerability management policy. What should you dc?

Options:

A.  

• Ensure that the Ops Agent Is Installed on the Compute Engine instance.• Create a custom metric in the Cloud Monitoring dashboard.• Provide the security team member with access to this dashboard.

B.  

• Ensure that the Ops Agent is installed on tie Compute Engine instance.• Provide the security team member roles/configure.inventoryViewer permission.

C.  

• Ensure that the OS Config agent Is Installed on the Compute Engine instance.• Provide the security team member roles/configure.vulnerabilityViewer permission.

D.  

• Ensure that the OS Config agent is installed on the Compute Engine instance• Create a log sink Co a BigQuery dataset.• Provide the security team member with access to this dataset.

Discussion 0
Get Associate-Cloud-Engineer dumps and pass your exam in 24 hours!

Free Exams Sample Questions