Spring Sale Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 65pass65

Associate-Cloud-Engineer Google Cloud Certified - Associate Cloud Engineer is now Stable and With Pass Result | Test Your Knowledge for Free

Associate-Cloud-Engineer Practice Questions

Google Cloud Certified - Associate Cloud Engineer

Last Update 4 days ago
Total Questions : 332

Dive into our fully updated and stable Associate-Cloud-Engineer practice test platform, featuring all the latest Google Cloud Certified exam questions added this week. Our preparation tool is more than just a Google study aid; it's a strategic advantage.

Our free Google Cloud Certified practice questions crafted to reflect the domains and difficulty of the actual exam. The detailed rationales explain the 'why' behind each answer, reinforcing key concepts about Associate-Cloud-Engineer. Use this test to pinpoint which areas you need to focus your study on.

Associate-Cloud-Engineer PDF

Associate-Cloud-Engineer PDF (Printable)
$43.75
$124.99

Associate-Cloud-Engineer Testing Engine

Associate-Cloud-Engineer PDF (Printable)
$50.75
$144.99

Associate-Cloud-Engineer PDF + Testing Engine

Associate-Cloud-Engineer PDF (Printable)
$63.7
$181.99
Question # 61

You have developed a containerized web application that will serve Internal colleagues during business hours. You want to ensure that no costs are incurred outside of the hours the application is used. You have just created a new Google Cloud project and want to deploy the application. What should you do?

Options:

A.  

Deploy the container on Cloud Run for Anthos, and set the minimum number of instances to zero

B.  

Deploy the container on Cloud Run (fully managed), and set the minimum number of instances to zero.

C.  

Deploy the container on App Engine flexible environment with autoscaling. and set the value min_instances to zero in the app yaml

D.  

Deploy the container on App Engine flexible environment with manual scaling, and set the value instances to zero in the app yaml

Discussion 0
Question # 62

You need to produce a list of the enabled Google Cloud Platform APIs for a GCP project using the gcloud command line in the Cloud Shell. The project name is my-project. What should you do?

Options:

A.  

Run gcloud projects list to get the project ID, and then run gcloud services list --project .

B.  

Run gcloud init to set the current project to my-project, and then run gcloud services list --available.

C.  

Run gcloud info to view the account value, and then run gcloud services list --account .

D.  

Run gcloud projects describe to verify the project value, and then run gcloud services list --available.

Discussion 0
Question # 63

Every employee of your company has a Google account. Your operational team needs to manage a large number of instances on Compute Engine. Each member of this team needs only administrative access to the servers. Your security team wants to ensure that the deployment of credentials is operationally efficient and must be able to determine who accessed a given instance. What should you do?

Options:

A.  

Generate a new SSH key pair. Give the private key to each member of your team. Configure the public key in the metadata of each instance.

B.  

Ask each member of the team to generate a new SSH key pair and to send you their public key. Use a configuration management tool to deploy those keys on each instance.

C.  

Ask each member of the team to generate a new SSH key pair and to add the public key to their Google account. Grant the “compute.osAdminLogin” role to the Google group corresponding to this team.

D.  

Generate a new SSH key pair. Give the private key to each member of your team. Configure the public key as a project-wide public SSH key in your Cloud Platform project and allow project-wide public SSH keys on each instance.

Discussion 0
Question # 64

You are building a new version of an application hosted in an App Engine environment. You want to test the new version with 1% of users before you completely switch your application over to the new version. What should you do?

Options:

A.  

Deploy a new version of your application in Google Kubernetes Engine instead of App Engine and then use GCP Console to split traffic.

B.  

Deploy a new version of your application in a Compute Engine instance instead of App Engine and then use GCP Console to split traffic.

C.  

Deploy a new version as a separate app in App Engine. Then configure App Engine using GCP Console to split traffic between the two apps.

D.  

Deploy a new version of your application in App Engine. Then go to App Engine settings in GCP Console and split traffic between the current version and newly deployed versions accordingly.

Discussion 0
Question # 65

You have an object in a Cloud Storage bucket that you want to share with an external company. The object contains sensitive data. You want access to the content to be removed after four hours. The external company does not have a Google account to which you can grant specific user-based access privileges. You want to use the most secure method that requires the fewest steps. What should you do?

Options:

A.  

Create a signed URL with a four-hour expiration and share the URL with the company.

B.  

Set object access to ‘public’ and use object lifecycle management to remove the object after four hours.

C.  

Configure the storage bucket as a static website and furnish the object’s URL to the company. Delete the object from the storage bucket after four hours.

D.  

Create a new Cloud Storage bucket specifically for the external company to access. Copy the object to that bucket. Delete the bucket after four hours have passed.

Discussion 0
Question # 66

You are in charge of provisioning access for all Google Cloud users in your organization. Your company recently acquired a startup company that has their own Google Cloud organization. You need to ensure that your Site Reliability Engineers (SREs) have the same project permissions in the startup company's organization as in your own organization. What should you do?

Options:

A.  

In the Google Cloud console for your organization, select Create role from selection, and choose destination as the startup company's organization

B.  

In the Google Cloud console for the startup company, select Create role from selection and choose source as the startup company's Google Cloud organization.

C.  

Use the gcloud iam roles copy command, and provide the Organization ID of the startup company'sGoogle Cloud Organization as the destination.

D.  

Use the gcloud iam roles copy command, and provide the project IDs of all projects in the startup company s organization as the destination.

Discussion 0
Question # 67

Your company requires all developers to have the same permissions, regardless of the Google Cloud project they are working on. Your company's security policy also restricts developer permissions to Compute Engine. Cloud Functions, and Cloud SQL. You want to implement the security policy with minimal effort. What should you do?

Options:

A.  

• Create a custom role with Compute Engine, Cloud Functions, and Cloud SQL permissions in one project within the Google Cloud organization.• Copy the role across all projects created within the organization with the gcloud iam roles copy command.• Assign the role to developers in those projects.

B.  

• Add all developers to a Google group in Google Groups for Workspace.• Assign the predefined role of Compute Admin to the Google group at the Google Cloud organization level.

C.  

• Add all developers to a Google group in Cloud Identity.• Assign predefined roles for Compute Engine, Cloud Functions, and Cloud SQL permissions to the Google group for each project in the Google Cloud organization.

D.  

• Add all developers to a Google group in Cloud Identity.• Create a custom role with Compute Engine, Cloud Functions, and Cloud SQL permissions at the Google Cloud organization level.• Assign the custom role to the Google group.

Discussion 0
Question # 68

You are planning to migrate your on-premises VMs to Google Cloud. You need to set up a landing zone in Google Cloud before migrating the VMs. You must ensure that all VMs in your production environment can communicate with each other through private IP addresses. You need to allow all VMs in your Google Cloud organization to accept connections on specific TCP ports. You want to follow Google-recommended practices, and you need to minimize your operational costs. What should you do?

Options:

A.  

Create individual VPCs per Google Cloud project. Peer all the VPCs together. Apply organization policies on the organization level.

B.  

Create individual VPCs for each Google Cloud project. Peer all the VPCs together. Apply hierarchical firewall policies on the organization level.

C.  

Create a host VPC project with each production project as its service project. Apply organization policies on the organization level.

D.  

Create a host VPC project with each production project as its service project. Apply hierarchical firewall policies on the organization level.

Discussion 0
Question # 69

For analysis purposes, you need to send all the logs from all of your Compute Engine instances to a BigQuery dataset called platform-logs. You have already installed the Stackdriver Logging agent on all the instances. You want to minimize cost. What should you do?

Options:

A.  

1. Give the BigQuery Data Editor role on the platform-logs dataset to the service accounts used by your instances.2. Update your instances’ metadata to add the following value: logs-destination: bq://platform-logs.

B.  

1. In Stackdriver Logging, create a logs export with a Cloud Pub/Sub topic called logs as a sink.2. Create a Cloud Function that is triggered by messages in the logs topic.3. Configure that Cloud Function to drop logs that are not from Compute Engine and to insert Compute Engine logs in the platform-logs dataset.

C.  

1. In Stackdriver Logging, create a filter to view only Compute Engine logs.2. Click Create Export.3. Choose BigQuery as Sink Service, and the platform-logs dataset as Sink Destination.

D.  

1. Create a Cloud Function that has the BigQuery User role on the platform-logs dataset.2. Configure this Cloud Function to create a BigQuery Job that executes this query:INSERT INTO dataset.platform-logs (timestamp, log)SELECT timestamp, log FROM compute.logsWHERE timestamp > DATE_SUB(CURRENT_DATE(), INTERVAL 1 DAY)3. Use Cloud Scheduler to trigger this Cloud Function once a day.

Discussion 0
Question # 70

You are performing a monthly security check of your Google Cloud environment and want to know who has access to view data stored in your Google Cloud

Project. What should you do?

Options:

A.  

Enable Audit Logs for all APIs that are related to data storage.

B.  

Review the IAM permissions for any role that allows for data access.

C.  

Review the Identity-Aware Proxy settings for each resource.

D.  

Create a Data Loss Prevention job.

Discussion 0
Get Associate-Cloud-Engineer dumps and pass your exam in 24 hours!

Free Exams Sample Questions