Pre-Summer Sale Special - Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: xmaspas7

Easiest Solution 2 Pass Your Certification Exams

Professional-Cloud-DevOps-Engineer Google Cloud Certified - Professional Cloud DevOps Engineer Exam Free Practice Exam Questions (2026 Updated)

Prepare effectively for your Google Professional-Cloud-DevOps-Engineer Google Cloud Certified - Professional Cloud DevOps Engineer Exam certification with our extensive collection of free, high-quality practice questions. Each question is designed to mirror the actual exam format and objectives, complete with comprehensive answers and detailed explanations. Our materials are regularly updated for 2026, ensuring you have the most current resources to build confidence and succeed on your first attempt.

You are designing a system with three different environments: development, quality assurance (QA), and production.

Each environment will be deployed with Terraform and has a Google Kubemetes Engine (GKE) cluster created so that application teams can deploy their applications. Anthos Config Management will be used and templated to deploy

infrastructure level resources in each GKE cluster. All users (for example, infrastructure operators and application owners) will use GitOps. How should you structure your source control repositories for both Infrastructure as Code (laC) and application code?

A.

Cloud Infrastructure (Terraform) repository is shared: different directories are different environmentsGKE Infrastructure (Anthos Config Management Kustomize manifests) repository is shared: differentoverlay directories are different environmentsApplication (app source code) repositories are separated: different branches are different features

B.

Cloud Infrastructure (Terraform) repository is shared: different directories are different environmentsGKE Infrastructure (Anthos Config Management Kustomize manifests) repositories are separated:different branches are different environmentsApplication (app source code) repositories are separated: different branches are different features

C.

Cloud Infrastructure (Terraform) repository is shared: different branches are different environmentsGKE Infrastructure (Anthos Config Management Kustomize manifests) repository is shared: differentoverlay directories are different environmentsApplication (app source code) repository is shared: different directories are different features

D.

Cloud Infrastructure (Terraform) repositories are separated: different branches are different environmentsGKE Infrastructure (Anthos Config Management Kustomize manifests) repositories are separated:different overlay directories are different environmentsApplication (app source code) repositories are separated: different branches are different features

You work for a global organization and are running a monolithic application on Compute Engine You need to select the machine type for the application to use that optimizes CPU utilization by using the fewest number of steps You want to use historical system metncs to identify the machine type for the application to use You want to follow Google-recommended practices What should you do?

A.

Use the Recommender API and apply the suggested recommendations

B.

Create an Agent Policy to automatically install Ops Agent in all VMs

C.

Install the Ops Agent in a fleet of VMs by using the gcloud CLI

D.

Review the Cloud Monitoring dashboard for the VM and choose the machine type with the lowest CPU utilization

You use Artifact Registry to store container images built with Cloud Build. You need to ensure that all existing and new images are continuously scanned for vulnerabilities. You also want to track who pushed each image to the registry. What should you do?

A.

Configure Artifact Registry to automatically trigger vulnerability scans for new image tags, and view scan results. Use Cloud Audit Logs to track image tag creation events.

B.

Configure Artifact Registry to automatically scan new images and periodically re-scan all images. Use Cloud Audit Logs to track image uploads and identify the user who pushed each image.

C.

Configure Artifact Registry to automatically re-scan images daily. Enable Cloud Audit Logs to track these scans, and use Logs Explorer to identify vulnerabilities.

D.

Configure Artifact Registry to send vulnerability scan results to a Cloud Storage bucket. Use a separate script to parse results and notify a security team.

You deploy a new release of an internal application during a weekend maintenance window when there is minimal user traffic. After the window ends, you learn that one of the new features isn't working as expected in the production environment. After an extended outage, you roll back the new release and deploy a fix. You want to modify your release process to reduce the mean time to recovery so you can avoid extended outages in the future. What should you do?

Choose 2 answers

A.

Before merging new code, require 2 different peers to review the code changes.

B.

Adopt the blue/green deployment strategy when releasing new code via a CD server.

C.

Integrate a code linting tool to validate coding standards before any code is accepted into the repository.

D.

Require developers to run automated integration tests on their local development environments before release.

E.

Configure a CI server.Add a suite of unit tests to your code and have your CI server run them on commit and verify any changes.

You have an application running in Google Kubernetes Engine. The application invokes multiple services per request but responds too slowly. You need to identify which downstream service or services are causing the delay. What should you do?

A.

Analyze VPC flow logs along the path of the request.

B.

Investigate the Liveness and Readiness probes for each service.

C.

Create a Dataflow pipeline to analyze service metrics in real time.

D.

Use a distributed tracing framework such as OpenTelemetry or Stackdriver Trace.

You are configuring a CI pipeline. The build step for your CI pipeline integration testing requires access to APIs inside your private VPC network. Your security team requires that you do not expose API traffic publicly. You need to implement a solution that minimizes management overhead. What should you do?

A.

Use Cloud Build private pools to connect to the private VPC.

B.

Use Cloud Build to create a Compute Engine instance in the private VPC. Run the integration tests on the VM by using a startup script.

C.

Use Cloud Build as a pipeline runner. Configure a cross-region internal Application Load Balancer for API access.

D.

Use Cloud Build as a pipeline runner. Configure a global external Application Load Balancer with a Google Cloud Armor policy for API access.

Your company follows Site Reliability Engineering principles. You are writing a postmortem for an incident, triggered by a software change, that severely affected users. You want to prevent severe incidents from happening in the future. What should you do?

A.

Identify engineers responsible for the incident and escalate to their senior management.

B.

Ensure that test cases that catch errors of this type are run successfully before new software releases.

C.

Follow up with the employees who reviewed the changes and prescribe practices they should follow in the future.

D.

Design a policy that will require on-call teams to immediately call engineers and management to discuss a plan of action if an incident occurs.

You are building an application that runs on Cloud Run The application needs to access a third-party API by using an API key You need to determine a secure way to store and use the API key in your application by following Google-recommended practices What should you do?

A.

Save the API key in Secret Manager as a secret Reference the secret as an environment variable in the Cloud Run application

B.

Save the API key in Secret Manager as a secret key Mount the secret key under the /sys/api_key directory and decrypt the key in the Cloud Run application

C.

Save the API key in Cloud Key Management Service (Cloud KMS) as a key Reference the key as an environment variable in the Cloud Run application

D.

Encrypt the API key by using Cloud Key Management Service (Cloud KMS) and pass the key to Cloud Run as an environment variable Decrypt and use the key in Cloud Run

Your company runs an e-commerce business. The application responsible for payment processing has structured JSON logging with the following schema:

Capture and access of logs from the payment processing application is mandatory for operations, but the jsonPayload.user_email field contains personally identifiable information (PII). Your security team does not want the entire engineering team to have access to PII. You need to stop exposing PII to the engineering team and restrict access to security team members only. What should you do?

A.

Apply a jsonPayload.user_email exclusion filter to the _Default bucket.

B.

Apply the conditional role binding resource.name.extract("locations/global/buckets/(bucket)/") == "_Default" to the _Default bucket.

C.

Apply a jsonPayload.user_email restricted field to the _Default bucket. Grant the Log Field Accessor role to the security team members.

D.

Modify the application to toggle inclusion of user_email when the log_user_email environment variable is set to true. Restrict the engineering team members who can change the production environment variable by using the CODEOWNERS file.

Your organization is starting to containerize with Google Cloud. You need a fully managed storage solution for container images and Helm charts. You need to identify a storage solution that has native integration into existing Google Cloud services, including Google Kubernetes Engine (GKE), Cloud Run, VPC Service Controls, and Identity and Access Management (IAM). What should you do?

A.

Use Docker to configure a Cloud Storage driver pointed at the bucket owned by your organization.

B.

Configure Container Registry as an OCI-based container registry for container images.

C.

Configure Artifact Registry as an OCI-based container registry for both Helm charts and container images.

D.

Configure an open source container registry server to run in GKE with a restrictive role-based access control (RBAC) configuration.

You are working with a government agency that requires you to archive application logs for seven years. You need to configure Stackdriver to export and store the logs while minimizing costs of storage. What should you do?

A.

Create a Cloud Storage bucket and develop your application to send logs directly to the bucket.

B.

Develop an App Engine application that pulls the logs from Stackdriver and saves them in BigQuery.

C.

Create an export in Stackdriver and configure Cloud Pub/Sub to store logs in permanent storage for seven years.

D.

Create a sink in Stackdriver, name it, create a bucket on Cloud Storage for storing archived logs, and then select the bucket as the log export destination.

Your organization uses a change advisory board (CAB) to approve all changes to an existing service You want to revise this process to eliminate any negative impact on the software delivery performance What should you do?

Choose 2 answers

A.

Replace the CAB with a senior manager to ensure continuous oversight from development to deployment

B.

Let developers merge their own changes but ensure that the team's deployment platform can roll back changes if any issues are discovered

C.

Move to a peer-review based process for individual changes that is enforced at code check-in time and supported by automated tests

D.

Batch changes into larger but less frequent software releases

E.

Ensure that the team's development platform enables developers to get fast feedback on the impact of their changes

You manage an application that runs in Google Kubernetes Engine (GKE) and uses the blue/green deployment methodology Extracts of the Kubernetes manifests are shown below:

The Deployment app-green was updated to use the new version of the application During post-deployment monitoring you notice that the majority of user requests are failing You did not observe this behavior in the testing environment You need to mitigate the incident impact on users and enable the developers to troubleshoot the issue What should you do?

A.

Update the Deployment app-blue to use the new version of the application

B.

Update the Deployment ape-green to use the previous version of the application

C.

Change the selector on the Service app-2vc to app: my-app.

D.

Change the selector on the Service app-svc to app: my-app, version: blue

You are building and running client applications in Cloud Run and Cloud Functions Your client requires that all logs must be available for one year so that the client can import the logs into their logging service You must minimize required code changes What should you do?

A.

Update all images in Cloud Run and all functions in Cloud Functions to send logs to both Cloud Logging andthe client's logging service Ensure that all the ports required to send logs are open in the VPC firewall

B.

Create a Pub/Sub topic subscription and logging sink Configure the logging sink to send all logs into thetopic Give your client access to the topic to retrieve the logs

C.

Create a storage bucket and appropriate VPC firewall rules Update all images in Cloud Run and allfunctions in Cloud Functions to send logs to a file within the storage bucket

D.

Create a logs bucket and logging sink. Set the retention on the logs bucket to 365 days Configure thelogging sink to send logs to the bucket Give your client access to the bucket to retrieve the logs

Your team is designing a new application for deployment both inside and outside Google Cloud Platform (GCP). You need to collect detailed metrics such as system resource utilization. You want to use centralized GCP services while minimizing the amount of work required to set up this collection system. What should you do?

A.

Import the Stackdriver Profiler package, and configure it to relay function timing data to Stackdriver for further analysis.

B.

Import the Stackdriver Debugger package, and configure the application to emit debug messages with timing information.

C.

Instrument the code using a timing library, and publish the metrics via a health check endpoint that is scraped by Stackdriver.

D.

Install an Application Performance Monitoring (APM) tool in both locations, and configure an export to a central data storage location for analysis.

You are investigating issues in your production application that runs on Google Kubernetes Engine (GKE). You determined that the source Of the issue is a recently updated container image, although the exact change in code was not identified. The deployment is currently pointing to the latest tag. You need to update your cluster to run a version of the container that functions as intended. What should you do?

A.

Create a new tag called stable that points to the previously working container, and change the deployment to point to the new tag.

B.

Apply the latest tag to the previous container image, and do a rolling update on the deployment.

C.

Build a new container from a previous Git tag, and do a rolling update on the deployment to the new container.

D.

Alter the deployment to point to the sha2 56 digest of the previously working container.

You are performing a semi-annual capacity planning exercise for your flagship service You expect a service user growth rate of 10% month-over-month for the next six months Your service is fully containerized and runs on a Google Kubemetes Engine (GKE) standard cluster across three zones with cluster autoscaling enabled You currently consume about 30% of your total deployed CPU capacity and you require resilience against the failure of a zone. You want to ensure that your users experience minimal negative impact as a result of this growth o' as a result of zone failure while you avoid unnecessary costs How should you prepare to handle the predicted growth?

A.

Verify the maximum node pool size enable a Horizontal Pod Autoscaler and then perform a load lest to verify your expected resource needs

B.

Because you deployed the service on GKE and are using a cluster autoscaler your GKE cluster will scale automatically regardless of growth rate

C.

Because you are only using 30% of deployed CPU capacity there is significant headroom and you do not need to add any additional capacity for this rate of growth

D.

Proactively add 80% more node capacity to account for six months of 10% growth rate and then perform a load test to ensure that you have enough capacity

You are responding to a high-priority incident where a critical, user-facing payment service is experiencing a 50% error rate. The cause is a non-critical, batch analytics Dataflow pipeline flooding a shared Memorystore for Redis instance with writes, which has spiked read latency for the payment service. A full rollback of the Dataflow pipeline's deployment will take 15 minutes to complete through your CI/CD process. You need to restore the payment service as quickly as possible. What should you do?

A.

Use Cloud Profiler to inspect the Dataflow pipeline's execution graph to pinpoint the source of the excessive writes.

B.

In the Google Cloud console, edit the Memorystore for Redis instance and increase its capacity tier.

C.

Initiate an automated rollback of the Dataflow pipeline's deployment to revert to the last stable version.

D.

Cancel the active Dataflow job.

You support a trading application written in Python and hosted on App Engine flexible environment. You want to customize the error information being sent to Stackdriver Error Reporting. What should you do?

A.

Install the Stackdriver Error Reporting library for Python, and then run your code on a Compute Engine VM.

B.

Install the Stackdriver Error Reporting library for Python, and then run your code on Google Kubernetes Engine.

C.

Install the Stackdriver Error Reporting library for Python, and then run your code on App Engine flexible environment.

D.

Use the Stackdriver Error Reporting API to write errors from your application to ReportedErrorEvent, and then generate log entries with properly formatted error messages in Stackdriver Logging.

Your company has recently experienced several production service issues. You need to create a Cloud Monitoring dashboard to troubleshoot the issues, and you want to use the dashboard to distinguish between failures in your own service and those caused by a Google Cloud service that you use. What should you do?

A.

Enable Personalized Service Health annotations on the dashboard.

B.

Create an alerting policy for the system error metrics.

C.

Create a log-based metric to track cloud service errors, and display the metric on the dashboard.

D.

Create a logs widget to display system errors from Cloud Logging on the dashboard.

Copyright © 2014-2026 Solution2Pass. All Rights Reserved