Summer Special - Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: top65certs

Free and Premium Google Professional-Cloud-Database-Engineer Dumps Questions Answers

Google Cloud Certified - Professional Cloud Database Engineer Questions and Answers

Question 1

Your organization has an existing app that just went viral. The app uses a Cloud SQL for MySQL backend database that is experiencing slow disk performance while using hard disk drives (HDDs). You need to improve performance and reduce disk I/O wait times. What should you do?

Options:

A.

Export the data from the existing instance, and import the data into a new instance with solid-state drives (SSDs).

B.

Edit the instance to change the storage type from HDD to SSD.

C.

Create a high availability (HA) failover instance with SSDs, and perform a failover to the new instance.

D.

Create a read replica of the instance with SSDs, and perform a failover to the new instance

Buy Now
Question 2

You are choosing a new database backend for an existing application. The current database is running PostgreSQL on an on-premises VM and is managed by a database administrator and operations team. The application data is relational and has light traffic. You want to minimize costs and the migration effort for this application. What should you do?

Options:

A.

Migrate the existing database to Firestore.

B.

Migrate the existing database to Cloud SQL for PostgreSQL.

C.

Migrate the existing database to Cloud Spanner.

D.

Migrate the existing database to PostgreSQL running on Compute Engine.

Question 3

You currently have a MySQL database running on Cloud SQL with a read replica in a different zone for non-mission critical analytics workloads. You want to enable high availability (HA) for the analytic workloads while keeping costs low. What should you do?

Options:

A.

Increase the size of the read replica Instance and enable MA.

B.

Enable HA on the current read replica.

C.

Create a new HA instance in the same zone db lie primary.

D.

Create a new MA Instance in a different region than the primary.

Question 4

Your customer is running a MySQL database on-premises with read replicas. The nightly incremental backups are expensive and add maintenance overhead. You want to follow Google-recommended practices to migrate the database to Google Cloud, and you need to ensure minimal downtime. What should you do?

Options:

A.

Create a Google Kubernetes Engine (GKE) cluster, install MySQL on the cluster, and then import the dump file.

B.

Use the mysqldump utility to take a backup of the existing on-premises database, and then import it into Cloud SQL.

C.

Create a Compute Engine VM, install MySQL on the VM, and then import the dump file.

D.

Create an external replica, and use Cloud SQL to synchronize the data to the replica.

Question 5

You are managing a Cloud SQL for PostgreSQL instance in Google Cloud. You have a primary instance in region 1 and a read replica in region 2. After a failure of region 1, you need to make the Cloud SQL instance available again. You want to minimize data loss and follow Google-recommended practices. What should you do?

Options:

A.

Restore the Cloud SQL instance from the automatic backups in region 3.

B.

Restore the Cloud SQL instance from the automatic backups in another zone in region 1.

C.

Check "Lag Bytes" in the monitoring dashboard for the primary instance in the read replica instance. Check the replication status usingpg_catalog.pg_last_wal_receive_lsn(). Then, fail over to region 2 by promoting the read replica instance.

D.

Check your instance operational log for the automatic failover status. Look for time, type, and status of the operations. If the failover operation is successful, no action is necessary. Otherwise, manually perform gcloud sql instances failover .

Question 6

You are managing a Cloud SQL for PostgreSQL instance in Google Cloud. You need to test the high availability of your Cloud SQL instance by performing a failover. You want to use the cloud command.

What should you do?

Options:

A.

Use gcloud sql instances failover .

B.

Use gcloud sql instances failover .

C.

Use gcloud sql instances promote-replica .

D.

Use gcloud sql instances promote-replica .

Question 7

Your application uses Cloud SQL for MySQL. Your users run reports on data that relies on near-real time; however, the additional analytics caused excessive load on the primary database. You created a read replica for the analytics workloads, but now your users are complaining about the lag in data changes and that their reports are still slow.You need to improve the report performance and shorten the lag in data replication without making changes to the current reports. Which two approaches should you implement? (Choose two.)

Options:

A.

Create secondary indexes on the replica.

B.

Create additional read replicas, and partition your analytics users to use different read replicas.

C.

Disable replication on the read replica, and set the flag for parallel replication on the read replica. Re-enable replication and optimize performance by setting flags on the primary instance.

D.

Disable replication on the primary instance, and set the flag for parallel replication on the primary instance. Re-enable replication and optimize performance by setting flags on the read replica.

E.

Move your analytics workloads to BigQuery, and set up a streaming pipeline to move data and update BigQuery.

Question 8

You are managing a Cloud SQL for MySQL environment in Google Cloud. You have deployed a primary instance in Zone A and a read replica instance in Zone B, both in the same region. You are notified that the replica instance in Zone B was unavailable for 10 minutes. You need to ensure that the read replica instance is still working. What should you do?

Options:

A.

Use the Google Cloud Console or gcloud CLI to manually create a new clone database.

B.

Use the Google Cloud Console or gcloud CLI to manually create a new failover replica from backup.

C.

Verify that the new replica is created automatically.

D.

Start the original primary instance and resume replication.

Question 9

Your company uses the Cloud SQL out-of-disk recommender to analyze the storage utilization trends of production databases over the last 30 days. Your database operations team uses these recommendations to proactively monitor storage utilization and implement corrective actions. You receive a recommendation that the instance is likely to run out of disk space. What should you do to address this storage alert?

Options:

A.

Normalize the database to the third normal form.

B.

Compress the data using a different compression algorithm.

C.

Manually or automatically increase the storage capacity.

D.

Create another schema to load older data.

Question 10

You are designing a physician portal app in Node.js. This application will be used in hospitals and clinics that might have intermittent internet connectivity. If a connectivity failure occurs, the app should be able to query the cached data. You need to ensure that the application has scalability, strong consistency, and multi-region replication. What should you do?

Options:

A.

Use Firestore and ensure that the PersistenceEnabled option is set to true.

B.

Use Memorystore for Memcached.

C.

Use Pub/Sub to synchronize the changes from the application to Cloud Spanner.

D.

Use Table.read with the exactStaleness option to perform a read of rows in Cloud Spanner.

Question 11

You are migrating critical production database from Amazon RDS for MySQL to Cloud SQL for MYSQL by using Google Cloud’s Migration Service.

You want to keep disruption to your production database to minimum and, at the same time, optimize migration performance. What should you do?

Options:

A.

Create and start multiple Database Migration Service jobs to migrate your database to the target Cloud SQL for MySQL instance.

B.

Upgrade the Amazon RDS for MySQL primary instance to an instance with more vCPUs and memory, and then run Google Cloud's Database Migration Service.

C.

Create a single Database Migration Service migration job with initial load parallelism configured to maximum on the source Amazon RDS for MySQL read replica.

D.

Create a single Database Migration Service migration job with initial Load Parallelism configured to Maximum on the Amazon RDS for MySQL primary instance.

Question 12

You are managing two different applications: Order Management and Sales Reporting. Both applications interact with the same Cloud SQL for MySQL database. The Order Management application reads and writes to the database 24/7, but the Sales Reporting application is read-only. Both applications need the latest data. You need to ensure that the Performance of the Order Management application is not affected by the Sales Reporting application. What should you do?

Options:

A.

Create a read replica for the Sales Reporting application.

B.

Create two separate databases in the instance, and perform dual writes from the Order Management application.

C.

Use a Cloud SQL federated query for the Sales Reporting application.

D.

Queue up all the requested reports in PubSub, and execute the reports at night.

Question 13

You are the primary DBA of a Cloud SQL for PostgreSQL database that supports 6 enterprise applications in production. You used Cloud SQL Insights to identify inefficient queries and now need to identify the application that is originating the inefficient queries. You want to follow Google-recommended practices. What should you do?

Options:

A.

Shut down and restart each application.

B.

Write a utility to scan database query logs.

C.

Write a utility to scan application logs.

D.

Use query tags to add application-centric database monitoring.

Question 14

Your team recently released a new version of a highly consumed application to accommodate additional user traffic. Shortly after the release, you received an alert from your production monitoring team that there is consistently high replication lag between your primary instance and the read replicas of your Cloud SQL for MySQL instances. You need to resolve the replication lag. What should you do?

Options:

A.

Identify and optimize slow running queries, or set parallel replication flags.

B.

Stop all running queries, and re-create the replicas.

C.

Edit the primary instance to upgrade to a larger disk, and increase vCPU count.

D.

Edit the primary instance to add additional memory.

Question 15

Your company is migrating their MySQL database to Cloud SQL and cannot afford any planned downtime during the month of December. The company is also concerned with cost, so you need the most cost-effective solution. What should you do?

Options:

A.

Open a support ticket in Google Cloud to prevent any maintenance in that MySQL instance during the month of December.

B.

Use Cloud SQL maintenance settings to prevent any maintenance during the month of December.

C.

Create MySQL read replicas in different zones so that, if any downtime occurs, the read replicas will act as the primary instance during the month of December.

D.

Create a MySQL regional instance so that, if any downtime occurs, the standby instance will act as the primary instance during the month of December.

Question 16

You are running an instance of Cloud Spanner as the backend of your ecommerce website. You learn that the quality assurance (QA) team has doubled the number of their test cases. You need to create a copy of your Cloud Spanner database in a new test environment to accommodate the additional test cases. You want to follow Google-recommended practices. What should you do?

Options:

A.

Use Cloud Functions to run the export in Avro format.

B.

Use Cloud Functions to run the export in text format.

C.

Use Dataflow to run the export in Avro format.

D.

Use Dataflow to run the export in text format.

Question 17

You work in the logistics department. Your data analysis team needs daily extracts from Cloud SQL for MySQL to train a machine learning model. The model will be used to optimize next-day routes. You need to export the data in CSV format. You want to follow Google-recommended practices. What should you do?

Options:

A.

Use Cloud Scheduler to trigger a Cloud Function that will run a select * from table(s) query to call the cloudsql.instances.export API.

B.

Use Cloud Scheduler to trigger a Cloud Function through Pub/Sub to call the cloudsql.instances.export API.

C.

Use Cloud Composer to orchestrate an export by calling the cloudsql.instances.export API.

D.

Use Cloud Composer to execute a select * from table(s) query and export results.

Question 18

You want to migrate an existing on-premises application to Google Cloud. Your application supports semi-structured data ingested from 100,000 sensors, and each sensor sends 10 readings per second from manufacturing plants. You need to make this data available for real-time monitoring and analysis. What should you do?

Options:

A.

Deploy the database using Cloud SQL.

B.

Use BigQuery, and load data in batches.

C.

Deploy the database using Bigtable.

D.

Deploy the database using Cloud Spanner.

Question 19

You are a DBA of Cloud SQL for PostgreSQL. You want the applications to have password-less authentication for read and write access to the database. Which authentication mechanism should you use?

Options:

A.

Use Identity and Access Management (IAM) authentication.

B.

Use Managed Active Directory authentication.

C.

Use Cloud SQL federated queries.

D.

Use PostgreSQL database's built-in authentication.

Question 20

You work for a large retail and ecommerce company that is starting to extend their business globally. Your company plans to migrate to Google Cloud. You want to use platforms that will scale easily, handle transactions with the least amount of latency, and provide a reliable customer experience. You need a storage layer for sales transactions and current inventory levels. You want to retain the same relational schema that your existing platform uses. What should you do?

Options:

A.

Store your data in Firestore in a multi-region location, and place your compute resources in one of the constituent regions.

B.

Deploy Cloud Spanner using a multi-region instance, and place your compute resources close to the default leader region.

C.

Build an in-memory cache in Memorystore, and deploy to the specific geographic regions where your application resides.

D.

Deploy a Bigtable instance with a cluster in one region and a replica cluster in another geographic region.

Question 21

Your online delivery business that primarily serves retail customers uses Cloud SQL for MySQL for its inventory and scheduling application. The required recovery time objective (RTO) and recovery point objective (RPO) must be in minutes rather than hours as a part of your high availability and disaster recovery design. You need a high availability configuration that can recover without data loss during a zonal or a regional failure. What should you do?

Options:

A.

Set up all read replicas in a different region using asynchronous replication.

B.

Set up all read replicas in the same region as the primary instance with synchronous replication.

C.

Set up read replicas in different zones of the same region as the primary instance with synchronous replication, and set up read replicas in different regions with asynchronous replication.

D.

Set up read replicas in different zones of the same region as the primary instance with asynchronous replication, and set up read replicas in different regions with synchronous replication.

Question 22

Your organization has a critical business app that is running with a Cloud SQL for MySQL backend database. Your company wants to build the most fault-tolerant and highly available solution possible. You need to ensure that the application database can survive a zonal and regional failure with a primary region of us-central1 and the backup region of us-east1. What should you do?

Options:

A.

Provision a Cloud SQL for MySQL instance in us-central1-a.

Create a multiple-zone instance in us-west1-b.

Create a read replica in us-east1-c.

B.

Provision a Cloud SQL for MySQL instance in us-central1-a.

Create a multiple-zone instance in us-central1-b.

Create a read replica in us-east1-b.

C.

Provision a Cloud SQL for MySQL instance in us-central1-a.

Create a multiple-zone instance in us-east-b.

Create a read replica in us-east1-c.

D.

Provision a Cloud SQL for MySQL instance in us-central1-a.

Create a multiple-zone instance in us-east1-b.

Create a read replica in us-central1-b.

Question 23

Your digital-native business runs its database workloads on Cloud SQL. Your website must be globally accessible 24/7. You need to prepare your Cloud SQL instance for high availability (HA). You want to follow Google-recommended practices. What should you do? (Choose two.)

Options:

A.

Set up manual backups.

B.

Create a PostgreSQL database on-premises as the HA option.

C.

Configure single zone availability for automated backups.

D.

Enable point-in-time recovery.

E.

Schedule automated backups.

Question 24

You are designing a highly available (HA) Cloud SQL for PostgreSQL instance that will be used by 100 databases. Each database contains 80 tables that were migrated from your on-premises environment to Google Cloud. The applications that use these databases are located in multiple regions in the US, and you need to ensure that read and write operations have low latency. What should you do?

Options:

A.

Deploy 2 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-east1 and us-west1.

B.

Deploy 2 Cloud SQL instances in the us-central1 region, and create read replicas in us-east1 and us-west1.

C.

Deploy 4 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-central1, us-east1, and us-west1.

D.

Deploy 4 Cloud SQL instances in the us-central1 region, and create read replicas in us-central1, us-east1 and us-west1.

Question 25

Your company's mission-critical, globally available application is supported by a Cloud Spanner database. Experienced users of the application have read and write access to the database, but new users are assigned read-only access to the database. You need to assign the appropriate Cloud Spanner Identity and Access Management (IAM) role to new users being onboarded soon. What roles should you set up?

Options:

A.

roles/spanner.databaseReader

B.

roles/spanner.databaseUser

C.

roles/spanner.viewer

D.

roles/spanner.backupWriter

Question 26

Your organization has a busy transactional Cloud SQL for MySQL instance. Your analytics team needs access to the data so they can build monthly sales reports. You need to provide data access to the analytics team without adversely affecting performance. What should you do?

Options:

A.

Create a read replica of the database, provide the database IP address, username, and password to the analytics team, and grant read access to required tables to the team.

B.

Create a read replica of the database, enable the cloudsql.iam_authentication flag on the replica, and grant read access to required tables to the analytics team.

C.

Enable the cloudsql.iam_authentication flag on the primary database instance, and grant read access to required tables to the analytics team.

D.

Provide the database IP address, username, and password of the primary database instance to the analytics, team, and grant read access to required tables to the team.

Question 27

You want to migrate an on-premises mission-critical PostgreSQL database to Cloud SQL. The database must be able to withstand a zonal failure with less than five minutes of downtime and still not lose any transactions. You want to follow Google-recommended practices for the migration. What should you do?

Options:

A.

Take nightly snapshots of the primary database instance, and restore them in a secondary zone.

B.

Build a change data capture (CDC) pipeline to read transactions from the primary instance, and replicate them to a secondary instance.

C.

Create a read replica in another region, and promote the read replica if a failure occurs.

D.

Enable high availability (HA) for the database to make it regional.

Question 28

Your company is launching a new globally distributed application with strict requirements for low latency, strong consistency, zero downtime, and high availability (HA). You need to configure a scalable database solution to support anticipated rapid growth and optimal application performance. What should you do?

Options:

A.

Create a Cloud SQL instance in HA mode with a cross-region read replica.

B.

Create an AlloyDB instance in HA mode with a cross-region read replica.

C.

Create a spanner instance across regions for optimal performance.

D.

Implement Bigtable with replication across multiple regions and configure to prioritize data accuracy.

Question 29

Your DevOps team is using Terraform to deploy applications and Cloud SQL databases. After every new application change is rolled out, the environment is torn down and recreated, and the persistent database layer is lost. You need to prevent the database from being dropped. What should you do?

Options:

A.

Set Terraform deletion_protection to true.

B.

Rerun terraform apply.

C.

Create a read replica.

D.

Use point-in-time-recovery (PITR) to recover the database.

Question 30

You need to redesign the architecture of an application that currently uses Cloud SQL for PostgreSQL. The users of the application complain about slow query response times. You want to enhance your application architecture to offer sub-millisecond query latency. What should you do?

Options:

A.

Configure Firestore, and modify your application to offload queries.

B.

Configure Bigtable, and modify your application to offload queries.

C.

Configure Cloud SQL for PostgreSQL read replicas to offload queries.

D.

Configure Memorystore, and modify your application to offload queries.

Question 31

You are designing a new gaming application that uses a highly transactional relational database to store player authentication and inventory data in Google Cloud. You want to launch the game in multiple regions. What should you do?

Options:

A.

Use Cloud Spanner to deploy the database.

B.

Use Bigtable with clusters in multiple regions to deploy the database

C.

Use BigQuery to deploy the database

D.

Use Cloud SQL with a regional read replica to deploy the database.

Question 32

Your retail organization is preparing for the holiday season. Use of catalog services is increasing, and your DevOps team is supporting the Cloud SQL databases that power a microservices-based application. The DevOps team has added instrumentation through Sqlcommenter. You need to identify the root cause of why certain microservice calls are failing. What should you do?

Options:

A.

Watch Query Insights for long running queries.

B.

Watch the Cloud SQL instance monitor for CPU utilization metrics.

C.

Watch the Cloud SQL recommenders for overprovisioned instances.

D.

Watch Cloud Trace for application requests that are failing.

Question 33

You are managing a mission-critical Cloud SQL for PostgreSQL instance. Your application team is running important transactions on the database when another DBA starts an on-demand backup. You want to verify the status of the backup. What should you do?

Options:

A.

Check the cloudsql.googleapis.com/postgres.log instance log.

B.

Perform the gcloud sql operations list command.

C.

Use Cloud Audit Logs to verify the status.

D.

Use the Google Cloud Console.

Question 34

You are migrating an on-premises application to Google Cloud. The application requires a high availability (HA) PostgreSQL database to support business-critical functions. Your company's disaster recovery strategy requires a recovery time objective (RTO) and recovery point objective (RPO) within 30 minutes of failure. You plan to use a Google Cloud managed service. What should you do to maximize uptime for your application?

Options:

A.

Deploy Cloud SQL for PostgreSQL in a regional configuration. Create a read replica in a different zone in the same region and a read replica in another region for disaster recovery.

B.

Deploy Cloud SQL for PostgreSQL in a regional configuration with HA enabled. Take periodic backups, and use this backup to restore to a new Cloud SQL for PostgreSQL instance in another region during a disaster recovery event.

C.

Deploy Cloud SQL for PostgreSQL in a regional configuration with HA enabled. Create a cross-region read replica, and promote the read replica as the primary node for disaster recovery.

D.

Migrate the PostgreSQL database to multi-regional Cloud Spanner so that a single region outage will not affect your application. Update the schema to support Cloud Spanner data types, and refactor the application.

Question 35

You are building a data warehouse on BigQuery. Sources of data include several MySQL databases located on-premises.

You need to transfer data from these databases into BigQuery for analytics. You want to use a managed solution that has low latency and is easy to set up. What should you do?

Options:

A.

Create extracts from your on-premises databases periodically, and push these extracts to Cloud Storage.

Upload the changes into BigQuery, and merge them with existing tables.

B.

Use Cloud Data Fusion and scheduled workflows to extract data from MySQL. Transform this data into the appropriate schema, and load this data into your BigQuery database.

C.

Use Datastream to connect to your on-premises database and create a stream. Have Datastream write to Cloud Storage. Then use Dataflow to process the data into BigQuery.

D.

Use Database Migration Service to replicate data to a Cloud SQL for MySQL instance. Create federated tables in BigQuery on top of the replicated instances to transform and load the data into your BigQuery database.

Question 36

You use Python scripts to generate weekly SQL reports to assess the state of your databases and determine whether you need to reorganize tables or run statistics. You want to automate this report but need to minimize operational costs and overhead. What should you do?

Options:

A.

Create a VM in Compute Engine, and run a cron job.

B.

Create a Cloud Composer instance, and create a directed acyclic graph (DAG).

C.

Create a Cloud Function, and call the Cloud Function using Cloud Scheduler.

D.

Create a Cloud Function, and call the Cloud Function from a Cloud Tasks queue.

Question 37

Your company has PostgreSQL databases on-premises and on Amazon Web Services (AWS). You are planning multiple database migrations to Cloud SQL in an effort to reduce costs and downtime. You want to follow Google-recommended practices anduse Google native data migration tools. You also want to closely monitor the migrations as part of the cutover strategy. What should you do?

Options:

A.

Use Database Migration Service to migrate all databases to Cloud SQL.

B.

Use Database Migration Service for one-time migrations, and use third-party or partner tools for change data capture (CDC) style migrations.

C.

Use data replication tools and CDC tools to enable migration.

D.

Use a combination of Database Migration Service and partner tools to support the data migration strategy.

Question 38

You need to provision several hundred Cloud SQL for MySQL instances for multiple project teams over a one-week period. You must ensure that all instances adhere to company standards such as instance naming conventions, database flags, and tags. What should you do?

Options:

A.

Automate instance creation by writing a Dataflow job.

B.

Automate instance creation by setting up Terraform scripts.

C.

Create the instances using the Google Cloud Console UI.

D.

Create clones from a template Cloud SQL instance.

Question 39

You need to migrate existing databases from Microsoft SQL Server 2016 Standard Edition on a single Windows Server 2019 Datacenter Edition to a single Cloud SQL for SQL Server instance. During the discovery phase of your project, you notice that your on-premises server peaks at around 25,000 read IOPS. You need to ensure that your Cloud SQL instance is sized appropriately to maximize read performance. What should you do?

Options:

A.

Create a SQL Server 2019 Standard on Standard machine type with 4 vCPUs, 15 GB of RAM, and 800 GB of solid-state drive (SSD).

B.

Create a SQL Server 2019 Standard on High Memory machine type with at least 16 vCPUs, 104 GB of RAM, and 200 GB of SSD.

C.

Create a SQL Server 2019 Standard on High Memory machine type with 16 vCPUs, 104 GB of RAM, and 4 TB of SSD.

D.

Create a SQL Server 2019 Enterprise on High Memory machine type with 16 vCPUs, 104 GB of RAM, and 500 GB of SSD.

Question 40

You support a consumer inventory application that runs on a multi-region instance of Cloud Spanner. A customer opened a support ticket to complain about slow response times. You notice a Cloud Monitoring alert about high CPU utilization. You want to follow Google-recommended practices to address the CPU performance issue. What should you do first?

Options:

A.

Increase the number of processing units.

B.

Modify the database schema, and add additional indexes.

C.

Shard data required by the application into multiple instances.

D.

Decrease the number of processing units.

Question 41

You are configuring a new application that has access to an existing Cloud Spanner database. The new application reads from this database to gather statistics for a dashboard. You want to follow Google-recommended practices when granting Identity and Access Management (IAM) permissions. What should you do?

Options:

A.

Reuse the existing service account that populates this database.

B.

Create a new service account, and grant it the Cloud Spanner Database Admin role.

C.

Create a new service account, and grant it the Cloud Spanner Database Reader role.

D.

Create a new service account, and grant it the spanner.databases.select permission.

Question 42

You are running a large, highly transactional application on Oracle Real Application Cluster (RAC) that is multi-tenant and uses shared storage. You need a solution that ensures high-performance throughput and a low-latency connection between applications and databases. The solution must also support existing Oracle features and provide ease of migration to Google Cloud. What should you do?

Options:

A.

Migrate to Compute Engine.

B.

Migrate to Bare Metal Solution for Oracle.

C.

Migrate to Google Kubernetes Engine (GKE)

D.

Migrate to Google Cloud VMware Engine