New Year Special Sale - Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: mxmas70

Home > Google > Cloud Database Engineer > Professional-Cloud-Database-Engineer

Professional-Cloud-Database-Engineer Google Cloud Certified - Professional Cloud Database Engineer Question and Answers

Question # 4

You are choosing a new database backend for an existing application. The current database is running PostgreSQL on an on-premises VM and is managed by a database administrator and operations team. The application data is relational and has light traffic. You want to minimize costs and the migration effort for this application. What should you do?

A.

Migrate the existing database to Firestore.

B.

Migrate the existing database to Cloud SQL for PostgreSQL.

C.

Migrate the existing database to Cloud Spanner.

D.

Migrate the existing database to PostgreSQL running on Compute Engine.

Full Access
Question # 5

You are the database administrator of a Cloud SQL for PostgreSQL instance that has pgaudit disabled. Users are complaining that their queries are taking longer to execute and performance has degraded over the past few months. You need to collect and analyze query performance data to help identity slow-running queries. What should you do?

A.

View Cloud SQL operations to view historical query information.

B.

White a Logs Explorer query to identify database queries with high execution times.

C.

Review application logs to identify database calls.

D.

Use the Query Insights dashboard to identify high execution times.

Full Access
Question # 6

Your company wants to migrate an Oracle-based application to Google Cloud. The application team currently uses Oracle Recovery Manager (RMAN) to back up the database to tape for long-term retention (LTR). You need a cost-effective backup and restore solution that meets a 2-hour recovery time objective (RTO) and a 15-minute recovery point objective (RPO). What should you do?

A.

Migrate the Oracle databases to Bare Metal Solution for Oracle, and store backups on tapes on-premises.

B.

Migrate the Oracle databases to Bare Metal Solution for Oracle, and use Actifio to store backup files on Cloud Storage using the Nearline Storage class.

C.

Migrate the Oracle databases to Bare Metal Solution for Oracle, and back up the Oracle databases to Cloud Storage using the Standard Storage class.

D.

Migrate the Oracle databases to Compute Engine, and store backups on tapes on-premises.

Full Access
Question # 7

You are migrating your data center to Google Cloud. You plan to migrate your applications to Compute Engine and your Oracle databases to Bare Metal Solution for Oracle. You must ensure that the applications in different projects can communicate securely and efficiently with the Oracle databases. What should you do?

A.

Set up a Shared VPC, configure multiple service projects, and create firewall rules.

B.

Set up Serverless VPC Access.

C.

Set up Private Service Connect.

D.

Set up Traffic Director.

Full Access
Question # 8

You are designing a highly available (HA) Cloud SQL for PostgreSQL instance that will be used by 100 databases. Each database contains 80 tables that were migrated from your on-premises environment to Google Cloud. The applications that use these databases are located in multiple regions in the US, and you need to ensure that read and write operations have low latency. What should you do?

A.

Deploy 2 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-east1 and us-west1.

B.

Deploy 2 Cloud SQL instances in the us-central1 region, and create read replicas in us-east1 and us-west1.

C.

Deploy 4 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-central1, us-east1, and us-west1.

D.

Deploy 4 Cloud SQL instances in the us-central1 region, and create read replicas in us-central1, us-east1 and us-west1.

Full Access
Question # 9

Your organization has a production Cloud SQL for MySQL instance. Your instance is configured with 16 vCPUs and 104 GB of RAM that is running between 90% and 100% CPU utilization for most of the day. You need to scale up the database and add vCPUs with minimal interruption and effort. What should you do?

A.

Issue a gcloud sql instances patch command to increase the number of vCPUs.

B.

Update a MySQL database flag to increase the number of vCPUs.

C.

Issue a gcloud compute instances update command to increase the number of vCPUs.

D.

Back up the database, create an instance with additional vCPUs, and restore the database.

Full Access
Question # 10

You are evaluating Cloud SQL for PostgreSQL as a possible destination for your on-premises PostgreSQL instances. Geography is becoming increasingly relevant to customer privacy worldwide. Your solution must support data residency requirements and include a strategy to:

configure where data is stored

control where the encryption keys are stored

govern the access to data

What should you do?

A.

Replicate Cloud SQL databases across different zones.

B.

Create a Cloud SQL for PostgreSQL instance on Google Cloud for the data that does not need to adhere to data residency requirements. Keep the data that must adhere to data residency requirements on-premises. Make application changes to support both databases.

C.

Allow application access to data only if the users are in the same region as the Google Cloud region for the Cloud SQL for PostgreSQL database.

D.

Use features like customer-managed encryption keys (CMEK), VPC Service Controls, and Identity and Access Management (IAM) policies.

Full Access
Question # 11

An analytics team needs to read data out of Cloud SQL for SQL Server and update a table in Cloud Spanner. You need to create a service account and grant least privilege access using predefined roles. What roles should you assign to the service account?

A.

roles/cloudsql.viewer and roles/spanner.databaseUser

B.

roles/cloudsql.editor and roles/spanner.admin

C.

roles/cloudsql.client and roles/spanner.databaseReader

D.

roles/cloudsql.instanceUser and roles/spanner.databaseUser

Full Access
Question # 12

Your team recently released a new version of a highly consumed application to accommodate additional user traffic. Shortly after the release, you received an alert from your production monitoring team that there is consistently high replication lag between your primary instance and the read replicas of your Cloud SQL for MySQL instances. You need to resolve the replication lag. What should you do?

A.

Identify and optimize slow running queries, or set parallel replication flags.

B.

Stop all running queries, and re-create the replicas.

C.

Edit the primary instance to upgrade to a larger disk, and increase vCPU count.

D.

Edit the primary instance to add additional memory.

Full Access
Question # 13

You work in the logistics department. Your data analysis team needs daily extracts from Cloud SQL for MySQL to train a machine learning model. The model will be used to optimize next-day routes. You need to export the data in CSV format. You want to follow Google-recommended practices. What should you do?

A.

Use Cloud Scheduler to trigger a Cloud Function that will run a select * from table(s) query to call the cloudsql.instances.export API.

B.

Use Cloud Scheduler to trigger a Cloud Function through Pub/Sub to call the cloudsql.instances.export API.

C.

Use Cloud Composer to orchestrate an export by calling the cloudsql.instances.export API.

D.

Use Cloud Composer to execute a select * from table(s) query and export results.

Full Access
Question # 14

Your organization has strict policies on tracking rollouts to production and periodically shares this information with external auditors to meet compliance requirements. You need to enable auditing on several Cloud Spanner databases. What should you do?

A.

Use replication to roll out changes to higher environments.

B.

Use backup and restore to roll out changes to higher environments.

C.

Use Liquibase to roll out changes to higher environments.

D.

Manually capture detailed DBA audit logs when changes are rolled out to higher environments.

Full Access
Question # 15

Your company is using Cloud SQL for MySQL with an internal (private) IP address and wants to replicate some tables into BigQuery in near-real time for analytics and machine learning. You need to ensure that replication is fast and reliable and uses Google-managed services. What should you do?

A.

Develop a custom data replication service to send data into BigQuery.

B.

Use Cloud SQL federated queries.

C.

Use Database Migration Service to replicate tables into BigQuery.

D.

Use Datastream to capture changes, and use Dataflow to write those changes to BigQuery.

Full Access
Question # 16

You want to migrate your on-premises PostgreSQL database to Compute Engine. You need to migrate this database with the minimum downtime possible. What should you do?

A.

Perform a full backup of your on-premises PostgreSQL, and then, in the migration window, perform an incremental backup.

B.

Create a read replica on Cloud SQL, and then promote it to a read/write standalone instance.

C.

Use Database Migration Service to migrate your database.

D.

Create a hot standby on Compute Engine, and use PgBouncer to switch over the connections.

Full Access
Question # 17

Your company is developing a new global transactional application that must be ACID-compliant and have 99.999% availability. You are responsible for selecting the appropriate Google Cloud database to serve as a datastore for this new application. What should you do?

A.

Use Firestore.

B.

Use Cloud Spanner.

C.

Use Cloud SQL.

D.

Use Bigtable.

Full Access