AWS Certified Database – Specialty – Question171

A database specialist is launching a test graph database using Amazon Neptune for the first time. The database specialist needs to insert millions of rows of test observations from a .csv file that is stored in Amazon S3. The database specialist has been using a series of API calls to upload the data to the Neptune DB instance.
Which combination of steps would allow the database specialist to upload the data faster? (Choose three.)

A.
Ensure Amazon Cognito returns the proper AWS STS tokens to authenticate the Neptune DB instance to the S3 bucket hosting the CSV file.
B. Ensure the vertices and edges are specified in different .csv files with proper header column formatting.
C. Use AWS DMS to move data from Amazon S3 to the Neptune Loader.
D. Curl the S3 URI while inside the Neptune DB instance and then run the addVertex or addEdge commands.
E. Ensure an IAM role for the Neptune DB instance is configured with the appropriate permissions to allow access to the file in the S3 bucket.
F. Create an S3 VPC endpoint and issue an HTTP POST to the database's loader endpoint.

AWS Certified Database – Specialty – Question170

A company uses an Amazon Redshift cluster to run its analytical workloads. Corporate policy requires that the company's data be encrypted at rest with customer managed keys. The company's disaster recovery plan requires that backups of the cluster be copied into another AWS Region on a regular basis.
How should a database specialist automate the process of backing up the cluster data in compliance with these policies?

A.
Copy the AWS Key Management Service (AWS KMS) customer managed key from the source Region to the destination Region. Set up an AWS Glue job in the source Region to copy the latest snapshot of the Amazon Redshift cluster from the source Region to the destination Region. Use a time-based schedule in AWS Glue to run the job on a daily basis.
B. Create a new AWS Key Management Service (AWS KMS) customer managed key in the destination Region. Create a snapshot copy grant in the destination Region specifying the new key. In the source Region, configure cross-Region snapshots for the Amazon Redshift cluster specifying the destination Region, the snapshot copy grant, and retention periods for the snapshot.
C. Copy the AWS Key Management Service (AWS KMS) customer-managed key from the source Region to the destination Region. Create Amazon S3 buckets in each Region using the keys from their respective Regions. Use Amazon EventBridge (Amazon CloudWatch Events) to schedule an AWS Lambda function in the source Region to copy the latest snapshot to the S3 bucket in that Region. Configure S3 Cross-Region Replication to copy the snapshots to the destination Region, specifying the source and destination KMS key IDs in the replication configuration.
D. Use the same customer-supplied key materials to create a CMK with the same private key in the destination Region. Configure cross-Region snapshots in the source Region targeting the destination Region. Specify the corresponding CMK in the destination Region to encrypt the snapshot.

Correct Answer: A

Explanation:

AWS Certified Database – Specialty – Question169

A company has an on-premises SQL Server database. The users access the database using Active Directory authentication. The company successfully migrated its database to Amazon RDS for SQL Server.
However, the company is concerned about user authentication in the AWS Cloud environment.
Which solution should a database specialist provide for the user to authenticate?

A.
Deploy Active Directory Federation Services (AD FS) on premises and configure it with an on-premises Active Directory. Set up delegation between the on-premises AD FS and AWS Security Token Service (AWS STS) to map user identities to a role using theAmazonRDSDirectoryServiceAccess managed IAM policy.
B. Establish a forest trust between the on-premises Active Directory and AWS Directory Service for Microsoft Active Directory. Use AWS SSO to configure an Active Directory user delegated to access the databases in RDS for SQL Server.
C. Use Active Directory Connector to redirect directory requests to the company's on-premises Active Directory without caching any information in the cloud. Use the RDS master user credentials to connect to the DB instance and configure SQL Server logins and users from the Active Directory users and groups.
D. Establish a forest trust between the on-premises Active Directory and AWS Directory Service for Microsoft Active Directory. Ensure RDS for SQL Server is using mixed mode authentication. Use the RDS master user credentials to connect to the DB instance and configure SQL Server logins and users from the Active Directory users and groups.

AWS Certified Database – Specialty – Question168

A company is running its production databases in a 3 TB Amazon Aurora MySQL DB cluster. The DB cluster is deployed to the us-east-1 Region. For disaster recovery (DR) purposes, the company's database specialist needs to make the DB cluster rapidly available in another AWS Region to cover the production load with an RTO of less than 2 hours.
What is the MOST operationally efficient solution to meet these requirements?

A.
Implement an AWS Lambda function to take a snapshot of the production DB cluster every 2 hours, and copy that snapshot to an Amazon S3 bucket in the DR Region. Restore the snapshot to an appropriately sized DB cluster in the DR Region.
B. Add a cross-Region read replica in the DR Region with the same instance type as the current primary instance. If the read replica in the DR Region needs to be used for production, promote the read replica to become a standalone DB cluster.
C. Create a smaller DB cluster in the DR Region. Configure an AWS Database Migration Service (AWS DMS) task with change data capture (CDC) enabled to replicate data from the current production DB cluster to the DB cluster in the DR Region.
D. Create an Aurora global database that spans two Regions. Use AWS Database Migration Service (AWS DMS) to migrate the existing database to the new global database.

AWS Certified Database – Specialty – Question167

A database specialist needs to configure an Amazon RDS for MySQL DB instance to close non-interactive connections that are inactive after 900 seconds.
What should the database specialist do to accomplish this task?

A.
Create a custom DB parameter group and set the wait_timeout parameter value to 900. Associate the DB instance with the custom parameter group.
B. Connect to the MySQL database and run the SET SESSION wait_timeout=900 command.
C. Edit the my.cnf file and set the wait_timeout parameter value to 900. Restart the DB instance.
D. Modify the default DB parameter group and set the wait_timeout parameter value to 900.

Correct Answer: B

Explanation:

Explanation:
If we set the wait_timeout variable for a session, it will valid only for a particular session. But when we set the wait_timeout variable globally it will valid for all the sessions.
Reference: https://dilsichandrasena.medium.com/changing-mysql-wait-timeout-var…

AWS Certified Database – Specialty – Question166

A company is migrating an IBM Informix database to a Multi-AZ deployment of Amazon RDS for SQL Server with Always On Availability Groups (AGs). SQL Server Agent jobs on the Always On AG listener run at 5-minute intervals to synchronize data between the Informix database and the SQL Server database.
Users experience hours of stale data after a successful failover to the secondary node with minimal latency.
What should a database specialist do to ensure that users see recent data after a failover?

A.
Set TTL to less than 30 seconds for cached DNS values on the Always On AG listener.
B. Break up large transactions into multiple smaller transactions that complete in less than 5 minutes.
C. Set the databases on the secondary node to read-only mode.
D. Create the SQL Server Agent jobs on the secondary node from a script when the secondary node takes over after a failure.

Correct Answer: C

Explanation:

Explanation:
After a failover, client applications that need to access the primary databases must connect to the new primary replica. Also, if the new secondary replica is configured to allow read-only access, read-only client applications can connect to it.
Reference: https://docs.microsoft.com/en-us/sql/database-engine/availability-g…

AWS Certified Database – Specialty – Question165

A company is planning on migrating a 500-GB database from Oracle to Amazon Aurora PostgreSQL using the AWS Schema Conversion Tool (AWS SCT) and AWS DMS. The database does not have any stored procedures to migrate but has some tables that are large or partitioned. The application is critical for business so a migration with minimal downtime is preferred.
Which combination of steps should a database specialist take to accelerate the migration process? (Choose three.)

A.
Use the AWS SCT data extraction agent to migrate the schema from Oracle to Aurora PostgreSQL.
B. For the large tables, change the setting for the maximum number of tables to load in parallel and perform a full load using AWS DMS.
C. For the large tables, create a table settings rule with a parallel load option in AWS DMS, then perform a full load using DMS.
D. Use AWS DMS to set up change data capture (CDC) for continuous replication until the cutover date.
E. Use AWS SCT to convert the schema from Oracle to Aurora PostgreSQL.
F. Use AWS DMS to convert the schema from Oracle to Aurora PostgreSQL and for continuous replication.

AWS Certified Database – Specialty – Question164

A database specialist has a fleet of Amazon RDS DB instances that use the default DB parameter group.
The database specialist needs to associate a custom parameter group with some of the DB instances.
After the database specialist makes this change, when will the instances be assigned to this new parameter group?

A.
Instantaneously after the change is made to the parameter group
B. In the next scheduled maintenance window of the DB instances
C. After the DB instances are manually rebooted
D. Within 24 hours after the change is made to the parameter group

Correct Answer: C

Explanation:

Explanation:
To apply the latest parameter changes to that DB instance, manually reboot the DB instance.
Reference: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_Working…

AWS Certified Database – Specialty – Question163

A bank plans to use an Amazon RDS for MySQL DB instance. The database should support read-intensive traffic with very few repeated queries.
Which solution meets these requirements?

A.
Create an Amazon ElastiCache cluster. Use a write-through strategy to populate the cache.
B. Create an Amazon ElastiCache cluster. Use a lazy loading strategy to populate the cache.
C. Change the DB instance to Multi-AZ with a standby instance in another AWS Region.
D. Create a read replica of the DB instance. Use the read replica to distribute the read traffic.

AWS Certified Database – Specialty – Question162

A company is running its customer feedback application on Amazon Aurora MySQL. The company runs a report every day to extract customer feedback, and a team reads the feedback to determine if the customer comments are positive or negative. It sometimes takes days before the company can contact unhappy customers and take corrective measures. The company wants to use machine learning to automate this workflow.
Which solution meets this requirement with the LEAST amount of effort?

A.
Export the Aurora MySQL database to Amazon S3 by using AWS Database Migration Service (AWS DMS). Use Amazon Comprehend to run sentiment analysis on the exported files.
B. Export the Aurora MySQL database to Amazon S3 by using AWS Database Migration Service (AWS DMS). Use Amazon SageMaker to run sentiment analysis on the exported files.
C. Set up Aurora native integration with Amazon Comprehend. Use SQL functions to extract sentiment analysis.
D. Set up Aurora native integration with Amazon SageMaker. Use SQL functions to extract sentiment analysis.

Correct Answer: C

Explanation:

Explanation:
Amazon Comprehend is a natural-language processing (NLP) service that uses machine learning to uncover valuable insights and connections in text.
Reference: https://aws.amazon.com/comprehend/