★ Pass on Your First TRY ★ 100% Money Back Guarantee ★ Realistic Practice Exam Questions

Free Instant Download NEW AWS-Certified-Database-Specialty Exam Dumps (PDF & VCE):
Available on: https://www.certleader.com/AWS-Certified-Database-Specialty-dumps.html


Act now and download your Amazon AWS-Certified-Database-Specialty test today! Do not waste time for the worthless Amazon AWS-Certified-Database-Specialty tutorials. Download Down to date Amazon AWS Certified Database - Specialty exam with real questions and answers and begin to learn Amazon AWS-Certified-Database-Specialty with a classic professional.

Free demo questions for Amazon AWS-Certified-Database-Specialty Exam Dumps Below:

NEW QUESTION 1
A business's production databases are housed on a 3 TB Amazon Aurora MySQL DB cluster. The database cluster is installed in the region us-east-1. For disaster recovery (DR) requirements, the company's database expert needs to fast deploy the DB cluster in another AWS Region to handle the production load with an RTO of less than two hours.
Which approach is the MOST OPERATIONALLY EFFECTIVE in meeting these requirements?

  • A. Implement an AWS Lambda function to take a snapshot of the production DB cluster every 2 hours, and copy that snapshot to an Amazon S3 bucket in the DR Regio
  • B. Restore the snapshot to an appropriately sized DB cluster in the DR Region.
  • C. Add a cross-Region read replica in the DR Region with the same instance type as the current primary instanc
  • D. If the read replica in the DR Region needs to be used for production, promote the read replica to become a standalone DB cluster.
  • E. Create a smaller DB cluster in the DR Regio
  • F. Configure an AWS Database Migration Service (AWS DMS) task with change data capture (CDC) enabled to replicate data from the current production DB cluster to the DB cluster in the DR Region.
  • G. Create an Aurora global database that spans two Region
  • H. Use AWS Database Migration Service (AWS DMS) to migrate the existing database to the new global database.

Answer: B

Explanation:
RTO is 2 hours. With 3 TB database, cross-region replica is a better option

NEW QUESTION 2
A company runs a customer relationship management (CRM) system that is hosted on-premises with a MySQL database as the backend. A custom stored procedure is used to send email notifications to another system when data is inserted into a table. The company has noticed that the performance of the CRM system has decreased due to database reporting applications used by various teams. The company requires an AWS solution that would reduce maintenance, improve performance, and accommodate the email notification feature.
Which AWS solution meets these requirements?

  • A. Use MySQL running on an Amazon EC2 instance with Auto Scaling to accommodate the reporting application
  • B. Configure a stored procedure and an AWS Lambda function that uses Amazon SES to send email notifications to the other system.
  • C. Use Amazon Aurora MySQL in a multi-master cluster to accommodate the reporting applications.Configure Amazon RDS event subscriptions to publish a message to an Amazon SNS topic and subscribe the other system's email address to the topic.
  • D. Use MySQL running on an Amazon EC2 instance with a read replica to accommodate the reporting application
  • E. Configure Amazon SES integration to send email notifications to the other system.
  • F. Use Amazon Aurora MySQL with a read replica for the reporting application
  • G. Configure a stored procedure and an AWS Lambda function to publish a message to an Amazon SNS topi
  • H. Subscribe the other system's email address to the topic.

Answer: D

Explanation:
RDS event subscriptions do not cover "data is inserted into a table" - see https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/USER_Events.Messages.html We can use stored procedure to invoke Lambda function - https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/AuroraMySQL.Integrating.Lambda.html

NEW QUESTION 3
A company has deployed an e-commerce web application in a new AWS account. An Amazon RDS for MySQL Multi-AZ DB instance is part of this deployment with a database-1.xxxxxxxxxxxx.us-east- 1.rds.amazonaws.com endpoint listening on port 3306. The company’s Database Specialist is able to log in to MySQL and run queries from the bastion host using these details.
When users try to utilize the application hosted in the AWS account, they are presented with a generic error message. The application servers are logging a “could not connect to server: Connection times out” error message to Amazon CloudWatch Logs.
What is the cause of this error?

  • A. The user name and password the application is using are incorrect.
  • B. The security group assigned to the application servers does not have the necessary rules to allow inbound connections from the DB instance.
  • C. The security group assigned to the DB instance does not have the necessary rules to allow inbound connections from the application servers.
  • D. The user name and password are correct, but the user is not authorized to use the DB instance.

Answer: C

NEW QUESTION 4
Developers have requested a new Amazon Redshift cluster so they can load new third-party marketing data. The new cluster is ready and the user credentials are given to the developers. The developers indicate that their copy jobs fail with the following error message:
“Amazon Invalid operation: S3ServiceException:Access Denied,Status 403,Error AccessDenied.”
The developers need to load this data soon, so a database specialist must act quickly to solve this issue. What is the MOST secure solution?

  • A. Create a new IAM role with the same user name as the Amazon Redshift developer user I
  • B. Provide the IAM role with read-only access to Amazon S3 with the assume role action.
  • C. Create a new IAM role with read-only access to the Amazon S3 bucket and include the assume role actio
  • D. Modify the Amazon Redshift cluster to add the IAM role.
  • E. Create a new IAM role with read-only access to the Amazon S3 bucket with the assume role actio
  • F. Add this role to the developer IAM user ID used for the copy job that ended with an error message.
  • G. Create a new IAM user with access keys and a new role with read-only access to the Amazon S3 bucket.Add this role to the Amazon Redshift cluste
  • H. Change the copy job to use the access keys created.

Answer: B

Explanation:
https://docs.aws.amazon.com/redshift/latest/gsg/rs-gsg-create-an-iam-role.html
"Now that you have created the new role, your next step is to attach it to your cluster. You can attach the role when you launch a new cluster or you can attach it to an existing cluster. In the next step, you attach the role to a new cluster."
https://docs.aws.amazon.com/redshift/latest/dg/copy-usage_notes-access-permissions.html

NEW QUESTION 5
A company migrated one of its business-critical database workloads to an Amazon Aurora Multi-AZ DB cluster. The company requires a very low RTO and needs to improve the application recovery time after database failovers.
Which approach meets these requirements?

  • A. Set the max_connections parameter to 16,000 in the instance-level parameter group.
  • B. Modify the client connection timeout to 300 seconds.
  • C. Create an Amazon RDS Proxy database proxy and update client connections to point to the proxy endpoint.
  • D. Enable the query cache at the instance level.

Answer: C

Explanation:
Amazon RDS Proxy allows applications to pool and share connections established with the database,
improving database efficiency and application scalability. With RDS Proxy, failover times for Aurora and RDS databases are reduced by up to 66% and database credentials, authentication, and access can be managed through integration with AWS Secrets Manager and AWS Identity and Access Management (IAM).
https://aws.amazon.com/rds/proxy/

NEW QUESTION 6
The Amazon CloudWatch metric for FreeLocalStorage on an Amazon Aurora MySQL DB instance shows that the amount of local storage is below 10 MB. A database engineer must increase the local storage available in the Aurora DB instance.
How should the database engineer meet this requirement?

  • A. Modify the DB instance to use an instance class that provides more local SSD storage.
  • B. Modify the Aurora DB cluster to enable automatic volume resizing.
  • C. Increase the local storage by upgrading the database engine version.
  • D. Modify the DB instance and configure the required storage volume in the configuration section.

Answer: A

Explanation:
https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/Aurora.AuroraMySQL.Monitoring.Metrics. Unlike for other DB engines, for Aurora DB instances this metric reports the amount of storage available to each DB instance. This value depends on the DB instance class (for pricing information, see the Amazon RDS product page). You can increase the amount of free storage space for an instance by choosing a larger DB instance class for your instance."

NEW QUESTION 7
A company maintains several databases using Amazon RDS for MySQL and PostgreSQL. Each RDS database generates log files with retention periods set to their default values. The company has now mandated that database logs be maintained for up to 90 days in a centralized repository to facilitate real-time and afterthe-fact analyses.
What should a Database Specialist do to meet these requirements with minimal effort?

  • A. Create an AWS Lambda function to pull logs from the RDS databases and consolidate the log files in an Amazon S3 bucke
  • B. Set a lifecycle policy to expire the objects after 90 days.
  • C. Modify the RDS databases to publish log to Amazon CloudWatch Log
  • D. Change the log retention policy for each log group to expire the events after 90 days.
  • E. Write a stored procedure in each RDS database to download the logs and consolidate the log files in an Amazon S3 bucke
  • F. Set a lifecycle policy to expire the objects after 90 days.
  • G. Create an AWS Lambda function to download the logs from the RDS databases and publish the logs to Amazon CloudWatch Log
  • H. Change the log retention policy for the log group to expire the events after 90 days.

Answer: B

Explanation:
https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_LogAccess.html https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_LogAccess.Procedural.UploadtoCloudWat https://aws.amazon.com/premiumsupport/knowledge-center/rds-aurora-mysql-logs-cloudwatch/ https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutRetentionPolicy.html

NEW QUESTION 8
A company is running an on-premises application comprised of a web tier, an application tier, and a MySQL database tier. The database is used primarily during business hours with random activity peaks throughout the day. A database specialist needs to improve the availability and reduce the cost of the MySQL database tier as part of the company’s migration to AWS.
Which MySQL database option would meet these requirements?

  • A. Amazon RDS for MySQL with Multi-AZ
  • B. Amazon Aurora Serverless MySQL cluster
  • C. Amazon Aurora MySQL cluster
  • D. Amazon RDS for MySQL with read replica

Answer: C

NEW QUESTION 9
A company has applications running on Amazon EC2 instances in a private subnet with no internet connectivity. The company deployed a new application that uses Amazon DynamoDB, but the application cannot connect to the DynamoDB tables. A developer already checked that all permissions are set correctly.
What should a database specialist do to resolve this issue while minimizing access to external resources?

  • A. Add a route to an internet gateway in the subnet’s route table.
  • B. Add a route to a NAT gateway in the subnet’s route table.
  • C. Assign a new security group to the EC2 instances with an outbound rule to ports 80 and 443.
  • D. Create a VPC endpoint for DynamoDB and add a route to the endpoint in the subnet’s route table.

Answer: D

Explanation:
https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/vpc-endpoints-dynamodb.html

NEW QUESTION 10
A Database Specialist is troubleshooting an application connection failure on an Amazon Aurora DB cluster with multiple Aurora Replicas that had been running with no issues for the past 2 months. The connection failure lasted for 5 minutes and corrected itself after that. The Database Specialist reviewed the Amazon RDS events and determined a failover event occurred at that time. The failover process took around 15 seconds to complete.
What is the MOST likely cause of the 5-minute connection outage?

  • A. After a database crash, Aurora needed to replay the redo log from the last database checkpoint
  • B. The client-side application is caching the DNS data and its TTL is set too high
  • C. After failover, the Aurora DB cluster needs time to warm up before accepting client connections
  • D. There were no active Aurora Replicas in the Aurora DB cluster

Answer: B

Explanation:
When your application tries to establish a connection after a failover, the new Aurora PostgreSQL writer will be a previous reader, which can be found using the Aurora read only endpoint before DNS updates have fully propagated. Setting the java DNS TTL to a low value helps cycle between reader nodes on subsequent connection attempts.
Amazon Aurora is designed to recover from a crash almost instantaneously and continue to serve your application data. Unlike other databases, after a crash Amazon Aurora does not need to replay the redo log from the last database checkpoint before making the database available for operations. Amazon Aurora performs crash recovery asynchronously on parallel threads, so your database is open and available immediately after a crash. Because the storage is organized in many small segments, each with its own redo log, the underlying storage can replay redo records on demand in parallel and asynchronously as part of a disk read after a crash. This approach reduces database restart times to less than 60 seconds in most cases

NEW QUESTION 11
AWS CloudFormation stack including an Amazon RDS database instance was mistakenly removed, resulting in the loss of recent data. A Database Specialist must apply RDS parameters to the CloudFormation template in order to minimize the possibility of future inadvertent instance data loss.
Which settings will satisfy this criterion? (Select three.)

  • A. Set DeletionProtection to True
  • B. Set MultiAZ to True
  • C. Set TerminationProtection to True
  • D. Set DeleteAutomatedBackups to False
  • E. Set DeletionPolicy to Delete
  • F. Set DeletionPolicy to Retain

Answer: ADF

Explanation:
A https://aws.amazon.com/about-aws/whats-new/2018/09/amazon-rds-now-provides-database-deletion-protection/
D https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_WorkingWithAutomatedBackups.html
F - https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-attribute-deletionpolicy.html

NEW QUESTION 12
A company is migrating a mission-critical 2-TB Oracle database from on premises to Amazon Aurora. The cost for the database migration must be kept to a minimum, and both the on-premises Oracle database and the Aurora DB cluster must remain open for write traffic until the company is ready to completely cut over to Aurora.
Which combination of actions should a database specialist take to accomplish this migration as quickly as possible? (Choose two.)

  • A. Use the AWS Schema Conversion Tool (AWS SCT) to convert the source database schem
  • B. Then restore the converted schema to the target Aurora DB cluster.
  • C. Use Oracle’s Data Pump tool to export a copy of the source database schema and manually edit the schema in a text editor to make it compatible with Aurora.
  • D. Create an AWS DMS task to migrate data from the Oracle database to the Aurora DB cluste
  • E. Select the migration type to replicate ongoing changes to keep the source and target databases in sync until the company is ready to move all user traffic to the Aurora DB cluster.
  • F. Create an AWS DMS task to migrate data from the Oracle database to the Aurora DB cluste
  • G. Once the initial load is complete, create an AWS Kinesis Data Firehose stream to perform change data capture (CDC) until the company is ready to move all user traffic to the Aurora DB cluster.
  • H. Create an AWS Glue job and related resources to migrate data from the Oracle database to the Aurora DB cluste
  • I. Once the initial load is complete, create an AWS DMS task to perform change data capture (CDC) until the company is ready to move all user traffic to the Aurora DB cluster.

Answer: AC

NEW QUESTION 13
A global digital advertising company captures browsing metadata to contextually display relevant images, pages, and links to targeted users. A single page load can generate multiple events that need to be stored individually. The maximum size of an event is 200 KB and the average size is 10 KB. Each page load must query the user’s browsing history to provide targeting recommendations. The advertising company expects over 1 billion page visits per day from users in the United States, Europe, Hong Kong, and India. The structure of the metadata varies depending on the event. Additionally, the browsing metadata must be written and read with very low latency to ensure a good viewing experience for the users.
Which database solution meets these requirements?

  • A. Amazon DocumentDB
  • B. Amazon RDS Multi-AZ deployment
  • C. Amazon DynamoDB global table
  • D. Amazon Aurora Global Database

Answer: C

NEW QUESTION 14
A company developed a new application that is deployed on Amazon EC2 instances behind an Application Load Balancer. The EC2 instances use the security group named sg-application-servers. The company needs a database to store the data from the application and decides to use an Amazon RDS for MySQL DB instance. The DB instance is deployed in private DB subnet.
What is the MOST restrictive configuration for the DB instance security group?

  • A. Only allow incoming traffic from the sg-application-servers security group on port 3306.
  • B. Only allow incoming traffic from the sg-application-servers security group on port 443.
  • C. Only allow incoming traffic from the subnet of the application servers on port 3306.
  • D. Only allow incoming traffic from the subnet of the application servers on port 443.

Answer: A

Explanation:
most restrictive approach is to allow only incoming connections from SG of EC2 instance on port 3306

NEW QUESTION 15
A manufacturing company’s website uses an Amazon Aurora PostgreSQL DB cluster.
Which configurations will result in the LEAST application downtime during a failover? (Choose three.)

  • A. Use the provided read and write Aurora endpoints to establish a connection to the Aurora DB cluster.
  • B. Create an Amazon CloudWatch alert triggering a restore in another Availability Zone when the primary Aurora DB cluster is unreachable.
  • C. Edit and enable Aurora DB cluster cache management in parameter groups.
  • D. Set TCP keepalive parameters to a high value.
  • E. Set JDBC connection string timeout variables to a low value.
  • F. Set Java DNS caching timeouts to a high value.

Answer: ABC

NEW QUESTION 16
Recently, an ecommerce business transferred one of its SQL Server databases to an Amazon RDS for SQL Server Enterprise Edition database instance. The corporation anticipates an increase in read traffic as a result of an approaching sale. To accommodate the projected read load, a database professional must establish a read replica of the database instance.
Which procedures should the database professional do prior to establishing the read replica? (Select two.)

  • A. Identify a potential downtime window and stop the application calls to the source DB instance.
  • B. Ensure that automatic backups are enabled for the source DB instance.
  • C. Ensure that the source DB instance is a Multi-AZ deployment with Always ON Availability Groups.
  • D. Ensure that the source DB instance is a Multi-AZ deployment with SQL Server Database Mirroring(DBM).
  • E. Modify the read replica parameter group setting and set the value to 1.

Answer: BC

Explanation:
https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/SQLServer.ReadReplicas.html

NEW QUESTION 17
A clothing company uses a custom ecommerce application and a PostgreSQL database to sell clothes to thousands of users from multiple countries. The company is migrating its application and database from its on- premises data center to the AWS Cloud. The company has selected Amazon EC2 for the application and Amazon RDS for PostgreSQL for the database. The company requires database passwords to be changed every 60 days. A Database Specialist needs to ensure that the credentials used by the web application to connect to the database are managed securely.
Which approach should the Database Specialist take to securely manage the database credentials?

  • A. Store the credentials in a text file in an Amazon S3 bucke
  • B. Restrict permissions on the bucket to the IAM role associated with the instance profile onl
  • C. Modify the application to download the text file and retrieve the credentials on start u
  • D. Update the text file every 60 days.
  • E. Configure IAM database authentication for the application to connect to the databas
  • F. Create an IAM user and map it to a separate database user for each ecommerce use
  • G. Require users to update their passwords every 60 days.
  • H. Store the credentials in AWS Secrets Manage
  • I. Restrict permissions on the secret to only the IAM role associated with the instance profil
  • J. Modify the application to retrieve the credentials from Secrets Manager on start u
  • K. Configure the rotation interval to 60 days.
  • L. Store the credentials in an encrypted text file in the application AM
  • M. Use AWS KMS to store the key for decrypting the text fil
  • N. Modify the application to decrypt the text file and retrieve the credentials on start u
  • O. Update the text file and publish a new AMI every 60 days.

Answer: C

NEW QUESTION 18
......

100% Valid and Newest Version AWS-Certified-Database-Specialty Questions & Answers shared by Allfreedumps.com, Get Full Dumps HERE: https://www.allfreedumps.com/AWS-Certified-Database-Specialty-dumps.html (New 270 Q&As)