Book Image

AWS Certified Database – Specialty (DBS-C01) Certification Guide

By : Kate Gawron
5 (1)
Book Image

AWS Certified Database – Specialty (DBS-C01) Certification Guide

5 (1)
By: Kate Gawron

Overview of this book

The AWS Certified Database – Specialty certification is one of the most challenging AWS certifications. It validates your comprehensive understanding of databases, including the concepts of design, migration, deployment, access, maintenance, automation, monitoring, security, and troubleshooting. With this guide, you'll understand how to use various AWS databases, such as Aurora Serverless and Global Database, and even services such as Redshift and Neptune. You’ll start with an introduction to the AWS databases, and then delve into workload-specific database design. As you advance through the chapters, you'll learn about migrating and deploying the databases, along with database security techniques such as encryption, auditing, and access controls. This AWS book will also cover monitoring, troubleshooting, and disaster recovery techniques, before testing all the knowledge you've gained throughout the book with the help of mock tests. By the end of this book, you'll have covered everything you need to pass the DBS-C01 AWS certification exam and have a handy, on-the-job desk reference guide.
Table of Contents (24 chapters)
1
Part 1: Introduction to Databases on AWS
Free Chapter
2
Chapter 1: AWS Certified Database – Specialty Overview
5
Part 2: Workload-Specific Database Design
12
Part 3: Deployment and Migration and Database Security
16
Part 4: Monitoring and Optimization
20
Part 5: Assessment
21
Chapter 16: Exam Practice

Review

To check your knowledge from this chapter, here are five questions that you should now be able to answer. Remember the exam techniques from Chapter 1, AWS Certified Database – Specialty Exam Overview, and remove the clearly incorrect answers first to help you:

  1. You are working as a developer for a small company with a DynamoDB table. The company is complaining of poor performance after increasing the number of records in the table and they say they are seeing "throttling errors." What is the most cost-efficient option for them to consider?
    1. Enable TTL against a timestamp attribute.
    2. Implement DAX.
    3. Turn on Dynamo Streams.
    4. Turn on autoscaling.
  2. You are designing a new DynamoDB table and need to calculate how many Capacity Units (CUs) to provision. Each item is 3 KB in size and you expect to read a maximum of 100 items and write 10 per second. You will only be using eventually consistent reads and standard writes.
    1. 50 RCUs and 30 WCUs
    2. 100 RCUs and 10 WCUs
    3. 50 RCUs...