CloudAcademy: Knowledge Check: High Availability (SAA-C03) Flashcards

1
Q

What is an AWS region?

A. A small group of AWS-managed data centers located relatively close together
B. A single physical data center hosting AWS virtual resources
C. A large data cache in a highly populated area that delivers CDN content
D. A large, defined geographic area that contains multiple availability zones.

A

D. A large, defined geographic area that contains multiple availability zones.

Explanation:
A region is a collection of availability zones that are geographically located close to one other. This is generally indicated by AZs within the same city. AWS has deployed them across the globe to allow its worldwide customer base to take advantage of low latency connections. Every Region will act independently of the others, and each will contain at least three Availability Zones.

The other answers describe an availability zone, a common misconception for an availability zone, and an edge location.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How can AWS Snowball assist you in migrating on-premises data to S3 if the process will take more than a week using your existing connection?

A. By using on-premises hardware that is then shipped to AWS for transfer to a Snowball device
B. By compressing data on one device for transfer to S3 via a high-speed network connection
C. By saving data to multiple AWS Snowball devices and shipping them to AWS for transfer to S3
D. By transferring data from S3 via a high-speed network connection

A

C. By saving data to multiple AWS Snowball devices and shipping them to AWS for transfer to S3

Explanation:
As a general rule, if your data retrieval will take longer than a week using your existing connection method, then you should consider using AWS Snowball. AWS snowball is a physical device shipped to you from AWS. It does not use Internet connections to migrate data to S3, and it is not a hard-drive component that you ship to AWS for data transfer to S3.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Which statement regarding DynamoDB’s point-in-time recovery is incorrect?

A. Tables can only be restored in the same region as the original table.
B. It allows you to restore data back to a specific date and time.
C. Point-in-time recovery is disabled by default.
D. Tables can only restore to a point in time within the last 35 days.

A

A. Tables can only be restored in the same region as the original table.

Explanation:
Point In Time Recovery operates at the table level, and when enabled provides you with the ability to perform a point in time recovery for any time between the current time and the last 35 days. This feature needs to be enabled as it is disabled by default. Once enabled you can request a recovery by specifying a date and time with second precision or granularity. The restoration will always be performed into a new table - of which you specify the new table name at the time of the restoration request. Table restoration can be performed in the same region as the original table, or into a different region altogether.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Amazon RDS offers read replicas. They also offer secondary databases through the multi-AZ configuration.

What are two distinct differences between read replicas and secondary databases? (Choose 2 answers)

A. Read replicas can process read requests for the primary database. Secondary databases cannot.
B. Secondary databases are designed to minimize data loss during failover. Read replicas are not.
C. Secondary databases can replace primary databases when they fail. Read replicas cannot.
D. Secondary bases can process write requests for the primary database. Read replicas cannot.

A

A. Read replicas can process read requests for the primary database. Secondary databases cannot.
B. Secondary databases are designed to minimize data loss during failover. Read replicas are not.

Explanation:
A secondary database’s entire purpose is to minimize data loss when the primary database fails. They do not process any traffic - neither reads nor writes.

A read replica only processes read traffic for the primary database, not write traffic. It can be promoted to replace the primary database, if the primary fails. However, it will not minimize the data loss or potential service outage as effectively as a secondary database would.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

When using Amazon Glacier Flexible Retrieval’s expedited retrieval option, which of the following is correct?

A. Amazon Glacier takes 1 hour to retrieve data.
B. Amazon Glacier takes 1-5 minutes to retrieve data.
C. Amazon Glacier takes 5-12 hours to retrieve data.
D. Amazon Glacier takes 3-5 hours to retrieve data.

A

B. Amazon Glacier takes 1-5 minutes to retrieve data.

Explanation:
Amazon Glacier is an extremely low-cost storage service that provides secure and durable storage for data archiving and backup. To keep costs low, Amazon Glacier is optimized for data that is infrequently accessed and for which retrieval times of several hours are suitable. The standard retrieval option, which is the default option, takes 3-5 hours to complete. The other options are expedited, which downloads a small amount of data (250 MB maximum) in 5 minutes, and bulk, which downloads large amounts of data (petabytes) in 5-12 hours.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Which of the following statements about DynamoDB Accelerator (DAX) is incorrect?

A. DAX nodes are located with VPC subnets, while standard DynamoDB endpoints are located outside the VPC.
B. Client read requests are directed to DAX first, and then DynamoDB.
C. DAX can process table operation requests, such as CreateTable or DeleteTable.
D. Client write requests are written to DynamoDB first, and then DAX.

A

C. DAX can process table operation requests, such as CreateTable or DeleteTable.

Explanation:
If a request received by DAX from your client is a read request, such as a GetItem, BatchGetItem, Query or Scan, then the DAX cluster will try and process the request if it has the data cached. If DAX does not have the request in its cache (a cache miss) then the request will be sent to DynamoDB for the results to be returned to the client. These results will also then be stored by DAX within its cache and distributed to the remaining read replicas in the DAX cluster.

With regards to any write requested made by the client, the data is first written to DynamoDB before it is written to the cache of the DAX cluster.

One final point I want to make is that DAX does not process any requests relating to table operations and management, for example, if you wanted to create, update or delete tables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Which AWS disaster recovery method has the lowest RTO and RPO, but the highest cost required to maintain resources?

A.Backup and restore
B. Pilot light
C. Warm standby
D. Multi-site

A

D. Multi-site

Explanation:
Multi-site is the quickest disaster recovery method, with minimal time to recover and minimal data loss, but with the highest overall cost in comparison to backup and restore, pilot light or warm standby methods. This method is ideal for highly transactional online businesses, whose potential losses in terms of revenue and reputation far outweigh any costs required to maintain a duplicate production environment in a separate AWS region.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Which of the following should you use as a backup solution in case an S3 bucket storage object is accidentally deleted?

A. Lifecycle Policies
B. Multi-Factor Authentication Delete
C. Versioning
D. Bucket Policies

A

C. Versioning

Explanation:
Enabling versioning on an S3 bucket ensures you can recover from misuse of an object or accidental deletion, and revert back to an older version of the same data object.

Multi-Factor Authentication Delete ensures that a user has to enter a 6 digit MFA code to delete an object, which prevents accidental deletion due to human error.

Lifecycle Policies allow you to automatically manage and move data between classes, allowing specific data to be relocated based on compliance and governance controls you might have in place.

Bucket Policies are JSON policies assigned to individual buckets. These Bucket Policies can also define who or what has access to that bucket’s contents, but they do not control versioning of objects within buckets.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

In regards to disaster recovery, what is a recovery point objective (RPO)?

A. The time it takes to scale a functioning resource vertically to adapt to increased demand.
B. The time it takes after a disruption to restore a system to its service level.
C. The time it takes to scale a functioning resource horizontally to adapt to increased demand.
D. It is the acceptable amount of data loss measured in time.

A

D. It is the acceptable amount of data loss measured in time.

Explanation:
The recovery point objective (RPO) is the amount of acceptable data loss measured in time. For example, a transactional business cannot afford a great deal of data loss, so its RPO is likely less than one hour because an hour is a great expense in terms of data loss. However, for a more isolated website with minimal customer interaction, an RPO of a day may be acceptable because this represents a tolerable amount of data loss.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Which disaster recovery method generally has the highest RTO and RPO, but requires the least cost, in terms of resources, to restore service?

A. Warm standby
B. Pilot light
C. Backup and restore
D. Multi-site

A

C. Backup and restore

Explanation:
With the backup and restore method of recovery, data is backed up to an AWS storage service, such as Amazon S3. The data can be imported into AWS using a variety of options such as storage gateway, AWS Snowball (import/export), Direct Connect, VPN or the internet. In the event of a disaster, archives can be recovered from Amazon S3, and the data can then be restored directly to cloud resources. This method takes the longest time, but requires the least investment because it is essentially a system of archiving data consistently and retrieving it when needed. Aside from data transfer and storage, no complex applications or smaller version of your service needs to be kept running offsite.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Which statement regarding the replication of DynamoDB partition data is correct?

A. The data stored on DynamoDB partitions is replicated asynchronously across Availability Zones.
B. The data stored on DynamoDB partitions is replicated synchronously across Availability Zones.
C. The data stored on DynamoDB partitions is not replicated across availability zones.
D. The data stored on DynamoDB partitions can be replicated synchronously or asynchronously across Availability Zones, depending on the database configuration.

A

A. The data stored on DynamoDB partitions is replicated asynchronously across Availability Zones.

Explanation:
A DynamoDb partition is a dedicated area of SSD storage allocated to a table and for which is automatically replicated synchronously across 3 availability zones within a particular region. DynamoDB being a managed service takes care of performing both the partition management and replication for you, therefore you can remain focused on your application design and not be distracted by the needs of data replication requirements. The synchronous AZ replication provides protection against any single node outage and/or a full availability zone outage - which although is a rare event should never be assumed to not happen. The synchronous replication takes place using low latency interconnects between each of the availability zones within a region and ensures high-speed sub second replication.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Which type of Amazon Aurora connection endpoint load balances connections across the read replica fleet within the cluster?

A. Custom Endpoints
B. Reader Endpoints
C. Instance Endpoints
D. Cluster Endpoints

A

B. Reader Endpoints

Explanation:

Cluster Endpoint: The cluster endpoint points to the current master database instance. Using the Cluster endpoint allows your application to perform read and writes against the master instance.
Reader Endpoint: The reader endpoint load balancers connections across the read replica fleet within the cluster.
Custom Endpoint: A custom endpoint load balancer's connections across a set of cluster instances that you choose and register within the custom endpoint. Custom endpoints can be used to group instances based on instance size or maybe group them on a particular db parameter group. You can then dedicate the custom endpoint for a specific role or task within your organization - for example, you may have a requirement to generate month-end reports - therefore you connect to a custom endpoint that has been specifically set up for this task. 
Instance Endpoint: An instance endpoint maps directly to a cluster instance. Each and every cluster instance has its own instance endpoint. You can use an instance endpoint when you want fine-grained control over which instance you need to service your requests.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is an AWS Availability Zone?

A. An isolated location that contains a single AWS-managed data center
B. A large cache located in highly populated areas to deliver CDN content
C. A single AWS-managed server that customers host on-premises
D. An isolated location that contains multiple AWS-managed data centers

A

A. An isolated location that contains a single AWS-managed data center

Explanation:
An availability zone is an isolated location containing multiple physical data centers located close together hosting AWS virtual resources.

An availability zone is often mistaken for a single physical data center hosting AWS virtual resources. A common misconception is that a single availability zone equals a single data center. This is not the case. Multiple data centers located close together can form a single availability zone.

AWS Outposts offers a single AWS-managed server that customers host on-premises.

An edge location is a cache in a highly populated area where AWS stores data for CDN services.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

When using Amazon Glacier’s bulk retrieval option for petabytes, which of the following is correct?

A. Amazon Glacier takes 5-12 hours to retrieve data.
B. Amazon Glacier takes 3-5 hours to retrieve data.
C. Amazon Glacier takes 1-5 minutes to retrieve data.
D. Amazon Glacier 1 hour to retrieve data.

A

A. Amazon Glacier takes 5-12 hours to retrieve data.

Explanation:
Amazon Glacier is an extremely low-cost storage service that provides secure and durable storage for data archiving and backup. To keep costs low, Amazon Glacier is optimized for data that is infrequently accessed and for which retrieval times of several hours are suitable. The standard retrieval option, which is the default option, takes 3-5 hours to complete. The other options are expedited, which downloads a small amount of data (250 MB maximum) in 5 minutes, and bulk, which downloads large amounts of data (petabytes) in 5-12 hours.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

When considering the four general disaster recovery methods, which choice lists them in order from slowest to fastest based on the time required to reach their recovery time objective (RTO)?
A . Multi-Site

Pilot Light
	
Warm Standby
	
Backup and Restore

B. Pilot Light

Backup and Restore
	
Multi-Site
	
Warm Standby

C Backup and Restore

Pilot Light
	
Warm Standby
	
Multi-Site

D Multi-Site

Warm Standby
	
Pilot Light
	
Backup and Restore
A

C Backup and Restore

Pilot Light
	
Warm Standby
	
Multi-Site

Explanation:
The general DR scenarios listed by their estimated RTO in descending order is as follows:

Backup and Restore - this option has the highest RTO, but the generally lowest maintenance cost.
Pilot Light - this option has a lower RTO than Backup and Restore, but will be more expensive to maintain.
Warm Standby - this option keeps a scaled-down version of your complete environment on standby, so it has a lower RTO than Pilot Light, but is more expensive to maintain.
Multi-Site - this option keeps a copy of your production environment live at all times, so you can failover to it very quickly, but it is the most expensive DR choice of the four listed here.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

_____ improves upon high availability to offer a greater level of protection should components begin to fail in your infrastructure.

A. Fault tolerance
B. A service level agreement
C. Automation
D. Redundancy

A

A. Fault tolerance

Explanation:
Let’s look at fault tolerance, which expands on high availability to offer a greater level of protection should components begin to fail in your infrastructure. However, there are usually additional cost implications due to the greater level of resiliency offered.

17
Q

Which statement best describes the AWS Snowball as a data migration service?
A. Data is transferred via a physical appliance to AWS.
B. Data is sent via hardware appliances shipped on dry ice.
C. Data is sent compressed via a high speed network connection.
D. Data is sent encoded (forward error correction) via a high speed network connection.

A

A. Data is transferred via a physical appliance to AWS.

Explanation:
The Snow family uses secure appliances to transfer large amounts of data into and out of the AWS cloud; this is fast and cheaper than high-speed Internet.

18
Q

What is beneficial about deploying resources to multiple availability zones in AWS?

A. It gives you the ability to separate out your cost controls.
B. Your app can be accessed anywhere in the world.
C. If one fails there is another still running.
D. It reduces latency for customers on different continents.

A

C. If one fails there is another still running.

Explanation:
The nice thing about having two separate availability zones is that if there really ever is a problem with AWS and one of the availability zones go down, then you’ve got that connectivity going to each one and you’re not gonna lose connectivity from those.

19
Q

Which of the following Storage Gateway configurations can assist you in recovering data in your local application and storage layers when using Amazon Machine Image templates?

A. Stored volume gateways
B. Cached volume gateways
C. File gateways
D. Virtual Tape Library

A

A. Stored volume gateways

Explanation:
Store volume gateways allow you to back up lost local data when using AMI templates which you could provision as EC2 instances within AWS. You could then attach EBS volumes to these instances which could be created from the storage gateway volume snapshots which would be stored on S3 giving you access to your production data required. Your applications storage infrastructure could potentially be up and running again in a matter of minutes within a VPC with connectivity from your on-premise data center.

Cached volume gateways are configured with the result that the primary data storage is actually Amazon S3 rather than your own local storage solution.

Virtual Tape Library is essentially a cloud-based tape backup solution replacing physical components with virtual ones, so they would not be applicable in a situation where you were using AMIs to recover local data.

File gateways allow you to securely store your files as objects within S3. Using as a type of file share which allows you to mount on map drives to an S3 bucket as if the share was held locally on your own corporate network.

20
Q

What are ideal uses for DynamoDB’s on-demand backup feature?

A. On-demand backups are ideal for each of these situations.
B. As part of a testing scenario
C. In case of table corruption
D. During long-term auditing requirements

A

A. On-demand backups are ideal for each of these situations.

Explanation:
On demand backups allow you to request a full backup of a table, as it is at the very moment the backup request is made. On demand backups are manually requested, and can be performed either through the AWS DynamoDB console or by using the AWS CLI.

On demand backups are useful in the following situations:

Table corruption - rare but possible
Long term regulatory, compliance, and/or auditing data requirements
Testing scenarios
21
Q

In regards to disaster recovery, what is the recovery time objective (RTO)?

A. The time it takes after a disruption to restore a system back to its desired service level.
B. The time it takes to scale a functioning resource horizontally to adapt to increased demand.
C. The time it takes to scale a functioning resource vertically to adapt to increased demand.
D. The acceptable amount of data loss measured in time.

A

A. The time it takes after a disruption to restore a system back to its desired service level.

Explanation:
The recovery time objective or RTO is the time it takes after a disruption to restore a system back to service level. Ideally, a business has established an official recovery procedure which ensures that the method of recovery allows the particular system to recover in a predetermined amount of time, and any lapse in service will have little to no negative effect on the business. AWS recommends multiple established methods of disaster recovery that balance cost with the time needed to re-establish service.

22
Q

What choice best describes how Amazon Aurora maintains current data between its master database and each database replica?

A. The master database and the replicas share the same storage layer, so no replication is needed.
B. The master database sends updates to the replica databases using asynchronous replication.
C. The master database sends updates to the replica databases using synchronous replication.
D. The master database sends updates to the replica databases using either synchronous or asynchronous replication, depending on the database configuration.

A

A. The master database and the replicas share the same storage layer, so no replication is needed.

Explanation:
The storage layer is presented to the compute layer as a single logical volume. This same single logical volume is shared across all compute instances involved in the compute layer whether it be a master or read replica - allowing the read replicas to accomplish the near-identical query performance as the master itself.

When compared with RDS - the management of data from a replication viewpoint is fundamentally different. With RDS data needs to be replicated from the master to each of its replicas. Aurora, on the other hand, has no need for replication since it uses and shares a single logical volume amongst all compute instances.

23
Q

_____ can be defined as maintaining a percentage of uptime, which maintains operational performance.

A. A service level agreement
B. Redundancy
C. High availability
D. Fault tolerance

A
24
Q

In which of the following circumstances will Amazon Aurora offer six-way replicated storage across three availability zones?

A. AWS provides this level of storage availability depending on the compute layer selected
B. AWS provides this when desired for a higher cost.
C. AWS provides this only for databases in regions with three availability zones. The storage availability will differ in a region with fewer zones.
D. AWS provides this for all Aurora databases.

A
25
Q

What is DynamoDB Accelerator (DAX)?

A. DAX is a managed data backup service that minimizes data loss exponentially for your applications using DynamoDB tables.
B. DAX is a managed, high-performance computing (HPC) cluster designed to improve the write performance of your DynamoDB service.
C. DAX is a data stream analysis service that processes incoming data to your DynamoDB tables in real time.
D. DAX is an in-memory cache designed to improve the read performance of your DynamoDB service.

A