Storage Flashcards

1
Q

Types of EBS Volumes are there?

A

Solid-State Drives (SSD):
General Purpose SSD (gp2)
Provisioned IOPS SSD (io1)

Hard Disk Drives (HDD):
Throughput Optimized HDD (st1)
Cold HDD (sc1)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the use cases for General Purpose SSD (gp2)

A
  • Recommended for most workloads
  • System boot volumes
  • Virtual desktops
  • Low-latency interactive apps
  • Development and test environments
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the use cases for Provisioned IOPS SSD (io1)

A
  • Critical business applications that require sustained IOPS performance, or more than 16,000 IOPS or 250 MiB/s of throughput per volume. When attached to Nitro system EC2 instances, peak performance can go up to 64,000 IOPS and 1,000 MB/s of throughput per volume.
  • Large database workloads.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are the use cases for Throughput Optimized HDD (st1) and Cold HDD (sc1)

A

Throughput Optimized HDD (st1):

  • Streaming workloads requiring consistent, fast throughput at a low price
  • Big data
  • Data warehouse
  • Log processing
  • Cannot be a boot volume

Cold HDD (sc1):

  • Throughput-oriented storage for large volumes of data that is infrequently accessed
  • Scenarios where the lowest storage cost is important
  • Cannot be a boot volume
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Some of the features of EBS are:

A
  • mount multiple volumes on same instance
  • can encrypt
  • can create point in time snapshots, persisted to Amazon S3, similar to AMI.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How can you make an EBS volume available outside of the AZ? or another region or account

A
  • you can create a snapshot and restore that snapshot to a new volume anywhere in that region.
  • You can copy snapshots to other regions and then restore them to new volumes there, making it easier to leverage multiple AWS regions for geographical expansion, data center migration, and disaster
  • You can share a snapshot across AWS accounts by modifying its access permissions. recovery.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Performance metrics, such as xx, xx, xx, and average xx length, provided by Amazon CloudWatch, allow you to monitor the performance of your volumes to make sure that you are providing enough performance for your applications without paying for resources you don’t need.

A

Performance metrics, such as bandwidth, throughput, latency, and average queue length, provided by Amazon CloudWatch, allow you to monitor the performance of your volumes to make sure that you are providing enough performance for your applications without paying for resources you don’t need.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Which tool to use for backing up of EBS volumes

A

Use AWS Backup, an automated and centralized backup service, to protect EBS volumes and your other AWS resources. AWS Backup is integrated with Amazon DynamoDB, Amazon EBS, Amazon RDS, Amazon EFS, and AWS Storage Gateway to give you a fully managed AWS backup solution.

With AWS Backup, you can configure backups for EBS volumes, automate backup scheduling, set retention policies, and monitor backup and restore activity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How do you detach an EBS volume from EC2 instance?

A
  • You can detach an EBS volume from an instance explicitly or by terminating the instance. However, if the instance is running, you must first unmount the volume from the instance.
  • If an EBS volume is the root device of an instance, you must stop the instance before you can detach the volume.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How do you backup EBS volumes?

A
  • Back up the data on your EBS volumes to S3 by taking point-in-time snapshots.
  • Snapshots are incremental backups, which means that only the blocks on the device that have changed after your most recent snapshot are saved. This minimizes the time required to create the snapshot and saves on storage costs by not duplicating data.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are some of the limitations of EBS snapshots?

A
  • User-defined tags are not copied from the source snapshot to the new snapshot.
  • Snapshots are constrained to the Region in which they were created. To share a snapshot with another Region, copy the snapshot to that Region.
  • Snapshots that you intend to share must instead be encrypted with a custom CMK.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How are you charged for EBS volumes?

A
  • You are charged by the amount you provision in GB per month until you release the storage.
    • Provisioned storage for gp2 volumes, provisioned storage and provisioned IOPS for io1 volumes, provisioned storage for st1 and sc1 volumes will be billed in per-second increments, with a 60 second minimum.
    • With Provisioned IOPS SSD (io1) volumes, you are also charged by the amount you provision in IOPS per month.
    • After you detach a volume, you are still charged for volume storage as long as the storage amount exceeds the limit of the AWS Free Tier. You must delete a volume to avoid incurring further charges.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Are you charged for EBS snapshot storage? Does copying it to a new region incur additional costs?

A
  • Snapshot storage is based on the amount of space your data consumes in Amazon S3.
    • Copying a snapshot to a new Region does incur new storage costs.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

On general purpose SSD (gp2) what it the max IOS and MB/s of throughput supported?

A
  • Base performance of 3 IOPS/GiB, with the ability to burst to 3,000 IOPS for extended periods of time.
  • Support up to 16,000 IOPS and 250 MB/s of throughput.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

On throughput Optimized HDD (st1) what is the max throughput

A

◦ Low-cost magnetic storage that focuses on throughput rather than IOPS.
◦ Throughput of up to 500 MiB/s.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What are the key differences between EBS and EFS in terms of Availability and durability, Access, Use cases

A

Amazon EFS:

Availability and durability
Data is stored redundantly across multiple AZs.

Access
Up to thousands of EC2 instances, from multiple AZs, can connect concurrently to a file system.

Use cases
Big data and analytics, media processing workflows, content management, web serving, and home directories.

Amazon EBS Provisioned IOPS

Data is stored redundantly in a single AZ.

A single EC2 instance in a single AZ can connect to a file system.

Boot volumes, transactional and NoSQL databases, data warehousing, and ETL.

17
Q

Fill in the blanks:

Amazon xxx Infrequent Access (xxx IA) is a new storage class for Amazon xxx that is cost-optimized for files that are accessed less frequently. Customers can use xxx IA by creating a new file system and enabling Lifecycle Management. With Lifecycle Management enabled, xxx automatically will move files that have not been accessed for 30 days from the Standard storage class to the Infrequent Access storage class.

A

Amazon EFS Infrequent Access (EFS IA) is a new storage class for Amazon EFS that is cost-optimized for files that are accessed less frequently. Customers can use EFS IA by creating a new file system and enabling Lifecycle Management. With Lifecycle Management enabled, EFS automatically will move files that have not been accessed for 30 days from the Standard storage class to the Infrequent Access storage class.

18
Q

How can data be moved between on premise storage and Amazon EFS

A

Moving your EFS file data can be managed simply with AWS DataSync – a managed data transfer service that makes it faster and simpler to move data between on-premises storage and Amazon EFS.

19
Q

What are the main advantages of using EFS ?

A

EFS file systems store data and metadata across multiple Availability Zones in an AWS Region.

Multiple Amazon EC2 instances can access an EFS file system at the same time, providing a common data source for workloads and applications running on more than one instance or server.

EFS file systems can grow to petabyte scale, drive high levels of throughput, and allow massively parallel access from EC2 instances to your data.

20
Q

What is POSIX?

A

Portable Operating System Interface, and is an IEEE standard designed to facilitate application portability.

21
Q

How can EFS be backed up?

A

You can schedule automatic incremental backups of your EFS file system using the EFS-to-EFS Backup solution.

22
Q

What are the performance modes of EFS?

A

General purpose performance mode (default)
◦ Ideal for latency-sensitive use cases.
Max I/O mode
◦ Can scale to higher levels of aggregate throughput and operations per second with a tradeoff of slightly higher latencies for file operations.

23
Q

What are the throughput modes of EFS?

A

Bursting Throughput mode (default)
◦ Throughput scales as your file system grows.
Provisioned Throughput mode
◦ You specify the throughput of your file system independent of the amount of data stored.

24
Q

In EFS whats the best way to automatically move files into EFS infrequent access (EFS IA) storage class to save cost?

A

You can choose from four EFS Lifecycle Management policies (14, 30, 60, or 90 days) to automatically move files into the EFS Infrequent Access (EFS IA) storage class and save up to 85% in cost.

25
Q

What are the max number of EFS file systems per customer account in an AWS region?

A

Number of file systems for each customer account in an AWS Region
1,000

26
Q

How do you uniquely identify objects in S3 bucket?

A

A key is the unique identifier for an object within a bucket.

27
Q

Whats the max number of buckets you can create in an AWS account?

A

By default, you can create up to 100 buckets in each of your AWS accounts.

28
Q

How do you modify the region of an S3 bucket?

A

You can’t change its Region after creation.

29
Q

What is the maximum upload size of files to S3 and how do you upload larger files?

A

• You can upload and copy objects of up to 5 GB in size in a single operation. For objects greater than 5 GB up to 5 TB, you must use the multipart upload API.

30
Q

What is the purpose in tagging within S3?

A

S3 provides the tagging subresource to store and manage tags on a bucket. AWS generates a cost allocation report with usage and costs aggregated by your tags.

◦ You can associate up to 10 tags with an object. Tags associated with an object must have unique tag keys.
◦ A tag key can be up to 128 Unicode characters in length and tag values can be up to 256 Unicode characters in length.
◦ Key and values are case sensitive.

31
Q

How can you provide another AWS account access to an object that is stored in an Amazon Simple Storage Service (Amazon S3) bucket?

A

These are the methods on how to grant cross-account access to objects that are stored in your own Amazon S3 bucket:

▪ Resource-based policies and AWS Identity and Access Management (IAM) policies for programmatic-only access to S3 bucket objects
▪ Resource-based Access Control List (ACL) and IAM policies for programmatic-only access to S3 bucket objects
▪ Cross-account IAM roles for programmatic and console access to S3 bucket objects

32
Q

MFA Delete grants additional authentication for which 2 operations in S3:

A

▪ Change the versioning state of your bucket

▪ Permanently delete an object version

33
Q

Why should you use versioning for S3 bucket objects?

A

◦ Use versioning to keep multiple versions of an object in one bucket.
◦ Versioning protects you from the consequences of unintended overwrites and deletions.
◦ You can also use versioning to archive objects so you have access to previous versions.
◦ Since versioning is disabled by default, need to EXPLICITLY enable.
◦ When you PUT an object in a versioning-enabled bucket, the non-current version is not overwritten.
◦ When you DELETE an object, all versions remain in the bucket and Amazon S3 inserts a delete marker.
◦ Performing a simple GET Object request when the current version is a delete marker returns a 404 Not Found error. You can, however, GET a non-current version of an object by specifying its version ID.
◦ You can permanently delete an object by specifying the version you want to delete. Only the owner of an Amazon S3 bucket can permanently delete a version.

34
Q

S3 Storage Classes for Frequently Accessed Objects are?

A
◦	S3 STANDARD for general-purpose storage of frequently accessed data.
	◦	S3 REDUCED_REDUNDANCY (RRS) for noncritical, reproducible data that can be stored with less redundancy than the STANDARD storage class. Not recommended because STANDARD storage class is more cost effective.
35
Q

S3 Storage Classes for Infrequently Accessed Objects are:

A

◦ S3 STANDARD_IA for long-lived, but less frequently accessed data. It stores the object data redundantly across multiple geographically separated AZs.
◦ S3 ONEZONE_IA stores the object data in only one AZ. Less expensive than STANDARD_IA, but data is not resilient to the physical loss of the AZ.
◦ These two storage classes are suitable for objects larger than 128 KB that you plan to store for at least 30 days. If an object is less than 128 KB, Amazon S3 charges you for 128 KB. If you delete an object before the 30-day minimum, you are charged for 30 days.

36
Q

Whats the best way to automate cost saving storage in Amazon S3?

A

Amazon S3 Intelligent Tiering:

◦ S3 Intelligent-Tiering is a storage class designed for customers who want to optimize storage costs automatically when data access patterns change, without performance impact or operational overhead.
◦ S3 Intelligent-Tiering is the first cloud object storage class that delivers automatic cost savings by moving data between two access tiers — frequent access and infrequent access — when access patterns change, and is ideal for data with unknown or changing access patterns.
◦ S3 Intelligent-Tiering monitors access patterns and moves objects that have not been accessed for 30 consecutive days to the infrequent access tier. If an object in the infrequent access tier is accessed later, it is automatically moved back to the frequent access tier.
◦ There are no retrieval fees in S3 Intelligent-Tiering.
37
Q

For long-term archiving S3 data what storage class is best suited?

A

GLACIER
◦ For long-term archive
◦ Archived objects are not available for real-time access. You must first restore the objects before you can access them.
◦ You cannot specify GLACIER as the storage class at the time that you create an object.
◦ Glacier objects are visible through S3 only

◦ Retrieval Options:
▪ Expedited – allows you to quickly access your data when occasional urgent requests for a subset of archives are required. For all but the largest archived objects, data accessed are typically made available within 1–5 minutes. There are two types of Expedited retrievals: On-Demand requests are similar to EC2 On-Demand instances and are available most of the time. Provisioned requests are guaranteed to be available when you need them.
▪ Standard – allows you to access any of your archived objects within several hours. Standard retrievals typically complete within 3–5 hours. This is the default option for retrieval requests that do not specify the retrieval option.
▪ Bulk – Glacier’s lowest-cost retrieval option, enabling you to retrieve large amounts, even petabytes, of data inexpensively in a day. Bulk retrievals typically complete within 5–12 hours.

◦ For S3 Standard, S3 Standard-IA, and Glacier storage classes, your objects are automatically stored across multiple devices spanning a minimum of three Availability Zones.

38
Q

For S3 data that is rarely accessed within a year but needs to be retrievable within 12 hours which storage class is best suited?

A

• Amazon S3 Glacier Deep Archive
◦ A new Amazon S3 storage class providing secure and durable object storage for long-term retention of data that is accessed rarely in a year.
◦ S3 Glacier Deep Archive offers the lowest cost storage in the cloud, at prices lower than storing and maintaining data in on-premises magnetic tape libraries or archiving data offsite.
◦ All objects stored in the S3 Glacier Deep Archive storage class are replicated and stored across at least three geographically-dispersed Availability Zones, protected by 99.999999999% durability, and can be restored within 12 hours or less.
◦ S3 Glacier Deep Archive also offers a bulk retrieval option, where you can retrieve petabytes of data within 48 hours.

39
Q

Which key would you use to decrypt and EBS snapshot?

A

The same key used to encrypt in the first place is taken into the snapshot. If you create a new EBS snapshot then will be a new key. Should use this my default

Exam power ups:

  • Accounts can be set to encrypt by default
  • Otherwise choose the CMK to use
  • Each volume uses 1 unique DEK
  • Can’t change a volume to not be encrypted
  • OS is not aware of the encryption so no performance hit