S3 Flashcards

1
Q

What is a bucket?

A

A bucket is a container of objects uploaded to S3.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are some rules that bucket names must follow?

A
  • A name must be unique across all the existing bucket names in Amazon S3.
  • Must comply with DNS naming conventions.
  • They cannot be ip addresses
  • Must not contain underscore and uppercase letters
  • Can start only with a lowercase letter or a number
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the default access of buckets?

A

Buckets are private by default.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are folders on Amazon S3?

A

Folders are saved as objects. Next to the specified name a forward slash “/” is appended and the object is displayed as a folder in the S3 console.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the maximum size of a file?

A

The maximum size of a file is 5TB.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the maximum file transfer size for a single PUT?

A

5 GB

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What content type is used for files larger than 100 MB?

A

Multipart

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is Amazon S3?

A
Amazon S3 is a simple key-based object store. 
Only files (objects) can be stored on S3.

It’s a simple storage service that offers an extremely durable, highly available, and infinitely scalable data storage infrastructure at very low costs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How do you communicate with Amazon S3?

A

Amazon S3 provides a simple, standards-based REST web services interface that is designed to work with any Internet-development toolkit.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the maximum storage capacity?

A

There is unlimited storage available.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the characteristic of updates to an object on S3?

A

Updates to objects on S3 are atomic. When reading an object you either get the old object or the new one. You will never get a partial, corrupted object.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What about event notifications?

A

Event notifications for specific actions, can send alerts or trigger actions. Notifications can be sent to:

  • SNS Topics.
  • SQS Queue.
  • Lambda functions.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is S3 data made up of?

A

S3 data is made up of:

  • key (name)
  • value (data)
  • version id
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What do you use to manipulate objects?

A

In order to manipulate objects (upload, download, delete, etc.) a REST API over HTTP is used.

S3 is accessible via a public endpoint of the form:

http://s3.aws-region.amazonaws.com/bucket (path style url)

or it can be also

http://bucket.s3.aws-region.amazonaws.com (virtual hosted style url)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are the states of a backet with regard to versioning?

A
  • unversioned (default)
  • versioning enabled
  • versioning-suspended
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What happens after versioning has been enabled?

A

Versioning can be only suspended on a bucket, a bucket cannot be unversioned anymore.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What do you need to know in order to delete a particular version of an object?

A

You need to know its version id.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is the version id of an object that has been added to a bucket before versioning was enabled?

A

In this case the version id is null.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What happens to an existing object if you deactivate bucket versioning and you update it?

A

The version id becomes null, and that version becomes the latest version.
The previous versions do not get deleted, and thereon all the changes are applied against this version.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What does Server Access Login provide?

A

Server Access Login provides detailed records for the requests that are made to a bucket (source bucket)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is Server Access Logging used for?

A

It’s used for security purposes and auditing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

What is Server Access Loging default state?

A

Server Access Login default state is disabled by default.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What happens when you enable Server Access Login?

A

When you enable it you will have to select a bucket where the logs will be delivered to (a target bucket) which can be the same as the source target. Source and target bucket should be in the same region.

If source and target bucket are the same you can select a different prefix to identify the logs.

Write permission should be granted to Amazon S3 Log Delivery Group on the target bucket.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What is the content of a Server Access Logging record?

A

A log record contains:

  • the name of the bucket which was accessed (source bucket)
  • the requester
  • the request time
  • the request action
  • the request status
  • the error code (if any)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

What is one of the possible ways to dump logs?

A

Logs of multiple source buckets can be dumped in a single target bucket

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

When are logs delivered?

A

Server access logs are delivered on a best effort basis.
Most records are delivered within a few hours from the time they were recorded but timeliness of server logging is not guaranted.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

What is the main difference between Server Access Logging and Object Level Logging?

A

Server Access Logging is performed at the bucket level while object level logging is executed at the object level.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

What happens if object-level logging is activated and a user on application tries to access an object?

A

If a Cloud Trail is set up for the S3 bucket Data Event, the logs will be created in the target bucket.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

What is the region where Cloud Trail must be activated?

A

Cloud Trail must be activated in the same region as the S3 bucket.

30
Q

How many kinds of encryption there are on S3?

A

Client-side encryption and Server Side Encryption.

31
Q

What is S3 server-side encryption?

A

In SSE, it’s the S3 server that is in charge of the encryption and decryption; whereas, in client-side encryption it is the customer who is in charge of that.

32
Q

How many options does S3 server-side encryption provide?

A

S3 SSE provides 3 options:

  • SSE-S3 in which S3 manges the data encryption key and the master key
  • SSE-KMS in which KMS creates and manages the data encryption key and the master key;
  • SSE-C in which S3 manages the keys that are provided by the user.
33
Q

How many options does S3 client-side encryption provide?

A

S3 CSE provides 2 options:

  • CSE with KMS-Managed Customer Master Key;
  • CSE with Client-side Master Key (S3 Encryption Client)
34
Q

What is an S3 ACL?

A
  • It enables you to manage access to buckets and objects. Each bucket and object has an ACL attached to it as a sub-resource.
  • It defines which AWS accounts or groups are granted access and the type of access.
  • When a request is received against a resource, S3 checks the corresponding ACl to verify that the requester has the necessary access permissions.
35
Q

What is a Bucket policy?

A

It grants other AWS accounts or IAM users permissions for the bucket and the objects in it.

Object permissions apply only to the objects that the bucket owner creates.

36
Q

What is the purpose of an S3 ACL?

A

To grant basic read/write permissions to other AWS accounts.

37
Q

What permissions can be granted through S3 ACL?

A

READ, WRITE, READ_ACP, WRITE_ACP, FULL CONTROL

38
Q

How do you handle objects with S3 ACL if you are not the owner?

A

You need to use the object ACL.

39
Q

What is a tag?

A

A tag is a way to label the AWS resources.

This way they can be easily managed, searched for and filtered.

40
Q

What is a tag made of?

A

A tag is made of a key-value pair.

41
Q

What is S3 Transfer acceleration?

A

It enables fast, easy and secure transfers of files over long distances between your client and S3 bucket.

It can upload objects over long distances in an accelerated manner.

42
Q

What is multipart upload?

A

It enables to upload large objects in parts for fast and efficient data transfer.

43
Q

Why should you use tags?

A

In AWS management console you can configure tags to be displayed with resources, and can search and filter by tag.

In AWS cost explorer you can create reports to get cumulative costs of the resources grouped by a given tag.

You can tag resources to filter them during automation activities: for example you can tag EC2 instances as “development”, “testing” and “production”.

In access control you can have conditions based on tags.

44
Q

What are the event notification types for Amazon S3?

A
  • new object created events
  • object removal events
  • Reduced Redundancy Storage (RRS) object lost events
  • Replication events
45
Q

Why do event notifications exist?

A

Event notifications can be sent in response to actions in S3 like PUT, POST, COPY, DELETE.

Amazon S3 event notifications enable you to run workflows, send alerts, or perform other actions in response to changes in your objects stored in S3.

46
Q

What destinations can be chosen for sending event notifications?

A
  • Publish event messages to an Amazon Simple Notification Service (SNS) topic;
  • Publish event messages to an Amazon Simple Queue Service (SQS) queue;
  • Publish event messages to AWS Lambda by invoking a Lambda function and passing the message as an argument.

You need to grant Amazon S3 permissions to post messages to an Amazon SNS topic or an Amazon SQS queue.

You need to also grant Amazon S3 permission to invoke an AWS Lambda function on your behalf.

47
Q

What is requester pays?

A

Instead of the bucket owner, the requester himself pays the cost of the request, and the data downloaded from the bucket.

The bucket owner just pays for the storage.

48
Q

When should requester pays be used?

A

When you host the data but you want to make custoemrs pay for downloading the data.

49
Q

How does “requester pays” technically work?

A

The requester must include “x-amz-requester-payer” in their requests either in the header (POST,GET,HEAD requests) or as a parameter (in a REST request). If the request does not include this header or parameter, the bucket owner gets charged for the request or data transfer.

50
Q

What is object lifecycle management?

A

Object Lifecycle Management manages the storage of the objects and the cost associated with it throughout the life othe objects.

51
Q

Why to using object lifecycle management?

A
  • To move the objects to a different storage class, depending on how frequently they are used, to reduce the cost (Transition).
  • To set an expiry date on the objects that would no longer be used after a certain period, and delete them (Expiration).
52
Q

How often can you transfer objects to another storage class?

A

In order to transfer an object to another storage class, you need to wait 30 days from the previous transition.

53
Q

What versions of an object can be transfered?

A

You can transfer to another storage class the current version or the previous versions of an object.

54
Q

What filters can be added for Object Lifecycle Management?

A

It’s possible to filter the objects by prefix.

55
Q

How long should you wait before deleting an object after its transition?

A

1 day

56
Q

How do you deal with incomplete multipart transfers?

A

Using the lifecycle rules you can clean up expired object delete markers (meaning there won’t be any current version any longer) and/or clean up incomplete multipart uploads.

57
Q

Is a new lifecycle configuration applied instantly against a bucket?

A

No, there’s usually some lag before the configuration is fully propagated to all the Amazon S3 systems.

58
Q

What objects are affected by a new or an updated Object Lifecycle Management Configuration?

A

Both existing objects and objects created after the entry into force of the new configuration.

59
Q

What is cross region replication (CRR)?

A

Automatic, asynchronous copying of objects from a bucket in one region (Source Bucket) to another bucket in another region (Destination Bucket).

60
Q

Why should you use Cross Region Replication (CRR)?

A

If you have an application where you are sending files from one bucket in one region to another in a different region for post processing or for disaster recovering.

61
Q

What are the requirements for Cross Region Replication (CRR)?

A
  • Source and destination buckets must be in different AWS regions;
  • Both the source and the destination buckets must have versioning enabled;
  • S3 must have permissions to replicate objects from that source bucket to the destination bucket on your behalf.
62
Q

What can be replicated with Amazon Cross Region Replication (CRR)?

A
  • Any new object created after you added a replication configuration.
  • Unencrypted objects as well as encrypted objects, but only if encrypted with SSE-S3 or SSE-KMS. If the object is encrypted in the source bucket, then also the replicated copy of the object is encrypted using the same type of server-side encryption.
  • Along with objects, S3 also replicates object metadata.
  • S3 replicates only objects in the source bucket for which the bucket owner has permissiosn to read objects and read access control lists (ACLs).
63
Q

What is not replicated with Amazon CRR (Cross Site Region)?

A
  • Objects that existed before enabling replication (can use the copy API).
  • Objects created with SSE-C.
  • Objects to which the bucket owner does not have permissions.
  • Updates to bucket-level subresources.
  • Actions from lifecycle rules are not replicated.
  • Objects in the source bucket that are replicated from another region are not replicated.
64
Q

How is deletion handled with Cross Region Replication?

A
  • if a delete request is made without specifying an object version ID a delete marker will be added and replicated.
  • If a delete request is made specifying an object version ID the object is deleted but the delete marker is not replicated.
65
Q

What is the purpose of S3 different storage classes?

A
  • S3 Standard: general-purpose storage of frequently accessed data.
  • S3 Standard-Infrequent Access: long-lived, but less frequently accessed data.
  • S3 Intelligent Tiering: cost optimization without performance impact.
  • S3 Glacier and S3 Glacier Deep Archive: long-term storage.
66
Q

What are the characteristics of S3 Standard storage class?

A
  • The default storage class for the objects.
  • The data is replicated across at least 3 different availability zones (AZs) for high durability and availability.
  • Durability of 99.99999999999% and availability of 99.99% (highest of all storage classes).
  • Offers low latency and high throughput, which makes it suitable for many applications such as cloud apps, dynamic websites, content distribution, mobile and gaming applications, Big Data analytics.
  • S3 Lifecycle management is used to configure the lifecycle policies to automatically migrate objects from this class to the other appropriate storage classes.
67
Q

What are the characteristics for S3 Standard-Infrequent Access (S3 Standard-IA)?

A
  • For data that is accessed less frequently, but requires rapid access when needed.
  • Data is replicated across (at least) 3 different availability zones (AZs) for high durability and availability.
  • Durability of 99.99999999999% and availability of 99.9% (lesser than that of S3-Standard).
  • Costs less than S3-Standard; charges you for retrieving the data per GB.
  • Supports lyfecicle policies to automatically migrate objects fro m this class to the other approapriate storage classes.
68
Q

What are the characteristics of S3 One Zone-Infrequent Access (S3 One Zone-IA) ?

A
  • It’s suited for data that is accessed less frequently, but requires rapid access when needed.
  • Data is stored in only a single availability zone (AZ).
  • Durability of 99.99999999999% and availability of 99.5% (lesser than that of S3 Standard IA).
  • Costs 20% less than S3 Standard IA; charges you for retrieving the data per GB.
  • Good choice for backup copies of on-premises data or easily re-creatable data.
  • Supports lifecycle policies to automatically migrate objects from this class to other appropriate storage classes.
69
Q
A
70
Q

What are the characteristics of S3 Intelligent Tiering?

A
  • It’s designed to optimize the costs by automatically moving data to the most cost-effective access tier, without performance impact or operational overhead.
  • Data is replicated across (at least) 3 different availability zones (AZs) for high durability and availability.
  • Same low latency and high throughput performance of S3 Standard.
  • Durability of 99.99999999999% and availability of 99.9 (lesser than S3-standard’s).
  • There are not retrieval fees and no additional tiering fees when objects are moved between access tiers, but charges a monthly monitoring and automation fee per object.
  • Ideal storage class for long-lived data with access patterns that are unknown or unpredictable.
  • Suitable for objects larget than 128 KB that you plan to store for at least 30 days. if you delete an object before the end of the 30-day minimum storage duration period, you are charged for 30 days.
  • Supports lifecycle policies to automatically migrate objects from this class to the other appropriate storare classes.
71
Q

What are the characteristics of S3 Glacier and S3 Glacier Deep archive?

A

Amazon S3 Glacier

  • Secure, durable, and extremely low-cost storage service for data archiving.
  • Three options for retrieving the archives: standard, expedited, and bulk.
  • Archived data can be retrieved in a few minutes to hours and the user will be charged accordignly to the option used.
  • Costs significantly less than all other S3 storage classes; charges you for retrieving the data per GB.

Amazon S3 Glacier Deep archive

* Lowest cost storage class designed for long-term retention of data that will be retained for 7-10 years.
*