What are the types of S3 storage in AWS?

Key takeaways:

  • Amazon S3 allows object-based storage for both structured and unstructured data.

  • There are multiple storage classes, such as:

    • S3 Standard: Frequent access, low latency.

    • Intelligent-Tiering: Auto-moves data to cost-effective tiers.

    • Glacier: Archival storage with delayed retrieval times.

  • All storage classes have 99.999999999% durability, with varying availability and costs based on access frequency.

  • Different storage classes balance cost and access needs, with Glacier Deep Archive being the most cost-effective option for long-term storage.

Simple Storage Service (S3) stores object-based data. Each object contains metadata and an ID. S3 stores both structured and unstructured data, and the latency for retrieving data varies depending on the type of data and the storage class used. Amazon S3 offers different types of storage for different purposes.

Types of S3 storage

Let’s discuss these storage types below:

  1. S3 Standard
  2. S3 Intelligent-Tiering
  3. S3 Standard-Infrequent Access
  4. S3 Glacier Deep Archive
  5. S3 Glacier Instant Retrieval
  6. S3 One Zone-Infrequent Access
  7. S3 Express One Zone
  8. S3 Glacier Flexible Retrieval
  9. S3 Outposts

1. S3 Standard

This is the most frequently used type of S3 storage. It is designed for general-purpose storage and is a good choice for storing frequently accessed data, such as website assets, application files, and user-generated content. S3 Standard has a durability of 99.999999999% and an availability of 99.9%. It supports data encryption with the help of server-side encryption (SSE). The cost of using S3 Standard is competitive with other S3 storage classes. The largest object that can be stored in an S3 bucket is 5 TB. Moreover, S3 Standard is designed to be highly durable and available, storing data across multiple Availability Zones. If one Availability Zone is deleted or disrupted, the data stored in S3 Standard will still be available and accessible from the other Availability Zones.

2. S3 Intelligent-Tiering

Amazon S3 Intelligent-Tiering is a storage class within Amazon S3 with a specific feature of moving data to the most cost-effective access tier based on the storage life cycle. It automatically stores objects in three access tiers:

Access tiers in S3 Intelligent-Tiering
Access tiers in S3 Intelligent-Tiering
  • Frequent Access(FA): This tier is optimized for frequently accessed data and provides low latency and high throughput.

  • Infrequent Access(IA): This tier is optimized for infrequently accessed data and is more cost-effective than the Frequent Access tier.

  • Archive Instant Access: This tier is the most cost-effective and optimized for rarely accessed data.

S3 Intelligent-Tiering observes access patterns and autonomously transitions objects between access tiers without compromising performance. This can help us optimize our storage costs without manually managing the placement of our data.

3. S3 Standard-Infrequent Access

This type of storage is useful when we do not want our data to be accessed frequently. It is cheaper than the S3 Standard and supports data encryption with the help of server-side encryption (SSE). S3 Standard-IA has a durability of 99.999999999% and an availability of 99.9%. Moreover, the S3 Standard-IA is designed in such a way that it is highly durable and available. It also stores data in more than one Availability Zone, so if one is deleted or disrupted, the data stored in others will still be available and accessible.

Did you know that we can also host static websites on Amazon S3? Our Cloud Lab offers hands-on experience configuring a static website with S3 and CloudFront.

4. S3 Glacier Deep Archive

The Amazon S3 Glacier Deep Archive storage option is typically best suited for data that remains largely untouched and isn’t expected to be accessed frequently. For instance, a vast repository of patient records and diagnostic images dating back many years in a medical clinic. While these records contain valuable historical data, we might rarely need to access them in our day-to-day operations. Storing this data in the Amazon S3 Glacier Deep Archive ensures it’s readily available for legal purposes while offering incredibly high durability and reliability. Because this data sees infrequent use, it can be efficiently stored using the Glacier Deep Archive option. It is the most cost-friendly storage type and also provides an impressive durability of 99.999999999% and an availability rate of 99.9%. However, the retrieval time is 12hrs.

5. S3 Glacier Instant Retrieval

S3 Glacier Instant Retrieval is an Amazon S3 storage class designed for archival data that is rarely accessed but requires immediate retrieval. It provides millisecond access to data at a lower cost compared to standard S3 storage classes, making it suitable for cases like regulatory archives or backups that must be quickly retrievable. It is designed to deliver to a 99.9% availability rate.

6. S3 One Zone-Infrequent access

This storage class within Amazon S3 is designed for storing data that is accessed infrequently and can be stored in a single Availability Zone. It is a good choice for archived data, such as backups, logs, and historical data.

S3 One Zone-IA has a durability of 99.999999999% and an availability of 99.5%. It is also cost-effective, charging a lower storage rate than S3 Standard-IA.

7. S3 Express One Zone

S3 Express One Zone is an Amazon S3 storage class optimized for the most frequently accessed data and designed to store data in a single availability zone. It is a lower-cost option for workloads that do not need the durability of multi-zone redundancy, such as secondary backups or easily re-creatable data. This class offers slightly lower availability and durability compared to multi-zone options but provides significant cost savings for specific use cases where data loss risk is minimal.

8. S3 Glacier Flexible Retrieval

S3 Glacier Flexible Retrieval is intended for long-term archiving, offering cost-effective storage with flexible retrieval options. Retrieval times can vary from minutes to hours depending on the option chosen (expedited, standard, or bulk). It’s most suited for historical data that is rarely accessed but must be retained for compliance or record-keeping purposes.

9. S3 Outposts

local S3 Outposts allows organizations to extend S3 storage to their on-premises environments using AWS Outposts. This enables data residency and low-latency access to S3 storage while benefiting from AWS’s services. It’s designed for industries where local data storage and compliance requirements are critical, such as healthcare or manufacturing.

The table below provides a comprehensive overview of the types of S3 storage"

S3 Storage Class

Cost

Durability

Availability

Data Access Time

S3 Standard

Higher cost for frequent access

99.999999999% (11 9’s)

99.99%

Milliseconds

S3 Intelligent-Tiering

Auto-tiering with higher cost in frequent access

99.999999999% (11 9’s)

99.9%-99.99% (depending on tier)

Milliseconds (frequent), Minutes (infrequent)

S3 Standard-Infrequent Access

Lower than Standard, but charged for retrieval

99.999999999% (11 9’s)

99.99%

Milliseconds

S3 Glacier Deep Archive

Lowest cost

99.999999999% (11 9’s)

99.99%

Hours (12-48 hours)

S3 Glacier Instant Retrieval

Low cost, higher than Deep Archive

99.999999999% (11 9’s)

99.9%

Milliseconds

S3 One Zone-Infrequent Access

Lower than Standard-IA

99.999999999% (11 9’s)

99.5%

Milliseconds

S3 Express One Zone

Moderate, cheaper than Standard

99.999999999% (11 9’s)

99.95%

Milliseconds (with restrictions)

S3 Glacier Flexible Retrieval

Very low, higher than Deep Archive

99.999999999% (11 9’s)

99.99%

Minutes to hours (depending on retrieval type)

Summary

Amazon S3 is an object storage service designed to store and retrieve large amounts of structured and unstructured data. It offers various storage classes tailored for different use cases, balancing cost, durability, and access requirements. Each storage class has different characteristics related to data access frequency, cost, durability, and availability.

Frequently asked questions

Haven’t found what you were looking for? Contact Us


Is Amazon S3 a blob storage?

Yes, Amazon S3 is a type of blob storage. It stores large amounts of unstructured data, known as objects or blobs, which can include files like images, videos, or backups. Each object contains data, metadata, and a unique identifier.


Why is S3 called object storage?

Amazon S3 is called object storage because it stores data as individual objects, rather than in a hierarchical file system. Each object consists of the data itself, metadata, and a unique identifier, making it easy to retrieve and manage large amounts of unstructured data. Unlike traditional storage systems, object storage doesn’t use folders or directories; instead, it uses a flat structure where each object is stored in a bucket and accessed via its unique identifier.


Is AWS S3 a data lake?

AWS S3 can function as a data lake by providing scalable storage for structured and unstructured data. While S3 itself is not a data lake, it serves as a foundational component when combined with other AWS services like AWS Glue for data cataloging and AWS Athena for querying, enabling effective analytics and machine learning.


What is the maximum storage size for S3?

Amazon S3 has no maximum storage limit; you can store virtually unlimited amounts of data. However, individual objects stored in S3 can be up to 5 TB in size. For larger datasets, multiple objects can be used to exceed this limit.


Free Resources

Copyright ©2024 Educative, Inc. All rights reserved