AWS助理架构师认证培训 | Advanced S3 Introduction

文章详细介绍了AWSS3的生命周期规则,包括如何在不同存储类之间转换对象,设置过期时间,以及适用于不同场景的示例。此外,还提到了S3的请求者付费模式,事件通知功能,以及S3的性能特性如多部分上传和字节范围获取。S3Select和GlacierSelect允许更高效的服务器端数据检索,而S3BatchOperations则支持对大量S3对象进行批量操作。

视频来源:B站《AWS 认证解决方案架构师 助理级 SAA-C03》

一边学习一边整理老师的课程内容及试验笔记,并与大家分享,侵权即删,谢谢支持!

附上汇总贴:AWS助理架构师认证培训 | 汇总_热爱编程的通信人的博客-优快云博客


Lifecycle Rules (with S3 Analitics)

Amazon S3 - Moving between Storage Classes

  • You can transition objects between storage classes
  • For infrequently accessed object, move them to Standard IA
  • For archive objects that you don't need fast access to, move them to Glacier or Glacier Deep Archive
  • Moving objects can be automated using a Lifecycle Rules

Amazon S3 - Lifecycle Rules

  • Transition Actions - configure objects to transition to another storage class Move objects to Standard IA class 60 days after creationMove to Glacier for archiving after 6 months
  • Expiration actions - configure objects to expire (delete) after some time Access log files can be set to delete after a 365 daysCan be used to delete old versions of files (if versioning is enabled)Can be used to delete incomplete Multi-Part uploads
  • Rules can be created for a certain prefix (example: s3://mybucket/mp3/*)
  • Rules can be created for certain objects Tags (example: Department: Finance)

Amazon S3 - Lifecycle Rules (Scenario 1)

  • Your application on EC2 creates images thumbnails after profile photos are uploaded to Amazon S3. These thumbnails can be easily recreated, and only need to be kept for 60 days. The source images should be able to be immediately retrieved for these 60 days, and afterwards, the user can wait up to 6 hours. How would you design this?
  • S3 source images can be on Standard, with a lifecycle configuration to transition them to Glacier after 60 days
  • S3 thumbnails can be on One-Zone IA, with a lifecycle configuration to expire them (delete them) after 60 days

Amazon S3 - Lifecycle Rules (Scenario 2)

  • A rule in your company states that you should be able to recover your deleted S3 objects immediately for 30 days, although this may happen rarely. After this time, and for up to 365 days, deleted objects should be recoverable within 48 hours.
  • Enable S3 Versioning in order to have object versions, so that "deleted objects" are in fact hidden by a "delete marker" and can be recovered
  • Transition the "noncurrent versions" of the object to Standard IA
  • Transition afterwards the "noncurrent versions" to Glacier Deep Archive

Amazon S3 Analytics - Storage Class Analysis

  • Help you decide when to transition objects to the right storage class
  • Recommendations for Standard and Standard IADoes NOT work for One-Zone IA or Glacier
  • Report is updated daily
  • 24 to 48 hours to start seeing data analysis
  • Good first step to put together Lifecycle Rules (or improve them)!

S3 Requester Pays

S3 - Requester Pays

  • In general, bucket owners pay for all Amazon S3 storage and data transfer costs associated with their bucket
  • With Requester Pays buckets, the requester instead of the bucket owner pays the cost of the request and the data download from the bucket
  • Helpful when you want to share large datasets with other accounts
  • The requester must be authenticated in AWS (can not be anonymous)

S3 Event Notifications

  • S3:ObjectCreated, S3:ObjectRemoved, S3:ObjectRestore, S3:Replication...
  • Object name filtering possible (*.jpg)
  • Use case: generate thumbnails of images uploaded to S3
  • Can create as many "S3 events" as desired
  • S3 event notifications typically deliver events in seconds but can sometimes take a minute or longer

S3 Event Notification with Amazon EventBridge

  • Advanced filtering options with JSON rules (metadata, object size, name...)
  • Multiple Destinations - ex Step Functions, Kinesis Streams / Firehose...
  • EventBridge Capabilities - Archive, Replay Events, Reliable delivery

S3 Performance

S3 - Baseline Performance

  • Amazon S3 automatically scales to high request rates,latency 100-200ms
  • Your application can achieve at least 3,500 PUT/COPY/POST/DELETE and 5,500 GET/HEAD requests per second per prefix in a bucket
  • There are no limits to the number of prefixes in a bucket
  • Example (object path => prefix):bucket/folder1/sub1/file => /folder1/sub1/bucket/folder1/sub2/file => /folder1/sub2/bucket/1/file => /1/bucket/2/file => /2/
  • If you spread reads across all four prefixes evenly, you can achieve 22,000 requests per second for GET and HEAD

S3 Performance

  • Multi-Part upload:recommended for files > 100MB, must use for files > 5GB Can help parallelize uploads (speed up transfers)
  • S3 Transfer AccelerationIncrease transfer speed by transferring file to an AWS edge location which will forward the data to the S3 bucket in the target regionCompatible with multi-part upload

S3 Performance - S3 Byte-Range Fetches

  • Parallelize GETs by requesting specific byte ranges
  • Better resilience in case of failures

Can be used to speed up downloads

Can be used to retrieve only partial data (for example the head of a file)

S3 Select & Glacier Select

S3 Select & Glacier Select

  • Retrieve less data using SQL by performing server-side filtering
  • Can filter by rows & columns (simple SQL statements)
  • Less network transfer, less CPU cost client-side

S3 Batch Operations

S3 Batch Operations

  • Perform bulk operations on existing S3 objects with a single request, example:Modify object metadata & propertiesCopy objects between S3 bucketsEncrypt un-encrypted objectsModify ACLs, tagsRestore objects from S3 GlacierInvoke Lambda function to perform custom action on each object
  • A job consists of a list of objects, the action to perform, and optional parameters
  • S3 Batch Operations manages retries, tracks progress, sends completion notifications, generate reports...
  • You can use S3 Inventory to get object list and use S3 Select to filter your objects

1.【模块一】1. 云计算与AWS .zip 10.【模块二】2.3 Amazon CloudFront 简介.zip 11.【模块二】3.1 Amazon Relational Database S.zip 12.【模块二】3.2 Amazon DynamoDB 简介.zip 13.【模块二】3.3 Amazon ElastiCache 简介.zip 14.【模块二】3.4 Amazon Elastic MapReduce (EM.zip 15.【模块二】3.5 Amazon Redshift 简介.zip 16.【模块二】4.1 Amazon Virtual Private Cloud.zip 17.【模块二】5.1 开始使用 AWS 上的 Microsoft SQL Se.zip 18.【模块二】5.2 开始使用 AWS 上的 Microsoft SQL Se.zip 19.【模块二】5.3 开始使用 AWS 上的 Microsoft SQL Se.zip 2.【模块一】2. 带你玩转AWS .zip 20.【模块二】5.4 开始使用 AWS 上的 Microsoft SQL Se.zip 21.【模块三】1.1 部署应用的那些事儿.zip 22.【模块三】1.2 关于 AWS Elastic Beanstalk.zip 23.【模块三】1.3 关于 AWS CodeDeploy.zip 24.【模块三】1.4 关于 Amazon EC2 Container Serv.zip 25.【模块三】2.1 导语:云数据迁移.zip 26.【模块三】2.2 什么是 AWS Direct Connect 光纤直连.zip 27.【模块三】2.3 Snowball 的速度有多快?.zip 28.【模块三】2.4 AWS Storage Gateway 的常见使用情形.zip 29.【模块三】2.5 CommVault 如何将本地和云数据战略紧密联系起来?.zip 3.【模块一】3. 来点有趣的东西.zip 30.【模块三】2.6 深入探讨——AWS Kinesis Firehose.zip 31.【模块三】3.1 AWS 云上大规模迁移的最佳实践(上).zip 32.【模块三】3.2 AWS 云上大规模迁移的最佳实践(下).zip 33.【模块三】4.1 AWS 云上数据库迁移(上).zip 34.【模块三】4.2 AWS 云上数据库迁移(下).zip 35.【模块三】5.1 迁移上云概览.zip 36.【模块三】5.2 加速数据迁移.zip 37.【模块三】5.3 demo演示和客户案例.zip 38.【模块三】6.1 使用 AWS 降低总拥有成本 (TCO) (一).zip 39.【模块三】6.2 使用 AWS 降低总拥有成本 (TCO) (二).zip 4.【模块一】4. 一起来AWS吧!.zip 40.【模块三】6.3使用 AWS 降低总拥有成本 (TCO) (三).zip 41.【模块三】6.4 使用 AWS 降低总拥有成本 (TCO) (四).zip 42.【模块四】1.1 AWS 云计算对于 AI 的影响及改变 (上).zip 43.【模块四】1.2 AWS 云计算对于 AI 的影响及改变 (下).zip 44.【模块四】2.1 AWS 上的人工智能和深度学习(上).zip 45.【模块四】2.2 AWS 上的人工智能和深度学习(下).zip 46.【模块四】3.1 人工智能新服务解析——Rekognition、Polly.zip 47.【模块四】3.2 人工智能新服务解析——Rekognition、Polly.zip 48.【模块四】4.1 第四章 AWS 云计算--Amazon ML 机器学习实.zip 49.【模块四】4.2 第四章 AWS 云计算--Amazon ML 机器学习实.zip 5.【模块二】1.1 Amazon EC2 简介.zip 50.【模块四】4.3 第四章 AWS 云计算--Amazon ML 机器学习实.zip 51.【模块四】4.4 第四章 AWS 云计算--Amazon ML 机器学习实.zip 52.【模块四】4.5 第四章 AWS 云计算--Amazon ML 机器学习实.zip 53.【模块四】4.6 第四章 AWS 云计算--Amazon ML 机器学习实.zip 54.【模块六】1.1 Amazon Elasticsearch服务概述.zip 55.【模块六】1.2 创建Amazon Elasticsearch集群.zip 56.【模块六】1.3 配置nginx反向代理实现kibana对ES的访问.zip 57.【模块六】1.4 使用curl进行ES的数据管理及kiban
评论
成就一亿技术人!
拼手气红包6.0元
还能输入1000个字符
 
红包 添加红包
表情包 插入表情
 条评论被折叠 查看
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值