Skip to content
This repository has been archived by the owner on Jul 12, 2021. It is now read-only.

Latest commit

 

History

History
1015 lines (791 loc) · 45.1 KB

File metadata and controls

1015 lines (791 loc) · 45.1 KB

Storage

S3 (Simple Storage Service)

S3 provides developers and IT teams with secure durable, highly scalable object storage. Amazon S3 is easy to use, with a simple web service interface to store and retrieve any amount of data from anywhere on the web.

What is S3?

  • One type of storage in AWS
  • S3 is a safe place to store your files
  • S3 buckets are PRIVATE by default
  • It is Object based storage - meaning you can upload files from 0 Bytes to 5 TB
    • Think of Objects just as files
    • Objects consist of the following:
      • Key - This is the name of the object
      • Value - This is the data and is made up of a sequence of bytes
      • Version ID (Important for versioning)
      • Metadata
      • Subresources
        • Access Control Lists
        • Torrent
  • The data is spread across multiple devices and facilities
  • Unlimited storage
  • Files are stored in Buckets (think of them like folders). S3 is a universal namespace. The Bucket names must be unique globally.
  • When you upload a file to S3 you receive a HTTP 200 code if the upload was successful
  • Because it's object based storage, it's not suitable to install an operating system or database on (for that you'd want block based storage)
  • S3 FAQ

S3 Data Consistency

  • Read after Write consistency for PUTS of new Objects
  • Eventual Consistency for overwrite PUTS and DELETES (can take some time to propagate)

In other words

  • If you write a new file and read it immediately afterwards you will be able to view that data.
  • If you update AN EXISTING file or delete a file and read it immediately, you may get the older version, or you may not. Basically the changes to objects take a bit of time to propagate.

S3 Guarantees from Amazon

  • Built for 99.99% availability for the S3 platform
  • Amazon guarantee 99.9% availability
  • Amazon guarantee 99.999999999% durability for S3 information (Remember 11 x 9's)

S3 Features

  • Lifecycle Management - ie. When an object is X days old have it on this tier and another tier when it hits Y days old
  • Versioning
  • Encryption (at rest)
  • MFA for Delete optional to safeguard deletes
  • Secure your data with Access Control Lists and Bucket Policies

S3 Tiered Storage Options

  • S3 Standard - 99.99 availability 99.999999999, durability, stored redundantly across multiple devices in multiple facilities, and is designed to sustain the loss of 2 facilities concurrently.
  • S3 IA (Infrequently Accessed) - For data that is accessed less frequently, but requires rapid access when needed. Lower fee than S3, but you are charged a retrieval fee.
  • S3 One Zone IA - For when you want a lower cost option for infrequently accessed data, but do not require the multiple Availability Zone data resilience.
  • Intelligent Tiering - Designed to cost optimize costs by automatically moving data to the most cost-effective access tier, without performance impact or operational overhead. Uses machine learning to automatically move your objects to different Storage Tiers based on use.
  • S3 Glacier - S3 Glacier is a secure, durable, and low cost storage class for data archiving. You can reliably store any amount of data at costs that are competitive with or cheaper than on-premises solutions. Retrieval times are configurable from minuets to hours.
  • S3 Glacier Deep Storage - S3 Glacier Deep Storage is Amazon S3's lowest cost storage class where a retrieval time of 12 hours is acceptable.

s3-info

S3 Charges

You are charged for S3 in the following ways:

  • Strorage
  • Requests
  • Storage Management Pricing
  • Data Transfer Pricing
  • Transfer Acceleration - Enables fast, easy, secure transfers of files over long distances between your end users and an S3 bucket. Transfer Acceleration takes advantage of Amazon CloudFront's globally distributed edge locations. As the data arrives at an edge location, data is routed to Amazon S3 over an optimized network path.
  • Cross Region Replication Pricing - Automatic replication to another bucket in a different region

S3 Access Control

  • Controlling access to S3 can be done using one of the following
    • bucket ACL
    • Bucket Policies
  • S3 buckets are PRIVATE by default. You can setup access control to your buckets using
    • Bucket Policies - Sets access at the bucket level
    • Access Control Lists - Sets access at the individual object level
  • S3 buckets can be configured to create access logs which log all requests made to the S3 bucket. This can be sent to another bucket and even another bucket in another account (👀).

S3 Encryption

Encryption in Transit is achieved by SSL/TLS Encryption at Rest (Server Side) is achieved by

  • Server Side
    • S3 Managed Keys - SSE-S3
    • AWS Key Management Service, Managed Keys - SSE-KMS
    • Server Side Encryption with Customer Provided Keys - SSE-C
  • Client Side

S3 Versioning

  • Stores all versions of an object (including all writes and even if you delete an object)
  • Great backup tool
  • Once enabled, Versioning cannot be disabled, only suspended. To remove versioning you'd have to delete the bucket and create a new one.
  • Integrates with Lifecycle rules
  • Versioning's MFA Delete capability, which uses multi-factor authentication, can be used to provide an additional layer of security.
  • If you have versioning enabled the size of your bucket is the sum of all versions.
  • Deleting an object doesn't remove the object but rather places a delete marker over the object in a new version. You can delete the delete marker in that last version to restore the version before it was deleted.

S3 Lifecycle Management

  • Automates moving your objects between the different storage tiers
  • Can be used in conjunction with versioning (but it can work without versioning also)
  • Can be applied to current version and if versioning is enabled previous versions

S3 Cross Regional Replication

  • Versioning must be enabled on both the source and destination bucket
  • Regions must be unique
  • Existing files in an existing bucket are not replicated automatically. When you add new files they will be then replicated atomically.
  • Delete markers are not replicated. This was a deliberate move by AWS to not have deletes auto replicated across all buckets.
  • Replication of deleting individual versions of a object (file) will not be replicated

Transfer Acceleration

Uses CloudFront Edge network to accelerate your uploads to S3. Instead of uploading directly to your S3 bucket, you can use a distinct URL to upload directly to an Edge Location which will then transfer that file to S3. You will get a distinct URL to upload to. For example: foo.s3-accelerate.amazonaws.com.

aws-s3-transfer-acceleration

image credit www.javatpoint.com

CloudFront

A content delivery network (CDN) is a system of distributed servers (network) that deliver webpages and other content to a user based on the geographic locations of the user, the origin of the webpage, and a content delivery server.

CloudFront can be used to deliver your entire website, including dynamic, static, streaming, and interactive content using a global network of edge locations. Requests for your content are automatically routed to the nearest edge location, so content is delivered with the best possible performance.

  • CloudFront is a Networking & Content Delivery solution in AWS
  • Edge Locations are not just READ only - you can write to them too (ie put an object on to them as seen in S3 Transfer Acceleration.)
  • Objects are cached for the TTL (Time To Live) which is configurable
  • You can clear cached objects, but you will be charged. This is called invalidating the cache.
  • You can restrict access to content served on CloudFront with signed URL's and or signed cookies

CloudFront Terminology

  • Edge Location - A location where content will be cached. This is separate to an AWS Region/ Availability Zone.
  • Origin - This is the origin of all files that the CDN will distribute. This can be an S3 Bucket, an EC2 Instance, an Elastic Load Balancer, or Route 53
  • Distribution - This is the name given to the CDN which consists of a collection of Edge Locations
  • Web Distribution - Typically used for websites
  • RTMP - Used for Media Streaming, common with Adobe Flash media streaming

Snowball

Snowball is a petabyte-scale data transport solution that uses secure appliances to transfer large ammounts of data into and out of AWS. Using Snowball addresses common challenges with large-scale data transfers including high network costs, long transfer times, and security concerns. Transferring data with Snowball is simple, fast, secure, and can be as little as one-fifth the cost of high speed Internet. Once the data transfer job has been processed and verified, AWS performs a software erasure of the Snowball appliance.

  • Snowball comes in a 50TB or 80TB size
  • Snowball uses multiple layers of security designed to protect your data including
    • Tamper-resistant enclosures
    • 256-bit encryption
    • Industry standard Trusted Platform Module (TPM) designed to ensure both security and full chain of custody of your data.
  • Snowball can import to S3
  • Snowball can export from S3

Snowball Edge

Snowball Edge is a 100TB data transfer device with on-board storage and compute capabilities. You can use Snowball Edge to move large ammounts of data into and out of AWS, as a temporary storage tier for large local datasets, or to support local workloads in remote or offline locations. In short it provides temporary storage of large data sets and local workloads in remote or offline locations. You could run AWS Lambda's off of it for example.

Snowball Edge connects to your existing applications and infrastructure using standard storage interfaces, streamlining the data transfer process and minimizing setup and integration. Snowball Edge can cluster together to form a local storage tier and process your data on-premises, helping ensure your applicatons continue to run even when they are not able to access the cloud.

Snowmobile

AWS Snowmobile is an Exabyte-scale data transfer service used to move extreamly large amounts of data to AWS. You can transfer up to 100PB per Snowmobile, a 45-foot long ruggedized shipping container, pulled by a semi-trailer truck. Snowmobile makes it easy to move massive volumes of data to the cloud, including video libraries, image repositories, or even complete data center migration. Transfering data with Snowmobile is secure, fast, and cost effective.

Storage Gateway

AWS Storage Gateway is a service that connects an on-premises software applicance with cloud-based storage to provide seamless and secure integration between an organization's on-premises IT environment and AWS storage infrastucture. The service enables you to securely store data to the AWS cloud for scalable and cost-effective storage.

Storage Gateway's Software appliance is available for download as a virtual machine (VM) image that you install on a host in your datacenter. Storage Gateway supports either VMWare ESXi or Microsoft Hyper-V. Once you've installed your gateway and associated it with your AWS account through the activation process, you can use the AWS Managment Console to create the storage gateway option that is right for you.

Storage Gateway Diagram

Storage Gateway Types

  • File Gateway (NFS & SMB) - Files are stored as objects in your S3 buckets, accessed through a Network File System (NFS) mount point. Ownership, permissions, and timestamps are durably stored in S3 in the user-metadata of the object associated with the file. Once objects are transfered to S3, they can be managed as native S3 objects, and bucket policies such as versioning, lifecycle management, and cross-region replication apply directly to objects stored in your bucket.
  • Volume Gateway (iSCSI) - The volume interface presents your applications with disk volumes using the iSCSI block protocol. Data written to these volumes can be asynchronous backed up as point-in-time shapshots of your volumes, and stored in the cloud as Amazon EBS snapshots. Snapshots are incremental backups that capture only your changed blocks. All snapshot storage is also compressed to minimize your storage charges. There are two types:
    • Stored Volumes - Enables you to store primary data locally, while asynchronous backing up that data to AWS. Stored Volumes provide your on-premises applications with low-latency access to their entire adatasets, while providing durable, off-site backups. You can create storage volumes and mount them as iSCSI devices from your on-premises application servers. Data written to your stored volumes is stroed on your on-premises storage hardware. This data is asynchronous backed up to Amazon S3 in the form of Amazon Elastic Block Stroage (Amazon EBS) snapshots.
    • Cached Volumes - Enables you to use Amazon S3 as your primary data storage while retaining frequently accessed data locally in your Storage Gateway. Cached volumes minimize the need to scale your on-premises storage infrastructure, while still providing your applications with low-latency access to their frequently accessed data. You can create storage volumes up to 32TiB in size and attach them as iSCSI devices from your on-premises Storage Gateway's cache and upload buffer storage.
  • Tape Gateway (VTL) - Offers a durable, cost effective solution to archive your data in the AWS cloud. The VTL interface it provides lets you leverage your existing tape-based backup application infrastructure to store on virtual tape cartridges that you create on your tape gateway. Each tape gateway is preconfigured with a media changer and tape drives which are available to your existing client backup applications as iSCSI devices. You add tape cartridges as you need to archive your data. Supported by NetBackup, Backup Exec, Veeam, etc.

Storage Gateway TL;DR

AWS offers virtual solutions for the following:

  • File Gateway - For flat files, stored directly on S3
  • Volume Gateway
    • Stored Volumes - Entire dataset is stored on site and is asynchronous backed up to S3
    • Cached Volumes - Entire dataset is stored on S3 and the most frequently accessed data is cached on site
  • Gateway Virtual Tape Library

EBS (Elastic Block Store)

Elastic Block Store provides persistent block storage volumes for use with Amazon EC2 instances in the AWS Cloud. Each Amazon EBS volume is automatically replicated within its Availability Zone to protect you from component failure, offering high availability and durability.

  • When adding additional volumes Delete on Termination is not checked automatically
  • You can change the type of EBS that the bootable OS is on to Provisioned IOPS (io1) without having downtime (crazy cool tech).
  • You can change EBS volume sizes and storage types on the fly

5 Different EBS Storage Types

  • General Purpose (gp2) (SSD) - General purpose SSD volume that balances price and performance for a wide variety of transactional workloads
  • Provisioned IOPS (io1) (SSD) - Really fast input/outputs per second. Highest performing SSD volume designed for mission-critical applications.
  • Throughput Optimised Hard Disk Drive (st1) - Low cost HDD volume designed for frequently accessed, throughput intensive workloads
  • Cold Hard Disk Drive (sc1) - Lowest cost HDD volume designed for less frequently accessed workloads
  • Magnetic (Standard) - Previous generation HDD

ebs-info

Volumes & Snapshots

  • Volumes will ALWAYS be in the same Availability Zone as the EC2 instance
  • AMI (Amazon Machine Image) provides the information required to launch an instance
  • Volumes exist on EBS. Think of EBS as virtual hard disk in the cloud.
  • Snapshots exist on S3. Think of snapshots as a photograph of the disk.
  • Snapshots are point in time copies of Volumes.
  • Snapshots are incremental - this means that only the blocks that have changed since your last snapshot are moved to S3
  • If this is your first snapshot, it may take some time to create
  • To create a snapshot for Amazon EBS volumes that serve as root devices, you should stop the instance before taking the snapshot. However, you can take a snapshot of the instance while running.
  • You can create AMI's from both volumes and Snapshots
  • Snapshots of encrypted volumes are encrypted automatically
  • Volumes restored from encrypted snapshots are encrypted automatically
  • You can share snapshots, but only if they are unencrypted
  • These snapshots can be shared with other AWS accounts or made public
  • You can now encrypt root device volumes upon creation of the EC2 instance
  • Process to encypt a root device volume if it was created unencrypted
    1. Create a snapshot of the unencrypted device volume
    2. Create a copy of the Snapshot and select the encrypt option
    3. Create an AMI from the encrypted Snapshot
    4. Use that AMI to launch a new encrypted instance(s)

Migration

To move an EC2 volume from one Availability Zone to another:

  1. Take a snapshot of it
  2. create an AMI from the snapshot
  3. Use the AMI to launch the EC2 instance in a new Availability Zone

To move an EC2 volume from one region to another:

  1. Take a snapshot of the volume
  2. Create an AMI from the snapshot
  3. Copy the AMI from one region to the other
  4. Use the copied AMI to launch the new EC2 instance in the new region

AMI

An Amazon Machine Image (AMI) provides the information required to launch an instance. You must specify an AMI when you launch an instance. You can launch multiple instances from a single AMI when you need multiple instances with the same configuration. You can use different AMIs to launch instances when you need instances with different configurations.

AMI Types

  • Region
  • Operating System
  • Architecture (32 bit or 64 bit)
  • Launch Permissions
  • Storage for the Root Device Volume
    • Instance Store (EPHEMERAL STORAGE)
    • EBS Backed Volumes

Instance Store Volumes

  • All AMIs are categorized as either backed by Amazon EBS or backed by instance store
  • For EBS Volumes: The root device for an instance launched from the AMI is an Amazon EBS volume created from an Amazon EBS snapshot
  • For Instance Store Volumes: The root device for an instance from the AMI is an instance store volume created from a template stored in Amazon S3
  • Instance Store Volumes are sometimes called Ephemeral Storage
  • Instance store volumes cannot be stopped. If the underlying host fails, you will lose your data.
  • EBS backed instances can be stopped. You will not lose the data on this instance if it is stopped.
  • You can reboot both, you will not lose your data
  • By default, both ROOT volumes will be deleted on termination. However, with EBS volumes, you can tell AWS to keep the root device volume.

EFS (Elastic File System)

Amazon Elastic File System (EFS) is a file storage service for EC2 (Amazon Elastic Compute Cloud) instances. Amazon EFS is easy to use and provides a simple interface that allows you to create and configure file systems quickly and easily. With Amazon EFS, storage capacity is elastic, growing as you add and remove files, so your applications have the storage they need, when they need it.

  • Supports the Network File System version 4 (NFSv4) protocol
  • You only pay for the storage you use (no pre-provisioning required) (With EBS we had to say how much we wanted)
  • Can scale up to the petabytes
  • Can support thousands of concurrent NFS connections
  • Data is stored across multiple Availability Zones within a region
  • Read After Write Consistency

Storage Review

  1. You have been asked to advise on a scaling concern. The client has an elegant solution that works well. As the information base grows they use CloudFormation to spin up another stack made up of an S3 bucket and supporting compute instances. The trigger for creating a new stack is when the PUT rate approaches 100 PUTs per second. The problem is that as the business grows that number of buckets is growing into the hundreds and will soon be in the thousands. You have been asked what can be done to reduce the number of buckets without changing the basic architecture.
  • a. Refine the key hashing to randomise the name Key to achieve the potential of 300 PUTs per second
  • b. Change the trigger level to around 3000 as S3 can now accomodate much higher PUT and GET levels
  • c. Upgrade all buckets to S3 provisioned IOPS to achieve better performance
  • d. Set up multiple accounts so that the per account hard limit on S3 buckets is avoided
Show answer

- b. Change the trigger level to around 3000 as S3 can now accomodate much higher PUT and GET levels

Until 2018 there was a hard limit on S3 puts of 100 PUTs per second. To achieve this care needed to be taken with the structure of the name Key to ensure parallel processing. As of July 2018 the limit was raised to 3500 and the need for the Key design was basically eliminated. Disk IOPS is not the issue with the problem. The account limit is not the issue with the problem.

  1. You run a meme creation website where users can create memes and then download them for use on their own sites. The original images are stored in S3 and each meme's metadata in DynamoDB. You need to decide upon a low-cost storage option for the memes, themselves. If a meme object is unavailable or lost, a Lambda function will automatically recreate it using the original file from S3 and the metadata from DynamoDB. Which storage solution should you use to store the non-critical, easily reproducible memes in the most cost-effective way?
  • a. S3
  • b. S3 - IA
  • c. S3 - 1 Zone - IA
  • d. S3 - RRS
Show answer

- c. S3 - 1 Zone - IA

S3 – OneZone-IA is the recommended storage for when you want cheaper storage for infrequently accessed objects. It has the same durability but less availability. There can be cost implications if you use it frequently or use it for short lived storage. Glacier is cheaper, but has a long retrieval time. RRS has effectively been deprecated. It still exists but is not a service that AWS want to sell anymore.

  1. You have uploaded a file to S3. Which HTTP code would indicate that the upload was successful?
  • a. HTTP 404
  • b. HTTP 501
  • c. HTTP 200
  • d. HTTP 307
Show answer

- c. HTTP 200

  1. S3 has an eventual consistency for which HTTP Methods?
  • a. PUTS of new Objects and DELETES
  • b. Overwrite PUTS and DELETES
  • c. PUTS of new objects and UPDATES
  • d. UPDATES and DELETES
Show answer

- b. Overwrite PUTS and DELETES

  1. What is Amazon Glacier?
  • a. A tool that allows you to "freeze" an EBS volume
  • b. An AWS service designed for long term data archival
  • c. A highly secure firewall designed to keeo everything out
  • d. It is a tool used to resurrect deleted EC2 snapshots
Show answer

- b. An AWS service designed for long term data archival

  1. You work for a health insurance company that amasses a large number of patients' health records. Each record will be used once when assessing a customer, and will then need to be securely stored for a period of 7 years. In some rare cases, you may need to retrieve this data within 24 hours of a claim being lodged. Given these requirements, which type of AWS storage would deliver the least expensive solution?
  • a. S3
  • b. S3 - IA
  • c. S3 - 1 Zone - IA
  • d. S3 - RRS
  • e. Glacier
Show answer

- e. Glacier

The recovery rate is a key decider. The record shortage must be; safe, durable, low cost, and the recovery can be slow. All features of Glacier.

  1. The difference between S3 and EBS is that EBS is object-based whereas S3 is block-based.
  • a. True
  • b. False
Show answer

- b. False

  1. What is the availability of S3 – OneZone-IA?
  • a. 99.90%
  • b. 99.50%
  • c. 99.99%
  • d. 100%
Show answer

- b. 99.50%

OneZone-IA is only stored in one Zone. While it has the same Durability, it may be less Available than normal S3 or S3-IA.

  1. One of your users is trying to upload a 7.5GB file to S3. However, they keep getting the following error message: "Your proposed upload exceeds the maximum allowed object size.". What solution to this problem does AWS recommend?
  • a. Design your application to use the Multipart Upload API for all objects
  • b. Design your application to use large object upload API for this object
  • c. Raise a ticket with AWS to increase your maxium object size
  • d. Log into the S3 console, click on the bucket and then click properties. You can thne increase your maxium object size to 1TB.
Show answer

- a. Design your application to use the Multipart Upload API for all objects

  1. You run a popular photo-sharing website that depends on S3 to store content. Paid advertising is your primary source of revenue. However, you have discovered that other websites are linking directly to the images in your buckets, not to the HTML pages that serve the content. This means that people are not seeing the paid advertising, and you are paying AWS unnecessarily to serve content directly from S3. How might you resolve this issue?
  • a. Use CloudFront to serve the static content
  • b. Remove the ability for images to be served publicly to the site and then use signed URLs with expiry dates
  • c. Use security groups to blacklist the IP addresses of the sites that link directly to your S3 bucket
  • d. Use EBS rather than S3 to store the content
Show answer

- b. Remove the ability for images to be served publicly to the site and then use signed URLs with expiry dates

  1. You have been asked by your company to create an S3 bucket with the name "acloudguru1234" in the EU West region. What would the URL for this bucket be?
  • a. https://s3-eu-west-1.amazonaws.com/acloudguru1234
  • b. https://s3-us-east-1.amazonaws.com/acloudguru1234
  • c. https://s3.acloudguru1234.amazonaws.com/eu-west-1
  • d. https://s3-acloudguru1234.amazonaws.com/
Show answer

- a. `https://s3-eu-west-1.amazonaws.com/acloudguru1234`

  1. You work for a major news network in Europe. They have just released a new mobile app that allows users to post their photos of newsworthy events in real-time, which are then reviewed by your editors before being copied to your website and made public. Your organization expects this app to grow very quickly, essentially doubling its user base each month. The app uses S3 to store the images, and you are expecting sudden and sizable increases in traffic to S3 when a major news event takes place (as users will be uploading large amounts of content.) You need to keep your storage costs to a minimum, and it does not matter if some objects are lost. With these factors in mind, which storage media should you use to keep costs as low as possible?
  • a. S3 - IA
  • b. S3 - 1 Zone - IA
  • c. S3 - RRS
  • d. Glacier
  • e. S3 Provisioned IOPS
Show answer

- b. S3 - 1 Zone - IA

The key driver here is cost, so an awareness of cost is necessary to answer this. Full S3 is quite expensive at around $0.023 per GB for the lowest band. S3 standard IA is $0.0125 per GB, S3 One-Zone-IA is $0.01 per GB, and Legacy S3-RRS is around $0.024 per GB for the lowest band. Of the offered solutions SS3 One-Zone-IA is the cheapest suitable option. Glacier cannot be considered as it is not intended for direct access, however it comes in at around $0.004 per GB. Of course you spotted that RRS is being deprecated, and there is no such thing as S3 – Provisioned IOPS. In this case OneZone

  1. How many S3 buckets can I have per account by default?
  • a. 10
  • b. 20
  • c. 50
  • d. 100
Show answer

- d. 100

  1. You work for a busy digital marketing company who currently store their data on-premise. They are looking to migrate to AWS S3 and to store their data in buckets. Each bucket will be named after their individual customers, followed by a random series of letters and numbers. Once written to S3 the data is rarely changed, as it has already been sent to the end customer for them to use as they see fit. However, on some occasions, customers may need certain files updated quickly, and this may be for work that has been done months or even years ago. You would need to be able to access this data immediately to make changes in that case, but you must also keep your storage costs extremely low. The data is not easily reproducible if lost. Which S3 storage class should you choose to minimize costs and to maximize retrieval times?
  • a. S3
  • b. S3 - IA
  • c. S3 - 1 Zone - IA
  • d. S3 - RRS
Show answer

- b. S3 - IA

The need to immediate access is an important requirement along with cost. Glacier has a long recovery time at a low cost or a shorter recovery time at a high cost, and 1Zone-IA has a lower Availability level which means that it may not be available when needed.

  1. What is the availability of objects stored in S3?
  • a. 99%
  • b. 99.90%
  • c. 99.99%
  • d. 100%
Show answer

- c. 99.99%

  1. S3 has what consistency model for PUTS of new objects?
  • a. Read After Write Consistency
  • b. Write After Read Consistency
  • c. Eventual Consistency
  • d. Usual Consistency
Show answer

- a. Read After Write Consistency

  1. What does S3 stand for?
  • a. Simple SQL Service
  • b. Simple Storage Service
  • c. Simplified Serial Sequence
  • d. Straight Storage Service
Show answer

- b. Simple Storage Service

  1. You are a solutions architect who works with a large digital media company. The company has decided that they want to operate within the Japanese region and they need a bucket called "testbucket" set up immediately to test their web application on. You log in to the AWS console and try to create this bucket in the Japanese region however you are told that the bucket name is already taken. What should you do to resolve this?
  • a. Change your region to Korea and then create the bucket "testbucket"
  • b. Raise a ticket with AWS and ask them to release the name "testbucket" to you
  • c. Bucket names are global, not regional. This is a popular bucket name and is already taken. You should choose another bucket name.
  • d. Run a WHOIS request on the bucket name and get the registered owners email address. Contact the owner and ask if you can purchase the rights to the bucket.
Show answer

- c. Bucket names are global, not regional. This is a popular bucket name and is already taken. You should choose another bucket name.

  1. What is the minimum file size that I can store on S3?
  • a. 1KB
  • b. 1MB
  • c. 0 bytes
  • d. 1 byte
Show answer

- c. 0 bytes

  1. What is AWS Storage Gateway?
  • a. It is a physical or virtual appliance that can be used to cache S3 locally at a customer's site.
  • b. It allows large scale import/exports into the AWS cloud withou the use of an internet connection
  • c. It allows direct MPLS connection to AWS
  • d. None of the above
Show answer

- a. It is a physical or virtual appliance that can be used to cache S3 locally at a customer's site.

At its heart it is a way of using AWS S3 managed storage to supplement on-premise storage. It can also be used within a VPC in a similar way.

  1. _________ is Object Based storage?
Show answer

S3

More info

  1. True or False, S3 is a universal namespace?
Show answer

True

More info

  1. What are the two access control mechanisms for controlling access to S3 buckets?
Show answer

  • Bucket Policies
  • Access Control Lists

More info

  1. What are the fundamental parts of S3?
Show answer

  • Key
  • Value
  • Version ID
  • Metadata
  • Subresources

More info

  1. What is the consistency model for S3?
Show answer

  • Read After Write consistency for PUTS of new Objects
  • Eventual Consistency for overwrite PUTS and DELETES (can take some time to propagate)

More info

  1. What are the types of S3?
Show answer

  • S3 Standard
  • S3 - IA
  • S3 - 1 Zone - IA
  • S3 - Intelligent Tiering
  • S3 Glacier
  • S3 Glacier Deep Archive

More info

  1. Encryption In Transit is achieved by?
Show answer

SSL/TLS

More info

  1. Encryption At Rest is achieved by?
Show answer

Server Side

  • S3 Managed Keys (SSE-S3)
  • AWS KMS (Key Management Service) (SSE-KMS)
  • Server Side Encryption with Customer Provided Keys (SSE-C)

Client Side Encryption

More info

  1. True or False, for S3 Cross Region Replication, Versioning must be enabled ob both the source and destination buckets??
Show answer

True

More info

  1. True or False, regions in Cross Region Replication must be unique?
Show answer

True

More info

  1. True or False, Existing files in an S3 bucket are not automatically replicated when you turn on Cross Region Replication?
Show answer

True, all subsequent updated files will be replicated automatically

More info

  1. True or False, with Cross Region Replication, delete markers are not replicated?
Show answer

True

More info

  1. True or False, with Cross Region Replication, deleting individual versions or delete markers will not be replicated?
Show answer

True

More info

  1. What AWS service automates moving your objects between the different S3 tiers??
Show answer

Lifecycle Management

More info

  1. True or False, Lifecycle Policies can be used with S3 Versioning?
Show answer

True

More info

  1. True or False, Lifecycle Policies can be applied to current versions and previous versions?
Show answer

True

More info

  1. What are the two types of distributions in CloudFront?
Show answer

  • Web - Used for websites
  • RTMP - Used for media streaming

More info

  1. CloudFront origin locations can be what type(s)?
Show answer

  • S3
  • EC2 Instance
  • ELB (Elastic Load Balancer)
  • Route53

More info

  1. True or False, Edge Locations are READ only?
Show answer

False, you can write to them too (ie put an object on to them)

More info

  1. True or False, you can clear cached items in CloudFront?
Show answer

True, you can invalidate a cached item but you will be charged

More info

  1. Instance Store Volumes are sometimes called ______________?
Show answer

Ephemeral Storage

More info

  1. What are the types of AMI storage for the Root Device Volume?
Show answer

  • Instance Store Volumes (EPHEMERAL STORAGE)
  • EBS Backed Volumes

More info

  1. Which AMI Root Device Volume type can be stopped without losing data?
Show answer

EBS backed instances can be stopped. You will not lise the data on this instance if it is stopped. Instance store volumes cannot be stopped. With Instance Store Volumes if the underlying host fails, you will lose your data.

More info

  1. Which AMI Storage for the Root Device Volume can be rebooted safely without data loss?
Show answer

You can reboot both Instance Store Volumes (EPHEMERAL STORAGE) and EBS Backed Volumes and you will not lose your data

More info

  1. Volumes exist on _________?
Show answer

EBS, think of EBS as virtual hard disk.

More info

  1. Snapshots exist on ___________?
Show answer

S3

More info

  1. Snapshots are __________ copies of Volumes.
Show answer

point in time

More info

  1. Why do initial snapshots take some time to create but subsequent snapshots areq more quick?
Show answer

Snapshots are incremental - this means that only the blocks that have changed since your last snapshot are moved to S3

More info

  1. What can you create AMI's from?
Show answer

EBS Volume or Snapshots

More info

  1. True or False, you can change EBS volume sizes and storage types on the fly?
Show answer

True

More info

  1. True or False, Volumes will always be in the same Availability Zone as the EC2 instance?
Show answer

True

More info

  1. What are the steps to move an EC2 volume from one Availability Zone to another?
Show answer

  1. Take a snapshot
  2. Create an AMI from the snapshot
  3. Use the AMI to launch the EC2 instance in the new Availability Zone

More info

  1. What are the steps to move an EC2 volume from one region to another?
Show answer

  1. Take a snapshot
  2. Create an AMI from the snapshot
  3. Copy the AMI from the orginal region to the desired region
  4. Use the AMI to launch the EC2 instance in the new region

More info

  1. True or False, Snapshots of encrypted volumes are encrypted automatically?
Show answer

True

More info

  1. True or False, Volumes restored from encrypted snapshots are encrypted automatically?
Show answer

True

More info

  1. What are the steps to encrypt an unencrypted root device volume that needs to be encrypted?
Show answer

  1. Create a Snapshot of the unencrypted root device volume
  2. Create a copy of the Snapshot and select the encrypt option
  3. Create an AMI from the encrypted Snapshot
  4. Use that AMI to launch new encrypted instances

More info

  1. What protocol does EFS support?
Show answer

Network File System version 4 (NFSv4)

More info

  1. True or False, with EFS, you only pay for the storage you use?
Show answer

True, no pre-provisioning required

More info

  1. True or False, NFS can support thousands of concurrent NFS connections?
Show answer

True

More info

  1. True or False, EFS data is stored in a single Availability Zone within a region?
Show answer

False

More info

  1. What is the consistency model for EFS?
Show answer

Read After Write Consistency

More info