There are a LOT of AWS services to get to grips with. Inspired by AWS in Plain English, I've created my own glossary of terminology to make sure I know my CloudFront from CloudTrail and Athena from Aurora.
On Demand Instance
Virtual Private Cloud
Amazon CloudWatch collects and tracks metrics for your AWS resources. If you are new to AWS and are using the free tier you may want to add a Billing Alarm to make sure you don't run into any unexpected charges. It's easy to forget something is running and get landed with a bill.
An S3 bucket is where objects are stored, similar to files and folders on your local machine. Each object consists of:
- Key - the name of the object
- Value - the data in the file itself made of bytes
The storage tiers are:
- S3 - Most expensive and reliable option
- S3:IA - For storing non-critical data that CANNOT be easily reproduced and needs to be retrieved quickly
- S3:IA-One Zone - For storing non-critical data that CAN be easily reproduced and needs to be retrieved quickly.
- Glacier - Extremely cheap long-term storage
- Deep Glacier - For long-term storage with a 12 hour retrieval time for 'cold' data
The AWS CLI allows you to issue commands from the command line. It's useful for uploading files to S3 buckets and launching EC2 instances.
Amazon RDS creates a storage volume snapshot of your entire instance. Creating this snapshot results in a brief I/O suspension that can last from a few seconds to a few minutes. Multi-AZ DB instances are not affected by this I/O suspension since the backup is taken on the standby.
When you create a DB snapshot, you need to identify which DB instance you are going to back up, and then give your DB snapshot a name so you can restore from it later. You can do this using the AWS Management Console, the AWS CLI, or the RDS API.
Amazon CloudFront is the AWS CDN. It caches information closest to the user to the next user can download a copy faster. CloudFront can distribute all website content including dynamic, static, streaming and interactive content from either AWS services like S3 or your own non-AWS server.
Kinesis allows data to be streamed in real-time from a Producer to a Processer or Storage option. This is a huge change from Batch Processing that has been the traditional way to land data from one location to another.
Users by default have no access until you assign them a role. Roles define a set of permissions for making AWS service requests and are most often used to assign Groups of Users permissions to perform tasks or access services.
Amazon Route 53 is Amazons Domain Name System (DNS) web service. It is designed to give developers a cost-effective way to route end users to Internet applications by translating names like www.example.com into the numeric IP addresses like 192.0.2.1 that computers use to connect to each other.
AWS named the service Route 53 because all DNS requests are handled through port 53.
EC2 is a service that provides virtual machines in the cloud where you only pay for the capacity you use and choose from ‘families’ of instance types that are good for different use cases.
General Purpose - a balance of compute, memory and networking resources
Compute Optimised -ideal for compute-bound applications that benefit from the high-performance processor
Memory-Optimised - fast performance for workloads that process large data sets in memory
Accelerated Optimised - hardware accelerators, or co-processors
Storage Optimised - high, sequential read and write access to very large data sets on local storage
Amazon EMR provides a scalable framework so you can run Spark and Hadoop processes over an S3 data lake. The Run Job on an EMR template launches an Amazon EMR cluster based on the parameters provided and starts running steps based on the specified schedule. Once the job completes, the EMR cluster is terminated.
The AWS KMS Service makes it easy to create and control encryption keys on AWS which can then be utilised to encrypt and decrypt data in a safe manner. The service leverages Hardware Security Modules (HSM) under the hood which in return guarantees security and integrity of the generated keys.
To manage your objects so that they are stored cost-effectively throughout their lifecycle, configure their lifecycle. A lifecycle configuration is a set of rules that define actions that Amazon S3 applies to a group of objects. For example, you might choose to transition objects to the S3:IA storage class 30 days after you created them, or archive objects to the Glacier storage class one year after creating them.
Amazon SNS allows applications to send time-critical messages to multiple subscribers through a “push” mechanism, eliminating the need to periodically check or “poll” for updates
Amazon SQS stores messages in a queue. SQS cannot deliver any messages, where an external service (lambda, EC2 etc) is needed to poll SQS and grab messages from SQS.
By using Amazon SNS and Amazon SQS together, messages can be delivered to applications that require immediate notification of an event, and also persisted in an Amazon SQS queue for other applications to process at a later time.
A virtual private cloud (VPC) is a virtual network dedicated to your AWS account. It is logically isolated from other virtual networks in the AWS Cloud. You can launch your AWS resources, such as Amazon EC2 instances, into your VPC.
You can use a NAT device to enable instances in a private subnet to connect to the internet (for example, for software updates) or other AWS services, but prevent the internet from initiating connections with the instances. A NAT device forwards traffic from the instances in the private subnet to the internet or other AWS services, and then sends the response back to the instances.
There are multiple ways to pay for Amazon EC2 instances:
- On-Demand - pay for capacity by per hour or per second depending on which instances you run.
- Reserved Instances - provide a reservation at 75% off the On-Demand price, giving you confidence in your ability to launch instances when you need them.
- Spot Instances - request spare Amazon EC2 computing capacity for up to 90% off the On-Demand price.
- Dedicated Hosts - provide EC2 instance capacity on physical servers dedicated for your use.
- Savings Plan - provides the benefits of Reserved Instances but with more flexibility to change instance type within the same family while taking advantage of savings
Amazon EBS is a persistent storage device that can be attached to a single EC2 instance to be used as a file system for databases, application hosting, and storage.
Amazon EFS is a managed network file system that can be shared across multiple Amazon EC2 instances and is scalable depending on workload.
Amazon RDS makes it easy to provision a managed database instance in the cloud. At the time of writing the following database engines were available.
- Amazon Aurora for MySQL and PostgreSQL
- MS SQL Server
Read replication can be part of your disaster recovery plan. Replication means that a secondary database is online and can be queried. This is not only good for disaster recovery but can be useful if you utilise one instance for reporting and one for live queries.
If you are using AWS setting this up takes just a few clicks. You can promote a read replica if the source database instance fails or route traffic here to reduce the load on the source database
Auto Scaling launches and terminates Amazon EC2 instances automatically according to user-defined policies, schedules, and alarms. You can use Auto Scaling to maintain a fleet of AWS EC2 instances that can adjust to any presented load. You can also use Auto Scaling to bring up multiple instances in a group at one time.
Metrics are the fundamental concept in CloudWatch. A metric represents a time-ordered set of data points that are published to CloudWatch. Think of a metric as a variable to monitor, and the data points represent the values of that variable over time.
Each data point has a timestamp and a unit of measure. When you request statistics, the returned data stream is identified by namespace, metric name, dimension, and the unit.
A Virtual Private Cloud (VPC) is a virtual data center which is a logically isolated section of AWS that can span availability zones. VPC’s are made of Internet Gateways/Virtual Private Gateways, route tables, network access control lists, subnets, and security groups.
AWS Web Application Firewall (WAF) protects web applications from attacks, like specific user-agents, bad bots, or content scrapers, by filtering traffic based on rules that you create.
Block IP addresses that exceed request limits
This lets you control access by IP address, country, blocking SQL injections, malicious scripts and the length of requests.
Block IP addresses that submit bad requests
This solution allows you to block IP addresses using Lambda, CloudWatch and AWS WAF to block requests after a threshold has been reached.
AWS WAF can be deployed on Amazon CloudFront, protecting your resources and content at the Edge locations. And the Application Load Balancer (ALB), to protect Internet-facing as well as internal load balancers.
You can use x.509 certificates in AWS Certificate Manager to identify users, computers, applications, services, servers, and other devices internally.
OK, I cheated here, but this is a really interesting post from Kevin and puts it all together.
Fun fact: a yobibyte is 2^80 or 1,208,925,819,614,629,174,706,176 bytes.
One of the most important introductory concepts to understand is that AWS hosts its infrastructure in data centres called Availability Zones (AZs). There are multiple AZs in a Region which means that if there is a problem in one AZ another can pick up the slack. For some services, you can host your application in multiple Regions.
This post first appeared on helenanderson.co.nz