0% found this document useful (0 votes)
8 views

Cloud computing activity - unit 3 (1)

Uploaded by

Tarana Ghauri
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views

Cloud computing activity - unit 3 (1)

Uploaded by

Tarana Ghauri
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 18

Activity – Unit 3

Introduction to Cloud Computing


(Skill Enhancement Course)
Session July 2024- December 2024

Name: Tarana Ghauri


Examination Roll No.- 22029504001
College Roll No.- 22/COM/146
Course: B.Com (Hons)
Semester: V
Section- A
AMAZON S3
Amazon Simple Storage Service (Amazon S3) is an
object storage service that offers industry-leading
scalability, data availability, security, and
performance. Customers of all sizes and industries
can use Amazon S3 to store and protect any amount
of data for a range of use cases, such as data lakes,
websites, mobile applications, backup and restore,
archive, enterprise applications, IoT devices, and big
data analytics. Amazon S3 provides management
features so that you can optimize, organize, and
configure access to your data to meet your specific
business, organizational, and compliance
requirements.

Features
1.Storage management and monitoring
Amazon S3’s flat, non-hierarchical structure and
various management features are helping
customers of all sizes and industries organize their
data in ways that are valuable to their businesses
and teams. All objects are stored in S3 buckets and
can be organized with shared names called prefixes.
You can also append up to 10 key-value pairs
called S3 object tags to each object, which can be
created, updated, and deleted throughout an
object’s lifecycle. To keep track of objects and their
respective tags, buckets, and prefixes, you can use
an S3 Inventory report that lists your stored objects
within an S3 bucket or with a specific prefix, and
their respective metadata and encryption status. S3
Inventory can be configured to generate reports on a
daily or a weekly basis.

2.Storage management

With S3 bucket names, prefixes, object tags, and S3


Inventory, you have a range of ways to categorize
and report on your data, and subsequently can
configure other S3 features to take action. Whether
you store thousands of objects or a billion, S3 Batch
Operations makes it simple to manage your data in
Amazon S3 at any scale. Amazon S3 also supports
features that help maintain data version control,
prevent accidental deletions, and replicate data to
the same or a different AWS Region.

3.Storage monitoring
In addition to these management capabilities,
Amazon S3 features and other AWS services can be
used to monitor and control your S3 resources.
Apply tags to S3 buckets to allocate costs across
multiple business dimensions (such as cost
centers, application names, or owners), then use
AWS Cost Allocation Reports to view the usage and
costs aggregated by the bucket tags. You can also
use Amazon CloudWatch to track the operational
health of your AWS resources and configure billing
alerts for estimated charges that reach a user-
defined threshold.

4.Storage analytics and insights

S3 storage lens delivers organization-wide visibility


into object storage usage, activity trends, and
makes actionable recommendations to improve
cost-efficiency and apply data protection best
practices.
Amazon S3 Storage Class Analysis analyzes storage
access patterns to help you decide when to
transition the right data to the right storage class.

5.Storage classes

With Amazon S3, you can store data across a range


of different S3 storage classes purpose-built for
specific use cases and access patterns: S3
Intelligent-Tiering, S3 Standard, S3 Express One
Zone, S3 Standard-Infrequent Access (S3 Standard-
IA), S3 One Zone-Infrequent Access (S3 One Zone-
IA), S3 Glacier Instant Retrieval, S3 Glacier Flexible
Retrieval, S3 Glacier Deep Archive, and S3
Outposts.
Every S3 storage class supports a specific data
access level at corresponding costs or geographic
location.

6.Access management and security

To protect your data in Amazon S3, by default, users


only have access to the S3 resources they create.
You can grant access to other users by using one or
a combination of access management features.
Amazon S3 offers flexible security features to block
unauthorized users from accessing your data.

7.Data processing

With S3 Object Lambda you can add your own code


to S3 GET, HEAD, and LIST requests to modify and
process data as it is returned to an application. You
can use custom code to modify the data returned by
standard S3 GET requests to filter rows, dynamically
resize images, redact confidential data, and much
more.

8.Query in place

Amazon S3 has complementary services that query


data without needing to copy and load it into a
separate analytics platform or data warehouse. This
means you can run data analytics directly on your
data stored in Amazon S3.

9.Data transfer

AWS provides a portfolio of data transfer services to


provide the right solution for any data migration
project. The level of connectivity is a major factor in
data migration, and AWS has offerings that can
address your hybrid cloud storage, online data
transfer, and offline data transfer needs.

10.Data exchange

AWS Data Exchange for Amazon S3 accelerates time


to insight with direct access to data providers’
Amazon S3 data. AWS Data Exchange for Amazon S3
helps you easily find, subscribe to, and use third-
party data files for storage cost optimization,
simplified data licensing management, and more. It
is intended for subscribers who want to easily use
third-party data files for data analysis with AWS
services without needing to create or manage data
copies. It is also helpful for data providers who want
to offer in-place access to data hosted in their
Amazon S3 buckets.

11.Performance

Amazon S3 provides industry leading performance


for cloud object storage. Amazon S3 supports
parallel requests, which means you can scale your
S3 performance by the factor of your compute
cluster, without making any customizations to your
application. Performance scales per prefix, so you
can use as many prefixes as you need in parallel to
achieve the required throughput. There are no limits
to the number of prefixes. Amazon S3 performance
supports at least 3,500 requests per second to add
data and 5,500 requests per second to retrieve data.
Each S3 prefix can support these request rates,
making it simple to increase performance
significantly.

12.Consistency

Amazon S3 delivers strong read-after-write


consistency automatically for all applications,
without changes to performance or availability,
without sacrificing regional isolation for
applications, and at no additional cost. With S3
Strong Consistency, S3 simplifies the migration of
on-premises analytics workloads by removing the
need to make changes to applications, and reduces
costs by removing the need for extra infrastructure
to provide strong consistency.

Benefits

Reliable Security
When created, Amazon S3 buckets are usable only
by the identity that created them. You can set
access permissions for each file, each bucket, or via
IAM(Identity access management), which provides a
complete control over how, where and by whom the
data can be frequently accessed. With these set of
rules and permissions, you can make sure that there
is no unauthorized access to your data.

All-time Availability
Amazon S3 gives every user, its service access to
the same highly scalable, reliable, fast, inexpensive
data storage infrastructure that Amazon uses to run
its own global network of websites. S3 Standard is
designed for 99.99% availability and Standard – IA is
designed for 99.9% availability. Both are backed by
the Amazon S3 Service Level Agreement, which is
strictly followed by Amazon.

Very Low cost


With Amazon S3, you only pay for the data you use,
which in itself is a very low price equivalent to
$0.022 / GB and ~$0.0125 / GB for infrequent
access. You can also define policies to migrate the
data automatically to the infrequent access which
further reduces the cost as Amazon Glacier is even
cheaper( ~$0.004 / GB).

Ease of Migration
With Amazon S3 you get multiple options (rsync, S3
command line interface and Glacier command line
interface) for Cloud Data Migration which are cost
effective and it is very simple to transfer a large
amount of data to Amazon S3 or out of Amazon S3.
Amazon S3 Storage also provides you with the
option to import or export data to any physical
device or on any network.

The Simplicity of Management


Amazon S3 has a very user friendly web interface
which takes out the usual hard work of maintaining
security, optimising storage classes and managing
the data transfer in the most suitable way. You can
define its own lifecycle policy, define replication
rules and configure the Amazon S3 inventory. It also
allows you to configure request metrics and storage
classes analysis with many filters to have a better
look at your storage.

Amazon relational database service


Amazon Relational Database Service (Amazon RDS)
is a web service that makes it easier to set up,
operate, and scale a relational database in the AWS
Cloud. It provides cost-efficient, resizable capacity
for an industry-standard relational database and
manages common database administration tasks.

Benefits
Amazon RDS is a managed database service. It’s
responsible for most management tasks. By
eliminating tedious manual processes, Amazon RDS
frees you to focus on your application and your
users.

Amazon RDS provides the following principal


advantages over database deployments that aren’t
fully managed:

1.You can use database engines that you are


already familiar with: IBM Db2, MariaDB, Microsoft
SQL Server, MySQL, Oracle Database, and
PostgreSQL.
2.Amazon RDS manages backups, software
patching, automatic failure detection, and recovery.

3.You can turn on automated backups, or manually


create your own backup snapshots. You can use
these backups to restore a database. The Amazon
RDS restore process works reliably and efficiently.

4.You can get high availability with a primary DB


instance and a synchronous secondary DB instance
that you can fail over to when problems occur. You
can also use read replicas to increase read scaling.

5.In addition to the security in your database


package, you can control access by using AWS
Identity and Access Management (IAM) to define
users and permissions. You can also help protect
your databases by putting them in a virtual private
cloud (VPC).

Amazon RDS vs Amazon S3


Amazon RDS (Relational Database Service) and
Amazon S3 (Simple Storage Service) are two popular
services offered by Amazon Web Services (AWS).
The key differences between them are-
Database vs. Storage: Amazon RDS is primarily a
managed relational database service that allows
users to set up, operate, and scale a relational
database in the cloud. It provides fully-managed
database instances for MySQL, PostgreSQL, Oracle,
SQL Server, and Amazon Aurora. On the other hand,
Amazon S3 is a scalable object storage service
designed for storing and retrieving large amounts of
unstructured data, such as images, videos,
documents, backups, logs, etc.

Data Structure: Amazon RDS deals with structured


data and follows a fixed schema defined by the
chosen database engine. It provides features like
ACID (Atomicity, Consistency, Isolation, Durability)
properties, data indexing, transactions, and SQL
querying. In contrast, Amazon S3 does not impose
any schema and allows storing any type of object in
its buckets, making it versatile for storing files,
objects, and media content of any format.

Storage Management: Amazon RDS automatically


manages database storage allocation, performance
optimization, backups, and software patching. It
provides features like automated backups,
replication, and point-in-time recovery for data
protection. On the other hand, Amazon S3 provides
simple storage management, enabling users to
create, delete, and organize buckets. It also offers
various storage classes for cost optimization, data
lifecycle policies, and server-side encryption for
data security.

Performance and Scalability: Amazon RDS allows


users to choose the database instance type and
provides options for read replicas and multi-AZ
deployments to enhance performance and
availability. It also supports auto-scaling based on
workload demands. In contrast, Amazon S3 is
designed for high-scalability, providing consistent
performance and virtually unlimited storage
capacity with automatic scaling. It is ideal for
applications that require storing and retrieving large
volumes of data with high throughput.

Data Access and Integration: Amazon RDS allows


direct access to the relational databases using
standard database connectivity protocols like JDBC,
ODBC, and network connectivity. It supports
integrations with other AWS services like AWS
Lambda, AWS Glue, and AWS Data Pipeline. On the
other hand, Amazon S3 provides a RESTful API and
SDKs for easy data access from applications. It
integrates well with other AWS services like AWS
Lambda, Amazon EMR, Amazon Athena, and
Amazon Redshift for data processing and analytics.
Cost Structure: Amazon RDS pricing is based on
the database instance type, allocated storage,
provisioned IOPS, backups, and data transfer. It
offers multiple pricing options similar to EC2
instances. On the contrary, Amazon S3 pricing is
based on the amount of storage used, data transfer,
and requests made. It offers different storage tiers
with varying costs based on data access patterns
and availability requirements.

In summary, Amazon RDS is a managed relational


database service for structured data, while Amazon
S3 is an object storage service for storing
unstructured data. RDS focuses on database
management, performance, and structured
querying, whereas S3 provides simple storage
management, scalability, and versatility for storing
large volumes of objects.

Amazon Dynamo database


Amazon DynamoDB is a serverless, NoSQL, fully
managed database with single-digit millisecond
performance at any scale.
DynamoDB addresses your needs to overcome
scaling and operational complexities of relational
databases. DynamoDB is purpose-built and
optimized for operational workloads that require
consistent performance at any scale. For example,
DynamoDB delivers consistent single-digit
millisecond performance for a shopping cart use
case, whether you’ve 10 or 100 million users.
Customers across all sizes, industries, and
geographies use DynamoDB to build modern,
serverless applications that can start small and
scale globally. DynamoDB scales to support tables
of virtually any size while providing consistent
single-digit millisecond performance and high
availability.
Characteristics
Server less
With DynamoDB, you don’t need to provision any
servers, or patch, manage, install, maintain, or
operate any software. DynamoDB provides zero
downtime maintenance. It has no versions (major,
minor, or patch), and there are no maintenance
windows.

NoSQL
As a NoSQL database, DynamoDB is purpose-built
to deliver improved performance, scalability,
manageability, and flexibility compared to
traditional relational databases. To support a wide
variety of use cases, DynamoDB supports both key-
value and document data models.

Fully managed
As a fully managed database service, DynamoDB
handles the undifferentiated heavy lifting of
managing a database so that you can focus on
building value for your customers. It handles setup,
configurations, maintenance, high availability,
hardware provisioning, security, backups,
monitoring, and more. This ensures that when you
create a DynamoDB table, it’s instantly ready for
production workloads. DynamoDB constantly
improves its availability, reliability, performance,
security, and functionality without requiring
upgrades or downtime.

Single-digit millisecond performance at any scale


DynamoDB was purpose-built to improve upon the
performance and scalability of relational databases
to deliver single-digit millisecond performance at
any scale. To achieve this scale and performance,
DynamoDB is optimized for high-performance
workloads and provides APIs that encourage
efficient database usage. It omits features that are
inefficient and non-performing at scale, for example,
JOIN operations. DynamoDB delivers consistent
single-digit millisecond performance for your
application, whether you’ve 100 or 100 million
users.

Amazon DynamoDB and Amazon S3


Amazon DynamoDB supports exporting table data to
Amazon S3 using the Export to S3 feature. You can
export data in DynamoDB JSON and Amazon Ion
formats. Exported data is compressed and can be
encrypted by using an Amazon S3 key or an AWS Key
Management Service (AWS KMS) key. Exporting a
table does not consume read capacity on the table,
and it has no impact on table performance and
availability during the export. You can export to an
S3 bucket within the account or to a different
account, even in a different AWS Region. Point-in-
time recovery (PITR) should be activated on the
source table before you perform an export to
Amazon S3.

Amazon DynamoDB recently added support to


import table data directly from Amazon S3 by using
the Import from S3 feature. Previously, after you
exported table data using Export to S3, you had to
rely on extract, transform, and load (ETL) tools to
parse the table data in the S3 bucket, infer the
schema, and load or copy to the target DynamoDB
table. This was a cumbersome process and didn’t
provide flexibility when table data structure changed
over time. Also, the use of ETL tools such as AWS
Glue incurred additional charges for infrastructure
and for write capacity consumed during the import.

The Import from S3 feature doesn’t consume write


capacity on the target table, and it supports different
data formats, including DynamoDB JSON, Amazon
Ion, and comma-separated values (CSV). Data can
also be in uncompressed or compressed (gzip or
zstd) format.

You might also like