Cloud Computingseurity
Cloud Computingseurity
Computing (CC)
from the paper
Abovce the Clouds: A
Berkeley View of Cloud
Computing (Feb. 2009)
Definitions (I)
Cloud Computing refers to both
the applications delivered as
services over the Internet and
the scalable hardware and
systems software that provide
those services.
The services themselves have long
been referred to as Software as a
Service (SaaS).
Definitions (II)
The scalable hardware and
software is what we will call a
Cloud.
When a Cloud is made available to
the general public, we call it a Public
Cloud.
This often means that one can
access a cloud for free or in a pay-asyou-go manner.
The service being sold is Utility
Definitions (III)
We use the term Private Cloud to
refer to scalable hardware and
software of a business or other
organization, not made available
to the general public.
Characteristics of clouds
illusion of infinite computing
resources (because of scalability)
elimination of an up-front
commitment by Cloud users,
thereby allowing companies to start
small and increase hardware
resources only when there is an
increase in their needs
The ability to pay for use of
computing resources on a short-
Why CC now?
large Internet companies (amazon,
Google, etc.) had to run huge data
centers anyway as their core
business with excess capacities
=> additional revenue stream
that also reduces excess
capacities
by having to run huge data centers
they significantly improved the
technology to manage them
Spectrum of CC
computational model (VM):
Spectrum of CC
storage model:
Amazon Web Services
Microsoft Azure
Google AppEngine
Spectrum of CC
networking model:
Amazon Web Services
Microsoft Azure
Google AppEngine
1) availability
down-time of dominant cloud
providers is extremely low
but: cloud is a single point of failure;
using multiple clouds (redundancy) is
not an option today (incompatibility)
Distributed Denial of Service attacks:
target shifted to cloud provider
2) data lock-in
clouds are currently proprietary
problem if more specific clouds are
used
speculation: standard APIs not to be
expected in the near future (5 years)
4) data transfer
bottlenecks
10 TB:
10 1012 Bytes / (20 106
bits/second) = (8 1013)/(2 107)
seconds = 4,000,000 seconds, which
is more than 45 days
remedy: express mailing of disks
5) performance
unpredictibility
main memory: no problem
I/O: 16% variability
remedy: flash memory
high performance computing:
requires synchronized scheduling of
tasks, which today's VMs and
operating systems do not provide
remedy: "gang scheduling" for CC
6) scalable storage
problem: varying richness of query
and storage API, of performance
guarantees and the complexity of
data structures
research opportunity: create CC
storage system that overcomes
these limitations
7) debugging in massively
distributed systems
problem: bugs often cannot be
reproduced in smaller configurations
research opportunity: create
appropriate VMs/debugging concepts
and tools