Cloud Computing Fundamentals Didgital Notes
Cloud Computing Fundamentals Didgital Notes
(R20A0554)
LECTURE NOTES
B. TECH
III YEAR – II SEM
(2022-23)
2023-2024
2023-2024
UNIT- II
Virtualization: Virtual Machines and Virtualization of Clusters and Data Centers- Implementation
Levels of Virtualization -Virtualization Structures/Tools and Mechanisms-Virtualization of CPU, Memory,
and I/O Devices-Virtual Clusters and Data-Centers
UNIT- III
Foundations: Introduction to Cloud Computing- Migrating into a Cloud-The Enterprise Cloud Computing
Paradigm.
UNIT- IV
Infrastructure as a Service (IAAS)& Platform (PAAS): Virtual machines provisioning and Migration
services-On the Management of Virtual machines for Cloud Infrastructures-Aneka—Integration of
Private and Public Clouds
UNIT- V
Software as a Service ( SAAS) &Data Security in the Cloud: Google App Engine – An Introduction to the
idea of Data Security- The Current State of Data Security in the Cloud- Cloud Computing and Data
Security Risk- Cloud Computing and Identity
TEXT BOOKS:
1. Distributed and Cloud Computing, Kaittwang Geoffrey C.Fox and Jack J Dongrra, Elsevier India 2012.
2.Mastering Cloud Computing- Raj Kumar Buyya, Christian Vecchiola and S.TanuraiSelvi, TMH, 2012.
3. Michael Miller, Cloud Computing: Web-Based Applications That Change the Way You Work and
Collaborate Online, Que Publishing, August 2008.
Reference Books:
1. Cloud Computing : A Practical Approach, Anthony T.Velte, Toby J.Velte, Robert Elsenpeter, Tata
McGraw Hill, rp2011.
2. Enterprise Cloud Computing, Gautam Shroff, Cambridge University Press, 2010.
3. Cloud Computing: Implementation, Management and Security, John W.Rittinghouse, James
F.Ransome, CRC Press, rp2012.
4. Cloud Application Architectures: Building Applications and Infrastructure in the Cloud, George
Reese, O’reilly, SPD, rp2011.
5. Cloud Security and Privacy: An Enterprise Perspective on Risks and Compliance, Tim Mather, Subra
Kumaraswamy, Shahed Latif, O’Reilly, SPD, rp2011.
Outcomes:
To distinguish the different models and computing paradigms.
To explain the levels of virtualization and resources virtulaization
To analyze the reasons for migrating into cloud
To effectively use the cloud services in terms of infrastructure and operating platforms.
To apply the services in the cloud for real world scenarios
INDEX
UNIT Topic Page No
Scalable computing over the Internet 1
System Models for Distributed and Cloud Computing 6
I Software Environments for distributed Systems and Clouds 11
Performance, Security & Energy Efficiency 13
Implementation Levels of Virtualization 18
Virtualization Structures/Tools and Mechanisms 21
High-Throughput Computing-HTC
HTC paradigm pays more attention to high-flux computing. The main application for high-flux
computing is in Internet searches and web services by millions or more users simultaneously. The
performance measures high throughput or the number of tasks completed per unit of time. HTC
technology needs to improve batch processing speed, and also address the acute problems of cost,
energy savings, security, and reliability at many data and enterprise computing centers
• Parallel computing
In parallel computing, all processors are either tightly coupled with centralized shared
memory or loosely coupled with distributed memory
Inter Process communication is accomplished through shared memory or via message
passing.
A computer system capable of parallel computing is commonly known as a parallel computer
Programs running in a parallel computer are called parallel programs. The process of writing
parallel programs is often referred to as parallel programming
• Distributed computing
A distributed system consists of multiple autonomous computers, each having its own
private memory, communicating through a computer network.
Cloud computing
An Internet cloud of resources can be either a centralized or a distributed computing system.
The cloud applies parallel or distributed computing, or both.
Clouds can be built with physical or virtualized resources over large data centers that are
centralized or distributed.
Cloud computing can also be a form of utility computing or service computing
Degrees of Parallelism
Bit-level parallelism (BLP) :
o converts bit-serial processing toword-level processing gradually.
Instruction-levelparallelism (ILP)
o the processor executes multiple instructions simultaneously rather thanonly one instruction
at a time.
o ILP is executed through pipelining, superscalarcomputing, VLIW (very long instruction
word) architectures, and multithreading.
o ILP requiresbranch prediction, dynamic scheduling, speculation, and compiler support to
work efficiently.
Data-level parallelism (DLP)
o DLP through SIMD (single instruction, multipledata) and vector machines using vector or
array types of instructions.
o DLP requires even more hardware support and compiler assistance to work properly.
Task-level parallelism (TLP):
o Ever since the introduction of multicoreprocessors and chip multiprocessors (CMPs), we
have been exploring TLP
o TLP is far from being very successful due to difficulty in programming and compilation of
Utility Computing
o Utility computing focuses on a business model in which customers receive computing
resources from a paid service provider. All grid/cloud platforms are regarded as utility
service providers.
Cyber-Physical Systems
o A cyber-physical system (CPS) is the result of interaction between computational processes
and the physical world.
o CPS integrates “cyber” (heterogeneous, asynchronous) with “physical” (concurrent and
information-dense) objects
o CPS merges the “3C” technologies of computation, communication, and control into an
intelligent closed feedback system
IoT emphasizes various networking connections among physical objects, while the CPS emphasizes
exploration of virtual reality (VR) applications in the physical world
Memory, Storage, and Wide-Area Networking: Memory chips have experienced a 4x increase in capacity
every three years. For hard drives, capacity increased from 260 MB in 1981 to 250 GB in 2004. Disks or disk
arrays have exceeded 3 TB in capacity. The rapid growth of flash memory and solid-state drives (SSDs) also
impacts the future of HPC and HTC systems.
System-Area Interconnects: The nodes in small clusters are mostly interconnected by an Ethernet switch or
a local area network(LAN).
As Figure shows, a LAN typically is used to connect client hosts to big servers.A storage area network (SAN)
connects servers to network storage such as disk arrays. Network attached storage (NAS) connects client hosts
directly to the disk arrays. All three types of networks often appear in a large cluster built with commercial
network components.
Wide-Area Networking: High-bandwidth networking increases the capability of building massively
distributed systems. The rapid growth of Ethernet bandwidth from 10 Mbps in1979 to 1 Gbps in 1999, and 40
~ 100 GE in 2011. It has been speculated that 1 Tbps network links will become available by 2013.
Virtual machines (VMs) offer novel solutions to underutilized resources, application inflexibility, software
manageability, and security concerns in existing physical machines. Today, to build large clusters, grids, and
clouds, we need to access large amounts of computing, storage, and networking resources in a virtualized
manner. We need to aggregate those resources, and hopefully, offer a single system image. In particular, a cloud
of provisioned resources must rely on virtualization of processors, memory, and I/O facilities dynamically.
The host machine is equipped with the physical hardware. The VM is built with virtual resources managed by
a guest OS to run a specific application. Between the VMs and the host platform, one needs to deploy a
middleware layer called a virtual machine monitor (VMM).
Figure shows a native VM installed with the use of a VMM called a hypervisor in privileged Mode.
The guest OS could be a Linux system and the hypervisor is the XEN system developed at
Cambridge University. This hypervisor approach is also called bare-metal VM, because the hypervisor handles
the bare hardware (CPU, memory, and I/O) directly. Architecture is the host VM shown in Figure(c). Here the
VMM runs in non-privileged mode. The host OS need not be modified. The VM can also be implemented with
a dual mode, as shown in Figure 1.12(d). Part of the VMM runs at the user level and another part runs at the
supervisor level. In this case, the host OS may have to be modified to some extent. Multiple VMs can be ported
to a given hardware system to support the virtualization process. The VM approach offers hardware
independence of the OS and applications.
These VM operations enable a VM to be provisioned to any available hardware platform. They also enable
flexibility in porting distributed application executions. Furthermore, the VM approach will significantly
enhance the utilization of server resources.
computer nodes. These node machines are interconnected by SANs, LANs, or WANs
o A massive system is with millions of computers connected to edge networks.
o Massive systems are considered highly scalable
o massive systems are classified into four groups: clusters, P2P networks, computing grids, and
Internet clouds
Cloud Computing Fundamentals Page 6
Computing cluster
o A computing cluster consists of interconnected stand-alone computers which work
cooperatively as a single integrated computing resource.
Cluster Architecture
o the architecture consists of a typical server cluster built around a low-latency, high bandwidth
interconnection network.
o build a larger cluster with more nodes, the interconnection network can be built with multiple
levels of Gigabit Ethernet, Myrinet, or InfiniBand switches.
o Through hierarchical construction using a SAN, LAN, or WAN, one can build scalable
clusters with an increasing number of nodes
o cluster is connected to the Internet via a virtual private network (VPN) gateway.
o gateway IP address locates the cluster
Grid Computing
A web service such as HTTP enables remote access of remote web pages
computing grid offers an infrastructure that couples computers, software/middleware, special
instruments, and people and sensors together
Enterprises or organizations present grids as integrated computing resources. They can also
beviewed as virtual platforms to support virtual organizations.
The computers used in a grid are primarilyworkstations, servers, clusters, and
supercomputers
Peer-to-Peer Network-P2P
The user can deploy and run on multiple VMs running guest OSes on specific applications.
The user does not manage or control the underlying cloud infrastructure, but can specify
when to request and release the needed resources.
Performance Metrics:
In a distributed system, performance is attributed to a large number of factors.
System throughput is often measured in MIPS, Flops (tera floating-point operationsper
second), or TPS (transactions per second).
System overhead is often attributed to OS boot time, compile time, I/O data rate, and the
runtime support system used.
Other performance-related metrics include the QoS for Internet and web services; system
availability and dependability; and security resilience for system defense against network
attacks
Dimensions of Scalability
Any resource upgrade Ina system should be backward compatible with existing hardware and
Size scalability
This refers to achieving higher performance or more functionality by increasing the machine
size.
The word “size” refers to adding processors, cache, memory, storage, or I/O Channels. The
most obvious way to determine size scalability is to simply count the number of processors
installed.
Not all parallel computer or distributed architectures are equally size scalable.
For example, the IBM S2 was scaled up to 512 processors in 1997. But in 2008,
theIBMBlueGene/L system scaled up to 65,000 processors.
• Software scalability
This refers to upgrades in the OS or compilers, adding mathematical and engineering
libraries, porting new application software, and installing more user- friendly
programming environments.
Some software upgrades may not work with large system configurations.
Testing and fine-tuning of new software on larger systems is a nontrivial job.
• Application scalability
This refers to matching problem size scalability with machine size scalability.
Problem size affects the size of the data set or the workload increase. Instead of increasing
machine size, users can enlarge the problem size to enhance system efficiency or cost-
effectiveness.
• Technology scalability
This refers to a system that can adapt to changes in building technologies, such as the
component and networking technologies
When scaling a system design with new technology one must consider three aspects: time,
space, and heterogeneity.
(1) Time refers to generation scalability. When changing to new-generation processors,
one must consider the impact to the motherboard, power supply, packaging and cooling,
and so forth. Based on past experience, most systems upgrade their commodity processors
every three to five years.
Cloud Computing Fundamentals Page 14
(2) Space is related to packaging and energy concerns. Technology scalability demands
harmony and portability among suppliers.
Amdahl’s Law
Let the program has been parallelized or partitioned for parallel execution on a cluster
of many processing nodes.
Assume that a fraction α of the code must be executed sequentially, called the
sequential bottleneck.
The total execution time of the program is calculated byα T + (1 − α)T/n, where the first
term is the sequential execution time on a single processor and the second term is the
parallel execution time on n processing nodes.
I/O time or exception handling time is also not included in the following speedup analysis.
Amdahl’s Law states that the speedup factor of using the n-processor system over the
use of a single processor is expressed by:
the code is fully parallelizable with α = 0. As the cluster becomes sufficiently large, that
this upper bound is independent of the cluster size n. The sequential bottleneck is
the portion of the code that cannot be parallelized.
Gustafson’s Law
To achieve higher efficiency when using a large cluster, we must consider scaling the
problem sizeto match the cluster capability. This leads to the following speedup
proposed by John Gustafson(1988), referred as scaled-workload speedup.
Primary performance goals in conventional parallel and distributed computing systems are
highperformance and high throughput, considering some form of performance reliability (e.g.,
fault toleranceand security). However, these systems recently encountered new challenging
issues includingenergy efficiency, and workload and resource outsourcing
Energy Consumption of Unused Servers: To run a server farm (data center) a company has to
spend a huge amount of money for hardware,software, operational support, and energy every year.
Therefore, companies should thoroughlyidentify whether their installed server farm (more
specifically, the volume of provisioned resources)is at an appropriate level, particularly in terms
of utilization.
Network Layer: Routing and transferring packets and enabling network services to the resource
layer are the mainresponsibility of the network layer in distributed computing systems. The major
challenge to buildenergy-efficient networks is, again, determining how to measure, predict, and
create a balancebetween energy consumption and performance.
function of the software layer for virtualization is to virtualize the physical hardware of a
host machine into virtual resources to be used by the VMs
Common virtualization layers include the instruction set architecture (ISA) level, hardware
level, operating system level, library support level, and application level
Instruction Set Architecture Level
At the ISA level, virtualization is performed by emulating a given ISA by the ISA of the host
machine. For example, MIPS binary code can run on an x86-based host machine with the help
of ISA emulation. With this approach, it is possible to run a large amount of legacy binary code
written for various processors on any given new hardware host machine.
Xen Architecture
Xen is an open source hypervisor program developed by Cambridge University.
Xen is a microkernel hypervisor
The core components of a Xen system are the hypervisor, kernel, and applications
The guest OS, which has control ability, is called Domain 0, and the others are called
Domain U
Domain 0 is designed to access hardware directly and manage devices
Memory Virtualization
• Memory Virtualization :the operating system maintains mappings of virtual memory to
machine memory using page table
• All modern x86 CPUs include a memory management unit (MMU) and a translation
lookaside buffer (TLB) to optimize virtual memory performance
• Two-stage mapping process should be maintained by the guest OS and the VMM,
Cloud Computing Fundamentals Page 23
respectively: virtual memory to physical memory and physical memory to machine
memory.
• The VMM is responsible for mapping the guest physical memory to the actual machine
memory.
I/O Virtualization
• I/O Virtualization managing the routing of I/O requests between virtual devices and the
shared physical hardware
• managing the routing of I/O requests between virtual devices and the shared physical
hardware
• Full device emulation emulates well-known, real-world devices All the functions of a
device or bus infrastructure, such as device enumeration, identification, interrupts, and
DMA, are replicated in software. This software is located in the VMM and acts as a virtual
device
• Two-stage mapping process should be maintained by the guest OS and the VMM,
respectively: virtual memory to physical memory and physical memory to machine
memory.
• The VMM is responsible for mapping the guest physical memory to the actual machine
memory.
Virtualization in Multi-Core Processors
• Muti-core virtualization has raised some new challenges
• Two difficulties: Application programs must be parallelized to use all cores fully, and
software must explicitly
• Assign tasks to the cores, which is a very complex problem
Cloud Computing Fundamentals Page 24
• The first challenge, new programming models, languages, and libraries are needed to
make parallel programming easier.
• The second challenge has spawned research involving scheduling algorithms and
resource management policies
• Dynamic heterogeneity is emerging to mix the fat CPU core and thin GPU cores on the
same chip
• The latest virtualization development highlights high availability (HA), backup services,
workload balancing, and further increases in client bases.
Server Consolidation in Data Centers
Deployment Models
• Public cloud as a “cloud made available in a pay-as-you-go manner to the general public”
and
• Private cloud as “internal data center of a business or other organization, not made available
to the general public.”
• A community cloud is “shared by several organizations and supports a specific community
that has shared concerns
• A hybrid cloud takes shape when a private cloud is supplemented with computing capacity
from public clouds.
• The approach of temporarily renting capacity to handle spikes in load is known as “cloud-
bursting”
• are essential to enable services that truly represent the cloud computing model
• Self-Service : clouds must allow self-service access so that customers can request,
customize, pay, and use services (expect on-demand, nearly instant access to resources)
without intervention of human operators
• Per-Usage Metering and Billing : Services must be priced on a shortterm basis (e.g., by
the hour), allowing users to release (and not pay for) resources as soon as they are not
needed
• Elasticity : users expect clouds to rapidly provide resources in any quantity at any time. In
particular, it is expected that the additional resources can be
• (a) provisioned, possibly automatically, when an application load increases and
Why Migrate?
• There are economic and business reasons why an enterprise application can be migrated
into the cloud, and there are also a number of technological reasons.
• Initiatives in adoption of cloud technologies in the enterprise,resulting in integration of
enterprise applications running off the captive data centers withthe new ones that have
been developed on the cloud.
Migration can happen at one of the five levels of
application,
code,
design,
architecture,
usage
The migration of an enterprise application is best captured by the following
where
P is the application before migration running in captive data center,
P’C is the application part after migration either into a (hybrid) cloud,
P’l is the part of application being run in the captive local data center, and
Iterative Step
• The biggest challenge to any cloud migration project is how effectively the migration
risks are identified and mitigated.
• Migration risks for migrating into the cloud fall under two broad categories:
– the general migration risks
– the security-related migration risks
• several issues identifying all possible production level deviants:
– the business continuity and disaster recovery in the world of cloud computing
service;
– the compliance with standards and governance issues; the IP and licensing issues;
– the quality of service (QoS) parameters as well as the corresponding SLAs
committed to;
– the ownership, transfer, and storage of data in the application;
– the portability and interoperability issues which could help mitigate potential
vendor lock-ins;
Challenges in the Cloud
• Security
• Costing model
• Charging model
• Service level agreement
• Cloud interoperability issue
• It is still unclear how safe out-sourced data is and when using these services
ownership of data is not always clear.
– There are also issues relating to policy and access:
• There have been cases of users being locked out of accounts and losing access to data.
• on-demand self-service
• resource pooling,
• rapid elasticity
• measured service
The ways in which these characteristics are manifested in an enterprise context vary according
to thedeployment model employed.
• Public clouds are provided by a designated service provider for general public under a utility-
based pay-per-use consumption model. The cloud resources are hosted generally on the service
provider’spremises
• Private clouds are built, operated, and managed by an organization for its internal use only to
support its business operations exclusively
• Virtual private clouds are a derivative of the private cloud deployment model but are further
characterized by an isolated and secure segment of resources, created as an overlay on top of public
cloud infrastructure using advanced network virtualization capabilities. Some of the public
cloud vendors that offer this capability include Amazon Virtual PrivateCloud, OpSource Cloud and Skytap
Virtual Lab
• Community clouds are shared by several organizations and support a specific community that has
shared concerns (e.g., mission, security requirements, policy, and compliance considerations).
They may be managed by the organizations or a third party and may exist on premise or off premise .
Oneexample of this is OpenCirrus formed by HP, Intel, Yahoo, and others
• Managed clouds arise when the physical infrastructure is owned by and/or physically located in
the organization’s data centers with an extension of management and security control plane
controlled by the managed service provider
• Hybrid clouds are a composition of two or more clouds (private, community,or public) that
remain unique entities but are bound together by standardized or proprietary technology that
enables data and
application portability. Some examples of these offerings include Amazon Virtual Private
Cloud, Skytap Virtual Lab, and CohesiveF T VPN-Cubed.
The selection of strategies for enterprise cloud computing is critical for IT capability as well as
for the earnings and costs the organization experiences, motivating efforts toward convergence of
business strategies and IT. Critical questions toward this convergence in the enterprise cloud paradigm
• Which areas of business and IT capability should be considered for the enterprise cloud?
• How can the process of transitioning to an enterprise cloud strategy be piloted and
systematically executed?
Adoption strategy : an organization makes a decision to adopt a cloud computing model based
on fundamental drivers for cloud computing—scalability, availability, cost and convenience
Scalability - Driven Strategy: The objective is to support increasing workloads of the organization
without investment and expenses exceeding returns. The conditions are that the effort, costs (CAPEX
and OPEX) and time involved in accessing and installing IT capability on a CDC are less than going
through a standard hardware and software procurement and licensing process
• Availability - Driven Strategy: Availability has close relations to scalability but is more concerned
with the assurance that IT capabilities and functions are accessible, usable and acceptable by the
standards of users. This is hence the objective of this basic enterprise cloud strategy.
• Market-Driven Strategy: This strategy is more attractive and viable for small, agile organizations that
do not have (or wish to have) massive investments in their IT infrastructure. The objective here is to
identify and acquire the “best deals” for IT capabilities as demand and supply change, enabling
ongoing reductions in OPEX and CAPEX.
• Convenience-Driven Strategy: The objective is to reduce the load and need for dedicated system
administrators and to make access to IT capabilities by users easier, regardless of their location and
connectivity (e.g. over the Internet). The expectation is that the cost of obtaining IT capabilities from
aCDC and making them accessible to users is significantly lower than the cost of having a dedicated
administrator
Consumption Strategy:
The consumption strategies make a distinction between data and application logic because there
are questions of programming models used, data sensitivity, software licensing and expected response
times that need to be considered.
Cloud Computing Fundamentals Page 36
There are four consumptions strategies identified, where the differences in objectives, conditions
and actionsreflect the decision of an organization to trade-off hosting costs, controllability and resource
elasticity of IT resources for software and data
• Software Provision. This strategy is relevant when the elasticity requirement is high for software and
low for data, the controllability concerns are low for software and high for data, and the cost reduction
concerns for software are high, while cost reduction is not a priority for data, given the high
controllability concerns for data, that is, data are highly sensitive.
• Storage Provision. This strategy is relevant when the elasticity requirements is high for data and low
for software, while the controllability of software is more critical than for data. This can be the case
for data intensive applications, where the results from processing in the application are more critical and
sensitive than the data itself. Furthermore, the cost reduction for data resources is a high concern, here as cost
for software, given its criticality, is not an issue for the organization within reasonable means.
• Solution Provision. This strategy is relevant when the elasticity and cost reduction requirements are
high for software and data, but the controllability requirements can be entrusted to the CDC. It is not
the case that controllability is an insignificant requirement; it is rather the case that the organization
trusts the CDC sufficiently to manage access and usage control of its software and data
• Redundancy Services. This strategy can be considered as a hybrid enterprise cloud strategy, where
the organization switches between traditional, software, storage or solution management based on
changes in its operational conditions and business demands
The strategy is referred to as the “redundancy strategy” because the CDC is used for situations
such as disaster recovery, fail-over and load balancing Software, storage or solution services can be
implemented using redundancy, such that users are redirected for the purpose of maintaining
availability of functionality or performance/response times experienced by the user of the service.
(iii) specialized components and services that aid particular applications (e.g.,loadbalancers,)
Ideally, a public IaaS provider must provide multiple access means to its cloud, thus
catering for various usersand their preferences.
Different types of user interfaces (UI) provide different levels of abstraction, the most
common being
• graphical user interfaces (GUI),
• Amazon EC2 states that “if the annual uptime Percentage for a customer drops
below 99.95% for the service year, that customer is eligible to receive a service
credit equal to 10% of their bill.3”
• the user pays only for the capacity of the provisioned resources at a particular time
Amazon Elastic Compute Cloud (EC2) is an IaaS service that provides elastic compute capacity
in the cloud
Private Cloud and Infrastructure Services
• A private cloud aims at providing public cloud functionality, but on private resources,
while maintaining control over an organization’s data and resources to meet security and
governance’s requirements in an organization.
• Private clouds exhibit the following characteristics:
• delivers broadly supported interoperability and portability standards for managing the
virtual computing lifecycle.
• Typical life cycle of VM and its major possible states of operation, which make the
management and automation of VMs in virtual and cloud environments easier Process. Here,
we describe the common and normal steps of provisioning a virtual server:
• Firstly, you need to select a server from a pool of available servers (physical servers with
enough capacity) along with the appropriate OS template you need to provision the virtual
machine.
• Secondly, you need to load the appropriate software (operating system you selected in the
previous step, device drivers, middleware, and the needed applications for the service
required).
• Thirdly, you need to customize and configure the machine (e.g., IP address, Gateway) to
configure an associated network and storage resources.
• Finally, the virtual server is ready to start with its newly loaded software
Migration service,
• in the context of virtual machines, is the process of moving a virtual machine from one
host server or storage location to another
• There are different techniques of VM migration,
• hot/life migration,
• to have the ability to migrate virtual machines from one platform to another
• For example, the VMware converter that handles migrations between ESX hosts;
• The VMware converter can also import from other virtualization platforms, such as
Microsoft virtual server machines
Deployment Scenario:
• whereConVirt is installed and ran, which provides the main console for managing the VM
life cycle, managing images, provisioning new VMs, monitoring machine resources, and
so on.
• There are two essential deployment scenarios for ConVirt:
• A basic configuration in which the Xen or KVM virtualization platform is on the local
machine, where ConVirt is already installed; B,
• An advanced configuration in which the Xen or KVM is on one or more remote servers.
• Once the installation is done and you are ready to manage your virtual infrastructure, then
you can start the ConVirt management console :
• Select any of servers’ pools existing (QA Lab in our scenario) and on its context menu,
select “Add Server.”
• Clark et al. did evaluate the above migration on an Apache 1.3 Web server; this served
static content at a high rate, as illustrated in Figure 5.6.
• The throughput is achieved when continuously serving a single 512-kB file to a set of one
hundred concurrent clients.
• This simple example demonstrates that a highly loaded server can be migrated with both
controlled impact on live services and a short downtime
VMware Vmotion.
• This allows users to (a) automatically optimize and allocate an entire pool of
resources for maximumhardware utilization, flexibility, and availability and
(b) perform hardware’s maintenance without scheduled downtime along with migrating
virtual machines away from failing or underperforming servers
Citrix XenServerXenMotion.
• This is a nice feature of the Citrix XenServer product, inherited from the Xen live migrate
utility, which provides the IT administrator with the facility to move a running VM from
one XenServer to another in the same pool without interrupting the service
Regular/Cold Migration.
• Main differences between live migration and cold migration are that
• 1) live migration needs a shared storage for virtual machines in the server’s pool, but
cold migration does not;
• 2) live migration for a virtual machine between two hosts, there would be certain CPU
compatibility checks to be applied; while in cold migration this checks do not apply
• The cold migration process (VMware ) can be summarized as follows:
Cloud Computing Fundamentals Page 45
• The configuration files, including the NVRAM file (BIOS settings), log files, as
well as the disks of the virtual machine, are moved from the source host to the
destination host’s associated storage area.
• The virtual machine is registered with the new host.
• After the migration is completed, the old version of the virtual machine is deleted
from the source host.
This kind of migration constitutes moving the virtual disks or configuration file of a
running virtual machine to a new data store without any interruption in the availability ofthe
virtual machine’s service
Aneka
• Manjrasoft Aneka is a .NET-based platform and framework designed for building and
deploying distributed applications on clouds.
• It provides a set of APIs for transparently exploiting distributed resources and expressing
the business logic of applications by using the preferred programming abstractions.
• Aneka also provides support for deploying and managing clouds.
• By using its Management Studio and a set ofWeb interfaces, it is possible to set up either
public or private clouds, monitor their status, update their configuration, and perform the
basic management operations.
SAAS
• Cloud-centric integration solutions are being developed and demonstrated for showcasing
their capabilities for integrating enterprise and cloud applications.
• Composition and collaboration will become critical and crucial for the mass adoption of
clouds
Jitterbit:
Jitterbit is a fully graphical integration solution that provides users a versatile platform suite
of productivity tools to reduce the integration efforts sharply. Jitterbit can be used standalone or with
existing EAI infrastructures, Help us quickly design, implement, test, deploy, and manage the
The app engine supports the delivery, testing and development of software on demand in
a Cloud computing environment that supports millions of users and is highly scalable.
The company extends its platform and infrastructure to the Cloud through its app engine.
It presents the platform to those who want to develop SaaS solutions at competitive costs.
Google is a leader in web-based applications and also offers cloud development services.
These services come in the form of the Google App Engine, which enables developers to
build their own web applications utilizing the same infrastructure that powers Google’s
powerful applications.
These are covered by the depreciation policy and the service-level agreement of the app engine.
Any changes made to such a feature are backward-compatible and implementation of such a
feature is usually stable. These include data storage, retrieval, and search; communications;
process management; computation; app configuration and management.
Data storage, retrieval, and search include features such as HRD migration tool, Google Cloud
SQL, logs, datastore, dedicated Memcache, blobstore, Memcache and search.
Communications include features such as XMPP. channel, URL fetch, mail, and Google Cloud
Endpoints.
Process management includes features like scheduled tasks and task queue Computation includes images.
App management and configuration cover app identity, users, capabilities, traffic splitting,
modules, SSL for custom domains, modules, remote access, and multitenancy
To check your home email from work, it took a bit of juggling and perhaps the use of your ISP’s
email access web page. That web page was never in sync with the messages on your home PC, of
course, which is just the start of the problems with trying to communicate in this fashion.
A better approach is to use a web-based email service, such as Google’s Gmail (mail.google.com),
Microsoft’s Windows Live Hotmail (mail.live.com), or Yahoo! Mail (mail.yahoo.com). These
services place your email inbox in the cloud; you can access it from any computer connected to the
Internet.
Collaborating via Web-Based Communication Tools
GMAIL
Gmail offers a few unique features that set it apart from the web-based email crowd.
First, Gmail doesn’t use folders. With Gmail you can’t organize your mail into folders, as
Cloud Computing Fundamentals Page 51
you can with the other services.
Instead, Gmail pushes the search paradigm as the way to find the messages you want— not
a surprise, given Google’s search-centric business model.
Gmail does, however, let you “tag” each message with one or more labels. This has the
effect of creating virtual folders, as you can search and sort your messages by any of their
labels.
In addition, Gmail groups together related email messages in what Google calls
conversations
Yahoo! Mail Yahoo! Mail (mail.yahoo.com)
is another web mail service, provided by the popular Yahoo! search site.
The basic Yahoo! Mail is free and can be accessed from any PC, using any web
browser.
Yahoo! also offers a paid service called Yahoo! Mail Plus that lets you send larger
messages and offers offline access to your messages via POP email clients
BigString (www.bigstring.com) E
FlashMail (www.flashmail.com)
Inbox.com (www.inbox.com)
Mail.com (www.mail.com)
Data Security
Information in a cloud environment has much more dynamism and fluidity than
information that is static on a desktop or in a network folder
Cloud Computing Fundamentals Page 52
Nature of cloud computing dictates that data are fluid objects, accessible froma multitude
of nodes and geographic locations and, as such, must have a datasecurity methodology that
takes this into account while ensuring that this fluidity is not compromised
The idea of content-centric or information-centric protection, being an inherent part of a
data object is a development out of the idea of the “de-perimerization” of the enterprise.
This idea was put forward by a group of Chief Information Officers (CIOs) who formed
an organization called the Jericho Forum
Cloud data security is the combination of technology solutions, policies, and procedures
that you implement to protect cloud-based applications and systems, along with the
associated data and user access.
When it comes to data, the cloud poses a variety of risks that need to be addressed as part
of security strategy. The biggest risks—as increasingly rely on the cloud for collecting,
storing, and processing critical data—are cyberattacks and data breaches.
5. financial issues
Current state of data security in Cloud:
Data privacy and security are placed under a lot of scrutiny these days thanks to the
numerous information leaks that became public knowledge in recent years. The independent study
from Sophos of 3,521 IT managers leveraging the public cloud across 26 countries and six
continents reveals many insights into the world of cloud security, with some of the key takeaways
being:
“[70%] of organizations reported they were hit by malware, ransomware, data theft,
account compromise attempts, or cryptojacking in the last year.”
Data loss/leakage is the topmost concern with 44% of organizations reporting “data loss as
Europe’s General Data Protection Regulation (GDPR) may be partly responsible for
European organizations seeing the “lowest attack rates of all regions.”
75% of organizations do not see staff expertise as a top priority despite the prevalence of
cyberattacks.
66% of attacks were exploitations of misconfigurations and 33% used stolen credentials to
gain access.
Digital identity
A digital identity represents who we are and how we interact with others on-line.
Access, identity, and risk are three variables that can become inherently connected when
applied to the security of data, because access and risk are directly proportional: As
access increases, so then risk to the security of the data increases.
Access controlled by identifying the actor attempting the access is the most logical
manner of performing this operation.
Ultimately, digital identity holds the key to securing data, if that digital identity can be
programmatically linked to security policies controlling the post-access usage of data.
Identity, Reputation, and Trust
Reputation is a real-world commodity; that is a basic requirement of human-to-human
relationships
Our basic societal communication structure is built upon the idea of reputation and trust.
Reputation and its counter value, trust, is easily transferable to a digital realm:
o eBay, for example, having partly built a successful business model on the strength
of a ratings system, builds up the reputation of its buyers and sellers through
successful (or unsuccessful) transactions.
Digital identities are a mechanism for identifying an individual, particularly within a cloud
environment ; identity ownership being placed upon the individual is known as user-
centric identity
It allows users to consent and control how their identity (and the individual identifiers
making up the identity, the claims) is used.
This reversal of ownership away from centrally managed identity platforms(enterprise-
centric) has many advantages.
This includes the potential to improve the privacy aspects of a digital identity, by giving an
individual the ability to apply permission policies based on their identity and to control
which aspects of that identity are divulged
An identity may be controllable by the end user, to the extent that the user can then
decide what information is given to the party relying on the identity
Information Card:
Information cards permit a user to present to a Web site or other service (relying party) one
or more claims, in the form of a software token, which may be used to uniquely identify
that user.
They can be used in place of user name/ passwords, digital certificates, and other
identification systems, when user identity needs to be established to control access to a
Web site or other resource, or to permit digital signing
Information cards are part of an identity meta-system consisting of:
1. Identity providers (IdP), who provision and manage information cards,with specific
claims, to users.
2. Users who own and utilize the cards to gain access to Web sites and other resources
that support information cards.
3. An identity selector/service, which is a piece of software on the user’s desktop or in
Cloud Computing Fundamentals Page 55
thecloud that allows a user to select and manage their cards.
4. Relying parties. These are the applications, services, and so on, that can use an
information card to authenticate a person and to then authorize an action such as loggingonto a
Web site, accessing a document, signing content, and so on
Each information card is associated with a set of claims which can be used to identify the user.
These claims include identifiers such as name, email address, post code
Using Information Cards to Protect Data
Information cards are built around a set of open standards devised by a consortium that
includes Microsoft, IBM, Novell, and so on.
The original remit of the cards was to create a type of single sign on system for the
Internet, to help users to move away from the need to remember multiple passwords.
However, the information card system can be used in many more ways.
Because an information card is a type of digital identity, it can be used in the same way
that other digital identities can be used.
For example, an information card can be used to digitally sign data and content and to control
access to data and content. One of the more sophisticated uses of an information card is the
advantage given to the cards by way of the claims system.
Data are uploaded into a cloud and stored in a data center, for access by users from that
data center; or in a more fully cloud-based model, the data themselves are created in the
cloud and stored and accessed from the cloud (again via a data center).
The most obvious risk in this scenario is that associated with the storage of that data. A
user uploading or creating cloud-based data include those data that are stored and
maintained by a third-party cloud provider such as Google, Amazon, Microsoft, and so on.
This action has several risks associated with it:
Firstly, it is necessary to protect the data during upload into the data center to ensure that
the data do not get hijacked on the way into the database.
• Secondly, it is necessary to the stores the data in the data center to ensure that they are
• Access control becomes a much more fundamental issue in cloud-based systems because
of the accessibility of the data
• Information-centric access control (as opposed to access control lists) can help to balance
improved accessibility with risk, by associating access rules with different data objects
within an open and accessible platform, without losing the Inherent usability of that
platform
• A further area of risk associated not only with cloud computing, but also with traditional
network computing, is the use of content after access.
• The risk is potentially higher in a cloud network, for the simple reason that the information
is outside of your corporate walls.
Data-centric mashups are those
• that are used to perform business processes around data creation and dissemination—by
their very nature, can be used to hijack data, leaking sensitive information and/or affecting
integrity of that data
• Cloud computing, more than any other form of digital communication technology, has
created a need to ensure that protection is applied at the inception of the information, in a
content centric manner, ensuring that a security policy becomes an integral part of that data
throughout its life cycle.
Encryption
• is a vital component of the protection policy, but further controls over the access of that
data and on the use of the data must be met.
• In the case of mashups, the controlling of access to data resources, can help toalleviate the
security concerns by ensuring that mashup access is authenticated.
• Linking security policies, as applied to the use of content, to the access control method