0% found this document useful (0 votes)
151 views35 pages

Itm Group of Institution: Lab File OF Cloud Computing (Cs-8002)

The document is a lab file submitted by Vishal Jain to his professor for his Cloud Computing course. It includes 8 experiments on topics related to cloud computing such as an introduction to cloud computing, cloud security, performance evaluation of clouds, Hadoop introduction, cloud resource management, deployment models, and using Google Drive and Justcloud. The first experiment defines cloud computing, provides examples like Gmail, and discusses Google Drive and Microsoft SkyDrive as examples of cloud storage services.

Uploaded by

Vishal Jain
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
151 views35 pages

Itm Group of Institution: Lab File OF Cloud Computing (Cs-8002)

The document is a lab file submitted by Vishal Jain to his professor for his Cloud Computing course. It includes 8 experiments on topics related to cloud computing such as an introduction to cloud computing, cloud security, performance evaluation of clouds, Hadoop introduction, cloud resource management, deployment models, and using Google Drive and Justcloud. The first experiment defines cloud computing, provides examples like Gmail, and discusses Google Drive and Microsoft SkyDrive as examples of cloud storage services.

Uploaded by

Vishal Jain
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 35

ITM GROUP OF INSTITUTION

Lab File
OF
Cloud Computing(Cs-8002)

SUBMITTED TO:- SUBMITTED BY:-


Mr. C P Bhargava Vishal Jain
Assistant prof. (0905CS161183)
(CSE department) CS-3 SEM-8

Vishal Jain
Contents

1. Introduction of Cloud Computing

2. Cloud Computing Security

3. Performance Evaluation of Cloud

4. Introduction of Hadoop

5. Management of Cloud Resources

6. Deployment Model

7. Working of Goggle Drive to make spreadsheet and notes.

8. Installation and Configuration of Justcloud.

Vishal Jain
Experiment No. 1
Objective:Introduction of Cloud Computing.

Definition: -

VMware, the global leader in virtualization and cloud infrastructure


define cloud computing as –
“Cloud computing is a new approach that reduces IT complexity by
leveraging the efficient pooling on-demand, self-managed virtual
infrastructure, consumed as a service.”
According to Wikipedia-
“Cloud computing is Internet based computing, whereby shared
resources, software and information are provided to computers and
other devices on demand, like the electricity grid.”

Examples: -
Web-based email services like Gmail and Hotmail deliver a cloud
computing service: users can access their email "in the cloud" from any
computer with a browser and Internet connection, regardless of what
kind of hardware is on that particular computer. The emails are hosted
on Google's and Microsoft's servers, rather than being stored locally on
the client computer.

Over the last few years we've seen tremendous growth in cloud
computing, as witnessed by the many popular Web apps used today,
including VoIP (e.g., Skype, Google Voice), social applications (e.g.,
Facebook, Twitter, LinkedIn), media services (e.g., Picasa, YouTube,
Flickr), content distribution (e.g., Bit Torrent) etc.

1. Google: -

Google Drive (Google Cloud Computing)

This was knowing as Google Docs until recently when it was


changed to Google Drive. Google Drive is an example of cloud
computing where videos, photographs, files and data can be
stored and organized whether for an individual or a corporate
organization. Google offers up to 15 GB storage for users.
There is a periodic payment plan for additional storage as the

Vishal Jain
demand grows for clients. Google account is required to access
this facility. For easy access Google Drive, can be downloaded
into a PC or a similar gadget.

2.Microsoft: -
Microsoft Sky Drive (Microsoft Cloud Computing)

What we are accustomed to in the world of computing before this


dispensation was the “Hard Drive”. Now before us is the SkyDrive as
provided by Microsoft Incorporated. Microsoft SkyDrive is a cloud based
storage device for personal and corporate files, data, information etc.
which is accessible from anywhere through an internet connection.
Microsoft word, power point, excel and a great deal of other products of
Microsoft can be accessed for free.
SkyDrive application can be downloaded and installed in a PC if the
system supports it. Otherwise it has to be accessed directly in the
Clouds. In order to access SkyDrive, you need a SkyDrive.com account,
through which the cloud is accessed. Microsoft provide up to 7GB free
storage for documents, photos, files, data etc.
Users of windows 8 are at advantage as SkyDrive application is
integrated into the operating system and by implication can be accessed
easily when the system is online. This is one of the advantages of
windows 8.

3.Salesforce: -

Examples of spreadsheets with salesforce custom objects

If you aren’t using an online accounting software like QuickBooks online


and want to keep track of expense trends, you do not need each
employee to make a personal expense report and manually compile
them for management. Instead, create a tab in salesforce that has fields
for employee, expense reason, expense amount, expense date. Then
create reports for various expense categories and time trends. Add
dashboard report for management or schedule and automatic email alert
with expense summaries. Employees don’t even need to be logged into
salesforce to record their expense if you create a Form Assembly from
to update the custom object.

Vishal Jain
ADVANTAGES

1.Easy Access- Cloud storage is convenient and easy to access without


distance barrier or tied to a single gadget. It is more advantageous if a
cloud based application is installed in a computer or a compatible
device.

2.Safe guard against loss of data- There could be a situation when a


physical computer develops a fault that can lead to losing the entire
stored documents. It is also possible to misplace USB storage device or
even an external hard drive, but in the case of cloud storage such data
will be safe and accessed as long as there is internet connection.

3.Security- A great measure has been taken to secure cloud storage. If


all the measures are judiciously followed by users there could be no
room for mistakes.

4.Simplicity-easy to deploy and use.

5.Pay as you use-pay only for the service you use.

6.Scalability

7.Backup and recovery

8.Easy to update

9.On demand availability

CHARACTERISTICS

Cloud computing builds on so many older concepts in computer


technology that it can be hard for people newly introduced to the concept
of grasps that it represents a paradigm shift in computing. It’san
revolutionary change new approach to how computing services are
produced and consumed.

Vishal Jain
PARADIGM SHIFT

When you choose a cloud service provider, you are renting or leasing a
part of an enormous infrastructure of data centers, computers, storage
and networking capacity. Many of these data centers are multi-million-
dollar investment by the companies that run them, to give you some
sense of scale, it has been established that a state-of-the-art microchip
fabrication facility can cost anywhere from $2 to $5 billion. By
comparison a state-of-the-art cloud computing data center run in the
range of $100 million. Most of the large cloud computing service
providers have multiple data centers located all over the world, an
accurate count can be difficult to obtain but the location of some 20
datacenters in Amazon Web Service’s cloud are detailed. Google cloud
includes perhaps some 35-data center worldwide. In the 1960’s, military
initiatives aimed at miniaturizing electronic funded many of the
semiconductor production lines that led to advanced microprocessors,
dense memory arrays, and the sophisticated integrated circuit
technology that makes computers, mobile devices, and so much more
possible today. the 1990s, the commercialization of the Internet gave
rise to some very large companies that were forced to build very large
computing infrastructures to support their businesses.
Amazon.com infrastructure was built to support elastic demand so the
system can accommodate peak traffic on a busy shopping day such as
“Black Monday” because much of the capacity was idle, Amazon.com
first open its network to partners then as Amazon Web Service to
customers. Google business has also grown exponentially and required
the building of data centers worldwide. One of its data center in Dallas,
Oregon built in 2006 on the banks of Columbia River. As these various
data centers grew in size, businesses have developed their data center
as “Greenfield” projects. Data centers have been sited to do the
following-
• Have access to low cost power
• Leverage renewable power source
• Be near abundant water
• Be sited where high speed network backbone connections can be
made
• Keep land cost modest and occupation unobtrusive
• Obtain tax breaks
• Optimize the overall system latency

Vishal Jain
These characteristics made cloud computing network highly
sufficient and capture enough margin to make utility computing
profitable.

Essential Characteristics

On-demand self-service. A consumer can unilaterally provision


computing capabilities, such as server time and network storage, as
needed automatically without requiring human interaction with each
service provider.

Broad network access. Capabilities are available over the network and
accessed through standard mechanisms that promote use by
heterogeneous thin or thick client platforms (e.g., mobile phones, tablets,
laptops, and workstations).

Resource pooling. The provider’s computing resources are pooled to


serve multiple consumers using a multi-tenant model, with different
physical and virtual resources dynamically assigned and reassigned
according to consumer demand. There is a sense of location
independence in that the customer generally has no control or
knowledge over the exact location of the provided resources but may be
able to specify location at a higher level of abstraction (e.g., country,
state, or datacenter). Examples of resources include storage,
processing, memory, and network bandwidth.
Rapid elasticity. Capabilities can be elastically provisioned and released,
in some cases automatically, to scale rapidly outward and inward
commensurate with demand. To the consumer, the capabilities available
for provisioning often appear to be unlimited and can be appropriated in
any quantity at any time.

Measured service. Cloud systems automatically control and optimize


resource use by leveraging a metering capability1 at some level of
abstraction appropriate to the type of service (e.g., storage, processing,
bandwidth, and active user accounts). Resource usage can be
monitored, controlled, and reported, providing transparency for both the
provider and consumer of the utilized service.

Vishal Jain
Service Models

Software as a Service (SaaS)

The capability provided to the consumer is to use the provider’s


applications running on a cloud infrastructure2. The applications are
accessible from various client devices through either a thin client
interface, such as a web browser (e.g., web-based email), or a program
interface. The consumer does not manage or control the underlying
cloud infrastructure including network, servers, operating systems,
storage, or even individual application capabilities, with the possible
exception of limited user-specific application configuration settings.

Platform as a Service (PaaS)

The capability provided to the consumer is to deploy onto the cloud


infrastructure consumer-created or acquired applications created using
programming languages, libraries, services, and tools supported by the
provider.3 The consumer does not manage or control the underlying
cloud infrastructure including network, servers, operating systems, or
storage, but has control over the deployed applications and possibly
configuration settings for the application-hosting environment.

Infrastructure as a Service (IaaS)

Vishal Jain
The capability provided to the consumer is to provision processing,
storage, networks, and other fundamental computing resources where
the consumer is able to deploy and run arbitrary software, which can
include operating systems and applications. The consumer does not
manage or control the underlying cloud infrastructure but has control
over operating systems, storage, and deployed applications; and
possibly limited control of select networking components (e.g., host
firewalls).

Vishal Jain
Experiment No. 2
Objective:Cloud Computing Security.

Cloud computing security or, more simply, cloud security is an evolving


sub domain of computer security, network security and more broadly
information security. It refers to a broad set of policies, technologies, and
controls deployed to protect data, application, and associated
infrastructure of cloud computing.
Cloud security architecture is effective only, if the correct defensive
implementation are in place.An efficient cloud security architecture
should recognize the issue that will arise with security management.
These controls are put in place to safeguard any weakness in the
system and reduce the effect of an attack.While there are many types of
control behind a cloud security architecture, they can usually be found in
one of the following categories.
Deterrent control
These controls are intended to reduce attack on cloud system. Much like
a warning sign ona fence, different controls typically reduce the threat
level by informing potential attacker that there will beadverse
consequences for them if proceed.
Preventive controls
Preventive controls strengthen the system against incidents, generally
by reducing, if not actually eliminating vulnerability.
Detective control
Detective control are intended to detect and react appropriately to any
incident that occur. In the event of an attack, a detective control will
signal the preventive or corrective control to address the issue.
Corrective control
Corrective control reduces the consequences of an incident, normally by
limiting the damage. They come into effect during or after an incident.

Vishal Jain
Cloud Security Services

A. Identification & authentication

In Cloud computing, depending on the type of cloud as well as he


delivery model, specified users must be established and
supplementary access and permissions may be granted
accordingly this process is:
Targeting at verifying and validating individual cloud users by
employing usernames and passwords protections in their cloud
profiles.

B. Authorization

Authorization is an important information security requirement in


cloud computing to ensure referential integrity is maintained. It
follows on in exerting control and privileges over process flows
within cloud computing. Authorization is maintained bt the system
administrator in a private cloud.

C. Confidentiality

In cloud computing, confidentiality plays a major part


especially in maintaining control over organization data
situated across multiple distributed database. It is a must
when employing a public cloud due to public cloud
accessibility nature. Asserting confidentiality of user’s
profile and protecting their data, that is virtually accessed,
allows for information security protocols to be enforced at
various different layers of cloud applications.

D. Integrity

The integrity requirement lies in applying the due diligence


within the cloud domain mainly when accessing data
therefore ACID (atomicity, consistency, isolation).
Properties of cloud data should without a doubt be
robustly imposed across all cloud computing deliver
models.

Vishal Jain
E. Non-Repudiation

In cloud computing non-repudiation can be obtained by


applying the traditional e-commerce security protocol and
token provisioning to data transmission within cloud
application such as digital signatures, timestamp and
confirmation receipts services.

F. Availability

It is one of the most critical information security requirements in


cloud computing because it is a key decision factor, the service
level agreement is most important document which highlights
the trepidation of availability in cloud services and resources
between the cloud provider and client. Therefore, by exploring
the information security requirements at each of the various
cloud deployment models set out by ISO.
• Workload(requests) to be serviced per second(Mbps) or a
unit of time.
• Throughput(Request/sec), this criterion will be recovered
recovery, buffering capacity and processing power factors
[15,20].
• The average time of processing(exe/sec).
• Percentage of CPU utilization [15,21]
• The number of requests executed per unit time.
• The number of requests per unit tie buffer.
• The number of rejected requests per unit time.

Vishal Jain
Experiment No. 3
Objective:Performance Evaluation of Cloud.

Cloud computing resources must be compatible, high performance and


powerful. High performance is one of the cloud advantages which must
be satisfactory for each service [1-5]. Higher performance of services
and anything related to cloud have influence on users and service
providers. Hence performance evaluation for cloud providers and users
is important. There are many methods for performance prediction and
evaluation we use following methods:

• Evaluation based on criteria and characteristics.


• Evaluation based on simulation
Factors Affective on Performance Nowadays

The term “performance” is more than a classic concept and includes


more extensive concepts such as reliability, energy efficiency,
scalability and soon. Due to the extent of cloud computing
environments and the large number of enterprises and normal users
who are using cloud environment, many factors can affect the
performance of cloud computing and its resources. Some of the
important factors considered in this paper are as follows:
• Security, the impact of security on cloud performance are seem
lightly strange, but the impact of security on network infrastructure
has been proven. For example, DoSattacks have wide impact on
network performance and if happen, it will greatly reduce network
performances and also be effective on response time too.
Therefore, if this risk and any same risks threaten cloud
environment, it will be a big concern for users and providers.
International journal on cloud computing: Services and architecture
(IJCCSA), Vol.3, No.5, October2013.
• Recovery, when data in cloud face errors and failures or data are
lost for any reason, the time required for data retrieval and
volumes of data which are
Other factors that can affected performance which are as follows-
• Usability
• Scalability
• Workload
• Repetition or redundancy

Vishal Jain
• Processor power
• Latency

PERFORMANCE EVALUATION CRITERIA

There is a series of criteria for evaluation of all factors affecting


performance of cloud computing some of which will be used in this
paper. The criteria are under development. Some of the criteria
have been selected considering the importance and criteria in
simulation.
• Average response time per unit time, this criterion will cover all the
factors completely.
• Network capacity per second or unit time, the most important
factor associated with this criterion is network bandwidth,
availability and scalability.
• The number of I/O commands per second (IOPS) or unit
international journal on cloud computing. Services and architecture
(IJCCSA), Vol.3, no.5, October2013
• Average waiting time per unit time [6,18]

Vishal Jain
Experiment No. 4
Objective:Introduction of Hadoop.

Hadoop is a rapidly evolving ecosystem of components for implementing


the Google MapReduce algorithms in a scalable fashion on commodity
hardware. Hadoop enables users to store and process large volumes of
data and analyze it in ways not previously possible with less scalable
solutions or standard SQL-based approaches. As an evolving
technology solutions, Hadoop design considerations are new to most
users and not common knowledge. As part of Dell | Hadoop solution,
Dell developed a series of best practices and architectural
considerations to use when designing and implementing Hadoop
solutions.

Hadoop is a highly scalable compute and storage platform. While most


users will not initially deploy servers numbered in the hundreds or
thousands. Dell recommends following the design principles that drive
large, hyper-scale deployments. This ensures that as you start with a
small Hadoop environment, you can easily scale that environment
without rework it exiting servers, software, deployments strategies, and
network connectivity.

• Apache Hadoop is an open source framework for developing


distributed applications that can process very large amounts of
data.
• It is a platform that provides both distributed storage and
computational capabilities.

Hadoop has two main layers:


1. Computation layer: The computation tier uses a framework called
MapReduce.
2. Distributed storage layer: A distributed filesystem called HDFS
provides storage.

Vishal Jain
Why Hadoop?

Building bigger and bigger servers is no longer necessarily the best


solution to larger-scale problems. Nowadays the approach is to tie
together many low-end machines together as a single functional
distributed system. For example:

• A high-end machine with four I/O channels each having a


throughout of 100-MB/sec will require three hours to read a 4 TB
data set! With Hadoop, this same data set will be divided into
smaller (typically 64 MB) blocks that are spread among many
machines in the cluster via the Hadoop Distributed File System
(HDFS).
• With a modest degree of replication, the cluster machines can
read the data set in parallel and provide a much higher
throughput. Moreover,its cheaper than one high-end ever!

For computationally intensive work,

• Most of the distributed systems (e.g. SET|@home) are having


approach moving the data to the place where computation will
take place and after the computation, the resulting data is moved
back for storage. This approach works fine for computationally
intensive work.

For data-intensive work,

• We need other better approach, Hadoop has better philosophy


toward that Because Hadoop focusses on moving code/algorithm
to data instead data to the code/algorithm.
• The move-code-to-data philosophy applies within the Hadoop
cluster itself, and data is broken up and distributed across the
cluster, and computation on a piece of data takes place on the
same machine where that piece of data resides.
• Hadoop philosophy of move-code-to-data makes more sense. As
we know the code/algorithm are always smaller than the Data
hence code/algorithm is easier to move around.

Vishal Jain
Hadoop Challenges

With all large environments, deployment of the servers and software is


an important consideration. Dell provides best practices for the
deployment of Hadoop solutions. These best practices are implemented
through a set of tools to automate the configuration of the hardware,
installation of the operating system (OS), and installation of the Hadoop
software stack from Cloudera.

As with many other types of information technology(IT) solutions,


change management and systems monitoring are a primary
consideration within Hadoop. The IT operations team needs to ensure
tools are in place to properly track and implement changes, and notify
staff when unexpected events occur within the Hadoop environment.

Hadoop is a constantly growing, complex ecosystem of software and


provide no guidance to the best platform for it to run on. The Hadoop
community leaves the platform decisions to end users, most of whom do
not have background in hardware or the necessary lab environment to
benchmark all possible design solutions. Hadoop is a complex set of
software with more than 200 tunable parameters. Each parameter
affects others as tuning is completed for a Hadoop environment and will
change over time as job structure changes, data layout evolves, and
data volumes grows. As data centers, have grown and the number of
servers under management for a given organization has expanded,
users are more conscious of the impact new hardware will have on
existing data centers and equipment.

Vishal Jain
Hadoop node types
Hadoop has a variety of types within each cluster, these include Data
Nodes, Name Nodes, and Edge Nodes.

Names of these nodes can vary from site to site, but the functionality is
common across the sites. Hadoop’s architecture is modular, allowing
individual components to be scaled up and down as the needs of the
environment changes. The base nodes types for a Hadoop cluster are:

• Names Node – The Names Node is the central for information


about the file system deployed in a Hadoop environment. An
environment can have one or two Name Nodes, configured to
provide minimal redundancy between the Name Node. The Name
Node is contacted by clients of the Hadoop Distributed File
System (HDFS) to locate information within the file system and
provide updates for data they have added, moved manipulated, or
deleted.

• DataNode – DataNodes make up the majority of the servers


contained in a Hadoop environment, CommonHadoop
environment they will number in the hundreds based on capacity
and performance needs. The DataNode servers two functions: It
contains a portion of the data in the HDFS and it acts as a
computer platform for running jobs, someof which will utilize the
local data within the HDFS.

• EdgeNode – The EdgeNode is the access point for the external


applications, tools and users that need to utilize the Hadoop
environment. The EdgeNode sits between the Hadoop cluster and
the corporate network to provide access control, policy
enforcement, logging, and gateway services to the Hadoop
environment. A typical Hadoop environment will have a minimum
of one EdgeNode environment and more based on performance
needs.

Vishal Jain
Hadoop Key Features

Distributed computing is the very vast field but following key


features has made Hadoop very distinctive and attractive.

1. Accessible

Hadoop runs on large cluster of commodity machines or on


cloud computing services such as Amazon’s Elastic
Compute Cloud (EC2).

2. Robust

As Hadoop is intended to run commodity hardware. It is


architecture with the assumption of frequent hardware
malfunctions. It can gracefully handle most such failures.

3. Scalable

Hadoop scales linearly to handle larger data by adding


nodes to the cluster.

4. Simple

Hadoop allows users to quickly write efficient parallel code.


Hadoop’s accessibility and simplicity give it an edge over
writing and running large distributed programs.

Vishal Jain
Hadoop Uses
Hadoop was originally developed to be an open implementation of
Google MapReduce and Google File System. As the ecosystem
around Hadoop has matured, a variety of tools have been
developed to streamline data access, data management, security,
and specialized additions for verticals and industries, Despite this
large ecosystem, there are several primary uses and workloads for
Hadoop that can be outlined as:

• Compute – A common use of Hadoop is a distributed


compute platform for analyzing or processing large amounts
of data. The compute use is characterized by the need for
large numbers of CPUs and large amounts of programming
to store in-process data. (APIs) necessary to distribute and
track workloads as they are run on large numbers of
individual machines.

• Storage – One primarily component of the Hadoop


ecosystem is HDFS – the Hadoop Distributed File System.
The HDFS allows users to have a single addressable
namespace, spread across many hundreds or thousands of
servers, creating a single large file system. HDFS manages
the replication of the data on this file system to ensure
hardware failures do not lead to data loss. Many users will
use this scalable file system as a place to store large
amounts of data that is then accessed within jobs run in
Hadoop or by external systems.

• Database – The Hadoop ecosystem contains components


that allow the data within the HDFSs to be presented in a
SQL-like interface. This allows standard tools to INSERT,
SELECT, and UPDATE data within the Hadoop environment,

Vishal Jain
with minimal code changes to existing applications. Users
will commonly employ this method for presenting data in a
SQL format for easy integration with existing system and
streamlined access by users.

Experiment No. 5
Objective:Management of Cloud Resources.

Cloud Computing becoming one of the most explosively expanding


technologies in the computing industry today. It enables users to migrate
their data and computation to a remote location with minimal impact on
system performance. Typically, this provides a number of benefits.
These benefits include:

1. Scalable – Clouds are designed to deliver as much computing


power as any user wants. While in practice the underlying infrastructure
is not infinite, the cloud resources are projected to case the developer’s
dependence on any specific hardware.

2. Quality of Service (QoS) – Unlike standard data centers and


advanced computing resources, a well-designed Cloud can project a
much higher (QoS) than typically possible. This is due to the lack of
dependence on specific hardware, so any physical machine failure can
be mitigated without the user’s knowledge.

3. Specialized Environment – Within a Cloud, the user can utilize


custom tools and services to meet their needs. This can be use the
latest library, toolkit, or to support legacy code within new infrastructure.

4. Cost Effective – Users finds only the hardware required for each
project. This greatly reduces the risk for institutions who may be looking
to build a scalable system. Thus providing greater flexibility since the
user is only paying for needed infrastructure while maintaining the option
to increase services as needed in the future.

5.Simplified Interface – Whether using a specific application, a set


of tools or Web services, Clouds provide access to a potentially vast
amount of computing resources in an easy and user-centric way. We
have investigated such an interface within Grid systems through the use
of the Cyber aide project.

Vishal Jain
Virtual Machine Abstraction
Virtualization is not specific to cloud computing. IBM originally pioneered
the concept in the 1960’s with the M44/44X systems. It has only recently
been reintroduced foe general use on X86 platforms. Today there are a
no. of clouds that offers Infrastructure as a Service (IaaS). The Amazon
Elastic Compute Cloud (EC2) is probably the most popular of which and
is used extensively in the IT industry. Eucalyptus is becoming popular in
both the scientific and industry communities. It provides the same
interface as EC2 and allows users to build and EC2- like space cloud
using their own internal resources. Other scientific cloud specific projects
exist such as OpenNebula. In-VIGO, and cluster-on-demand. They
provide their own interpretation of private cloud services within a data
center. Using a cloud deployment overlaid on a grid computing system
has been explored by the Nimbus project with the Globus Toolkit. All of
these cloud leverage the power of virtualization to create an enhanced
data center.

Vishal Jain
Resource Management Challenges
A successful resource management at the cloud scale requires a
management system to provide a rich set of resource controls and
enforce the controls by providing services such as resource pool
management, initial placement and load balancing. The problems with
scaling the existing solutions can be broadly classified as the following:

1.Decomposition of the management space:

As clusters increase in size, the overhead in collecting, analyzing, and


acting upon the associated data grows, and the need to develop an
appropriate decomposition of the management space to compute a
timely response to changes in resource demands and settings
increases. Note that management of elasticity also becomes more
important, so that approaches that support rapid increase and decrease
of available resources are attractive.

2.Hetrogeneity of clusters:

As cluster scale increases the amount of heterogeneity in the cluster


tends to increase. It is difficult to cloud providers to produce a large no.
of machines that are identical. Also, cloud providers typically add
machines overtime; there could be several generations of hardware co-
existing in the cloud, possibly impacting the ability to live-migrate
between hosts. In addition, the storage, networking, and other hardware
needs of the VM may limit the host to which it may live-migrate.as
cluster scale increases, the hosts to which a VM can be migrated
becomes a small percentage of the clusters.

3.Frequency of operations:

As the scale increases the no. of users and the frequency of


management operations increase. It is important for the system to keep
providing low latency as the cluster size increases.

Vishal Jain
4.Resitance to failures:

As scale increases the impact of failure of the resource management


component increases; its failure would mean users not be able to power
on more VMs or change resource pool settings, resources not flowing
between resource pools when the VMs need them, and VMs not being
moved out of a host when it becomes overloaded. This layer needs to be
increasingly robust and to handle host failures robustly.

Techniques to resolve Resource Management


Challenges

In this section, we discuss various techniques to handle the challenges


mentioned above while supporting the resource management model
discussed. The goal of this is not to advocate a single approach but to
promote debate and research projects in different directions. We
envision three:

(1) Hierarchical-Scaling (2) Flat-Scaling and (3) Statistical-Scaling.

Hierarchical-Scaling

In this approach, resource management systems are built on top of each


other in order to reach scale. For instance, current tools like DRS
provide cluster level resource management using a cluster of up to 32
hosts and 3000 VMs. Similar solution has been proposed by other
companies like Microsoft PRO. In order to scale these solutions, one can
build a layer on top of such solutions that interacts with these solutions
and does operations like initial placement and load balancing across
cluster. this hierarchy can be built within a data center and even across
data centers. Some of these issues in building are such a hierarchy:

Flat-Scaling

In this approach, a completely distributed and decentralized resource


management layer is built, creating a single domain for resource
management. Decisions are made using data collected and aggregated
over the large number of hosts and VMs. One way to do this is using

Vishal Jain
structured Peer-to-Peer technologies where all hosts are part of a P2P
network. An aggregation layer similar to SDIMS or Astrolab can collect
and aggregate stats across all hosts. A query infrastructure on top of this
P2P substrate can find hosts that are overloaded and under-loaded
using technologies like any cast and multicast. This layer has the benefit
that it can do optimization at a global scale rather than limiting itself to
local optimization as done by hierarchical solution. For example, power
management may yield more benefit in this case because hosts spare
capacity is not binned into smaller sized clusters.
Statistical-Scaling

In this approach, large scale resource management is achieved by doing


smarter operations at small scale. The idea is to create dynamic clusters
based on the current operation by querying a set of relevant hosts that
are needed. for example, if the operation is VM power on, a query
service will ask for top K lightly loaded hosts that also satisfy other
constraints for the VM. It will then create a dynamic cluster of those K
hosts and invoke DRS like placement algorithm on that cluster.one key
property need to show for the success of this approach is that one can
attain large scale load balancing and optimal placement by doing
repeated small scale optimization. This intuitive property is supported by
the well-known research field related to the power of two choices.
This main implication of the key result in this area is that having a small
amount of information in the form of one more choice improves the
overall load balancing by a large factor. This result has been used
extensively in many other areas for load balancing.

Vishal Jain
Experiment No. 6
Objective:Deployment Model.

There are four common deployment models for cloud services loosely
determined by who can access to the cloud services:

• Public Cloud
• Private Cloud
• Community Cloud
• Hybrid Cloud

Public Cloud:

A “Public” Cloud infrastructure is available to the general public and is


owned by a third party cloud service provider (CSP). In a public cloud,
an agency dynamically provisions computing resources over the internet
from a CSP who shares its resources with other organization. Similar to
that of an electric utility billing system, the CSP bills the agency for its
share of resources.

Vishal Jain
Private Cloud:

A “Private” Cloud infrastructure us operated solely for single organization


or agency, the CSP dedicates specific cloud services to that agency and
no other clients. The agency specifies, architects, and controls a pool of
computing resources that the CSP delivers as a standardized set of
services. A common reason for agency to provide private clouds is their
ability to enforce their own data security standards and controls.
An agency will typically host a private cloud on-premises, connect to it
through private network links, and only shares its resources within the
agency.

Vishal Jain
Community Cloud:

A “Community” cloud infrastructure is procured jointly by several


agencies or programs that share specific needs such as security,
compliance, or jurisdiction considerations. The agencies or CSP may
manage the community cloud and may keep it on-premises off-
premises.
When agencies have common set of requirements and customers, a
community cloud enables them to combine assets and share computing
resources, data, and capabilities. By eliminating the duplication of similar
systems, agencies can say money and allocate their resources more
efficiently procuring a community cloud is also a way that an agency can
advance the Federal IT Shared Service Strategy.

Vishal Jain
Hybrid Cloud:

A “Hybrid” cloud comprises two or more clouds (private, community, or


public) with a mix of both internally and externally hosted services.
Agencies will likely not limit themselves to one cloud but will rather
incorporate different and overlapping cloud services to meet their unique
requirements. Hybrid deployments models are complex and require
careful planning to execute and manage especially when communication
between two different cloud deployment is necessary.

Vishal Jain
Experiment No. 7
Objective:Working of Google Drive to make spreadsheet and notes.

Requirement:Google account, Internet Connection.

Theory:

Google Docs is a free cloud-based suite of tools for creating


documents, spreadsheets, presentations, and more. This tutorial will
cover the Spreadsheets application in Google Docs, in addition to
showing you how to access and store your Docs from Google Drive.

Google Docs, Sheets, and Slides are productivity apps that let you
create different kinds of online documents, work on them in real time
with other people, and store them in your Google Drive online — all for
free. You can access the documents, spreadsheets, and presentations
you create from any computer, anywhere in the world. (There's even
some work you can do without an Internet connection!) This guide will
give you a quick overview of the many things that you can do with
Google Docs, Sheets, and Slides.

Google Docs
Google Docs is an online word processor that lets you create and format
text documents and collaborate with other people in real time. Here's
what you can do with Google Docs:

• Upload a Word document and convert it to a Google document


• Add flair and formatting to your documents by adjusting margins,
spacing, fonts, and colors — all that fun stuff
• Invite other people to collaborate on a document with you, giving
them edit, comment or view access

Vishal Jain
• Collaborate online in real time and chat with other collaborators —
right from inside the document
• View your document's revision history and roll back to any
previous version
• Download a Google document to your desktop as a Word,
OpenOffice, RTF, PDF, HTML or zip file
• Translate a document to a different language
• Email your documents to other people as attachments

Google Sheets
Google Sheets is an online spreadsheet app that lets you create and
format spreadsheets and simultaneously work with other people. Here's
what you can do with Google Sheets:

• Import and convert Excel, .csv, .txt and .ods formatted data to a
Google spreadsheet
• Export Excel, .csv, .txt and .ods formatted data, as well as PDF
and HTML files
• Use formula editing to perform calculations on your data, and use
formatting make it look the way you'd like
• Chat in real time with others who are editing your spreadsheet
• Create charts with your data
• Embed a spreadsheet — or individual sheets of your spreadsheet
— on your blog or website

Google Slides
Google Slides is an online presentations app that allows you to show off
your work in a visual way. Here's what you can do with Google Slides:

• Create and edit presentations


• Edit a presentation with friends or coworkers, and share it with
others effortlessly
• Import .pptx and .pps files and convert them to Google
presentations
• Download your presentations as a PDF, a PPT, or a .txt file
• Insert images and videos into your presentation
• Publish and embed your presentations in a website

Vishal Jain
Create, name or delete a Google document
Create a Google document
To create a new document, go to your Drive, click the Create button,
and select Document.

A window with a new Google document will open, and you'll be able to
edit the document, share it with other people, and collaborate on it in
real-time. Google Docs saves your document automatically, and you can
always access it from your Drive.

Name a document
When you create a new document, Google Docs will name it Untitled by
default.

To choose a name other than Untitled, click the File menu, and select
Rename. From here you can choose and confirm your document's title.
You can also edit the name by clicking the title displayed at the top of
the page, and making your changes in the dialog that appears. Titles
can be up to 255 characters long.

Delete a document
Delete an item that you own from your Drive

1. From your Drive, select the item(s) you want to delete.


2. From the More menu, choose Move to trash.
3. If you're deleting a shared document that you own, you'll see an
option to change the ownership of the document.
4. The item will be moved to the Trash.
5. To purge individual items from Trash, select them and choose
Delete forever. To purge all your items, click Empty Trash in the
upper left.

Create and save a document


There are different ways of getting started using Google documents: you
can create a new online document, you can upload an existing one, or
you can use a template from our templates gallery.

Vishal Jain
To create a new document, go to your Drive, click the red Create button,
and select Document from the drop-down menu.

As soon as you name the document or start typing, Google Docs will
automatically save your work every few seconds. At the top of the
document, you'll see text that indicates when your document was last
saved. You can access your document at any time by opening your
Drive at http://drive.google.com.

To save a copy of a document to your computer, you can download it. In


your document, go to the File menu and point your mouse to the
Download as option. Select one of the following file types: HTML
(zipped), RTF, Word, Open Office, PDF, and plain text. Your document
will download to your computer.

Upload a document
You can upload existing documents to Google documents at any time.
When you're uploading, you can either keep your document in its original
file type or convert it to Google Docs format. Converting your document
to Google Docs format allows you to edit and collaborate online from any
computer.

Note: When uploaded, images within a document are left as images


(rather than being converted to text by Optical Character Recognition
technology).

You can upload the following file types:

• .html
• .txt
• .odt
• .rtf
• .doc and .docx
• .pdf

Follow these steps to upload a document:

1. Click the Upload icon in the top left of your Documents List.
2. Click Files..., and select the document you'd like to upload.
3. Click Open.
4. Check the box next to 'Convert documents, presentations,
spreadsheets, and drawings to the corresponding Google Docs
format' if you'd like to be able to edit and collaborate on the

Vishal Jain
document online. Uploaded document files that are converted to
Google documents format can't be larger than 1 MB.
5. Click Start upload. The uploaded file will appear in your
Documents List.

Experiment No. 8
Objective:Installation and Configuration of Justcloud.

Requirement:Justcloud exe File.

Theory:
Professional Cloud Storage from JustCloud is Simple, Fast and Secure.
Just Cloud will automatically backup the documents, photos, music and
videos stored on your computer, to the cloud so you are never without
files again.

Installation:
1. Download Software this link

Vishal Jain
http://www.justcloud.com/download/

2. By following these steps, you will download and install the


JustCloud software application on this computer. This software will
automatically start backing up files from your computer and saving
them securely in an online cloud user account. Your free account
gives you 15MB storage space or 50 files for 14 days. Once
installed a sync folder will be added to your desktop for you to
easily drag and drop files you wish to backup.

Vishal Jain

You might also like