601_2_Cloud_Computing
601_2_Cloud_Computing
1
What is Cloud?
In other words, we can say that Cloud is something, which is present at remote location. Cloud
can provide services over public and private networks, i.e., WAN, LAN or VPN.
Applications such as e-mail, web conferencing, customer relationship management (CRM) execute
on cloud.
Cloud Computing refers to manipulating, configuring, and accessing the hardware and software
resources remotely. It offers online data storage, infrastructure, and application.
Cloud computing offers platform independency, as the software is not required to be installed
locally on the PC. Hence, the Cloud Computing is making our business
applications mobile and collaborative.
Cloud Computing provides us resources of accessing the applications as utilities over the Internet.
It allows us to create, configure, and customize the applications online.
Basic Concepts
There are certain services and models working behind the scene making the cloud computing
feasible and accessible to end users.
Deployment Models
Service Models
1
Cloud Computing Unit.1
Deployment Models:
Deployment models define the type of access to the cloud, i.e., how the cloud is located?
Cloud can have any of the four types of access: Public, Private, Hybrid, and Community.
Public Cloud:
The public cloud allows systems and services to be easily accessible to the general public. Public
cloud may be less secure because of its openness.
Private Cloud:
The private cloud allows systems and services to be accessible within an organization. It is more
secured because of its private nature.
Community Cloud:
The community cloud allows systems and services to be accessible by a group of organizations.
Hybrid Cloud:
The hybrid cloud is a mixture of public and private cloud, in which the critical activities are
performed using private cloud while the non-critical activities are performed using public cloud.
Service Models:
Cloud computing is based on service models. These are categorized into three basic service
models:-
Infrastructure-as–a-Service (IaaS)
Platform-as-a-Service (PaaS)
Software-as-a-Service (SaaS)
2
Cloud Computing Unit.1
Anything-as-a-Service (XaaS) is yet another service model, which includes Network-as-a-Service,
Business-as-a-Service, Identity-as-a-Service, Database-as-a-Service or Strategy-as-a-Service.
The Infrastructure-as-a-Service (IaaS) is the most basic level of service. Each of the service models
inherit the security and management mechanism from the underlying model, as shown in the
following diagram:
Infrastructure-as-a-Service (IaaS):
IaaS provides access to fundamental resources such as physical machines, virtual machines, virtual
storage, etc.
Platform-as-a-Service (PaaS):
PaaS provides the runtime environment for applications, development and deployment tools, etc.
Software-as-a-Service (SaaS):
The concept of Cloud Computing came into existence in the year 1950 with implementation of
mainframe computers, accessible via thin/static clients.
Since then, cloud computing has been evolved from static clients to dynamic ones and from
software to services.
3
Cloud Computing Unit.1
[A thin client is a computer that runs from resources stored on a central server instead of a
localized hard drive. Thin clients work by connecting remotely to a server-based
computing environment where most applications, sensitive data, and memory, are stored.]
Benefits
Cloud Computing has numerous advantages. Some of them are listed below -
4
Cloud Computing Unit.1
Although cloud Computing is a promising innovation with various benefits in the world of
computing, it comes with risks as below:
It is the biggest concern about cloud computing. Since data management and infrastructure
management in cloud is provided by third-party, it is always a risk to handover the sensitive
information to cloud service providers.
Although the cloud computing vendors ensure highly secured password protected accounts, any
sign of security breach may result in loss of customers and businesses.
Lock In:
It is very difficult for the customers to switch from one Cloud Service Provider (CSP) to another. It
results in dependency on a particular CSP for service.
Isolation Failure:
This risk involves the failure of isolation mechanism that separates storage, memory, and routing
between the different occupiers.
In case of public cloud provider, the customer management interfaces are accessible through the
Internet.
5
Cloud Computing Unit.1
Insecure or Incomplete Data Deletion:
It is possible that the data requested for deletion may not get deleted. It happens because either
of the following reasons
Extra copies of data are stored but are not available at the time of deletion
Disk that stores data of multiple occupiers is destroyed.
Cloud Computing allows the users to use web services and resources on demand. One can logon
to a website at any time and use them.
The Cloud computing services does not require any human administrators, user themselves are
able to provision, monitor and manage computing resources as needed.
6
Cloud Computing Unit.1
Broad Network Access:
Since cloud computing is completely web based, it can be accessed from anywhere and at any
time.
The Computing services are generally provided over standard networks and heterogeneous
devices.
Resource Pooling:
Cloud computing allows multiple residents to share a pool of resources. One can share single
physical instance of hardware, database and basic infrastructure.
The IT resource (e.g., networks, servers, storage, applications, and services) present are shared
across multiple applications and occupant in an uncommitted manner. Multiple clients are
provided service from a same physical resource.
Rapid Elasticity:
It is very easy to scale the resources vertically or horizontally at any time. Scaling of resources
means the ability of resources to deal with increasing or decreasing demand.
The resources being used by customers at any given point of time are automatically monitored.
The Computing services should have IT resources that are able to scale out and in quickly and on
as needed basis.
Whenever the user require services it is provided to him and it is scale out as soon as its
requirement gets over
Measured Service:
In this service cloud provider controls and monitors all the aspects of cloud service. Resource
optimization, billing, and capacity planning etc. depend on it.
The resource utilization is tracked for each application and occupant, it will provide both the user
and the resource provider with an account of what has been used. This is done for various reasons
like monitoring billing and effective use of resource
Cloud Vulnerability
Vulnerability is a cyber-security term that refers to a flaw in a system that can leave it open to
attack. A vulnerability may also refer to any type of weakness in a computer system itself, in a
set of procedures, or in anything that leaves information security exposed to a threat.
7
Cloud Computing Unit.1
Denial of Service (DoS) and Distributed Denial of Service (DDoS) attacks are cloud security
vulnerabilities that make servers inaccessible for users by flooding the network's traffic. Data
Loss.
Cloud computing, an emergent technology, has placed many challenges in different aspects of
data and information handling. Some of these are shown in the following diagram:
Security and Privacy of information is the biggest challenge to cloud computing. Security and
privacy issues can be overcome by employing encryption, security hardware and security
applications.
Portability:
This is another challenge to cloud computing that applications should easily be migrated from one
cloud provider to another. There must not be vendor lock-in. However, it is not yet made possible
because each of the cloud provider uses different standard languages for their platforms.
Interoperability:
It means the application on one platform should be able to incorporate services from the other
platforms. It is made possible via web services, but developing such web services is very complex.
8
Cloud Computing Unit.1
Computing Performance:
Data intensive applications on cloud requires high network bandwidth, which results in high cost.
Low bandwidth does not meet the desired computing performance of cloud application.
It is necessary for cloud systems to be reliable and robust because most of the businesses are now
becoming dependent on services provided by third-party.
Cloud Migration
Cloud migration is the process of moving data, applications or other business elements to a
cloud computing environment. ... However, a cloud migration could also involve moving data
and applications from one cloud platform or provider to another -- a model known as cloud-
to-cloud migration.
There are various types of cloud migrations an enterprise can perform. One common model is
the transfer of data and applications from a local, on-premises data center to the public cloud.
The most important part of any cloud migration is making sure the migration gets your
company where it needs to be. The four types of cloud migration are called lift and shift, shift
to Software-as-a-Service (SaaS), application refactoring, and replatforming.
The lift and shift approach is for organizations looking to get out of the data center and stop
managing hardware.
Lift and shift provides the same software that your company used in the data center, but now
in the cloud. There isn’t any learning curve for the cloud applications, since they work exactly
the same as before. This is the fastest method for migrating applications to the cloud, and the
one that causes the least disruption. It only requires the involvement of the infrastructure and
security teams, leaving everyone else free to pursue their work uninterrupted.
It’s also the option with the least upfront cost. Moving the application to the cloud allows it to
handle peak performance, without your company having to pay for it.
Lift and shift comes with its drawbacks, however. This cloud migration can’t take full
advantage of the speed and versatility the cloud can provide.
Since the process doesn’t change the application – it just moves the code to a new location –
the shifted version of the application doesn’t usually have better performance than the
original.
It’s also unlikely to lead to long-term savings. This model is best suited to companies with a
regular peak schedule and slow, predictable changes in the market.
9
Cloud Computing Unit.1
Food delivery companies with their regular peaks on Monday, Saturday, and Sunday are a
good example of the former.
Tax companies are a good example of the latter, since tax rules change on a yearly basis and
new rules are released a year in advance, giving the companies plenty of time to prepare.
Lift and shift cloud migration involves copying, bit by bit, the workloads, tasks, and
applications housed within an organization’s virtual machines, and storing the replicated
version in a cloud-based location. Because the data is simply a replication, no code is modified
and no costly, time-consuming redesign is required during migration.
The method is certainly less complex than other migration techniques, IT teams and their
cloud provider will want to be sure data sets are properly matched with handling systems in
the new environment. Additionally, they need to ensure that all applications have the
resources needed to operate effectively and efficiently.
Companies that want to stop allocating time and resources to applications outside their core
business should think about shifting to SaaS.
Shifting to SaaS means outsourcing one or more applications to a cloud services company that
specializes in managing those applications. Companies do this on an application by application
basis and only shift the applications they need to.Static applications can remain on-premises.
Shifting to SaaS frees employees up to focus on core competencies and the things that make a
business unique and competitive.
It’s extremely important when shifting an application to SaaS that you pick the right service.
The main drawback of shifting to SaaS, while you can personalize it, is that customizing it can lead
to problems.
Shifting to SaaS should only be used for routine functions – not for anything that needs to be
unique.
Email is a good example of a routine business function that can be shifted to SaaS. For example,
we worked with a manufacturing company that built engine components. The company was sick
of managing its own email. While it needed to have email for all its employees, the customers
didn’t care what the email service looked like – so long as it worked. We helped them find a
hosting service that freed the company of needing to worry about their email so it could focus on
manufacturing.
App modernization is a preferred approach for organizations that have specific applications which
could benefit from the cloud.
10
Cloud Computing Unit.1
With refactoring, organizations can copy their inheritance applications whole and unbroken onto
a cloud platform.
What makes it low risk is that inheritance applications can run in parallel while new applications
are constructed, with the immediate benefits of quickness and speed to market. This approach
focuses on the applications that benefit the most from a cloud platform.
You can save money on the platform itself as well by switching to cloud native services that cost
less than the ones you use on-premises.
Refactoring is not just about cutting costs. It also allows you to make changes to your enterprise
very quickly, which means that you can keep up with your customers.
Refactoring lets you respond faster and prioritize updates. One big box store we worked with
started out with its applications so hard coded that it took months to do simple things like
changing the font or background color. Refactoring got them moving.
Refactoring usually requires outside help. We worked with a company that wanted to refactor its
entire application suite into modern technology. The project would have taken four years for the
company working on their own, but we were able to cut that time down to eighteen months. We
provided the company making the migration with its unique model for retraining.
When introducing the new technology, we rotate through everyone who needs training. Slowly,
over time, we dial back how many consultants they have helping until the local IT people are able
to handle everything on their own. This means that the IT staff learns by doing, so there is no
awkward learning curve. At every point in the process, there is someone trained to handle
difficulties on-site.
These companies want their core capabilities to be scalable, flexible, robust, redundant, and
available.
This is the hardest option to implement, requires the most planning for the future, and comes
with the most upfront cost, but it’s the only option that lets you utilize the full strength and
flexibility of the cloud.
Replatforming is replacing the application at the code level to make it cloud native. This is a
complete reimagination of the application and usually requires a complete rewrite.
When considering replatforming, think about how fast your company can change. Then think
about how fast it needs to change to keep up with customers and the market.
11
Cloud Computing Unit.1
By making applications truly cloud native, they can be updated and those updates pushed out at
the speed of the cloud. This boosts the speed to change across the board for all aspects of the
business.
Replatformed applications can also be designed to be more modular and thus easier to maintain.
Refactoring the application can save development time, since modules of code from your first
refactored application can be used to extend the capabilities of new applications.
Unlike refactored applications, refactored platforms can work across multiple cloud providers.
This makes it easier to port from one mobile platform to another, which positions replatforming
as the ideal strategy for those looking to develop mobile applications.
We worked with a healthcare company building a net-new mobile application for patient
management, which included sending push notifications on appointments and subscriptions. The
company wanted to build this application as a cloud-native company. However, they already had
internal patient managing applications for the nurses and doctors that complemented the new
applications. The company used replatforming to get their inheritance patient management
application to the cloud, so the application would be ready and waiting while it developed its net-
new mobile application.
Determine your goals before you select your cloud migration model:
Each of the four primary methods we identified comes with its share of advantages, but also with
its disadvantages. Finding the method that matches your organizations goals and needs is the first
step to a successful migration.
AWS is Amazon's cloud web hosting platform which offers fast, flexible, reliable and cost-
effective solutions. It offers a service in the form of building block which can be used to create
and deploy any kind of application in the cloud. It is the most popular as it was the first to
enter the cloud computing space.
Cloudways:
The platform has partnered with top cloud providers including AWS, Google Cloud,
DigitalOcean,etc.
12
Cloud Computing Unit.1
Experience the freedom to build, deploy and manage applications including PHP, WordPress, etc
without requiring any knowledge of cloud server management.
Cloudways users can focus on business growth without worrying about the technical complexities
of server management, security, and maintenance.
DigitalOcean:
Digitalocean's droplet is a scalable computer service. It is more than just virtual machines.
This cloud platform offers add-on storage, security, and monitoring capabilities to run production
applications easily.
Rackspace”
Rackspace is another useful cloud computer service tool. It offers services like hosting web
applications, cloud files, cloud backup, database, and cloud server, etc.
Alibaba Cloud:
Alibaba is the largest Chinese cloud computing company. It is a new platform which created a
global footprint with over 1500 Nodes worldwide of 19 regions and 56 availability zones across
more than 200 countries.
[A content delivery network (CDN) is a system of distributed servers (network) that deliver pages
and other web content to a user, based on the geographic locations of the user, the origin of the
webpage and the content delivery server. ... CDNs also provide protection from large flows in
traffic.]
Microsoft Azure:
Azure is a cloud computing platform which is launched by Microsoft in February 2010. This open
source and flexible cloud platform which helps in development, data storage, service
management & hosting solutions.
Google Cloud is a set of solution and products which includes GCP & G suite. It helps you to solve
all kind of business challenges with ease.
Oracle Cloud:
Oracle Cloud offers innovative and integrated cloud services. It helps you to build, deploy, and
manage workloads in the cloud or on premises. Oracle Cloud also helps companies to transform
their business and reduce complexity.
13
Cloud Computing Unit.1
IBM Cloud:
IBM cloud is a full stack cloud platform which spans public, private and hybrid environments. It is
built with a robust suite of advanced and AI tools.
There is a important growth of cloud adoption across small as well as large enterprises. This has
resulted in a large spectrum of cloud offerings including cloud delivery models and a variety of
cloud computing services that are being provided by cloud hosting companies.
Cloud adoption not only helps improve business processes and enhances the efficiency of IT
infrastructures but also brings down costs of running, upgrading, and maintaining on-site IT
facilities.
Your business-critical data is armed with added security in the cloud environment. In reality, the
data is not actually being placed up in the cloud but is distributed to a number of remote data
center facilities that are owned and operated by third-party service providers. These
establishments consist of climate-controlled rooms to house enterprise-grade servers for
seamless protection and easy accessibility for maintaining business continuity in spite of any tragic
event that may impact the main office of your enterprise.
The cloud data centers are designed to house a multitude of servers for storing data under strict
security controls. The arrangement is aimed at enabling uninterrupted connectivity among vast
networks comprising of millions of machines. Cloud computing is controled by end users as well as
cloud hosting companies for the enhancement of their services.
In order to understand the precise reasons for increased cloud adoption in enterprise setups, we
should have in-depth knowledge about of cloud’s attributes that boost business processes.
Cloud services are designed to set your IT staff free from boring(ordinary) and time-consuming
tasks of maintaining, repairing, and upgrading hardware equipment such as servers. On-site IT
infrastructure in enterprises will be thinner after moving workloads to cloud data center. In the
majority of cases, there will be no need to allocate separate space for housing servers and other IT
equipment.
The direct benefit of cloud computing is associated with reduced capital expenditure as
companies need not invest funds in purchasing costly hardware equipment. improvement of
hardware costs is also backed by freedom from maintenance and repair costs of web servers.
There is a definite reduction in upfront costs of ownership of cost-intensive software as well as
hardware.
14
Cloud Computing Unit.1
Performance with a promise of security:
In comparison with a physical server, a cloud hosting delivers better performance. This is because
established web hosting service providers are in a better position to afford enterprise-grade cloud
servers as against small or medium-sized enterprises.
Cloud hosting providers attach great importance to the security of customers’ digital assets by
spending a significant amount of financial and manpower resources. These providers harden the
defences by the implementation of strict measures such as firewalls, anti-malware and anti-virus
deployments. In addition to this, the host data centers are armed with barrier-like security for
safeguarding physical as well as networking assets.
Greater affordability:
By provisioning top of the line hardware and software resources to customers at affordable prices,
cloud hosting service providers help business enterprises reduce their capital as well as operating
costs without impacting performance.
Cloud services go all out by investing huge sums of money to offer world-class resources to
customers at economical prices. Their efficient staffs are well equipped to look after the routine
tasks as well as technical problems irrespective of the time of the day for all weekdays.
Users of cloud services are allowed to access the optimum amount of resources in response to
resource requirements. This not only assures guaranteed resource availability but also helps
businesses achieve resource optimization for reduction of operating costs.
Cloud-based infrastructure also enables users to access a variety of resources such as applications
or platforms via any internet enabled device, from any location. These services are always
available on round the clock basis for improved efficiency of enterprises. Employees can use a
number of devices including smart-phones, tablets, and laptops to get their hands on a huge
number of files and folders without the need to make a trip to the office. Cloud-based solutions
are inherently flexible and accessible and businesses can easily keep their employees well-
connected with each other for greater efficiency.
On-site IT infrastructures are resource intensive and need to be regularly upgraded and
maintained. In contrast, cloud service providers shoulder the entire responsibility of looking after
the performance of servers, bandwidth, network, and software applications. This also includes
periodic upgrades and security patching of operating systems and other business-critical
applications.
15
Cloud Computing Unit.1
driven by the need to have consistently available, flexible, secure, and well managed IT
infrastructure in the absence of any on-premise facility.
These are some of the valuable benefits of cloud computing that signify the role of cloud
service providers. Therefore, the future of faultless data management is secure in the hands of
established cloud service providers.
A cloud consumer browses the service catalog from a cloud provider, requests the appropriate
service, sets up service contracts with the cloud provider, and uses the service.
To satisfy consumer's expectation several Data centres are established all over the world and each
Data centre contain thousands of servers.
The idle servers and resources in data center wastes huge amount of energy.
An appealing customer experience is essential for consumer business. If you can’t manage
expectations for consumer fast enough, they too may go elsewhere.
1. Flexibility:
Change is the only constant in the digital economy. A hybrid cloud model gives you the flexibility
to adjust and grow.
This flexibility allows you to go after the big opportunities knowing that the location of your
company’s data won’t be an obstacle.
The value is in more than just cost and speed, though these remain key principle for the public
cloud.
The hybrid model also enables you to realize new insights across your entire ecosystem and
quickly move priorities and resources to meet opportunities.
2. Freedom to choose:
Never forget you have options. You need to be able to easily change where an application runs
based on your business needs. Which cloud helps you realize the most value: public, private?
16
Cloud Computing Unit.1
What if that changes? If you’re locked in with a public vendor, for example, moving data without
disturbance can be very costly and time-consuming.
Look for solutions that give you the freedom to choose, with easy application portability
regardless of your architectural environment across any cloud.
Solutions like IBM WebSphere Application Server Version 9 are built to put clients in control, not
cloud providers.
3. Cognitive insights:
Cognitive is the way to outthink the competition and make sense of information. What can your
data do? Bring new customer experiences, new applications and even new business models, for
starters.
IBM Cloud offers a host of accessible cognitive capabilities which you can build into your
applications.
You can rapidly infuse apps with cognitive capabilities to gain operational insights and dazzle your
customers. Use these cognitive capabilities to breathe new life into your existing investments and
extend their value while still putting the customer first.
Earlier, in cloud computing all Service Level Agreements were negotiated between a client and the
service consumer. Nowadays, with the initiation of large utility-like cloud computing providers,
most Service Level Agreements are standardized until a client becomes a large consumer of cloud
services.
Service level agreements are also defined at different levels which are mentioned below:
Customer-based SLA
Service-based SLA
Multilevel SLA
Few Service Level Agreements are enforceable as contracts, but mostly are agreements or
contracts which are more along the lines of an Operating Level Agreement (OLA) and may not
have the restriction of law.
It is fine to have an advocate review the documents before making a major agreement to the
cloud service provider.
17
Cloud Computing Unit.1
Service Level Agreements usually specify some parameters which are mentioned below:
In any case, if a cloud service provider fails to meet the stated targets of minimums then the
provider has to pay the penalty to the cloud service consumer as per the agreement. So, Service
Level Agreements are like insurance policies in which the corporation has to pay as per the
agreements if any casualty occurs.
Microsoft publishes the Service Level Agreements linked with the Windows Azure Platform
components, which is demonstrative of industry practice for cloud service vendors.
Window Azure has different SLA’s for compute and storage. For compute, there is a
guarantee that when a client deploys two or more role instances in separate fault and
upgrade domains, client’s internet facing roles will have external connectivity minimum
99.95% of the time. Moreover, all of the role instances of the client are monitored and
there is guarantee of detection 99.9% of the time when a role instance’s process is not
runs and initiates properly.
2. SQL Azure SLA –
SQL Azure clients will have connectivity between the database and internet gateway of
SQL Azure. SQL Azure will handle a “Monthly Availability” of 99.9% within a month.
Monthly Availability Proportion for a particular tenant database is the ratio of the time the
database was available to customers to the total time in a month. Time is measured in
some intervals of minutes in a 30-day monthly cycle. Availability is always remunerated for
a complete month. A portion of time is marked as unavailable if the customer’s attempts
to connect to a database are denied by the SQL Azure gateway.
Service Level Agreements are based on the usage model. Frequently, cloud providers charge their
pay-as-per-use resources at a premium and deploy standards Service Level Agreements only for
that purpose.
Clients can also subscribe at different levels that guarantees access to a particular amount of
purchased resources.
18
Cloud Computing Unit.1
The Service Level Agreements (SLAs) attached to a subscription many times offer various terms
and conditions. If client requires access to a particular level of resources, then the client need to
subscribe to a service.
19
Advantages and Disadvantages of Cloud Computing
Advantages:
Business operations are being transformed by the transformational technology known as cloud
computing. With its extensive advantages and possibilities, cloud computing has emerged as a crucial
strategic tool for businesses in a range of sectors. Businesses can take advantage of various benefits
that promote development, innovation, and operational effectiveness by leveraging the power of the
cloud.
Cloud computing offers a quick and easy method for data backup and restoration. Businesses may
simply access and restore their data in the event of any data loss or system failure by keeping it in the
cloud.
Improved Collaboration:
Collaboration is improved because cloud technologies make it possible for teams to share
information easily. Multiple users may work together on documents, projects, and data thanks to
shared storage in the cloud, enhancing productivity and teamwork.
Excellent Accessibility:
Access to information stored in the cloud is made possible. Users can access their data from
anywhere in the world with an internet connection, making remote work, flexibility, and effective
operations possible.
Cost-effective Maintenance:
Organizations using cloud computing can save money on both hardware and software upkeep.
Because cloud service providers manage the maintenance and updates, businesses no longer need to
make costly infrastructure investments or set aside resources for continuous maintenance.
Cloud service providers take care of infrastructure upkeep, security patches, and updates, freeing
organizations from having to handle these duties themselves.
This frees up IT teams' time and resources to work on higher-value projects like application
development, data analysis, or strategic initiatives rather than wasting them on rote upkeep and
updates.
Mobility:
Cloud computing makes it simple for mobile devices to access data. Utilizing smartphones and
tablets, users can easily access and control their cloud-based applications and data, increasing their
mobility and productivity.
Pay-per-use Model:
Cloud computing uses a pay-per-use business model that enables companies to only pay for the
services they really utilize. This method is affordable, eliminates the need for up-front investments,
and offers budget management flexibility for IT.
Businesses can virtually store and manage a limitless amount of data in the cloud. The cloud offers a
scalable and centralized storage option for all types of data, including documents, photos, audio,
video, and other kinds of files.
Cloud computing places a high focus on data security. To guarantee that data is handled and stored
safely, cloud service providers offer cutting-edge security features like encryption, access limits, and
regular security audits. Businesses can rest easy knowing that their important data is secure.
Cloud computing provides reliable options for these two issues. Businesses can quickly bounce back
from any unforeseen disasters or disruptions thanks to data redundancy, backup systems, and
geographically dispersed data centers.
Agility and Innovation:
Businesses can continue to be innovative and nimble thanks to cloud computing. Organizations may
quickly embrace new solutions, test out emerging trends, and promote corporate growth with access
to a variety of cloud-based tools, services, and technology.
Green Computing:
By maximizing the use of computer resources, lowering energy use, and minimizing e-waste, cloud
computing may support environmental sustainability.
By utilizing technologies like virtualization and load balancing to maximize the use of computer
resources, cloud providers can operate large-scale data centers built for energy efficiency, resulting in
lower energy usage and a smaller carbon footprint.
These benefits of cloud computing give companies the ability to use cutting-edge technology offered
by cloud service providers while maximizing productivity, cost savings, scalability, and data security.
They also enable them to concentrate on their core capabilities.
When we talk about the "disadvantages of cloud computing," we're talking about any potential
drawbacks or difficulties that businesses might have when utilizing cloud computing services. These
drawbacks draw attention to some restrictions or risks related to cloud computing that businesses
should take into account before making a choice.
o Internet Dependency:
A dependable and fast internet connection is essential for cloud computing. Business operations may
be delayed or interrupted if there are connectivity problems or interruptions in the internet service
that affect access to cloud services and data.
Using standardized services and platforms offered by the cloud service provider is a common part of
cloud computing. As a result, organizations may have less ability to customize and control their
infrastructure, applications, and security measures. It may be difficult for some organizations to
modify cloud services to precisely match their needs if they have special requirements or compliance
requirements.
Concerns about data security and privacy arise when sensitive data is stored on the cloud. Businesses
must have faith in the cloud service provider's security procedures, data encryption, access controls,
and regulatory compliance. Unauthorized access to data or data breaches can have serious
repercussions, including financial loss, reputational harm, and legal obligations.
o Hidden Costs and Pricing Models:
Although pay-as-you-go models and lower upfront costs make cloud computing more affordable,
businesses should be wary of hidden charges. Data transfer fees, additional storage costs, fees for
specialized support or technical assistance, and expenses related to regulatory compliance are a few
examples.
When data is stored in the cloud, it frequently sits in numerous data centers around the globe that
may be governed by multiple legal systems and data protection laws. This may pose compliance
issues, especially if some sectors of the economy or nations have stringent data sovereignty laws.
Organizations should carry out a comprehensive risk assessment, thoroughly examine the
dependability and security procedures of possible cloud service providers, and build backup and
disaster recovery strategies to counteract these drawbacks.
In the ever-evolving scene of networking administration and cloud computing, organizations are
progressively turning to innovative solution for smooth out their tasks, improve productivity, and
reduce costs. Network as a Service (NaaS) is one such solution that has gained popularity recently.
NaaS addresses a change in perspective in the way businesses manage and use their networking
infrastructure. By utilizing cloud-based technologies, NaaS offers organizations the capacity to
access, provision, and manage network assets on-request, all through an incorporated and
effectively open platform.
In this complete guide, we dig into the complexities of Network as a Service, exploring its core
ideas, primary terminologies, functional cycles, and practical applications, from understanding the
basic standards of NaaS to examining real-world and use cases, this guide expects to furnish you
with the information expected to successfully explore and harness the power of NaaS.
In order to assist you in comprehending how Network as a Service (NaaS) functions and its
potential advantages for businesses, this guide will cover key terminologies, operational
procedures, and practical examples.
What is NaaS ?
Network as a Service (NaaS) is a cloud computing model that provides organizations with on-
demand access to networking resources over the Internet. In NaaS, traditional systems networking
parts like routers, switches, firewalls, and different devices are virtualized and given as services
through cloud service providers. This makes it possible for businesses to access and manage their
network infrastructure without having to purchase any physical hardware. As a result, there is a
reduction in operational complexity, costs associated with maintenance, and initial investment.
Pay-per-use billing, in which organizations are charged based on how much they actually use
networking services, and centralized management, where users can monitor and manage their
network resources from a single dashboard, are two important aspects of NaaS, this utilization-
based evaluating model offers cost proficiency and adaptability, as associations just compensate
for the assets they consume.
Through a cloud-based platform, Network as a Service (NaaS) provides organizations with internet
access to networking resources, this is the way NaaS typically works:
• Security and Compliance: Network infrastructure and data are protected from
unauthorized access, data breaches, and other security threats by NaaS platforms'
extensive security measures. They additionally offer consistence confirmations and
adherence to industry principles to ensure data privacy, respectability, and accessibility.
• Scalability and Flexibility: Organizations can dynamically scale their network infrastructure
in response to shifting demand with NaaS. They can add or remove virtual networking
components, adjust configurations, and allocate resources depending on the situation to
support business development, seasonal fluctuations, or new initiative.
Network as a Service (NaaS) offers an extensive variety of networking resources that organizations
can access and use over the internet. Some of the key types of NaaS resources include:
Virtual Networks
• NaaS platforms give virtual routers and switches that work with the routing and switching
of network traffic inside virtualized conditions, these product based routing and switching
parts empower associations to define and manage network traffic streams, carry out
routing policies, and ensure proficient data transmission.
Load Balancers
• Virtual private networks given by NaaS platform empower secure correspondence and
data exchange between remote users, branch offices, and cloud-based resources, VPNs
encrypt network traffic, verify users, and lay out private and secure associations over open
network like the internet.
Firewalls
• Scalability: Businesses can easily up or down their network infrastructure using NaaS,
depending on their requirements, businesses that are experiencing rapid growth or have
varying demands especially benefit from this flexibility.
• Cost-Efficiency: Businesses can save a lot of money upfront by using NaaS, which typically
operates on a subscription or pay-per-use model, because businesses only pay for the
services they use, this also makes it easier to predict costs.
• Accessibility: NaaS makes network resources accessible from anywhere with an internet
connection, making it possible to work remotely and for teams with different locations to
work together.
• Reliability: Service level agreements (SLAs) that ensures a certain level of uptime and
performance are provided by many NaaS providers. Businesses that heavily rely on their
network infrastructure for critical operations may particularly benefit from this.
• Global Reach: NaaS providers often have a global presence, businesses can easily expand
their network's reach to new locations without making significant infrastructure
investment.
• Compliance: When it comes to data protection and privacy, NaaS providers assist
businesses by adhering to industry standards and regulations.
• Vendor Lock-In: Adopting NaaS solutions from a single service provider may result in
vendor lock-in, making it difficult or costly for businesses to switch providers or return to
in-house network management. Organizations that lack flexibility may be unable to take
advantage of emerging technologies or adapt to shifting business requirements.
• Reliability and Availability: NaaS solutions may still experience downtime due to network
issues, maintenance activities, or failures in the service provider's infrastructure, despite
the promises of high availability and redundancy. To reduce the likelihood of disruptions to
their operations, businesses must evaluate the provider's reliability and availability history.
• Potential Performance Issues: Limitations on bandwidth, latency, and congestion can have
an impact on network performance, which can vary depending on the service provider and
the underlying infrastructure, the NaaS solution's performance should be evaluated to see
if it meets the needs of requirements
• Limited Control: With NaaS, organizations relinquish a level of control over their network
infrastructure to the service provider. Concerns regarding customization, configuration,
and troubleshooting may result from this lack of control, particularly for businesses with
particular network requirements
• Integration Challenges: If the NaaS solution does not seamlessly integrate with other
technologies utilized by the business, it can be difficult to integrate NaaS with existing IT
systems, applications, and workflows. Compatibility issues may arise, necessitating
additional integration efforts and resources.
A Some of the top Network as a Service (NaaS) providers in the cloud computing industry include:
PROVIDER DESCRIPTION
AWS offers various networking services like Virtual Private Cloud (VPC),
Amazon Web AWS Transit Gateway, and AWS Global Accelerator to provide networking
Services (AWS) solutions
Google Cloud GCP offers various networking services like Cloud VPN, Cloud Load
Platform (GCP) Balancing and Virtual Private Cloud (VPC) to create global networks
IBM Cloud provides networking services like VPC, IBM Cloud Load
Balancer and IBM Cloud VPN offering to secure and reliable networks
IBM Cloud options in cloud
OCI offers networking services like VCN that is Virtual Cloud Network,
Oracle Cloud VPN Connect, FastConnect and load balancing these are enabling high
Infrastructure (OCI) available and scalable networks in the cloud.
PROVIDER DESCRIPTION
Cisco SD-WAN Cisco offers SD-WAN solutions this was provide secure connection.
Juniper Contrail SD- Juniper Network offers contrail SD-WAN this provide automated
WAN networking and security to simply network infrastructure.
Conclusion on NaaS
Network as a Service (NaaS) provides businesses with a compelling set of benefits, which include
simplified management, scalability, accessibility, and cost efficiency. businesses can build agile,
secure, and high-performance networks in the cloud by utilizing NaaS solutions from leading
providers like Amazon Web Services (AWS), Microsoft Azure, Google Cloud Platform (GCP), and
others.
However, it is essential to acknowledge that NaaS has its own set of difficulties and considerations.
Data security concerns, reliance on service providers, limited control, potential performance
issues, and the need to comply with regulatory requirements are just a few examples, when
adopting NaaS solutions, businesses should also carefully consider the risks, integration difficulties,
and long-term costs of vendor lock-in.
In this complete guide, we dig into the complexities of Network as a Service, exploring its core
ideas, primary terminologies, functional cycles, and practical applications, from understanding the
basic standards of NaaS to examining real-world and use cases, this guide expects to furnish you
with the information expected to successfully explore and harness the power of NaaS.
Like SaaS, PaaS, and IaaS of cloud computing, we can consider DBaaS (also known as Managed
Database Service) as a cloud computing service. It allows users associated with database activities to
access and use a cloud database system without purchasing it.
1. DBaaS and cloud databases come under Software as a Service (SaaS) whose demand is
growing so fast
3. As we know cloud computing services are like pay-per-use so DBaaS is also based on the
same payment structure like how much you will use just pay for your usage.
4. This DBaaS provides the same function as standard traditional and relational database
models.
6. DBaaS consists of an info manager element, that controls all underlying info instances via
API.
7. This API is accessible to the user through a management console, typically an online
application, that the user might use to manage and assemble the info and even provision or
deprovision info instances.
Features of DBaaS
1. Scalability: DBaaS platforms supports scaling resources automatically up and down based on
the demand and the amount of storage and processing can be easily adjusted by the users
without downtime.
2. High Availability: In DBaaS, data is replicated over multiple servers to ensure continuous
availability.
4. Caching: Frequently accessed data is cache to speed up the query response time.
5. Regular Updates: The service provider handles regular updates, security fixes, and patches.
6. Shared Infrastructure: Multiple tenants share the same infrastructure, thus reducing the
costs.
7. Support for Multiple Database Engines: DBaaS provides support for multiple database
engines like MySQL, PostgreSQL, MongoDB, etc.
8. Cost-efficiency: The charges are based on the actual usage, thus this is more cost-effective
than the traditional database solutions.
DBaaS is utilized by a wide range of organizations and users. Below are some of the user groups that
utilize DBaaS:
1. SaaS Providers: They can leverage DBaaS efficiently to manage multi-tenant architectures,
ensuring high availability and reliability.
2. Data Analysts and Scientists: They can utilize the features of DBaaS like performance and
scalability while working with large datasets for analyzing and machine learning.
3. Startups: Startups have limited IT resources and can benefit from cost efficiency of DBaaS.
4. Small and Medium-Sized Enterprises (SMEs): SMEs can use DBaaS to avoid cost and
maintenance expenses of the traditional database infrastructure.
5. Large Enterprises: Large enterprises can benefit from the performance optimization and
scalability of DBaaS with complex and diverse workloads.
DBaaS providers host data and database infrastructure while enabling access through API endpoints.
They offer features like alerts, notifications, monitoring, constant support, and geo-replication for
backups and availability.
1. DBaaS is delivered over the internet and users can access it through a web-based interface or
API.
2. It can help to reduce the complexity and cost of managing databases.
3. It allows the development team to deploy and access a database without worrying about
hardware purchase, hardware setup, database installation, and configuration, database
maintenance and administration.
4. The organizations can free up the resources to focus on strategic initiatives by offloading
tasks like data backup, recovery, etc. to a service provider.
5. It provides disaster recovery capability and helps to improve the availability and performance
of databases.
Infrastructure as a
Database as a Service Platform as a Service
Full Form Service
It is delivered as managed
In this access to cloud It is a complete
database services where
infrastructure resources development and
the provider takes care of
is rented as individual deployment
the patching, upgrading,
services from a cloud environment in the
and backing up the
service provider. cloud.
Definition database.
Security is managed by
Security is managed by Security is managed by
provider and is
provider and it is focussed user and have high
focussed mainly on
on DB security. control over security.
Security application security.
Below are some of the key factors to consider when choosing a DBaaS:
1. Database Type: Determine the type of database needed. If you need a relational database
like MySQL or a non-relational database like MongoDB.
3. Uptime: Check provider’s Service Level Agreement to check for uptime guarantees.
4. Security: Check to ensure that DBaaS offers robust security features like data encryption,
Identity and Access Management, and network security.
5. Cost: Consider the cost related to storage, data transfer, understand the pricing model and
ensure that it aligns with your budget.
6. Documentation: Look for comprehensive and detailed documentation to help with setup
and troubleshooting.
There are numerous DBaaS tools and vendors available, each catering to different needs. Here are
some of prominent DBaaS tools and vendors:
1. Amazon RDS: This offers managed relational databases with support for multiple database
engines, including MySQL, PostgreSQL, Oracle, MariaDB, and Microsoft SQL server. It offers
automated backups, scaling, and patching.
2. Google Cloud SQL: It is fully managed relational relational database service for MySQL,
PostgreSQL, and SQL server. It is best for businesses looking for seamless integration with
Google Cloud Sevices. It offers services like automated backups, failover, scaling, and
maintenance.
3. MongoDB Atlas: It is a fully managed cloud database service for MongoDB, offering flexible
NoSQL database capabilities. It is suited for applications requiring flexible schema designs,
high performance, and scalability, specially for unstructured data.
5. Microsoft Azure SQL Database: It is a fully managed relational database service with AI-
powered features for optimization and performance tuning. It is well-suited for enterprises
seeking a robust, scalable SQL database solution.
Benefits of DBaaS
DBaaS offers numerous benefits to the organizations, enhancing efficiency, cost-efficiency, and many
more. Below are some of the benefits of using DBaaS:
1. Scalability: DBaaS can automatically scale resources up and down based on the demand,
ensuring cost-efficiency and optimal performance.
2. High Availability: DBaaS services ensures high availability as they often include built-in
redundancy and failover mechanisms.
3. Reliability: DBaaS services ensures high reliability and business continuity with automated
backup and disaster recovery solutions to help protect data against loss.
4. Cost-efficiency: DBaaS services offer pay-as-you-go pricing that allow organizations to pay
only for the resources they use.
6. Flexibility: Many DBaaS providers support a variety of database engines like SQL, NoSQL,
thus giving organizations the flexibility to choose the best tool for their needs.
Limitations of DBaaS
Database as a Service (DBaaS) offers many benefits but it also hs several limitations. Here are some
of the limitations of DBaaS:
2. Complex Customizations: Sometimes it can be difficult to optimize the database for specific
needs as users have limited control over the underlying hardware and software.
3. Hidden Costs: Additional storage charges and charges for premium features can lead to
unexpected expenses.
4. Complex Data Portability: Transferring huge volumes of data between providers or back to
on-premises systems can be complex and costly.
5. Time-consuming Data Transfer: Migrating large datasets to a DBaaS platform can be time-
consuming and may require downtime.
CLOUD COMPUTING Unit.2
Types of Cloud
Public Cloud:
Public Cloud allows systems and services to be easily accessible to general public. The IT
giants such as Google, Amazon and Microsoft offer cloud services via Internet.
Benefits:
There are many benefits of deploying cloud as public cloud model. The following diagram
shows some of those benefits:
Cost Effective
Since public cloud shares same resources with large number of customers it turns out
inexpensive.
Reliability
The public cloud employs large number of resources from different locations. If any of the
resources fails, public cloud can employ another one.
Flexibility
The public cloud can smoothly integrate with private cloud, which gives customers a flexible
approach.
1
CLOUD COMPUTING Unit.2
Location Independence
Public cloud services are delivered through Internet, ensuring location independence.
Public cloud is also based on pay-per-use model and resources are accessible whenever
customer needs them.
High Scalability
Cloud resources are made available on demand from a pool of resources, i.e., they can be
scaled up or down according the requirement.
Disadvantages:
Low Security
In public cloud model, data is hosted off-site and resources are shared publicly, therefore
does not ensure higher level of security.
Less Customizable
Private Cloud:
Private Cloud allows systems and services to be accessible within an organization.
The Private Cloud is operated only within a single organization. However, it may be
managed internally by the organization itself or by third-party.
2
CLOUD COMPUTING Unit.2
Benefits:
There are many benefits of deploying cloud as private cloud model. The following diagram
shows some of those benefits:
Private cloud operations are not available to general public and resources are shared from
distinct pool of resources. Therefore, it ensures high security and privacy.
More Control
The private cloud has more control on its resources and hardware than public cloud because
it is accessed only within an organization.
The private cloud resources are not as cost effective as resources in public clouds but they
offer more efficiency than public cloud resources.
Disadvantages:
The private cloud is only accessible locally and is very difficult to deploy globally.
High Priced
Limited Scalability
The private cloud can be scaled only within capacity of internal hosted resources.
Additional Skills
Hybrid Cloud:
Non-critical activities are performed using public cloud while the critical activities are
performed using private cloud. The Hybrid Cloud Model is shown in the diagram below.
3
CLOUD COMPUTING Unit.2
Benefits:
There are many benefits of deploying cloud as hybrid cloud model. The following diagram
shows some of those benefits:
Scalability
It offers features of both, the public cloud scalability and the private cloud scalability.
Flexibility
Cost Efficiency
Public clouds are more cost effective than private ones. Therefore, hybrid clouds can be cost
saving.
Security
Disadvantages:
Networking Issues
Security Compliance
It is necessary to ensure that cloud services are compliant with security policies of the
organization.
Infrastructure Dependency
4
CLOUD COMPUTING Unit.2
Community Cloud:
Benefits:
Cost Effective
Community cloud offers same advantages as that of private cloud at low cost.
Security
The community cloud is comparatively more secure than the public cloud but less secured
than the private cloud.
Issues:
Since all data is located at one place, one must be careful in storing data in
community cloud because it might be accessible to others.
It is also challenging to allocate responsibilities of governance, security and cost
among organizations.
5
CLOUD COMPUTING Unit.2
[A fat client (also called heavy, rich or thick client) is a computer (clients), in client–server
architecture or networks, that typically provides rich functionality independent of the
central server. It is Originally known as just a "client" or "thick client," the name is contrasted
to thin client, which describes a computer heavily dependent on a server's applications.]
Cloud Computing architecture comprises of many cloud components, which are loosely
coupled. We can broadly divide the cloud architecture into two parts:
Front End
Back End
The following diagram shows the graphical view of cloud computing architecture:
Front End:
The front end refers to the client part of cloud computing system. It consists of interfaces
and applications that are required to access the cloud computing platforms, Example - Web
Browser.
6
CLOUD COMPUTING Unit.2
Back End:
The back End refers to the cloud itself. It consists of all the resources required to provide
cloud computing services. It comprises of huge data storage, virtual machines, security
mechanism, services, deployment models, servers, etc.
Note:
It is the responsibility of the back end to provide built-in security mechanism, traffic
control and protocols.
The server employs certain protocols known as middleware, which help the
connected devices to communicate with each other.
Hypervisor:
Management Software:
Deployment Software:
7
CLOUD COMPUTING Unit.2
Network:
It is the key component of cloud infrastructure. It allows to connect cloud services over the
Internet. It is also possible to deliver network as a utility over the Internet, which means, the
customer can customize the network route and protocol.
Server:
The server helps to compute the resource sharing and offers other services such as resource
allocation and de-allocation, monitoring the resources, providing security etc.
Storage:
Cloud keeps multiple replicas of storage. If one of the storage resources fails, then it can be
extracted from another one, which makes cloud computing more reliable.
Infrastructural Constraints
Fundamental constraints that cloud infrastructure should implement are shown in the
following diagram:
Transparency:
Virtualization is the key to share resources in cloud environment. But it is not possible to
satisfy the demand with single resource or server. Therefore, there must be transparency in
resources, load balancing and application, so that we can scale them on demand.
Scalability:
Intelligent Monitoring:
To achieve transparency and scalability, application solution delivery will need to be capable
of intelligent monitoring.
Security:
The mega data center in the cloud should be securely architected. Also the control node, an
entry point in mega data center, also needs to be secure.
8
CLOUD COMPUTING Unit.2
Virtualization
Virtualization in Cloud Computing is a process in which the user of cloud shares the data
present in the cloud which can be application software etc. It provides a virtual environment
in the cloud which can be software hardware or any other thing.
Virtualization is a technique how to separate a service from the underlying physical delivery
of that service. It is the process of creating a virtual version of something like computer
hardware.
With the help of Virtualization multiple operating systems and applications can run on same
Machine and its same hardware at the same time increasing the utilization and flexibility of
hardware.
In other words, One of the main cost effective, hardware reducing, energy saving
techniques used by cloud providers is virtualization.
The term virtualization is often synonymous with hardware virtualization, which plays a
fundamental role in efficiently delivering Infrastructure-as-a-Service (IaaS) solutions for
cloud computing. Moreover, virtualization technologies provide a virtual environment for
not only executing applications but also for storage, memory, and networking.
9
CLOUD COMPUTING Unit.2
The machine on which the virtual machine is going to be build is known as Host Machine
and that virtual machine is referred as a Guest Machine.
BENEFITS OF VIRTUALIZATION:
Types of Virtualization:
1. Application Virtualization.
2. Network Virtualization.
3. Desktop Virtualization.
4. Storage Virtualization.
1. Application Virtualization:
Application virtualization helps user to have a remote access of an application from a server.
The server stores all personal information and other characteristics of the application, but
can still run on a local workstation through internet. Example of this would be a user who
needs to run two different versions of the same software. Technologies that use application
virtualization are hosted applications and packaged applications.
2. Network Virtualization:
10
CLOUD COMPUTING Unit.2
3. Desktop Virtualization:
Desktop virtualization allows the users’ OS to be remotely stored on a server in the data
center. It allows the user to access their desktop virtually, from any location by different
machine. Users who wants specific operating systems other than Windows Server will need
to have a virtual desktop. Main benefits of desktop virtualization are user mobility,
portability, easy management of software installation, updates and patches.
4. Storage Virtualization:
Storage virtualization is an array of servers that are managed by a virtual storage system.
The servers aren’t aware of exactly where their data is stored. It makes managing storage
from multiple sources to be managed and utilized as a single repository. storage
virtualization software maintains smooth operations, consistent performance and a
continuous suite of advanced functions despite changes, break down and differences in the
underlying equipment.
Hypervisor
Hypervisor is a form of virtualization software used in Cloud hosting to divide and allocate
the resources on various pieces of hardware.
Types of Hypervisor:-
TYPE-1 Hypervisor:
Hypervisor runs directly on underlying host system. It is also known as “Native Hypervisor”
or “Bare metal hypervisor”. It dose not require any base server operating system. It has
direct access to hardware resources.
TYPE-2 Hypervisor:
A Host operating system runs on undrlying host system. It is also known as ‘Hosted
Hypervisor”.
11
CLOUD COMPUTING Unit.2
Type 1 hypervisors offer much better performance than Type 2 because there’s no middle
layer, making them the logical choice for mission-critical applications and workloads.
But that’s not to say that hosted hypervisors don’t have their place – they’re much simpler
to set up, so they’re a good.
One of the best ways to determine which hypervisor meets your needs is to compare their
performance metrics. These include CPU overhead, amount of maximum host and guest
memory, and support for virtual processors.
1. Understand your needs: The company and its applications are the reason for the data
centre (and your job). Besides your company’s needs, you (and your co-workers in IT) also
have your own needs.
a. Flexibility
b. Scalability
c. Usability
d. Availability
e. Reliability
f. Efficiency
g. Reliable support
2. The cost of a hypervisor: For many buyers, the toughest part of choosing a hypervisor is
striking the right balance between cost and functionality. While a number of entry-level
solutions are free, or practically free, the prices at the opposite end of the market can be
overwhelming. Licensing frameworks also vary, so it’s important to be aware of exactly what
you’re getting for your money.
3. Virtual machine performance: Virtual systems should meet or exceed the performance of
their physical counterparts, at least in relation to the applications within each server.
Everything beyond meeting this benchmark is profit.
4. Ecosystem: It’s tempting to overlook the role of a hypervisor’s ecosystem – that is, the
availability of documentation, support, training, third-party developers and consultancies,
and so on – in determining whether or not a solution is cost-effective in the long term.
5. Test for yourself: You can gain basic experience from your existing desktop or laptop.
You can run both Hypervisor to create a nice virtual learning and testing environment.
12
CLOUD COMPUTING Unit.2
There are 3 main modules coordinate in order to follow the fundamental hardware:
1. Dispatcher
2. Allocator
3. Interpreter
DISPATCHER:
The dispatcher behaves like the entry point of the monitor and reroutes the instructions of
the virtual machine instance to one of the other two modules.
ALLOCATOR:
The allocator is responsible for deciding the system resources to be provided to the virtual
machine instance. It means whenever virtual machine tries to execute an instruction that
results in changing the machine resources associated with the virtual machine, the allocator
is invoked by the dispatcher.
INTERPRETER:
The interpreter module consists of interpreter routines. These are executed, whenever
virtual machine executes a privileged instruction.
13
CLOUD COMPUTING Unit.2
CPU Virtualization
Virtualization of the hardware is known as CPU Virtualization. This is where any hardware
platform that can be controlled by the user or a guest software over a virtual machine on a
platform, virtually. This is not limited to guest software but also several operating systems.
The critical instructions are divided into three categories: privileged instructions, control-
sensitive instructions, and behavior-sensitive instructions.
Privileged instructions execute in a privileged mode and will be attentive if executed outside
this mode. Control-sensitive instructions attempt to change the configuration of resources
used. Behavior-sensitive instructions have different behaviors depending on the
configuration of resources, including the load and store operations over the virtual memory.
Memory Virtualization
Virtual memory virtualization is similar to the virtual memory support provided by modern
operating systems.
All modern x86 CPUs include a memory management unit (MMU) and a translation look
aside buffer (TLB) to optimize virtual memory performance. However, in a virtual execution
environment, virtual memory virtualization involves sharing the physical system memory in
RAM and dynamically allocating it to the physical memory of the VMs.
That means a two-stage mapping process should be maintained by the guest OS and the
VMM, respectively: virtual memory to physical memory and physical memory to machine
memory.
14
CLOUD COMPUTING Unit.2
The VMM is responsible for mapping the guest physical memory to the actual machine
memory.
Since each page table of the guest OS has a separate page table in the VMM corresponding
to it, the VMM page table is called the shadow page table.
Nested page tables add another layer of indirection to virtual memory. The MMU already
handles virtual-to-physical translations as defined by the OS. Then the physical memory
addresses are translated to machine addresses using another set of page tables defined by
the hypervisor.
Since modern operating systems maintain a set of page tables for every process, the shadow
page tables will get flooded. Consequently, the performance overhead and cost of memory
will be very high.
When the guest OS changes the virtual memory to a physical memory mapping, the VMM
updates the shadow page tables to enable a direct lookup.
I/O Virtualization
I/O virtualization involves managing the routing of I/O requests between virtual devices and
the shared physical hardware.
There are three ways to implement I/O virtualization: full device emulation, para-
virtualization, and direct I/O.
Full device emulation is the first approach for I/O virtualization. Generally, this approach
emulates well-known, real-world devices.
The I/O access requests of the guest OS are trapped in the VMM which interacts with the
I/O devices.
A single hardware device can be shared by multiple VMs that run concurrently. However,
software emulation runs much slower than the hardware.
The para-virtualization method of I/O virtualization is typically used in Xen. [Xen Project is a
type-1 hypervisor, providing services that allow multiple computer operating systems to execute on
the same computer hardware concurrently.]
It is also known as the split driver model consisting of a front end driver and a backend
driver. The frontend driver is running in Domain U and the backend driver is running in
Domain 0. They interact with each other via a block of shared memory.
15
CLOUD COMPUTING Unit.2
The frontend driver manages the I/O requests of the guest OS and the backend driver is
responsible for managing the real I/O devices and multiplexing the I/O data of different
VMs.
Although para-I/O-virtualization achieves better device performance than full device
emulation, it comes with a higher CPU overhead.
Direct I/O virtualization lets the VM access devices directly. It can achieve close-to-native
performance without high CPU costs. However, current direct I/O virtualization
implementations focus on networking for mainframes.
There are a lot of challenges for commodity hardware devices. For example, when a physical
device is reclaimed (required by workload migration) for later reassignment, it may have
been set to an arbitrary state that can function incorrectly or even crash the whole system.
Since software-based I/O virtualization requires a very high overhead of device emulation,
hardware-assisted I/O virtualization is critical.
Amazon’s Elastic Compute Cloud (EC2) is a good example of a web service that provides
elastic computing power in a cloud. EC2 permits customers to create VMs and to manage
user accounts over the time of their use.
Most virtualization platforms, including XenServer and VMware ESX Server, support a
bridging mode which allows all domains to appear on the network as individual hosts. By
using this mode, VMs can communicate with one another freely through the virtual network
interface card and configure the network automatically.
Virtual clusters are built with VMs installed at distributed servers from one or more physical
clusters.
The VMs in a virtual cluster are interconnected logically by a virtual network across several
physical networks.
Each virtual cluster is formed with physical machines or a VM hosted by multiple physical
clusters. The virtual cluster boundaries are shown as distinct boundaries.
16
CLOUD COMPUTING Unit.2
The provisioning of VMs to a virtual cluster is done dynamically to have the following
interesting properties:
The virtual cluster nodes can be either physical or virtual machines. Multiple VMs
running with different OS can be deployed on the same physical node.
• A VM runs with a guest OS, which is often different from the host OS, that manages the
resources in the physical machine, where the VM is implemented.
• The purpose of using VMs is to consolidate multiple functionalities on the same server.
This will greatly enhance server utilization and application flexibility.
• VMs can be colonized (replicated) in multiple servers for the purpose of promoting
distributed parallelism, fault tolerance, and disaster recovery.
• The size (number of nodes) of a virtual cluster can grow or shrink dynamically, similar to
the way an overlay network varies in size in a peer-to-peer (P2P) network.
• The failure of any physical nodes may disable some VMs installed on the failing nodes.
But the failure of VMs will not pull down the host system.
The system should have the capability of fast deployment. Here, deployment means two
things: to construct and distribute software stacks (OS, libraries, applications) to a physical
node inside clusters as fast as possible, and to quickly switch runtime environments from
one user’s virtual cluster to another user’s virtual cluster.
17
CLOUD COMPUTING Unit.2
If one user finishes using his system, the corresponding virtual cluster should shut down or
suspend quickly to save the resources to run other VMs for other users.
The concept of “green computing” has attracted much attention recently. However,
previous approaches have focused on saving the energy cost of components in a single
workstation without a global vision. Consequently, they do not necessarily reduce the
power consumption of the whole cluster.
The live migration of VMs allows workloads of one node to transfer to another node.
However, it does not guarantee that VMs can randomly migrate among themselves. In fact,
the probable overhead caused by live migrations of VMs cannot be ignored.
The overhead may have serious negative effects on cluster utilization and throughput issues.
Therefore, the challenge is to determine how to design migration strategies to implement
green computing without influencing the performance of clusters.
Mapping VMs onto the most appropriate physical node should promote performance.
Dynamically adjusting loads among nodes by live migration of VMs is desired, when the
loads on cluster nodes become quite unbalanced.
The template VM can be distributed to several physical hosts in the cluster to customize the
VMs. In addition, existing software packages reduce the time for customization as well as
switching virtual environments.
Basically, there are four steps to deploy a group of VMs onto a target cluster: preparing the
disk image, configuring the VMs, choosing the destination nodes, and executing the VM
deployment command on every host.
Many systems use templates to simplify the disk image preparation process. A template is a
disk image that includes a preinstalled operating system with or without certain application
software.
Users choose a proper template according to their requirements and make a duplicate of it
as their own disk image. Templates could implement the COW (Copy on Write) format. A
new COW backup file is very small and easy to create and transfer. Therefore, it definitely
18
CLOUD COMPUTING Unit.2
reduces disk space consumption. In addition, VM deployment time is much shorter than
that of copying the whole raw image file.
Every VM is configured with a name, disk image, network setting, and allocated CPU and
memory. One needs to record each VM configuration into a file. However, this method is
inefficient when managing a large group of VMs.
In a cluster built with mixed nodes of host and guest systems, the normal method of
operation is to run everything on the physical machine. When a VM fails, its role could be
replaced by another VM on a different node, as long as they both run with the same guest
OS.
In other words, a physical node can fail over to a VM on another host. This is different from
physical-to-physical failover in a traditional physical cluster. The advantage is enhanced
failover flexibility.
The potential drawback is that a VM must stop playing its role if its residing host node fails.
However, this problem can be mitigated with VM life migration. The migration copies the
VM state file from the storage area to the host machine.
First, you can use a guest-based manager, by which the cluster manager resides on a guest
system. In this case, multiple VMs form a virtual cluster. For example, openMosix is an open
source Linux cluster running different guest systems on top of the Xen hypervisor. Another
example is Sun’s cluster Oasis, an experimental Solaris cluster of VMs supported by a
VMware VMM.
Second, you can build a cluster manager on the host systems. The host-based manager
supervises the guest systems and can restart the guest system on another physical machine.
A good example is the VMware HA system that can restart a guest system after failure.
These two cluster management systems are either guest-only or host-only, but they do not
mix.
A third way to manage a virtual cluster is to use an independent cluster manager on both
the host and guest systems. This will make infrastructure management more complex,
however.
Finally, you can use an integrated cluster on the guest and host systems. This means the
manager must be designed to distinguish between virtualized resources and physical
resources. Various cluster management schemes can be greatly enhanced when VM life
migration is enabled with minimal overhead.
19
CLOUD COMPUTING Unit.2
VMs can be live-migrated from one physical machine to another; in case of failure, one VM
can be replaced by another VM. Virtual clusters can be applied in computational grids, cloud
platforms, and high-performance computing (HPC) systems.
The major attraction of this scenario is that virtual clustering provides dynamic resources
that can be quickly put together upon user demand or after a node failure. In particular,
virtual clustering plays a key role in cloud computing.
When a VM runs a live service, it is necessary to make an exchange to ensure that the
migration occurs in a manner that minimizes all three metrics.
Furthermore, we should ensure that the migration will not disrupt other active services
residing in the same host through resource contention (e.g., CPU, network bandwidth).
An inactive state is defined by the virtualization platform, under which the VM is not
enabled.
An active state refers to a VM that has been instantiated at the virtualization platform to
perform a real task.
A paused state corresponds to a VM that has been instantiated but disabled to process a
task or paused in a waiting state.
A VM enters the suspended state if its machine file and virtual resources are stored back to
the disk.
This step makes preparations for the migration, including determining the migrating VM and
the destination host. Although users could manually make a VM migrate to an appointed
host, in most circumstances, the migration is automatically started by strategies such as
load balancing and server consolidation.
Since the whole execution state of the VM is stored in memory, sending the VM’s memory
to the destination node ensures continuity of the service provided by the VM. All of the
memory data is transferred in the first round, and then the migration controller recopies the
memory data which is changed in the last round. These steps keep iterating until the dirty
portion of the memory is small enough to handle the final copy.
Step 3: Suspend the VM and copy the last portion of the data.
20
CLOUD COMPUTING Unit.2
The migrating VM’s execution is suspended when the last round’s memory data is
transferred. Other non memory data such as CPU and network states should be sent as well.
During this step, the VM is stopped and its applications will no longer run. This “service
unavailable” time is called the “downtime” of migration, which should be as short as
possible so that it can be negligible to users.
After all the needed data is copied, on the destination host, the VM reloads the states and
recovers the execution of programs in it, and the service provided by this VM continues.
Then the network connection is redirected to the new VM and the dependency to the
source host is cleared. The whole migration process finishes by removing the original VM
from the source host.
Since clusters have a high initial cost of ownership, including space, power conditioning, and
cooling equipment, leasing or sharing access to a common cluster is an attractive solution
when demands vary over time.
Shared clusters offer economies of scale and more effective utilization of resources by
multiplexing. When one system migrates to another physical node, we should consider the
following issues.
Memory Migration:
This is one of the most important aspects of VM migration. Moving the memory instance of
a VM from one physical host to another can be approached in any number of ways. But
traditionally, the concepts behind the techniques tend to share common implementation
paradigms. The techniques employed for this purpose depend upon the characteristics of
application/workloads supported by the guest OS.
21
CLOUD COMPUTING Unit.2
Another way is to have a global file system across all machines where a VM could be
located. This way removes the need to copy files from one machine to another because all
files are network-accessible.
The relevant VM files are explicitly copied into the local file system for a resume operation
and taken out of the local file system for a suspend operation. This approach relieves
developers from the complexities of implementing several different file system calls for
different distributed file systems.
Network Migration:
To enable remote systems to locate and communicate with a VM, each VM must be
assigned a virtual IP address known to other entities. This address can be distinct from the
IP address of the host machine where the VM is currently located.
Each VM can also have its own distinct virtual MAC address. The VMM maintains a mapping
of the virtual IP and MAC addresses to their corresponding VMs.
In general, a migrating VM includes all the protocol states and carries its IP address with it.
Live migration means moving a VM from one physical node to another while keeping its OS
environment and applications unbroken. This capability is being increasingly utilized in
today’s enter-prise environments to provide efficient online system maintenance,
reconfiguration, load balancing, and proactive fault tolerance.
Live migration is a key feature of system virtualization technologies. Only memory and CPU
status needs to be transferred from the source node to the target node.
22
CLOUD COMPUTING Unit.3
In traditional hosting services, IT infrastructure was rented out for a specific period of time,
with pre-determined hardware configuration. The client paid for the configuration and time,
regardless of the actual use. With the help of the IaaS cloud computing platform layer,
clients can dynamically scale the configuration to meet changing requirements and are
billed only for the services actually used.
IaaS cloud computing platform layer eliminates the need for every organization to maintain
the IT infrastructure.
The private cloud implies that the infrastructure resides at the customer-premise.
In the case of public cloud, it is located at the cloud computing platform vendor's data
center.
The hybrid cloud is a combination of the two in which the customer selects the best of both
public cloud or private cloud.
1
CLOUD COMPUTING Unit.3
Advantages of IaaS cloud computing layer:
Shared infrastructure:
Pay-as-per-use model:
IaaS providers provide services based on the pay-as-per-use basis. The users are required to
pay for what they have used.
IaaS providers focus on the organization's core business rather than on IT infrastructure.
On-demand scalability:
On-demand scalability is one of the biggest advantages of IaaS. Using IaaS, users do not
worry about to upgrade software and troubleshoot the issues related to hardware
components.
Security:
Security is one of the biggest issues in IaaS. Most of the IaaS providers are not able to
provide 100% security.
Although IaaS service providers maintain the software, but they do not upgrade the
software for some organizations.
Interoperability issues:
It is difficult to migrate VM from one IaaS provider to the other, so the customers might face
problem related to vendor lock-in.
2
CLOUD COMPUTING Unit.3
Some important point about IaaS cloud computing layer:
IaaS cloud computing platform cannot replace the traditional hosting method, but it
provides more than that, and each resource which are used are predictable as per the
usage.
IaaS cloud computing platform may not eliminate the need for an in-house IT department. It
will be needed to monitor or control the IaaS setup. IT salary expenditure might not reduce
significantly, but other IT expenses can be reduced.
Breakdowns at the IaaS cloud computing platform vendor's can bring your business to the
halt stage. Assess the IaaS cloud computing platform vendor's stability and finances. Make
sure that SLAs (i.e., Service Level Agreement) provide backups for data, hardware, network,
and application failures. Image portability and third-party support is a plus point.
The IaaS cloud computing platform vendor can get access to your sensitive data. So, engage
with credible companies or organizations. Study their security policies and precautions.
Top Iaas Providers who are providing IaaS cloud computing platform:
3
CLOUD COMPUTING Unit.3
PaaS includes infrastructure (servers, storage, and networking) and platform (middleware,
development tools, database management systems, business intelligence, and more) to
support the web application life cycle.
Programming languages:
PaaS providers provide various programming languages for the developers to develop the
applications. Some popular programming languages provided by PaaS providers are Java,
PHP, Perl, etc.
Application frameworks:
Databases:
PaaS providers provide various databases such as ClearDB, PostgreSQL, MongoDB, and Redis
to communicate with the applications.
4
CLOUD COMPUTING Unit.3
Other tools:
PaaS providers provide various other tools that are required to develop, test, and deploy the
applications.
Advantages of PaaS:
Simplified Development:
PaaS allows developers to focus on development and innovation without worrying about
infrastructure management.
Lower risk:
No need for up-front investment in hardware and software. Developers only need a PC and
an internet connection to start building applications.
Some PaaS vendors also provide already defined business functionality so that users can
avoid building everything from very scratch and hence can directly start the projects only.
Instant community:
PaaS vendors frequently provide online communities where the developer can get the ideas
to share experiences and seek advice from others.
Scalability:
Applications deployed can scale from one to thousands of users without any changes to the
applications.
Vendor lock-in:
One has to write the applications according to the platform provided by the PaaS vendor, so
the migration of an application to another PaaS vendor would be a problem.
Data Privacy:
Corporate data, whether it can be critical or not, will be private, so if it is not located within
the walls of the company, there can be a risk in terms of privacy of data.
5
CLOUD COMPUTING Unit.3
Integration with the rest of the systems applications:
It may happen that some applications are local, and some are in the cloud. So there will be
chances of increased complexity when we want to use data which in the cloud with the local
data.
The below table shows some popular PaaS providers and services that are provided by
them :-
Providers Services
Google App Engine
App Identity, URL Fetch, Cloud storage client library, Logservice
(GAE)
Faster implementation, Rapid scalability, CRM Services, Sales cloud,
Salesforce.com
Mobile connectivity, Chatter.
Windows Azure Compute, security, IoT, Data Storage.
AppFog Justcloud.com, SkyDrive, GoogleDocs
Openshift RedHat, Microsoft Azure.
Cloud Foundry from
Data, Messaging, and other services.
VMware
6
CLOUD COMPUTING Unit.3
Business Services - SaaS Provider provides various business services to start-up the
business. The SaaS business services include ERP (Enterprise Resource Planning), CRM
(Customer Relationship Management), billing, and sales.
Example: Slack.
Social Networks - As we all know, social networking sites are used by the general public, so
social networking service providers use SaaS for their convenience and handle the general
public's information.
Mail Services - To handle the unpredictable number of users and load on e-mail services,
many e-mail providers offering their services using SaaS.
SaaS pricing is based on a monthly fee or annual fee subscription, so it allows organizations
to access business functionality at a low cost, which is less than licensed applications.
7
CLOUD COMPUTING Unit.3
Unlike traditional software, which is sold as a licensed based with an up-front cost (and
often an optional ongoing support fee), SaaS providers are generally pricing the applications
using a subscription fee, most commonly a monthly or annually fee.
One to Many:
SaaS services are offered as a one-to-many model means a single instance of the application
is shared by multiple users.
Software as a service removes the need for installation, set-up, and daily maintenance for
the organizations. The initial set-up cost for SaaS is typically less than the enterprise
software. SaaS vendors are pricing their applications based on some usage parameters, such
as a number of users using the application. So SaaS does easy to monitor and automatic
updates.
All users will have the same version of the software and typically access it through the web
browser. SaaS reduces IT support costs by outsourcing hardware and software maintenance
and support to the IaaS provider.
Multidevice support:
SaaS services can be accessed from any device such as desktops, laptops, tablets, phones,
and thin clients.
API Integration:
SaaS services easily integrate with other software or services through standard APIs.
No client-side installation:
SaaS services are accessed directly from the service provider using the internet connection,
so do not need to require any software installation.
Security:
8
CLOUD COMPUTING Unit.3
Actually, data is stored in the cloud, so security may be an issue for some users. However,
cloud computing is not more secure than in-house deployment.
Latency issue:
Since data and applications are stored in the cloud at a variable distance from the end-user,
there is a possibility that there may be greater latency when interacting with the application
compared to local deployment. Therefore, the SaaS model is not suitable for applications
whose demand response time is in milliseconds.
Switching SaaS vendors involves the difficult and slow task of transferring the very large
data files over the internet and then converting and importing them into another SaaS also.
9
CLOUD COMPUTING Unit.3
The below table shows some popular SaaS providers and services that are provided by
them -
Provider Services
Salseforce.com On-demand CRM solutions
Microsoft Office
Online office suite
365
Google Apps Gmail, Google Calendar, Docs, and sites
ERP, accounting, order management, CRM, Professionals Services
NetSuite
Automation (PSA), and e-commerce applications.
GoToMeeting Online meeting and video-conferencing software
Constant Contact E-mail marketing, online survey, and event marketing
Oracle CRM CRM applications
Workday, Inc Human capital management, payroll, and financial management.
The most widely used cloud computing applications are given below -
10
CLOUD COMPUTING Unit.3
Art Applications:
Cloud computing offers various art applications for quickly and easily design attractive
cards, booklets, and images. Some most commonly used cloud art applications are given
below:
i Moo
Moo is one of the best cloud art applications. It is used for designing and printing
business cards, postcards, and mini cards.
ii. Vistaprint
Adobe creative cloud is made for designers, artists, filmmakers, and other creative
professionals. It is a suite of apps which includes PhotoShop image editing
programming, Illustrator, InDesign, TypeKit, Dreamweaver, XD, and Audition.
Business Applications:
Business applications are based on cloud service providers. Today, every organization
requires the cloud business application to grow their business. It also ensures that business
applications are 24*7 available to users.
i. MailChimp
ii. Salesforce
Salesforce platform provides tools for sales, service, marketing, e-commerce, and
more. It also provides a cloud development platform.
iii. Chatter
Chatter helps us to share important information about the organization in real time.
11
CLOUD COMPUTING Unit.3
iv. Paypal
Paypal offers the simplest and easiest online payment mode using a secure internet
account. Paypal accepts the payment through debit cards, credit cards, and also
from Paypal account holders.
v. Slack
Slack stands for Searchable Log of all Conversation and Knowledge. It provides a
user-friendly interface that helps us to create public and private channels for
communication.
vi. Quickbooks
Cloud computing allows us to store information (data, files, images, audios, and videos) on
the cloud and access this information using an internet connection. As the cloud provider is
responsible for providing security, so they offer various backup recovery application for
retrieving the lost data.
A list of data storage and backup applications in the cloud are given below -
i. Box.com
ii. Mozy
Mozy provides powerful online backup solutions for our personal and business data.
It schedules automatically back up for each day at a specific time.
Google G Suite is one of the best cloud storage and backup application. It includes
Google Calendar, Docs, Forms, Google+, Hangouts, as well as cloud storage and tools
for managing cloud apps. The most popular app in the Google G Suite is Gmail. Gmail
offers free email services to users.
12
CLOUD COMPUTING Unit.3
Education Applications:
Cloud computing in the education sector becomes very popular. It offers various online
distance learning platforms and student information portals to the students. The
advantage of using cloud in the field of education is that it offers strong virtual classroom
environments, Ease of accessibility, secure data storage, scalability, greater reach for the
students, and minimal hardware requirements for the applications.
Google Apps for Education is the most widely used platform for free web-based
email, calendar, documents, and collaborative study.
It allows educators to quickly implement the latest technology solutions into the
classroom and make it available to their students.
Entertainment Applications:
Entertainment industries use a multi-cloud strategy to interact with the target audience.
Cloud computing offers various entertainment applications such as online games and video
conferencing.
i. Online games
Today, cloud gaming becomes one of the most important entertainment media. It
offers various online games that run remotely from the cloud. The best cloud gaming
services are Shaow, GeForce Now, Vortex, Project xCloud, and PlayStation Now.
13
CLOUD COMPUTING Unit.3
cloud-based video conferencing. The benefits of using video conferencing are that it
reduces cost, increases efficiency, and removes interoperability.
Management Applications
Cloud computing offers various cloud management tools which help admins to manage all
types of cloud activities, such as resource deployment, data integration, and disaster
recovery. These management tools also provide administrative control over the platforms,
applications, and infrastructure.
i. Toggl
Toggl helps users to track allocated time period for a particular project.
ii. Evernote
Evernote allows you to sync and save your recorded notes, typed notes, and other
notes in one convenient place. It is available for both free as well as a paid version.
It uses platforms like Windows, macOS, Android, iOS, Browser, and Unix.
iii. Outright
Outright is used by management users for the purpose of accounts. It helps to track
income, expenses, profits, and losses in real-time environment.
iv. GoToMeeting
GoToMeeting provides Video Conferencing and online meeting apps, which allows
you to start a meeting with your business partners from anytime, anywhere using
mobile phones or tablets. Using GoToMeeting app, you can perform the tasks
related to the management such as join meetings in seconds, view presentations on
the shared screen, get alerts for upcoming meetings, etc.
Social Applications:
Social cloud applications allow a large number of users to connect with each other using
social networking applications such as Facebook, Twitter, Linkedln, etc.
i. Facebook
Facebook is a social networking website which allows active users to share files,
photos, videos, status, more to their friends, relatives, and business partners using
14
CLOUD COMPUTING Unit.3
the cloud storage system. On Facebook, we will always get notifications when our
friends like and comment on the posts.
ii. Twitter
iii. LinkedIn
15
CLOUD COMPUTING Unit.4
SOA allows users to combine a large number of facilities from existing services to
form applications.
SOA encompasses a set of design principles that structure system development and
provide means for integrating components into a coherent and decentralized
system.
SOA based computing packages functionalities into a set of interoperable services,
which can be integrated into different software systems belonging to separate
business domains.
1. Service provider: The service provider is the maintainer of the service and the
organization that makes available one or more services for others to use. To
advertise services, the provider can publish them in a registry, together with a
service contract that specifies the nature of the service, how to use it, the
requirements for the service, and the fees charged.
2. Service consumer: The service consumer can locate the service metadata in the
registry and develop the required client components to bind and use the service.
1
CLOUD COMPUTING Unit.4
3. Abstraction: A service is completely defined by service contracts and description
documents. They hide their logic, which is encapsulated within their
implementation.
4. Reusability: Designed as components, services can be reused more effectively, thus
reducing development time and the associated costs.
5. Autonomy: Services have control over the logic they encapsulate and, from a service
consumer point of view, there is no need to know about their implementation.
6. Discoverability: Services are defined by description documents that constitute
supplemental metadata through which they can be effectively discovered. Service
discovery provides an effective means for utilizing third-party resources.
7. Composability: Using services as building blocks, sophisticated and complex
operations can be implemented.
Advantages of SOA:
Service reusability: In SOA, applications are made from existing services. Thus,
services can be reused to make many applications.
Easy maintenance: As services are independent of each other they can be updated
and modified easily without affecting other services.
Platform independent: SOA allows making a complex application by combining
services picked from different sources, independent of the platform.
Availability: SOA facilities are easily available to anyone on request.
Reliability: SOA applications are more reliable because it is easy to debug small
services rather than huge codes
Scalability: Services can run on different servers within an environment, this
increases scalability
Disadvantages of SOA:
Cloud computing is a model used for enabling convenient and usage-based network access
to a configurable computing resources (eg. networks, servers etc) that can be provided and
used rapidly.
2
CLOUD COMPUTING Unit.4
It provides an abstraction layer between computing resources and its technical
implementation details and sequentially enables computational resources to be used
while avoiding efforts in infrastructure management.
The below figure shows the SOA cloud computing along with the models:
Below are the models that are differentiated on the horizontal scaling basis in cloud
computing:
The distinctions are made according to availability and the location of installation in the
deployment models. Private clouds are internal company services whereas public clouds are
the services that are available to the public on internet.
In the large companies where IT plays an important role, internal company cloud solutions
are often built in their own data centers. Small and medium companies often use public
cloud services.
Cloud Computing provides a very flexible and scalable platform through processing external
services and also has the ability to connect with customers, suppliers etc.
3
CLOUD COMPUTING Unit.4
Diversified Services
Diversified is a full-service information technology service provider delivering data.
deploying a data centre on virtualization and cloud computing technologies.
Virtualization:
Data centre virtualization is the process of designing, developing and deploying a data
centre on virtualization and cloud computing technologies. It primarily enables virtualizing
physical servers in a data centre facility along with storage, networking and other
infrastructure devices and equipment. Diversified’s expertise in virtualization helps data
centres achieve the following benefits:
The overwhelming number of vendors, products and hybrid options make it challenging to
develop your future IT roadmap. Diversified aides clients in navigating the pros and cons of
hosting locally, in a private/public cloud, or a combination thereof to create a strategy that
works.
Storage:
A proper strategy for storage is essential for quality and cost efficiency. Diversified assists
clients in selecting the proper storage to fit their needs and budget.
Design for migrating inheritance storage solutions from drives, tapes, and arrays.
Deployment of complete storage solutions consisting of internal and external
storage, backup and storage software, storage networking, hyperconverged, cloud,
and policies.
Hyperconverged Infrastructure:
The drive to reduce complexity while increasing scalability is also becoming increasingly
popular in the data centre. Diversified’s hyperconverged infrastructure solutions provide a
software-centric architecture that tightly integrates computing, storage
and virtualization resources in a single system that runs on off the shelf servers.
4
CLOUD COMPUTING Unit.4
Not all the applications are suitable for cloud. It is extremely important to identify the most
suitable applications for migration and identify any potential problems. Create a checklist to
ensure a complete and successful migration.
If you are managing application performance in the cloud, you need a topological map of
service delivery across all tiers. Although cloud computing offers numerous benefits,
performance issues can complicate or reduce the benefits.
While moving to cloud, various businesses need to face the impact of moving from a
primarily static to dynamic network architecture. Firewall, load balancing and security
services are still required for the network architecture.
While transitioning from a resource focused cost center to a business service focused profit
center, it demands assessing the resource consumption. Unfortunately, the traditional
chargeback and AMP (Advanced Malware Protection) tools lack the ability to enable the
business aligned costing and chargeback paradigms. This means that you need to come up
with a solution to monitor consumption for every service across multiple applications and
tiers.
In order to make it sure that SLA’s in cloud are met, you need to prioritize the allocation of
resources based on the measurement of the end user performance. It demands a clear
picture of the resource consumption at the transaction level.
The lack of knowledge and infrastructure configuration for the service deployment has
limited the ability of researchers to study the impact of resource management inside the
cloud infrastructures on the service performance through measurement-based evaluations.
This makes it difficult for a service customer to use a measurement-based method to get
insight about the performance behaviours of new Cloud services.
Depending upon the size and the type of business you are running, the cloud offers you
various benefits. For a startup, cloud computing provides essential differentiators to keep
the business up and running quickly with minimal up-front costs. Larger businesses often
5
CLOUD COMPUTING Unit.4
face complex challenges to ensure the availability and performance of the high traffic
websites.
You will face various risks along the way along that can be complex to manage. Transitioning
to the cloud is a non-trivial decision that demands a proper evaluation of both the data and
services. It is important to perform a thorough evaluation on the cloud service performance
as this is what will be beneficial for both service providers as well as customers.
Overall, cloud performance issues are considerable and demands innovation to overcome
the challenges.
Data Centre
Data centres are simply centralized locations where computing and networking equipment
is concentrated for the purpose of collecting, storing, processing, distributing or allowing
access to large amounts of data. They have existed in one form or another since the
introduction of computers.
Data centre refers to on-premise hardware while the cloud refers to off-premise computing.
The cloud stores your data in the public cloud, while a data centre stores your data on your
own hardware.
They are also responsible for data backup and recovery, as well as networking. These
centres also host websites, manage e-mails, and instant messaging services. They support
cloud storage applications and e-commerce transactions. Even online gaming communities
require data centres to manage their online activities.
Decades ago, early computers were massive machines that can occupy whole rooms. But as
technology evolved, equipment shrunk and became cheaper than before. However, with
this progress, data processing demands have also begun to increase exponentially.
Unlike before, where data centres are just one big supercomputer, modern aged data
centres functions using multiple servers to optimize further and boost its processing power.
Now data centres consist of thousands of potent and tiny servers that run non-stop around
the clock.
6
CLOUD COMPUTING Unit.4
Importance of Data Centres:
Almost every modern business and government offices need their very own data centre, or
they may decide to lease data centr. Big corporations and government institutions may
choose to build and manage them in-house if they have the resources. While others choose
to rent servers at ‘colos’ or colocation facilities. Some business owners also have the choice
to use public cloud-based services.
Now, we must remember that all of this information needs to be housed somewhere. The
idea of running or storing our data and resources at home or work computers is getting
replaced by faraway storing mentality. Many firms are also migrating their professional
applications to data centre services to minimize the cost of running an in-house server.
That is why data centres are an essential resource for a business that wants to run their
operation without worries. The importance of data centres in the modern world has
increased ten times due to the rising demand of information trading.
The data that is stored on a data centre server is distributed into packets before
transmission and is sent via routers that decide the most suitable path for that data to
progress.
It then uses a series of wired and wireless networks to reach the user’s Internet service
provider and finally arrive at the end user’s computer. Every time a Web address is enters
into a browser, it automatically requests information from a server. If the end-user wants to
upload information, then the process will be reversed.
With how data centres are essential in running big corporations and with even small-
medium enterprises joining the trend, choosing one to fit a business model is essential.
There are different types of data centres and service models.
Colocation data centers or most commonly known as “colo” is a company that rents space
within a data centre that they do not own and is housed outside the company’s premises.
The colocation data centre provides the infrastructure like the building itself, cooling,
7
CLOUD COMPUTING Unit.4
bandwidth, and security, among others. While the company produces and maintains the
components, which include the servers, storage system, and security firewalls.
Enterprise data centres are established, owned, and managed by companies. These data
canters are operated under a single purpose and that this optimized service for their end-
user clients. Enterprise data canters are often located inside corporate compounds.
These data canters are operated by a third-party entity or a managed services provider
instead of the company. The company rents the equipment and infrastructure to cut costs.
Cloud data canters are an off-premises form of a data centre. The most common cloud
hosting services are Amazon Web Services (AWS), Microsoft (Azure), and IBM Cloud.
Business owners are in constant need of reliability when in terms of maintaining a smooth
operation. Good thing that a Data centre is built to withstand a 24/7 service easily.
However, the components require a significant amount of infrastructure support in both
hardware and software areas.
These include power subsystems, stable and uninterruptible power supplies, proper
ventilation, high-quality cooling systems, fire control, reliable backup generators, and
connections to external networks.
The business world is moving at an incredibly fast pace that matches the overwhelming
demand for information. With the ever-changing requirements of the modern business
model, many companies place their confidence in data canters, as these facilities play a
crucial role in reaching their IT specifications.
Data centre service providers are capable of handling higher volumes of traffic without
making compromises on security and storage capacity of data. Generally, a typical data
centre carries the responsibility of managing significant characteristics like data workloads,
operating conditions, data protection and security fulfilment.
Data canters are more than just a safe and secure facility with space that is equipped with
reliable power, and network. They are becoming a valuable addition to many businesses as
they prove to be a dependable extension of their IT team. That is why data canters in
modern business set up are increasingly becoming an essential factor for success.
8
CLOUD COMPUTING Unit.4
Cloud computing is bringing amazing advantages and benefits companies. But it also brings
some challenges. There are several legal issues that must be taken into consideration when
moving into the cloud. Let’s see which are the most challenging legal issues around cloud
computing.
Security procedures:
The majority of companies which implemented cloud solutions and services do not have
security procedures in place. Also, they lack measures to approve or evaluate cloud
applications. When adopting the BYOD trend for example, organizations needed these
security procedures more than ever. General data security trainings, multiple levels of
security, rigorous procedures to use one’s own device and to transfer or copy data are some
of the options available to protect data in organizations. The bottom line is that security
procedures must be established according to every company’s objectives and work flow.
[Bring your own device (BYOD) refers to the trend of employees using personal devices to
connect to their organizational networks and access work-related systems and potentially
sensitive or confidential data. Personal devices could include smartphones, personal computers,
tablets, or USB drives. ]
Third-party involvement could be a risk. All third parties using a multi-tenant shared cloud
are using the same administration interface, so make sure multi-factor authentication and
enhanced security is present. Also, look for HIPAA (Health Insurance Portability and
Accountability Act) compliant providers – a business associate agreement (BAA) with third-
party vendor who access Protected Health Information (PHI) is necessary to ensure privacy
and security requirements. A partnership with a HIPAA solutions provider that signs a BAA is
an efficient method to make sure this goes smoothly and everything is secure. And don’t
forget to read carefully the terms and conditions before signing up for a cloud based
services.
9
CLOUD COMPUTING Unit.4
Intellectual Property Rights:
Intellectual Property Rights differ from one country to another, so it is not very clear what
intellectual property laws will apply in the cloud computing environment. Make sure you are
aware of the regulations and rights from the country you store your intellectual work. The
provider you choose should know how to protect intellectual property it stores and how to
avoid potential infringement pitfalls.
Data stored in the cloud might be compromised or breached. Therefore, most cloud
computing providers also offer the customer different levels of security protection, which
allows for more enhanced security. Encryption might seem to have failed in protecting data
from theft attacks, but other methods have been discovered and implemented, including
monitoring data access in the cloud to detect abnormal data access patterns. The customer
has to understand the cloud provider’s disclosure policy and how quickly the breach would
be disclosed to them. Most of the U.S. states have security breach disclosure laws requiring
the provider to inform the customers when their data has been compromised.
Many of these legal issues and the methods to inform about them or to solve them should
be mentioned in the Service Level Agreement. It is essential to understand all the terms of
the cloud’s provider and to consider the needs and objectives of the enterprise before
signing an agreement.
10
What is Big Data? Introduction, Types,
Characteristics, Examples
What is Data?
The quantities, characters, or symbols on which operations are
performed by a computer, which may be stored and transmitted in the
form of electrical signals and recorded on magnetic, optical, or
mechanical recording media.
The New York Stock Exchange is an example of Big Data that generates
about one terabyte of new trade data per day.
Social Media
The statistic shows that 500+terabytes of new data get ingested into
the databases of social media site Facebook, every day. This data is
mainly generated in terms of photo and video uploads, message
exchanges, putting comments etc.
1. Structured
2. Unstructured
3. Semi-structured
Structured
Any data that can be stored, accessed and processed in the form of fixed
format is termed as a ‘structured’ data. Over the period of time, talent in
computer science has achieved greater success in developing
techniques for working with such kind of data (where the format is well
known in advance) and also deriving value out of it. However,
nowadays, we are foreseeing issues when a size of such data grows to a
huge extent, typical sizes are being in the rage of multiple zettabytes.
Looking at these figures one can easily understand why the name Big
Data is given and imagine the challenges involved in its storage and
processing.
• Volume
• Variety
• Velocity
• Variability
(i) Volume – The name Big Data itself is related to a size which is
enormous. Size of data plays a very crucial role in determining value out
of data. Also, whether a particular data can actually be considered as a
Big Data or not, is dependent upon the volume of data.
Hence, ‘Volume’ is one characteristic which needs to be considered
while dealing with Big Data solutions.
Big Data Velocity deals with the speed at which data flows in from
sources like business processes, application logs, networks, and social
media sites, sensors, Mobile devices, etc. The flow of data is massive
and continuous.
Big Data technologies can be used for creating a staging area or landing
zone for new data before identifying what data should be moved to
the data warehouse. In addition, such integration of Big Data
technologies and data warehouse helps an organization to offload
infrequently accessed data.
Summary
• Big Data definition : Big Data meaning a data that is huge in
size. Bigdata is a term used to describe a collection of data that
is huge in size and yet growing exponentially with time.
• Big Data analytics examples includes stock exchanges, social
media sites, jet engines, etc.
• Big Data could be 1) Structured, 2) Unstructured, 3) Semi-
structured
• Volume, Variety, Velocity, and Variability are few Big Data
characteristics
• Improved customer service, better operational efficiency,
Better Decision Making are few advantages of Bigdata
Difference between Traditional data and Big
data
••
1. Traditional data: Traditional data is the structured data that is being majorly
maintained by all types of businesses starting from very small to big organizations.
In a traditional database system, a centralized database architecture used to store
and maintain the data in a fixed format or fields in a file. For managing and
accessing the data Structured Query Language (SQL) is used.
Traditional data is characterized by its high level of organization and structure,
which makes it easy to store, manage, and analyze. Traditional data analysis
techniques involve using statistical methods and visualizations to identify
patterns and trends in the data.
Traditional data is often collected and managed by enterprise resource planning
(ERP) systems and other enterprise-level applications. This data is critical for
businesses to make informed decisions and drive performance improvements.
2. Big data: We can consider big data an upper version of traditional data. Big data
deal with too large or complex data sets which is difficult to manage in traditional
data-processing application software. It deals with large volume of both
structured, semi structured and unstructured data. Volume, Velocity and Variety,
Veracity and Value refer to the 5’V characteristics of big data. Big data not only
refers to large amount of data it refers to extracting meaningful data by analyzing
the huge amount of complex data sets. semi-structured
Big data is characterized by the three Vs: volume, velocity, and variety. Volume
refers to the vast amount of data that is generated and collected; velocity refers
to the speed at which data is generated and must be processed; and variety refers
to the many different types and formats of data that must be analyzed, including
structured, semi-structured, and unstructured data.
Due to the size and complexity of big data sets, traditional data management
tools and techniques are often inadequate for processing and analyzing the data.
Big data technologies, such as Hadoop, Spark, and NoSQL databases, have
emerged to help organizations store, manage, and analyze large volumes of data.
The main differences between traditional data and big data as follows:
• Volume: Traditional data typically refers to small to medium-sized
datasets that can be easily stored and analyzed using traditional data
processing technologies. In contrast, big data refers to extremely large
datasets that cannot be easily managed or processed using traditional
technologies.
• Variety: Traditional data is typically structured, meaning it is organized
in a predefined manner such as tables, columns, and rows. Big data, on
the other hand, can be structured, unstructured, or semi-structured,
meaning it may contain text, images, videos, or other types of data.
• Velocity: Traditional data is usually static and updated on a periodic
basis. In contrast, big data is constantly changing and updated in real-
time or near real-time.
• Complexity: Traditional data is relatively simple to manage and
analyze. Big data, on the other hand, is complex and requires
specialized tools and techniques to manage, process, and analyze.
• Value: Traditional data typically has a lower potential value than big
data because it is limited in scope and size. Big data, on the other hand,
can provide valuable insights into customer behavior, market trends,
and other business-critical information.
Some similarities between them, including:
• Data Quality: The quality of data is essential in both traditional and big
data environments. Accurate and reliable data is necessary for making
informed business decisions.
• Data Analysis: Both traditional and big data require some form of
analysis to derive insights and knowledge from the data. Traditional
data analysis methods typically involve statistical techniques and
visualizations, while big data analysis may require machine learning
and other advanced techniques.
• Data Storage: In both traditional and big data environments, data needs
to be stored and managed effectively. Traditional data is typically
stored in relational databases, while big data may require specialized
technologies such as Hadoop, NoSQL, or cloud-based storage systems.
• Data Security: Data security is a critical consideration in both
traditional and big data environments. Protecting sensitive information
from unauthorized access, theft, or misuse is essential in both contexts.
• Business Value: Both traditional and big data can provide significant
value to organizations. Traditional data can provide insights into
historical trends and patterns, while big data can uncover new
opportunities and help organizations make more informed decisions.
The difference between Traditional data and Big data are as follows:
Traditional Data Big Data
Its volume ranges from Gigabytes to Its volume ranges from Petabytes to
Terabytes. Zettabytes or Exabytes.
Traditional data is generated per hour or per But big data is generated more
day or more. frequently mainly per seconds.
Traditional data source is centralized and it Big data source is distributed and it is
is managed in centralized form. managed in distributed form.
Its data model is strict schema based and it is Its data model is a flat schema based
static. and it is dynamic.
Traditional data is stable and inter Big data is not stable and unknown
relationship. relationship.
Its data sources includes ERP transaction Its data sources includes social media,
data, CRM transaction data, financial data, device data, sensor data, video,
organizational data, web transaction data etc. images, audio etc.
Conclusion :
The key differences between traditional data and big data are related to the
volume, variety, velocity, complexity, and potential value of the data. Traditional
data is typically small in size, structured, and static, while big data is large,
complex, and constantly changing. As a result, big data requires specialized tools
and techniques to manage and analize effectively.
Data Warehousing
A Database Management System (DBMS) stores data in the form of tables and uses
an ER model and the goal is ACID properties. For example, a DBMS of a college has
tables for students, faculty, etc.
A Data Warehouse is separate from DBMS, it stores a huge amount of data,
which is typically collected from multiple heterogeneous sources like files, DBMS,
etc. The goal is to produce statistical results that may help in decision-making.
For example, a college might want to see quick different results, like how the
placement of CS students has improved over the last 10 years, in terms of
salaries, counts, etc.
Need for Data Warehouse
An ordinary Database can store MBs to GBs of data and that too for a specific
purpose. For storing data of TB size, the storage shifted to the Data Warehouse.
Besides this, a transactional database doesn’t offer itself to analytics. To effectively
perform analytics, an organization keeps a central Data Warehouse to closely
study its business by organizing, understanding, and using its historical data for
making strategic decisions and analyzing trends.
Benefits of Data Warehouse
• Better business analytics: Data warehouse plays an important role in
every business to store and analysis of all the past data and records of
the company. which can further increase the understanding or analysis
of data for the company.
• Faster Queries: The data warehouse is designed to handle large
queries that’s why it runs queries faster than the database.
• Improved data Quality: In the data warehouse the data you gathered
from different sources is being stored and analyzed it does not interfere
with or add data by itself so your quality of data is maintained and if
you get any issue regarding data quality then the data warehouse team
will solve this.
• Historical Insight: The warehouse stores all your historical data which
contains details about the business so that one can analyze it at any
time and extract insights from it.
Data Warehouse vs DBMS
Database Data Warehouse
There are five basic analytical operations that can be performed on an OLAP
cube:
1. Drill down: In drill-down operation, the less detailed data is converted
into highly detailed data. It can be done by:
• Moving down in the concept hierarchy
• Adding a new dimension
In the cube given in overview section, the drill down operation is
performed by moving down in the concept hierarchy
of Time dimension (Quarter -> Month).
4. Slice: It selects a single dimension from the OLAP cube which results in
a new sub-cube creation. In the cube given in the overview section,
Slice is performed on the dimension Time = “Q1”.
What is OLTP?
OLTP is an operational system that supports transaction-oriented
applications in a 3-tier architecture. It administers the day to day
transaction of an organization. OLTP is basically focused on query
processing, maintaining data integrity in multi-access environments as
well as effectiveness that is measured by the total number of
transactions per second.
The full form of OLTP is Online Transaction Processing.
Characteristics of OLTP
Following are important characteristics of OLTP:
Architecture of OLTP
Here is the architecture of OLTP:
OLTP Architecture
• Online banking
• Online airline ticket booking
• Sending a text message
• Order entry
• Add a book to shopping cart
Advantages of OLTP
Following are the pros/benefits of OLTP system:
Disadvantages of OLTP
Here are cons/drawbacks of OLTP system:
Summary
• OLTP is defined as an operational system that supports
transaction-oriented applications in a 3-tier architecture.
• OLTP uses transactions that include small amounts of data.
• OLTP system is an online database changing system.
• The architecture of OLTP contains 1) Business / Enterprise
Strategy, 2) Business Process, 3) Customers, Orders, and Products,
4) ETL Processes, 5) Data Mart and Data warehouse, and 6) Data
Mining, Analytics, and Decision Making.
• OLTP is an online transactional system, whereas OLAP is an online
analysis and data retrieving process.
• OLTP provides a solid foundation for a stable business
/organization due to timely modification of all transactions.
• OLTP systems allow multiple users to access and change the same
data at the same time, which many times created an
unprecedented situation.
Data Lake
The Data Lake democratizes data and is a cost-effective way to store all
data of an organization for later processing. Research Analyst can focus
on finding meaning patterns in data and not data itself.
1. Ingestion Tier: The tiers on the left side depict the data sources.
The data could be loaded into the data lake in batches or in real-
time
2. Insights Tier: The tiers on the right represent the research side
where insights from the system are used. SQL, NoSQL queries, or
even excel could be used for data analysis.
3. HDFS is a cost-effective solution for both structured and
unstructured data. It is a landing zone for all data that is at rest in
the system.
4. Distillation tier takes data from the storage tire and converts it to
structured data for easier analysis.
5. Processing tier run analytical algorithms and users queries with
varying real time, interactive, batch to generate structured data
for easier analysis.
6. Unified operations tier governs system management and
monitoring. It includes auditing and proficiency management,
data management, workflow management.
Data Ingestion
Data Ingestion allows connectors to get data from a different data
sources and load into the Data lake.
Data Governance
Data governance is a process of managing availability, usability,
security, and integrity of data used in an organization.
Security
Security needs to be implemented in every layer of the Data lake. It
starts with Storage, Unearthing, and Consumption. The basic need is to
stop access for unauthorized users. It should support different tools to
access data with easy to navigate GUI and Dashboards.
Data Quality
Data quality is an essential component of Data Lake architecture. Data
is used to exact business value. Extracting insights from poor quality
data will lead to poor quality insights.
Data Discovery
Data Discovery is another important stage before you can begin
preparing data or analysis. In this stage, tagging technique is used to
express the data understanding, by organizing and interpreting the data
ingested in the Data lake.
Data Auditing
Two major Data auditing tasks are tracking changes to the key dataset.
Data Lineage
This component deals with data’s origins. It mainly deals with where it
movers over time and what happens to it. It eases errors corrections in a
data analytics process from origin to destination.
Data Exploration
It is the beginning stage of data analysis. It helps to identify right
dataset is vital before starting Data Exploration.
• After some time, Data Lake may lose significance and momentum
• There is larger amount risk involved while designing Data Lake
• Unstructured Data may lead to Ungoverned Chao, Unusable Data,
Disparate & Complex Tools, Enterprise-Wide Collaboration,
United, Consistent, and Common
• It also increases storage & computes costs
• There is no way to get insights from others who have worked with
the data because there is no account of the lineage of findings by
previous analysts
• The biggest risk of data lakes is security and access control.
Sometimes data can be placed into a lake without any oversight,
as some of the data may have privacy and regulatory need
Summary
• A Data Lake is a storage repository that can store large amount of
structured, semi-structured, and unstructured data.
• The main objective of building a data lake is to offer an unrefined
view of data to data scientists.
• Unified operations tier, Processing tier, Distillation tier and HDFS
are important layers of Data Lake Architecture
• Data Ingestion, Data storage, Data quality, Data Auditing, Data
exploration, Data discover are some important components of
Data Lake Architecture
• Design of Data Lake should be driven by what is available instead
of what is required.
• Data Lake reduces long-term cost of ownership and allows
economic storage of files
• The biggest risk of data lakes is security and access control.
Sometimes data can be placed into a lake without any oversight,
as some of the data may have privacy and regulatory need.