Cloud computing question paper Midsem Ques-Ans KIIT
Cloud computing question paper Midsem Ques-Ans KIIT
Ans: It brought about a change in the architecture of the web which enables features like word
press, social media, OTT platforms etc.
2.What is Multi-tenancy?
Ans: In cloud computing, multitenancy means that multiple customers of a cloud vendor are
using the same computing resources. Despite the fact that they share resources, cloud
customers aren't aware of each other, and their data is kept totally separate. Multitenancy is a
crucial component of cloud computing; without it, cloud services would be far less practical.
Multitenant architecture is a feature in many types of public cloud computing, including IaaS,
PaaS, SaaS, containers, and serverless computing.
5.Define SSL.
Ans: Secure Socket Layer (SSL) provides security to the data that is transferred between web
browser and server. SSL encrypts the link between a web server and a browser which ensures
that all data passed between them remains private and free from attack.
8.What is Eucalyptus?
Ans: A virtual private cloud (VPC) is a private cloud computing environment contained within a
public cloud. Essentially, a VPC provisions logically isolated sections of a public cloud in order
to provide a virtual private environment.
Like all cloud environments, VPC resources are available on-demand to scale up as needed
and are highly configurable.
Ans: Microsoft Azure, commonly referred to as Azure, is a cloud computing service created by
Microsoft for building, testing, deploying, and managing applications and services through
Microsoft-managed data centers.
Ans: Community Cloud is a hybrid form of private cloud. They are multi-tenant platforms that
enable different organizations to work on a shared platform.
Community Cloud computing facilitates its users to identify and analyze their business
demands better. Community Cloud may be hosted in a data center, owned by one of the
tenants, or by a third-party cloud services provider and can be either on-site or off-site.
Example:
1)Example of using a community cloud would be to test-drive some high-end security products
or even test out some features of a public cloud environment. This is great for organizations that
are driven by compliance and regulatory measures. Government, healthcare, and some
regulated private industries are leveraging the added security features within a community cloud
environment. Instead of just provisioning space in a public cloud, organizations can test and
work on a cloud platform which is secure, “dedicated,” and even compliant with certain
regulations. The really interesting part is that with a community cloud, the presence can be
either onsite or offsite.
Ans: A virtual machine (or "VM") is an emulated computer system created using software. It
uses physical system resources, such as the CPU, RAM, and disk storage, but is isolated from
other software on the computer. It can easily be created, modified, or destroyed without
affecting the host computer.
As cloud services have grown in popularity, cloud-based VMs have become increasingly
popular as well. "Cloud instances," as they are often called, run on a computer that is accessed
over the Internet. The VM is often controlled through a web browser or a remote access utility.
Cloud-based VMs are a common way for companies to test software deployments since they
can test on dozens of machines without hosting the VMs locally.
Ans: Full virtualization is a virtualization technique used to provide a VME that completely
simulates the underlying hardware. In this type of environment, any software capable of
execution on the physical hardware can be run in the VM, and any OS supported by the
underlying hardware can be run in each individual VM. Users can run multiple different guest
OSes simultaneously.
In full virtualization, the VM simulates enough hardware to allow an unmodified guest OS to
be run in isolation. This is particularly helpful in a number of situations. For example, in OS
development, experimental new code can be run at the same time as older versions, each in a
separate VM. The hypervisor provides each VM with all the services of the physical system,
including a virtual BIOS, virtual devices, and virtualized memory management. The guest OS is
fully disengaged from the underlying hardware by the virtualization layer.
Ans: The purpose of elasticity is to match the resources allocated with actual amount of
resources needed at any given point in time. Scalability handles the changing needs of an
application within the confines of the infrastructure via statically adding or removing resources to
meet applications demands if needed.
15.Write a short note on Xen Hypervisor.
Ans: Xen is an open source hypervisor based on paravirtualization. Xen has been extended to
compatible with full virtualization using hardware-assisted virtualization. It enables high
performance to execute guest operating systems.
Ans:
The Internet is a network of networks, which provides software/hardware infrastructure to
establish and maintain connectivity of the computers around the word, while Cloud computing is
a new technology that delivers many types of resources over the Internet. Therefore Cloud
computing could be identified as a technology that uses the Internet as the communication
medium to deliver its services. Cloud services can be offered within enterprises through LANs
but in reality, Cloud computing cannot operate globally without the Internet.
Ans: A multi-tenant cloud is a cloud computing architecture that allows customers to share
computing resources in a public or private cloud. Each tenant's data is isolated and remains
invisible to other tenants.
In a multi-tenant cloud system, users have individualized space for storing their projects and
data.
Advantage:
1)Helps a great Deal in Cutting the Cost of Investment
2)Adding a New Customer is Easy
3)It Becomes More Convenient to Maintain the Same Application
4)Maximising the Resource Usage
5)Holds Multiple Tenants all at the Same Time
Ans:
On-demand self service:
Cloud computing provides resources on demand, i.e. when the consumer wants it.
For eg:
The consumer’s request is then automatically processed by the cloud infrastructure, without
human intervention on the provider’s side.
Resource pooling:
The provider’s computing resources are pooled to serve multiple consumers using a
multi-tenant model, with different physical and virtual resources dynamically assigned and
reassigned according to consumer demand.
Examples of resources include storage, processing, memory, and network bandwidth.
Ans: . In Full virtualization, virtual machines permit the execution of the instructions with running
of unmodified OS in an entirely isolated way. In paravirtualization, virtual machines do not
implement full isolation of OS but rather provide a different API which is utilized when OS is
subjected to alteration.
Public cloud:
1)The public cloud is the first deployment model. In this model, users have many options to opt
for and decide on any service provider as per requirement.
2)This model assists in the reduction of capital expenses and removes equipped IT expenses.
Ans: A hypervisor is computer software or hardware that enables you to host multiple virtual
machines. Each virtual machine is able to run its own programs.
It is important because A hypervisor allows you to access several virtual machines that are all
working optimally on a single piece of computer hardware. For the most part, cloud computing
entails you being able to access a virtual machine for you to be able to do what you need to do
anywhere. A hypervisor manaches these virtual machines.
Ans: The SaaS maturity model is broken down into four levels, and each of them brings certain
opportunities and challenges you should be aware of when accessing SaaS vendors.
Level 1 (Single-Tenant, Custom Instances)
Level 2 (Single-Tenant, Configurable Instances)
Level 3 (Multi-Tenant, Configurable)
Level 4 (Multi-Tenant Configurable & Scalable)
Ans: Self-service cloud computing is a form of private cloud service where the customer
provisions storage and launches applications without going through an external cloud service
provider. With a self-service cloud, users access a web-based portal, where they can request or
configure servers and launch applications.
Ans: In cloud computing, multi-tenancy means that a SaaS (Software as a Service) vendor
provides a single version of its software for all its customers. ... Shared infrastructure leads to
lower costs: SaaS allows companies of all sizes to share infrastructure and data center
operational costs.
Rapid elasticity:
The capabilities of the cloud should appear unlimited to the user. These capabilities should be
elastically scalable both outward and inward in accordance with demand, whatever the quantity
of resources required, and at any time.
Ans: A cloud API enables end users to access a cloud provider's application or service, such as
compute infrastructure, storage resources or monitoring tools. APIs define the possible features
and functions of that app or service, along with the details needed to execute them.
Ultimate scalability – vast on demand resources are available so that applications can respond
seamlessly to fluctuations in activity.
Cost effective - public cloud brings together a greater level of resources so users can benefit
from the largest economies of scale. The centralized operation and management of the
underlying resources is shared across all of the subsequent services.
Utility style payment model - public cloud services employ a pay-as-you-go model whereby the
consumer is able to access the resource they need, when they need them, and only pay for
what they use; avoiding wasted capacity.
Reliability - the sheer number of servers and networks involved in creating a public cloud mean
that if one physical component fails, the service would still run unaffected on the remaining
components. In some cases, where clouds draw resources from multiple data centers, an entire
data center could go offline and individual services would suffer no ill effect. There is, in other
words, no single point of failure.
Flexibility - there are a myriad of IaaS, PaaS and SaaS services available on the market which
follow the public cloud model and that are ready to be accessed as a service from any internet
enabled device. These services can fulfill most computing requirements and can deliver their
benefits to private and enterprise clients alike. Businesses can even integrate their public cloud
services with private clouds, where they need to perform sensitive business functions, to create
hybrid clouds.
Location independence - the availability of public cloud services through an internet connection
ensures that the services are available wherever the client is located. This provides remote
access to IT infrastructure (in case of emergencies, etc.) and online document collaboration
from multiple locations.
Ans: Resource scheduling assigns the precise and accurate task to CPU, network, and storage.
organized scheduling is needed for both cloud providers and cloud users.
Ans:
Some services offered by IBM Smart Cloud are as follows:
Compute Infrastructure
Compute Services
Storage
Watson (IBM’s artificial intelligence and machine learning service)
Data and Analytics
Security
DevOps
Application services
33.What is SSL?
Ans: Secure Sockets Layer (SSL) is a standard protocol used for the secure transmission of
documents over a network. Developed by Netscape, SSL technology creates a secure link
between a Web server and browser to ensure private and integral data transmission. SSL uses
Transport Control Protocol (TCP) for communication.
34.What is the selection criterion for private cloud deployment model?
Ans: A private cloud offers flexibility, cost savings, security, and control benefits. These benefits
are particularly valuable for businesses with predictable workloads or customization
requirements, and businesses in regulated industries.
Ans: Cloud computing is the on-demand availability of computer system resources, especially
data storage (cloud storage) and computing power, without direct active management by the
user.
38.What are the different things to be taken into consideration before migrating into a cloud
platform?
Ans: https://whitehatsme.com/factors-consider-when-moving-to-cloud
Ans: Cluster computing refers to the process of sharing the computation task to multiple
computers of the cluster. The number of computers are connected on a network and they
perform a single task by forming a Cluster of computers where the process of computing is
called cluster computing.
Cluster Computing is a high performance computing framework which helps in solving more
complex operations more efficiently with a faster processing speed and better data integrity.
Cluster Computing is a networking technology that performs its operations based on the
principle of distributed systems.
Cloud Computing refers to the on demand delivery of the IT resources especially computing
power and data storage through the internet with pay per use pricing. It generally refers to the
data centers available to the users over internet. Cloud Computing is the virtualized pool of
resources. It allows us to create, configure and customize our applications online. The user can
access any resource at any time and anywhere without worrying about the management and
maintenance of actual resources. Cloud computing delivers both a combination of hardware and
software based computing resources over the network.
Ans: https://www.geeksforgeeks.org/difference-between-cloud-computing-and-grid-computing
Ans: Reduced IT costs. Moving to cloud computing may reduce the cost of managing and
maintaining your IT systems. ...
Scalability.
Business continuity.
Collaboration efficiency.
Flexibility of work practices.
Access to automatic updates.
43.Discuss the functionality of various components of cloud infrastructure.
Ans: https://www.javatpoint.com/cloud-computing-architecture
44.Discuss various cloud service models and prepare a comparative report of these models with
example.
Ans:
https://www.javatpoint.com/cloud-service-models
https://mrcet.com/pdf/Lab%20Manuals/IT/R15A0529_CloudComputing_Notes-converted.pdf (pg
40)
45.Explain cloud computing architecture and its various components with diagrams.
Ans: https://intellipaat.com/blog/cloud-computing-architecture/
Ans: https://www.slideshare.net/lersmethasakul/nist-cloud-computing-reference-architecture
Long Questions
1.What are the major distributed computing technologies that led to cloud computing?
Ans: The major distributed computing technologies which led to cloud computing:-
At the ISA level, virtualization is performed by emulating a given ISA by the ISA of the host
machine. For example, MIPS binary code can run on an x86-based host machine with the help
of ISA emulation. With this approach, it is possible to run a large amount of legacy binary code
writ-ten for various processors on any given new hardware host machine. Instruction set
emulation leads to virtual ISAs created on any hardware machine.
The basic emulation method is through code interpretation. An interpreter program interprets the
source instructions to target instructions one by one. One source instruction may require tens or
hundreds of native target instructions to perform its function. Obviously, this process is relatively
slow. For better performance, dynamic binary translation is desired. This approach translates
basic blocks of dynamic source instructions to target instructions. The basic blocks can also be
extended to program traces or super blocks to increase translation efficiency. Instruction set
emulation requires binary translation and optimization. A virtual instruction set architecture
(V-ISA) thus requires adding a processor-specific software translation layer to the compiler.
Hardware-level virtualization is performed right on top of the bare hardware. On the one hand,
this approach generates a virtual hardware environment for a VM. On the other hand, the
process manages the underlying hardware through virtualization. The idea is to virtualize a
computer’s resources, such as its processors, memory, and I/O devices. The intention is to
upgrade the hardware utilization rate by multiple users concurrently. The idea was implemented
in the IBM VM/370 in the 1960s. More recently, the Xen hypervisor has been applied to
virtualize x86-based machines to run Linux or other guest OS applications. We will discuss
hardware virtualization approaches in more detail in Section 3.3.
Most applications use APIs exported by user-level libraries rather than using lengthy system
calls by the OS. Since most systems provide well-documented APIs, such an interface becomes
another candidate for virtualization. Virtualization with library interfaces is possible by controlling
the communication link between applications and the rest of a system through API hooks. The
software tool WINE has implemented this approach to support Windows applications on top of
UNIX hosts. Another example is the vCUDA which allows applications executing within VMs to
leverage GPU hardware acceleration. This approach is detailed in Section 3.1.4.
5. User-Application Level
Virtualization at the application level virtualizes an application as a VM. On a traditional OS, an
application often runs as a process. Therefore, application-level virtualization is also known as
process-level virtualization. The most popular approach is to deploy high level language (HLL)
VMs. In this scenario, the virtualization layer sits as an application program on top of the
operating system, and the layer exports an abstraction of a VM that can run programs written
and compiled to a particular abstract machine definition. Any program written in the HLL and
compiled for this VM will be able to run on it. The Microsoft .NET CLR and Java Virtual Machine
(JVM) are two good examples of this class of VM.
Packaged software: This is the biggest area of the SaaS market. Packaged software comes in
many different flavors: customer relationship management, supply chain management, financial
management, and human resources, to name the most common.
These integrated offers focus on a specific process, such as managing employees’ benefits,
salaries, and annual performance reviews. These products tend to have several characteristics
in common: They’re designed with specific business processes built in that customers can
modify. They have moved in great numbers to the cloud because customers were finding the
platforms too hard to manage.
Collaborative software: This increasingly vibrant area of the market is driven by the ubiquitous
availability of the Internet, combined with the fact that teams are located all over the world. This
area is dominated by software that focuses on all sorts of collaborative efforts including Web
conferencing, document collaboration, project planning, instant messaging, and even email. In a
sense, it was inevitable that these platforms would move to the cloud: These tasks occur
throughout the organization and need to be easily accessed from many locations.
Enabling and management tools: What’s in this category? Think about the development tools
that developers need when creating and extending a SaaS platform; also think about the
testing, monitoring, and measuring that a customer and the developer need. Also consider the
compliance issues related to the use of this type of software in the real world. These issues are
included in this third category.
4.What are the security risk aspects associated with cloud?Classify the various types of clouds.
Ans:
1. Data Breaches — Data Theft and Data Loss
Risk Factors
A data breach can result in data theft or data loss and damage data confidentiality, availability
and integrity.
Causes of cloud data breaches include:
Insufficient identity and credential management
Easy registration systems, phishing and pretexting
Insecure APIs
2)Misconfigurations
Risk Factors
Common types of misconfiguration include:
Human error
Allowing excessive permissions
Maintaining unused and stale accounts
Allowing excessive sharing settings, which can lead to sensitive data being overexposed
Leaving default settings unchanged, including admin credentials and port numbers
Disabling standard security controls
Disabling encryption
3)Insider Threats
Risk Factors
Insider threats can be intentional, such as a disgruntled employee taking revenge, or accidental,
like an admin making a mistake. Employees aren’t the only insiders either. Contractors,
suppliers and partners can also access data inappropriately, expose it or allow it to be stolen.
Many enterprises lack visibility into user and admin activity and application usage across their
cloud storage systems.
Specific insider threats include privilege abuse, compromised routers and VPNs, shared
accounts, privileged accounts, and service accounts.
Best Practices for Risk Mitigation
De-provision access to resources immediately whenever you have personnel changes.
Implement data discovery and classification technology. Identify all sensitive and
business-critical data you have; know which users, contractors and partners have access to it;
and track their activities concerning Look for signs of suspicious activity trends, such as an
increased number of failed access attempts. Staying on top of user activities around sensitive
and business-critical data helps you identify malicious operations before they cause real
damage.
Monitor privileged users. Track service and privileged accounts separately from other user
accounts. These accounts should be used sparingly for specific tasks that other accounts do not
have sufficient rights to perform.
Implement user behavior analytics. Create a baseline behavioral profile of each user and watch
for actions atypical for that user or others with the same role. Track attempts to access disabled
accounts, along with any other anomalous attempts to access data or gain elevated
permissions.
4)Account Hijacking
Account hijacking is the use of stolen credentials for various purposes, such as to gain access
to sensitive data.
Risk Factors
Hackers use password cracking, phishing emails and cross-site scripting, among other
industry-known tricks, to guess credentials and gain access to staff accounts.
While the volume of DDoS attacks has declined, new forms of DoS attacks are being
discovered that integrate AI and machine learning.
6)Malware
Risk Factors
Malware infects a cloud provider’s servers just as it does on-prem systems: The attacker entices
a user to click on a malicious email attachment or social media link, enabling them to download
malware encoded to bypass detection and designed to eavesdrop, steal data stored in cloud
service applications or otherwise compromise data security.
Antivirus solutions
Regular comprehensive data backups
Employee training on safe browsing and downloading habits
Advanced web application firewalls
Constant activity monitoring
Other Risks:-
Compliance
Loss of data
Data storage
Business continuity
Uptime
Data integrity in cloud computing
5.Before going for a cloud computing platform what are the essential things to be taken in
concern by users?
Ans: https://whitehatsme.com/factors-consider-when-moving-to-cloud
Features of AWS
AWS provides various powerful features for building scalable, cost-effective, enterprise
applications. Some important features of AWS is given below-
AWS is scalable because it has an ability to scale the computing resources up or down
according to the organization's demand.
AWS is cost-effective as it works on a pay-as-you-go pricing model.
It provides various flexible storage options.
It offers various security services such as infrastructure security, data encryption, monitoring &
logging, identity & access control, penetration testing, and DDoS attacks.
It can efficiently manage and secure Windows workloads.
AWS Compute Services
Here, are Cloud Compute Services offered by Amazon:
1.EC2(Elastic Compute Cloud) - EC2 is a virtual machine in the cloud on which you have OS
level control. You can run this cloud server whenever you want.
2.LightSail -This cloud computing tool automatically deploys and manages the computer,
storage, and networking capabilities required to run your applications.
3.Elastic Beanstalk — The tool offers automated deployment and provisioning of resources
like a highly scalable production website.
4.EKS (Elastic Container Service for Kubernetes) — The tool allows you to Kubernetes on
Amazon cloud environment without installation.
5.AWS Lambda — This AWS service allows you to run functions in the cloud. The tool is a big
cost saver for you as you to pay only when your functions execute.
Migration
Migration services used to transfer data physically between your datacenter and AWS.
1.DMS (Database Migration Service) -DMS service can be used to migrate on-site databases to
AWS. It helps you to migrate from one type of database to another — for example, Oracle to
MySQL.
2.SMS (Server Migration Service) - SMS migration services allows you to migrate on-site
servers to AWS easily and quickly.
3.Snowball — Snowball is a small application which allows you to transfer terabytes of data
inside and outside of AWS environment.
Storage
1.Amazon Glacier- It is an extremely low-cost storage service. It offers secure and fast storage
for data archiving and backup.
2.Amazon Elastic Block Store (EBS)- It provides block-level storage to use with Amazon EC2
instances. Amazon Elastic Block Store volumes are network-attached and remain independent
from the life of an instance.
3.AWS Storage Gateway- This AWS service is connecting on-premises software applications
with cloud-based storage. It offers secure integration between the company's on-premises and
AWS's storage infrastructure.
Security Services
1.IAM (Identity and Access Management) — IAM is a secure cloud security service which
helps you to manage users, assign policies, form groups to manage multiple users.
2.Inspector — It is an agent that you can install on your virtual machines, which reports any
security vulnerabilities.
3.Certificate Manager — The service offers free SSL certificates for your domains that are
managed by Route53.
4.WAF (Web Application Firewall) — WAF security service offers application-level protection
and allows you to block SQL injection and helps you to block cross-site scripting attacks.
5.Cloud Directory — This service allows you to create flexible, cloud-native directories for
managing hierarchies of data along multiple dimensions.
6.KMS (Key Management Service) — It is a managed service. This security service helps you
to create and control the encryption keys which allows you to encrypt your data.
7.Organizations — You can create groups of AWS accounts using this service to manages
security and automation settings.
8.Shield — Shield is managed DDoS (Distributed Denial of Service protection service). It offers
safeguards against web applications running on AWS.
9.Macie — It offers a data visibility security service which helps classify and protect your
sensitive critical content.
10.GuardDuty —It offers threat detection to protect your AWS accounts and workloads.
Database Services
1.Amazon RDS- This Database AWS service is easy to set up, operate, and scale a relational
database in the cloud.
2.Amazon DynamoDB- It is a fast, fully managed NoSQL database service. It is a simple service
which allow cost-effective storage and retrieval of data. It also allows you to serve any level of
request traffic.
3.Amazon ElastiCache- It is a web service which makes it easy to deploy, operate, and scale an
in-memory cache in the cloud.
4.Neptune- It is a fast, reliable and scalable graph database service.
5.Amazon RedShift - It is Amazon's data warehousing solution which you can use to perform
complex OLAP queries.
Analytics
1.Athena — This analytics service allows perm SQL queries on your S3 bucket to find files.
2.CloudSearch — You should use this AWS service to create a fully managed search engine
for your website.
3.ElasticSearch — It is similar to CloudSearch. However, it offers more features like application
monitoring.
4.Kinesis — This AWS analytics service helps you to stream and analyzing real-time data at
massive scale.
5.QuickSight —It is a business analytics tool. It helps you to create visualizations in a
dashboard for data in Amazon Web Services. For example, S3, DynamoDB, etc.
6.EMR (Elastic Map Reduce) —This AWS analytics service mainly used for big data processing
like Spark, Splunk, Hadoop, etc.
7.Data Pipeline — Allows you to move data from one place to another. For example from
DynamoDB to S3.
Management Services
1.CloudWatch — Cloud watch helps you to monitor AWS environments like EC2, RDS
instances, and CPU utilization. It also triggers alarms depends on various metrics.
2.CloudFormation — It is a way of turning infrastructure into the cloud. You can use templates
for providing a whole production environment in minutes.
3.CloudTrail — It offers an easy method of auditing AWS resources. It helps you to log all
changes.
4.OpsWorks — The service allows you to automated Chef/Puppet deployments on AWS
environment.
5.Config — This AWS service monitors your environment. The tool sends alerts about changes
when you break certain defined configurations.
6.Service Catalog — This service helps large enterprises to authorize which services user will
be used and which won't.
7.AWS Auto Scaling — The service allows you to automatically scale your resources up and
down based on given CloudWatch metrics.
8.Systems Manager — This AWS service allows you to group your resources. It allows you to
identify issues and act on them.
9.Managed Services—It offers management of your AWS infrastructure which allows you to
focus on your applications.
Internet of Things
1.IoT Core— It is a managed cloud AWS service. The service allows connected devices like
cars, light bulbs, sensor grids, to securely interact with cloud applications and other devices.
2.IoT Device Management — It allows you to manage your IoT devices at any scale.
3.IoT Analytics — This AWS IOT service is helpful to perform analysis on data collected by
your IoT devices.
4.Amazon FreeRTOS — This real-time operating system for microcontrollers helps you to
connect IoT devices in the local server or into the cloud.
Application Services
1.Step Functions — It is a way of visualizing what's going inside your application and what
different microservices it is using.
2.SWF (Simple Workflow Service) — The service helps you to coordinate both automated
tasks and human-led tasks.
3.SNS (Simple Notification Service) — You can use this service to send you notifications in the
form of email and SMS based on given AWS services.
4.SQS (Simple Queue Service) — Use this AWS service to decouple your applications. It is a
pull-based service.
5.Elastic Transcoder — This AWS service tool helps you to changes a video's format and
resolution to support various devices like tablets, smartphones, and laptops of different
resolutions.
Deployment and Management
1.AWS CloudTrail: The services records AWS API calls and send backlog files to you.
2.Amazon CloudWatch: The tools monitor AWS resources like Amazon EC2 and Amazon RDS
DB Instances. It also allows you to monitor custom metrics created by user's applications and
services.
3.AWS CloudHSM: This AWS service helps you meet corporate, regulatory, and contractual,
compliance requirements for maintaining data security by using the Hardware Security
Module(HSM) appliances inside the AWS environment.
Developer Tools
1.CodeStar — Codestar is a cloud-based service for creating, managing, and working with
various software development projects on AWS.
2.CodeCommit — It is AWS's version control service which allows you to store your code and
other assets privately in the cloud.
3.CodeBuild — This Amazon developer service help you to automates the process of building
and compiling your code.
4.CodeDeploy — It is a way of deploying your code in EC2 instances automatically.
5.CodePipeline — It helps you create a deployment pipeline like testing, building, testing,
authentication, deployment on development and production environments.
6.Cloud9 —It is an Integrated Development Environment for writing, running, and debugging
code in the cloud.
Mobile Services
1.Mobile Hub — Allows you to add, configure and design features for mobile apps.
2.Cognito — Allows users to signup using his or her social identity.
3.Device Farm — Device farm helps you to improve the quality of apps by quickly testing
hundreds of mobile devices.
4.AWS AppSync —It is a fully managed GraphQL service that offers real-time data
synchronization and offline programming features.
Business Productivity
1.Alexa for Business — It empowers your organization with voice, using Alexa. It will help you
to Allows you to build custom voice skills for your organization.
2.Chime — Can be used for online meeting and video conferencing.
3.WorkDocs — Helps to store documents in the cloud
4.WorkMail — Allows you to send and receive business emails.
Desktop & App Streaming
1.WorkSpaces — Workspace is a VDI (Virtual Desktop Infrastructure). It allows you to use
remote desktops in the cloud.
2.AppStream — A way of streaming desktop applications to your users in the web browser. For
example, using MS Word in Google Chrome.
Artificial Intelligence
1.Lex — Lex tool helps you to build chatbots quickly.
2.Polly — It is AWS's text-to-speech service allows you to create audio versions of your notes.
3.Rekognition — It is AWS's face recognition service. This AWS service helps you to
recognize faces and object in images and videos.
4.SageMaker — Sagemaker allows you to build, train, and deploy machine learning models at
any scale.
5.Transcribe — It is AWS's speech-to-text service that offers high-quality and affordable
transcriptions.
6.Translate — It is a very similar tool to Google Translate which allows you to translate text in
one language to another.
AR & VR (Augmented Reality & Virtual Reality)
1.Sumerian — Sumerian is a set of tool for offering high-quality virtual reality (VR) experiences
on the web. The service allows you to create interactive 3D scenes and publish it as a website
for users to access.
Customer Engagement
1.Amazon Connect — Amazon Connect allows you to create your customer care center in the
cloud.
2.Pinpoint — Pinpoint helps you to understand your users and engage with them.
3.SES (Simple Email Service) — Helps you to send bulk emails to your customers at a
relatively cost-effective price.
Game Development
1.GameLift- It is a service which is managed by AWS. You can use this service to host
dedicated game servers. It allows you to scale seamlessly without taking your game offline.
Applications of AWS services
Amazon Web services are widely used for various computing purposes like:
Web site hosting
Application hosting/SaaS hosting
Media Sharing (Image/ Video)
Mobile and Social Applications
Content delivery and Media Distribution
Storage, backup, and disaster recovery
Development and test environments
Academic Computing
Search Engines
Social Networking
Companies using AWS
Instagram
Zoopla
Smugmug
Pinterest
Netflix
Dropbox
Etsy
Talkbox
Playfish
Ftopia
Advantages of AWS
Following are the pros of using AWS services:
AWS allows organizations to use the already familiar programming models, operating systems,
databases, and architectures.
It is a cost-effective service that allows you to pay only for what you use, without any up-front or
long-term commitments.
You will not require to spend money on running and maintaining data centers.
Offers fast deployments
You can easily add or remove capacity.
You are allowed cloud access quickly with limitless capacity.
Total Cost of Ownership is very low compared to any private/dedicated servers.
Offers Centralized Billing and management
Offers Hybrid Capabilities
Allows you to deploy your application in multiple regions around the world with just a few clicks
Disadvantages of AWS
If you need more immediate or intensive assistance, you'll have to opt for paid support
packages.
Amazon Web Services may have some common cloud computing issues when you move to a
cloud. For example, downtime, limited control, and backup protection.
AWS sets default limits on resources which differ from region to region. These resources consist
of images, volumes, and snapshots.
Hardware-level changes happen to your application which may not offer the best performance
and usage of your applications.
Best practices of AWS
You need to design for failure, but nothing will fail.
It's important to decouple all your components before using AWS services.
You need to keep dynamic data closer to compute and static data closer to the user.
It's important to know security and performance tradeoffs.
Pay for computing capacity by the hourly payment method.
Make a habit of a one-time payment for each instance you want to reserve and to receive a
significant discount on the hourly charge.
Next
Compute Services
Google App Engine: Platform as a Service to deploy Java, PHP, and other applications. It is a
Cloud Computing platform for developing and hosting web applications in Google-managed
data centers. It offers the automatic scaling feature, i.e., as the number of requests for an
application increases, the App Engine automatically allocates more resources for the application
to handle additional demand.
Compute Engine: Infrastructure as a Service to run Microsoft Windows and Linux virtual
machines. It is a component of the Google Cloud platform which is built on the same
infrastructure that runs Google’s search engine, YouTube, and other services.
Kubernetes Engine: It aims at providing a platform for automating deployment, scaling, and
operations of application containers across clusters of hosts. It works with a wide range of
container tools including docker.
Storage Services
Google Cloud Storage: An online file storage web service for storing and accessing data on a
Google Cloud platform infrastructure. The service combines the performance and scalability of
Google Cloud with advanced security and sharing capabilities.
Cloud SQL: A web service that allows you to create, configure, and use relational databases
that live in Google Cloud. It maintains, manages, and administers your databases allowing you
to focus on your applications and services.
Cloud Bigtable: A fast, fully managed, and a highly scalable NoSQL database service. It is
designed for the collection and retention of data from 1 TB to hundreds of PB.
Networking
VPC: Virtual Private Cloud provides a private network with IP allocation, routing, and network
firewall policies to create a secure environment for your deployments.
Cloud Load Balancing: It is a process of distributing workloads across multiple computing
resources. This reduces the cost and maximizes the availability of the resources.
Content Delivery Network: A geographically distributed network of proxy servers and their data
centers. The goal here is to provide high availability and high performance by spatially
distributing the service relating to end users.
Big Data
BigQuery: Google BigQuery Service is a fully managed data analysis service that enables
businesses to analyse Big Data. It features highly scalable data storage, the ability to perform
ad-hoc queries, and the ability to share data insights via the web.
Google Cloud Datastore: A fully managed, schema less, non-relational datastore. It supports
atomic transactions and a rich set of query capabilities and can automatically scale up and
down depending on the load.
Google Cloud Dataproc: A fast, easy-to-use and manage Spark and Hadoop service for
distributed data processing. With Cloud Dataproc, you can create Spark or Hadoop clusters,
sized for your workloads precisely when you need them.
Cloud AI
Cloud Machine Learning Engine: A managed service that will enable you to build Machine
Learning models based on mainstream frameworks.
Cloud AutoML: A Machine Learning product that enables developers to provide their data sets
and obtain access to quality trained models by Google’s transfer learning and Neural
Architecture Search.
Management Tools
Google Stackdriver: Provides performance and diagnostics data in the form of monitoring,
logging, tracing, error reporting, and alerting it to public cloud users.
Google Cloud Console App: A native mobile application that enables customers to manage the
key Google Cloud services. It provides monitoring, altering, and the ability to take actions on
resources.
Identity and Security
Cloud Data Loss Prevention API: It helps you manage sensitive data. It provides a fast and
scalable classification for sensitive data elements like credit card numbers, names, passport
numbers, and more.
Cloud IAM: Cloud Identity and Access Management refers to a framework of policies and
technologies for ensuring that proper people in an enterprise have the appropriate access to
technology resources. It is also called identity management (IdM).
Check out Intellipaat’s GCP Training Course to get ahead in your career!
IoT
Cloud IoT Core: It is a fully managed service that allows you to easily and securely connect,
manage, and ingest data from devices that are connected to the Internet. It permits utilization of
other Google Cloud services for collecting, processing, analysing, and visualizing IoT data in
real time.
Cloud IoT Edge: Edge computing brings memory and computing power closer to the location
where it is needed.
Top Users of Google Cloud
Now that we are well aware of the Google Cloud benefits and services, let’s now have a look on
the top users of this cloud platform.
Twitter: A well-known application, which lets people share information. With people tweeting
more and more every day, the data produced is enormously large. Google Cloud is used for
storing and computing purpose.
20th Century Fox: Data scientists at 20th Century Fox and Google Cloud have developed a
Machine Learning software that can analyse movie trailers and predict how likely people are to
see those movies in theaters.
PayPal: PayPal partners with Google Cloud to increase security, build a faster network, and
develop services for its customers.
eBay: eBay uses Google Cloud to innovate in image search, improve customer experiences in
China, and train translation models.
Chevron: Chevron uses Google AutoML Vision to find information that is always challenging to
get when you need it.
HSBC: HSBC brings a new level of security, compliance, and governance to its banks using
Google Cloud.
LG CNS: LG CNS data analytics solution, with Google AI and Edge TPU, will provide a great
value for LG CNS customers in the smart factory arena.
8.What pros and cons of Cloud computing in comparisons of Distributed Grid computing?
Ans:
Cloud Computing Distributed Grid Computing
Used for business and public needs. Used for specific purposes.
Clouds evolve faster than grid. Grids evolve slower than cloud.
Level of expertise to use a cloud is lower than Level of expertise to use a grid is higher than
grid. cloud.
Cloud offers more services than grid It is the base concept of cloud computing
computing.
A cloud would usually use cloud or part of a Grids tends to be more loosely
grid.In fact almost all services on the Internet coupled,heterogeneous, and geographically
can be obtained from cloud. dispersed compared to conventional cluster
computing systems
Disaster Recovery
With many services offering options such as regular backups and snapshots, important data
loss is highly unlikely.
Increased Collaboration and Flexibility
Moving to the cloud increases opportunities for collaboration amongst employees. Colleagues
can sync documents workflows or share apps with ease. Often these apps will allow
collaborators to simultaneously receive updates in real time. Additionally, cloud computing
allows for each team member to work from anywhere, at any time. The cloud centralizes data,
meaning that business owners, employees, and clients can access company data from
anywhere that has Internet access (this is often referred to as “tunneling-in”, via some sort of
Remote Desktop Protocol).
Eco Friendly
Cloud computing can decrease a business’ carbon footprint, reducing energy consumption and
carbon emissions by more than 30%. For small companies, the decreased energy usage can
reach 90%! It can also help a business project an environmentally sound image.
Cons of Cloud Computing
Internet Connectivity
Running business applications in the cloud is great, as long as a consistent Internet connection
is maintained. If a cloud-based service provider loses connectivity, there’s not much to be done
until that Internet connection returns. Even the best servers go down occasionally. Because of
this, most cloud providers maintain several redundancies to ensure up time. It is important when
selecting a cloud provider, to review their SLA.
Learning Curve
Cloud computing platforms are great tools for developers, but for newcomers to the industry, it
might not be as simple as it seems. For those unfamiliar with Linux, SSH Key Generation, and
other ins-and-outs of basic server administration, cloud computing platforms that deploy virtual
machines or dedicated VPS (Virtual Private Servers) may seem much more different than what
is anticipated.
Pros and Cons of Grid Computing
Pros of Grid Computing
Cheaper Servers
No need to buy large SMP servers! Applications would be able to break apart and run across
smaller servers. Those servers cost far less than SMP servers.
More Efficient
Much more efficient use of idle resources. Idle servers and desktops would be able to accept
jobs! Many resources sit idle, especially during off business hours. This is not the case anymore
with a grid computing setup.
Fail-safe
Grid computer environments are modular and don’t have just one fail point. Hence if one of the
machines within the grid fails, there are plenty of others able to pick the load. Jobs can
automatically restart if a failure occurs.
Cons of Grid Computing
May Still Require Large SMP
Will be forced to run on a large SMP for memory hungry applications that can’t take advantage
of MPI
Requires Fast Interconnect
You may need to have a fast interconnect between compute resources (gigabit ethernet at a
minimum). Infiniband for MPI intense applications
Some Applications Require Customization
Applications would need tweaking to take full advantage of new models.
Licensing
Licensing across many servers may make it prohibitive for some apps.
Advantages and Benefits of Distributed Computing
Scalability and Modular Growth
Distributed systems are inherently scalable as they work across different machines and scale
horizontally. This means a user can add another machine to handle the increasing workload
instead of having to update a single system over and over again. There is virtually no cap on
how much a user can scale. A system under high demand can run each machine to its full
capacity and take machines offline when workload is low.
Fault Tolerance and Redundancy
Distributed systems are also inherently more fault tolerant than single machines. A business
running a cluster of 8 machines across two data centers means its apps would work even if one
data center goes offline. This translates into more reliability as in case of a single machine
everything goes down with it. Distributed systems stay put even if one or more nodes/sites stop
working (performance demand on the remaining nodes would go up).
Low Latency
Since users can have a node in multiple geographical locations, distributed systems allow the
traffic to hit a node that’s closest, resulting in low latency and better performance. However, the
software also has to be designed for running on multiple nodes at the same time, which can
result in higher cost and more complexity.
Cost Effectiveness
Distributed systems are much more cost effective compared to very large centralized systems.
Their initial cost is higher than standalone systems, but only up to a certain point after which
they are more about economies of scale. A distributed system made up of many mini computers
can be more cost effective than a mainframe machine.
Efficiency
Distributed systems allow breaking complex problems/data into smaller pieces and have
multiple computers work on them in parallel, which can help cut down on the time needed to
solve/compute those problems.
Disadvantages of Distributed Computing
Complexity
Distributed computing systems are more difficult to deploy, maintain and troubleshoot/debug
than their centralized counterparts. The increased complexity is not only limited to the hardware
as distributed systems also need software capable of handling the security and
communications.
Higher Initial Cost
The deployment cost of a distribution is higher than a single system. Increased processing
overhead due to additional computation and exchange of information also adds up to the overall
cost.
Security Concerns
Data access can be controlled fairly easily in a centralized computing system, but it’s not an
easy job to manage security of distributed systems. Not only the network itself has to be
secured, users also need to control replicated data across multiple locations.
9.What are different types of distributed computing systems before cloud computing?
Ans: Types of Distributed Computing before Cloud Computing:-
Grid computing
Cluster computing
Grid computing is based on the idea of a supercomputer with enormous computing power.
However, computing tasks are performed by many instances rather than just one. Servers and
computers can thus perform different tasks independently of one another. Grid computing can
access resources in a very flexible manner when performing tasks. Normally, participants will
allocate specific resources to an entire project at night when the technical infrastructure tends to
be less heavily used.
One advantage of this is that highly powerful systems can be quickly used and the computing
power can be scaled as needed. There is no need to replace or upgrade an expensive
supercomputer with another pricey one to improve performance.
Since grid computing can create a virtual supercomputer from a cluster of loosely
interconnected computers, it is specialized in solving problems that are particularly
computationally intensive. This method is often used for ambitious scientific projects and
decrypting cryptographic codes.
Cluster computing cannot be clearly differentiated from cloud and grid computing. It is a more
general approach and refers to all the ways in which individual computers and their computing
power can be combined together in clusters. Examples of this include server clusters, clusters in
big data and in cloud environments, database clusters, and application clusters. Computer
networks are also increasingly being used in high-performance computing which can solve
particularly demanding computing problems.
10.What are the advantages of "Software as a Service" (SaaS)? Explain with example
Ans:
SaaS offers many potential advantages over the traditional models of business software
installation, including:
Lower up-front cost - SaaS is generally subscription-based and has no up-front licence fees
resulting in lower initial costs. The SaaS provider manages the IT infrastructure that is running
the software, which brings down fees for hardware and software maintenance.
Quick set up and deployment - SaaS application is already installed and configured in the cloud.
This minimises common delays resulting from often lengthy traditional software deployment.
Easy upgrades - The SaaS providers deal with hardware and software updates, deploying
upgrades centrally to the hosted applications and removing this workload and responsibility from
you.
Accessibility - All you need to access a SaaS application is a browser and an internet
connection. This is generally available on a wide range of devices and from anywhere in the
world, making SaaS more accessible than the traditional business software installation.
Scalability - SaaS providers generally offer many subscription options and flexibility to change
subscriptions as and when needed, eg when your business grows, or more users need to
access the service.
If you’re considering moving your ecommerce business to the Cloud, here are three important
acronyms to wrap your head around:
IaaS.
PaaS.
SaaS.
These are the three main types of cloud computing.
You’ve probably heard of them before; they’re all experiencing a surge in popularity as more
businesses move to the Cloud.
SaaS accounts for approximately 24% of all enterprise workloads (up from 14% in 2016)
IaaS is hovering around 12% (up from 6%).
PaaS is currently the most popular model, hovering around 32% and expected to grow in 2020.
With uptake rates like these, cloud computing is becoming the norm, and many businesses are
phasing out on-premise software altogether.
The Key Differences Between On-Premise, SaaS, PaaS, IaaS
Not so long ago, all of a company’s IT systems were on-premise, and clouds were just white
fluffy things in the sky.
Now, you can utilize the Cloud platform for nearly all your systems and processes.
SaaS, PaaS, and IaaS are simply three ways to describe how you can use the cloud for your
business.
IaaS: cloud-based services, pay-as-you-go for services such as storage, networking, and
virtualization.
PaaS: hardware and software tools available over the internet.
SaaS: software that’s available via a third-party over the internet.
On-premise: software that’s installed in the same building as your business.
Here’s a great visual breakdown from Hosting Advice:
Private Cloud: Cloud services are used by a single organization, so only the organization has
access to its data and can manage it.
Public Cloud:Cloud services are accessible via a network like internet and can be used by
multiple clients. MS Azure, AWS are some of the examples of public cloud hosting.
Hybrid Cloud:Cloud services are distributed among public and private clouds. Its users can use
them as per their requirement i.e.private as well as public cloud services in day to day
operations.
A comparison: Private cloud vs Public cloud vs Hybrid cloud
Comparing Cloud Deployment models Public cloud Private cloud Hybrid Cloud
Cloud environment Multi-Tenancy-Shared environment. Single tenancy-only for single use of
an organization. Both single tenancy and multi –tenancy. When data is stored in the public
cloud, it delivers multi- tenant environment meaning the data from multiple organizations is
stored in a shared environment whereas when data is stored in private cloud, it is kept private
for the use of a single organization.
Data center location Anywhere – where the cloud service provider’s services are located.
Inside the organization’s network. Inside the organization’s network for private cloud services
as well as wherever service provider’s services are there for public cloud services.
Resource sharing Server hardware, network and storage are shared by multiple users in the
cloud. No sharing of resources. Hardware, storage and network are dedicated to the use of a
single client or company. Very secure; integration options add an additional layer of
security.
Cloud storage Public cloud delivers storage as a service on a pay per use basis. Best for
backups as a part of a disaster recovery plan as well as archiving email and static non-core
application data. OneDrive is an example of public cloud storage. Private cloud delivers internal
cloud storage that runs on a dedicated infrastructure in a data center. Hybrid cloud
manages streamlined storage that uses both local and off-site resources and serves as a
gateway between on premise and public cloud storage.
Scalability Instant and unlimited. Sacrifices scalability but provides greater control and
security. On demand unlimited resources.
Pricing structure Prices charged on the usage basis. Comparatively expensive. High
but delivers competitive advantage.
Cloud SecurityGood, but depends on the security measures of the service provider. Most
secure. Secure.
Performance Low to medium. Very High. Very High.
Ans: The Windows Azure platform is a cloud computing platform supporting the de-velopment
and deployment of Windows applications that provide services to both businesses and
consumers. These applications can be created using the .NET frame-work in languages such as
C# and Visual Basic, or they can be built without the NET framework in C++, Java, PHP, Ruby
and Python. Windows Azure is originally a PaaS platform but since November 2010 it is also
possible to deploy virtual ma-chine (VM) images of Windows Server (2008), which is a form of
IaaS. Notice that Windows Azure provides not necessarily an execution environment for all
supported languages. For example, Java APIs are available for the different Azure services, but
the developer has to install a JVM and an application server in an Azure VM himself.
The Windows Azure platform is a group of cloud technologies, each providing a specific set of
services to application developers.
– Windows Azure: A Windows environment for running applications (Compute) and storing data
(Storage) in the cloud. A Windows Azure application can be created using three kinds of
compute service types, called roles: (i) a Web role for running web-based applications, (ii) a
Worker role for (background) processing, and (iii) a VM role to run VM images. An application is
structured as any combination of role instances. Web and worker role instances consist of a
Windows Server 2008 R2 VM with the .NET 4.0 framework pre-installed. In a web role, there is
also an IIS web server available.
The storage service consists of blob storage (coarse-grained, unstructured data), table storage
(set of entities with properties, NoSQL), and queues (asynchronous communication between
different roles). All data is accessible via a REST API and the OData query language for tables.
The Fabric Controller makes abstrac-tion of the underlying infrastructure and offers a pool of
processing power to the compute and storage services.
– SQL Azure: A cloud-based service for relational data, built on Microsoft SQL Server.
The Windows Azure platform provides an environment that strongly corresponds to a traditional
on-premise setup (cf. Category 1), this especially applies to the Windows Azure Compute
service and SQL Azure. In this comparison we focus on Windows Azure using traditional .NET
frameworks in the C# programming language and the Ninject dependency injection framework .
We also include SQL Azure into the comparison. The case study application is implemented
using Azure SDKs 1.2 - 1.6.
Ans: Software as a Service (SaaS) – It provides a complete product that is run and managed
by the service provider. One need not to worry about how the service is maintained or how
the underlying infrastructure is managed.
Categories:-
The SaaS maturity model is broken down into four levels, and each of them brings certain
opportunities and challenges you should be aware of when accessing SaaS vendors.
Level 1 (Single-Tenant, Custom Instances)
At this level of the SaaS maturity model, the only way to support multiple customers (tenants) is
to provide each of them a separate copy of the software. Because the provided copies can be
customized by writing custom code, each customer is required to run a different instance of the
software and scalability is non-existent, even though the software is technically delivered as a
service. As such, no economies of scale can be harnessed, making this level the least
cost-effective and sustainable when managing a larger number of customers.
Level 2 (Single-Tenant, Configurable Instances)
At level 2, software can be customized by changing its configuration instead of writing custom
code. In other words, all tenants interact with the same code configured in different ways, with
each tenant running their own copy on a separate virtual or physical machine. Consequently,
scalability and multi-tenancy are still not achieved. What’s more, the provider is at a competitive
disadvantage because individual instances don’t share the same pool of computing power,
which would make it possible to achieve economies of scale.
16.What is Service-Level Agreement (SLA)? Explain about the security controls classified
in cloud computing.
Ans: A cloud SLA (cloud service-level agreement) is an agreement between a cloud service
provider(IaaS, PaaS, SaaS) and a customer(Cloud consumer) that ensures a minimum level of
service is maintained.
It guarantees levels of reliability, availability and responsiveness to systems and applications;
specifies who governs when there is a service interruption; and describes penalties if service
levels are not met.
Service Level Agreements usually specify some parameters which are mentioned below:
1.Availability of the Service (uptime)
2.Latency or the response time
3.Service components reliability
4.Each party accountability
5.Warranties
Cloud security control primarily helps in addressing, evaluating and implementing security in the
cloud. The Cloud Security Alliance (CSA) has created a Cloud Control Matrix (CCM), which is
designed to help prospective cloud buyers evaluate a cloud solution’s overall security. Although
there are limitless cloud security controls, they are similar to the standard information security
controls and can be categorized in different domains including:
Deterrent Controls: Don’t protect the cloud architecture/infrastructure/environment but serve as
warning to a potential perpetrator of an attack.
Preventative Controls: Used for managing, strengthening and protecting the vulnerabilities
within a cloud.
Corrective Controls: Help reduce the after-effects of an attack.
Detective Controls: Used to identify or detect an attack.
17.Explain the cloud deployment models in detail. Also, outline their advantages and
disadvantages while implementing an application.
Ans: Private cloud – It allows systems and services to be accessible within an organization.
•Ex -CityCloud
Public cloud – It allows systems and services to be accessible by the general public.
•A Public Cloud is a huge data centre that offers the same services to all its users.
•Globally distributed
•Hybrid cloud - It is a mixture of public and private cloud. The critical activities are
performed by private cloud and the non- critical activities are performed by public cloud.
Ans: Middleware is software which lies between an operating system and the applications
running on it. Essentially functioning as a hidden translation layer, middleware enables
communication and data management for distributed applications. It is sometimes called
plumbing, as it connects two applications together so data and databases can be easily passed
between the “pipe.” Using middleware allows users to perform such requests as submitting
forms on a web browser or allowing the web server to return dynamic web pages based on a
user’s profile.
Middleware helps developers build applications more efficiently. It acts like the connective tissue
between applications, data, and users. For organizations with multi-cloud and containerized
environments, middleware can make it cost-effective to develop and run applications at scale.
Middleware can help developers transform legacy monolithic applications into cloud-native
applications, keeping valuable tools active with better performance and more portability
For all the benefits cloud-native development provides, it also brings added complexity.
Applications can be deployed across multiple infrastructures, from on-premises systems to
public clouds. Architectures can vary widely. Developers are juggling multiple tools, languages,
and frameworks. And the pressure is on to do more in less time and at a lower cost.
Organizations turn to middleware as a way to manage this complexity and to keep application
development quick and cost-effective. Middleware can support application environments that
work smoothly and consistently across a highly distributed platform.
19.Design the architecture of PaaS and SaaS in cloud computing? Also, explain the different
categories of PaaS and SaaS with example.
Ans: To help you make sense of this complicated world, SaaS can be divided into three
categories:
Packaged software: This is the biggest area of the SaaS market. Packaged software comes in
many different flavors: customer relationship management, supply chain management, financial
management, and human resources, to name the most common.
These integrated offers focus on a specific process, such as managing employees’ benefits,
salaries, and annual performance reviews. These products tend to have several characteristics
in common: They’re designed with specific business processes built in that customers can
modify. They have moved in great numbers to the cloud because customers were finding the
platforms too hard to manage.
Collaborative software: This increasingly vibrant area of the market is driven by the ubiquitous
availability of the Internet, combined with the fact that teams are located all over the world. This
area is dominated by software that focuses on all sorts of collaborative efforts including Web
conferencing, document collaboration, project planning, instant messaging, and even email. In a
sense, it was inevitable that these platforms would move to the cloud: These tasks occur
throughout the organization and need to be easily accessed from many locations.
Enabling and management tools: What’s in this category? Think about the development tools
that developers need when creating and extending a SaaS platform; also think about the
testing, monitoring, and measuring that a customer and the developer need. Also consider the
compliance issues related to the use of this type of software in the real world. These issues are
included in this third category.
PaaS Types :
Various sorts of PaaS are presently accessible to engineers. They are :
1. Public PaaS2. Private PaaS3. Hybrid PaaS4. Communication PaaS5. Mobile PaaS6. Open
PaaS
These are explained as follows below.
1.Public PaaS –
Is appropriate for use in public clouds. Public PaaS permits clients to control programming
extension, while cloud supplier deals with circulation of all other key IT segments expected to
have applications, including working frameworks, databases, workers, and capacity framework
systems.
Public dice merchants offer middleware, which permits engineers to introduce, design, and
control workers and databases without the need to introduce a framework. Therefore, Public
Pass and Infrastructure (IaaS) run all together. PaaS with exercises on merchant’s IaaS
foundation using public cloud. Tragically, this implies clients will adhere to similar public cloud
choices that they would prefer not to utilize.
Some small and medium-sized organizations have received public PaaS, however huge
enterprises and associations wouldn't acknowledge it because of their cozy relationship with the
public cloud. This is principally a consequence of an enormous number of guidelines and
consistency issues on organization application advancement in the public cloud.
2.Private PaaS –
Plans to give dexterity of public pass while keeping up security, consistency, advantages and
ease of private security community. A private pass is normally circulated as gadget or
programming in a client's firewall, which is regularly kept up in a server farm on an
organization's premises. A private PaaS can be created on a framework and works inside an
organization's particular private cloud.
Private dice permit an association to all more likely serve designers, improve inside asset
usage, and decrease expensive cloud entrance that numerous organizations face. Likewise,
private dice permit designers to run and deal with their organization’s applications in
consistency with severe security and protection necessities.
3.Hybrid PaaS –
Consolidates organizations with Public PaaS and Private PaaS, with accommodation of
unbounded limits offered by Public PaaS and cost-adequacy of having inside framework in
Private PaaS. Hybrid PaaS utilizes hybrid cloud.
CPaaS gives a completely evolved system to making ongoing communication highlights without
requirement for engineers to assemble their own structure, including standard-based application
programming interfaces, programming apparatuses, prebuilt applications, and test code.
CPaaS suppliers help clients in improvement procedures by offering help and item
documentation. A few suppliers offer programming advancement packs just as libraries that
help assemble applications on different work areas and mobile stages. Improvement groups
picking to utilize CPaaS spares time for foundation, HR and market.
6.Open PaaS –
Is a free, open-source, business-situated community oriented stage that is alluring on all
gadgets and gives a helpful web application including schedule, contacts and mail applications.
It is intended to permit clients to immediately run new applications. One of its assets is to create
innovation sent for big business synergistic applications, particularly half and half mists.
Ans: Virtualization is a technique of how to separate a service from the underlying physical
delivery of that service. It is the process of creating a virtual version of something like computer
hardware. It was initially developed during the mainframe era. It involves using specialized
software to create a virtual or software-created version of a computing resource rather than the
actual version of the same resource. With the help of Virtualization, multiple operating systems
and applications can run on the same machine and its same hardware at the same time,
increasing the utilization and flexibility of hardware.
1.Application Virtualization:
Application virtualization helps a user to have remote access to an application from a server.
The server stores all personal information and other characteristics of the application but can
still run on a local workstation through the internet. Example of this would be a user who needs
to run two different versions of the same software. Technologies that use application
virtualization are hosted applications and packaged applications.
Partitioning: In virtualization, many applications and operating systems (OSes) are supported in
a single physical system by partitioning (separating) the available resources.
Isolation: Each virtual machine is isolated from its host physical system and other virtualized
machines. Because of this isolation, if one virtual-instance crashes, it doesn’t affect the other
virtual machines. In addition, data isn’t shared between one virtual container and another.
Encapsulation: A virtual machine can be represented (and even stored) as a single file, so you
can identify it easily based on the service it provides. In essence, the encapsulated process
could be a business service. This encapsulated virtual machine can be presented to an
application as a complete entity. Therefore, encapsulation can protect each application so that it
doesn’t interfere with another application.
2.Server Virtualization:
Server virtualization is a technology for partitioning one physical server into multiple virtual
servers. Each of these virtual servers can run its own operating system and applications, and
perform as if it is an individual server. This makes it possible, for example, to complete
development using various operating systems on one physical server or to consolidate servers
used by multiple business divisions.
Among the various virtualization methods available, NEC primarily focuses on virtualization
software solutions. Because the virtualization software, or hypervisor, used by NEC runs directly
on bare hardware (physical servers), our virtualized environments have little overhead. NEC’s
proven, reliable solutions are built upon years of experience with virtualization.
Benefits
Benefits
Primary advantages of server virtualization
Reduce number of servers
Partitioning and isolation, the characteristics of server virtualization, enable simple and safe
server consolidation.
Through consolidating, the number of physical servers can be greatly reduced. This alone
brings benefits such as reduced floor space, power consumption and air conditioning costs.
However, it is essential to note that even though the number of physical servers is greatly
reduced, the number of virtual servers to be managed does not change. Therefore, when
virtualizing servers, installation of operation management tools for efficient server management
is recommended.
Reduce TCO
Server consolidation with virtualization reduces costs of hardware, maintenance, power, and air
conditioning. In addition, it lowers the Total Cost of Ownership (TCO) by increasing the
efficiency of server resources and operational changes, as well as virtualization-specific
features. As a result of today’s improved server CPU performance, a few servers have high
resource-usage rates but most are often underutilized. Virtualization can eliminate such
ineffective use of CPU resources, plus optimize resources throughout the server environment.
Furthermore, because servers managed by each business division’s staff can be centrally
managed by a single administrator, operation management costs can be greatly reduced.
Improve availability and business continuity
One beneficial feature of virtualized servers not available in physical server environments is live
migration. With live migration, virtual servers can be migrated to another physical server for
tasks such as performing maintenance on the physical servers without shutting them down.
Thus there is no impact on the end user. Another great advantage of virtualization technology is
that its encapsulation and hardware-independence features enhance availability and business
continuity.
Increase efficiency for development and test environments
At system development sites, servers are often used inefficiently. When different physical
servers are used by each business division’s development team, the number of servers can
easily increase. Conversely, when physical servers are shared by teams, reconfiguring
development and test environments can be time and labor consuming.
Such issues can be resolved by using server virtualization to simultaneously run various
operating system environments on one physical server, thereby enabling concurrent
development and testing of multiple environments. In addition, because development and test
environments can be encapsulated and saved, reconfiguration is extremely simple.
NEC specific advantages on server virtualization
NEC specific advantages on server virtualization
The impact of a failure in a physical server can increase after virtualizing and consolidating
servers. To prevent such problems, NEC provides ExpressCluster and other products for
improving the availability of virtual servers.
22.Draw the framework of common cloud management platform reference architecture and
explain its components.
Ans: Same as Q26. Explain about the transactional process between Data Owner, Third Party
Auditor and Cloud Service Provider and Application Users.
23.Write different types of internal security breaches in cloud computing. Also, explain the steps
to reduce cloud security breaches.
24.Discuss, in detail, the cloud security reference model and explain how cloud security is
integrated into the design of application.
26.Explain cloud computing architecture and its various components with a diagram.
Ans:
30.Explain the technologies that work behind the cloud computing platform.
Ans: https://www.tutorialride.com/cloud-computing/cloud-computing-technologies.htm
33.Discuss various cloud service models and prepare a comparative report of these models
with examples.
Ans:- Advantages:-
Controls: Better controls for data, users and information assets.
Security: The cloud belongs to a single client. Hence, the infrastructure and
systems can be configured to provide high levels of security.
Superior Performance: Normally private clouds are deployed inside the firewall
of the organization’s intranet which ensures efficiency and good network
performance.
Easy Customization: The hardware and other resources can be customized easily
by the company.
Compliance: Compliance is achieved easily in private clouds.
Disadvantages:-
Cost: Costs are substantial in the case of building an on-premise private cloud. The
running cost would include personnel cost and periodic hardware upgrade costs. In
the case of outsourced private cloud, operating cost will include per resource usage
and subject to change at the discretion of the service provider.
Capacity ceiling: Due to physical hardware limitations with the service provider,
there could be a capacity ceiling
to handle only a certain amount of servers or storage.
Vendor lock-in: This can be a major impediment in private cloud adoption especially
when the hardware and infrastructure is outsourced. This is a service delivery
technique where the client company is forced to continue with the same service
provider, thus preventing the client to migrate to another vendor.
Ans:-
https://www.sectorlink.com/article/pros-and-cons-between-public-private-and-community-cloud
Ans:- Virtualization relies on software to simulate hardware functionality and create a virtual
computer system. This enables IT organizations to run more than one virtual system – and
multiple operating systems and applications – on a single server. The resulting benefits include
economies of scale and greater efficiency.
Hardware Virtualization:- In hardware virtualization, software called hypervisor is used.
With the help of a hypervisor virtual machine, software embedded into the hardware component
of the server. The work of hypervisor is that it manages the physical hardware resource which is
shared between the customer and the provider.
Hardware virtualization can be done by extracting the physical hardware with the help of the
virtual machine monitor (VVM). There are several extensions in the processes, which help to
accelerate virtualization activities and boost the performance of hypervisors. If this virtualization
is done for a server platform it is known as server socialization.
Ans:-
Ans:- https://en.wikipedia.org/wiki/Google_App_Engine
Ans:- https://en.wikipedia.org/wiki/Google_App_Engine
44.Identify the stages of live migration of Virtual Machine.
Ans:-
Ans:- Min-min scheduling is based on Minimum Completion Time (MCT) that is used to assign
tasks to the resources to have minimum expected completion time. It will work in two Phases. In
the first phase, the expected completion time will be calculated for each task in a metatask list.
In the second phase, the task with the overall minimum expected completion time from the
metatask list is selected and assigned to the corresponding resource. Then this task is removed
from the metatask list and the process is repeated until all tasks in the metatask list are mapped
to the corresponding resources However, the Min- min algorithm is unable to balance the load
well as it usually does the scheduling of small tasks initially.
Min-min scheduling is based on Minimum Completion Time (MCT) that is used to assign tasks
to the resources to have minimum expected completion time. It will work in two Phases, In the
first phase, the expected completion time will be calculated for each task in a metatask list. In
the second phase, the task with the overall minimum expected completion time from the
metatask list is select and assigned to the corresponding resource. Then this task is removed
from metatask list and the process is repeated until all tasks in the metatask list are mapped to
the corresponding resources However, the Min- min algorithm is unable to balance the load well
as it usually does the scheduling of small tasks initially
The Max-min algorithm is commonly used in a distributed environment which begins with a set
of unscheduled tasks. Then calculate the expected execution matrix and expected completion
time of each task on the available resources. Next, choose the task with overall maximum
expected completion time and assign it to the resource with minimum overall execution time.
Finally recently scheduled task is removed from the metatasks set, update all calculated times,
then repeat until meta-tasks set become empty
Ans:- Concept:- Virtualization relies on software to simulate hardware functionality and create a
virtual computer system. This enables IT organizations to run more than one virtual system –
and multiple operating systems and applications – on a single server. The resulting benefits
include economies of scale and greater efficiency.
Hardware Virtualization:- In hardware virtualization, software called hypervisor is used.
With the help of a hypervisor virtual machine, software embedded into the hardware component
of the server. The work of hypervisor is that it manages the physical hardware resource which is
shared between the customer and the provider.
Hardware virtualization can be done by extracting the physical hardware with the help of the
virtual machine monitor (VVM). There are several extensions in the processes, which help to
accelerate virtualization activities and boost the performance of hypervisors. If this virtualization
is done for a server platform it is known as server socialization.
Ans- Virtualization uses software to create an abstraction layer over computer hardware that
allows the hardware elements of a single computer—processors, memory, storage and
more—to be divided into multiple virtual computers, commonly called virtual machines (VMs).
A Virtual Machine is a software implementation of a
computing environment where an operating system (OS) or
program may be installed and run.
Virtual Machine Monitor (VMM)/ Hypervisor
Virtual Machine Monitor mediates access to the physical
hardware presenting to each guest OS a Virtual Machine.
52.Examine the various privacy and security issues of cloud computing.
Ans:- A cloud database is a database service built and accessed through a cloud platform. It
serves many of the same functions as a traditional database with the added flexibility of cloud
computing. Users install software on a cloud infrastructure to implement the database.
Ans :- The CAP theorem is called the Brewer’s Theorem. It states that a
distributed computing environment can only have 2 of the 3: Consistency,
Availability and Partition Tolerance – one must be sacrificed.
Consistency implies that every read fetches the last write
Availability implies that reads and writes always succeed. In other words,
each non-failing node will return a response in a reasonable amount of time
Partition Tolerance implies that the system will continue to function when
network partition occurs.
58.Describe the framework of mobile cloud computing architecture with schematic diagram.
59.Explain in brief the meaning of Data Integrity, Data confidentiality, Data availability and data
privacy.
Ans-. Data Integrity involves maintaining the consistency and trustworthiness of data over its
entire life cycle. Data must not be changed in transit, and precautionary steps must be taken to
ensure that data cannot be altered by unauthorized people.
Data availability means that information is accessible to authorized users. It provides an
assurance that your system and data can be accessed by authenticated users whenever they’re
needed. Similar to confidentiality and integrity, availability also holds great value.
Confidentiality ensures that data exchanged is not accessible to unauthorized users. The
users could be applications, processes, other systems and/or humans. When designing a
system, adequate control mechanisms to enforce confidentiality should be in place, as well as
policies that dictate what authorized users can and cannot do with the data.
Data privacy or information privacy is a branch of data security concerned with the
proper handling of data – consent, notice, and regulatory obligations. More specifically, practical
data privacy concerns often revolve around: Whether or how data is shared with third parties.
60.What is a cloud database? Explain SQL based and NoSQL based data model in cloud
database.
Ans:- A cloud database is a database service built and accessed through a cloud platform. It
serves many of the same functions as a traditional database with the added flexibility of cloud
computing. Users install software on a cloud infrastructure to implement the database.
NoSQL,which stands for “not only SQL,” is an approach to database design that provides
flexible schemas for the storage and retrieval of data beyond the traditional table structures
found in relational databases. While NoSQL databases have existed for many years, NoSQL
databases have only recently become more popular in the era of cloud, big data and
high-volume web and mobile applications. They are chosen today for their attributes around
scale, performance and ease of use. The most common types of NoSQL databases are
key-value, document, column and graph databases.
SQL or relational traditional databases. are one type of table based database which can run in
the cloud, either in a virtual machine or as a service, depending on the vendor. While SQL
databases are easily vertically scalable, horizontal scalability poses a challenge that cloud
database services based on SQL have started to address.
61.Explain the risks and security issues that arise in a cloud environment.